Skip to content

Uncertainty & limitation reporting

This protocol standardizes how to write uncertainty and limitation sections so readers and downstream agents can distinguish evidence-backed conclusions, open assumptions, and unresolved gaps.

Summary

Use this protocol whenever writing or revising uncertainty and limitation content in: - wiki/papers/*.md pages (publication-level interpretation boundaries). - Synthesis pages (wiki/concepts/, wiki/materials/, wiki/forcefields/, wiki/debates/) where cross-paper claims can overgeneralize.

Core objective: - Make uncertainty explicit, scoped, and actionable. - Tie each non-trivial caveat to a traceable source context. - Prevent hidden assumptions from being mistaken as conclusions.

Inputs and prerequisites

Required inputs: - Target page draft with frontmatter complete. - Relevant source_refs and linked paper: pages. - Any available protocol context from ## Methods and ## Findings on cited paper pages.

Preconditions before drafting uncertainty text: - Claims in the target page are already grouped by topic. - Each group has at least one evidence anchor candidate (paper id plus locator context). - You know whether the page is: - Paper-focused (single-study interpretation constraints), or - Synthesis-focused (cross-paper consistency and transferability constraints).

Minimum evidence anchor unit: - paper_id plus at least one of: - section label, - page range, - short locator note tied to method/result context.

Procedure

  1. Build a claim inventory first.
  2. List each substantive claim in the page in one sentence.
  3. Mark each claim as:
  4. Directly measured/reported,
  5. Derived inference,
  6. Hypothesis-level interpretation.

  7. Classify uncertainty type for each claim.

  8. Method uncertainty: model choice, functional/formulation, force-field transferability, sampling limitations.
  9. Data uncertainty: sparse points, missing controls, low extraction quality, SI-only visibility.
  10. Scope uncertainty: domain shift across materials, temperatures, fields, interfaces, time scales.
  11. Reporting uncertainty: missing parameters, unclear validation, unresolved disagreement in corpus.

  12. Attach evidence anchors.

  13. For every uncertainty statement, attach at least one anchor path:
  14. For paper pages: anchor to the same publication sections backing Methods/Findings.
  15. For synthesis pages: anchor to specific supporting papers in source_refs.
  16. If uncertainty is due to absence of evidence, state that explicitly ("not reported in indexed source") and identify what source tier was checked (PDF, SI, extract, metadata).

  17. Write uncertainty statements in reproducible format.

  18. Use the pattern:
  19. Observation: what is known.
  20. Uncertainty: what is not constrained.
  21. Impact: what interpretation may change.
  22. Check/next evidence: what would reduce uncertainty.
  23. Keep one uncertainty per paragraph or bullet; avoid merged caveats with mixed causes.

  24. Add limitations with boundary conditions.

  25. For each major conclusion, add at least one "valid under..." boundary:
  26. method regime,
  27. composition/system class,
  28. thermodynamic window,
  29. sampling/time horizon.
  30. State non-applicability where needed ("do not transfer to X without validation").

  31. Encode confidence and residual risk.

  32. Align wording with frontmatter confidence:
  33. high: anchored claims with narrow residual uncertainty.
  34. med: evidence present but transferability or parameter dependence remains.
  35. low: thin or indirect evidence; interpretation primarily provisional.
  36. Include a short residual-risk line for decision-critical claims.

  37. Add explicit corpus-honesty note when evidence is incomplete.

  38. Required when page relies on SI-only, galley/proof, or extract-thin sources.
  39. Distinguish:
  40. "Not in this corpus/source" from
  41. "Unknown to science."

Validation checks and acceptance criteria

Quality gates (all required): - Coverage gate: every major claim has at least one associated uncertainty or explicit "no material uncertainty identified under current scope" statement. - Anchor gate: every uncertainty/limitation statement has a traceable evidence anchor or explicit absence-of-evidence declaration. - Scope gate: each limitation specifies where it applies (system, regime, method, or data boundary). - Impact gate: each uncertainty states why it matters for interpretation, comparison, or protocol choice. - Actionability gate: each high-impact uncertainty includes a concrete follow-up check.

Reproducibility checklist: - Another curator can reconstruct why each limitation exists from cited anchors. - A retrieval agent can parse uncertainty type without guessing intent. - No limitation sentence depends on implicit background knowledge.

Reject and revise if any of the following occur: - Generic caveats with no evidence path ("results may vary" without conditions). - Mixed certainty language ("proves" and "possibly" in the same claim scope). - Overreach beyond cited pages or source quality. - Limitations that repeat Methods details without interpretation consequences.

Failure modes and mitigations

Failure mode: boilerplate uncertainty text reused across unrelated pages. - Mitigation: require claim-specific anchors and impact statements.

Failure mode: uncertainty phrased as disclaimer only, not analysis. - Mitigation: enforce Observation -> Uncertainty -> Impact -> Check pattern.

Failure mode: synthesis page hides cross-paper disagreement. - Mitigation: add explicit disagreement bullets and link to a debate page where relevant.

Failure mode: paper page claims transferability not shown in source. - Mitigation: add "transferability untested beyond reported domain" unless direct evidence exists.

Failure mode: confidence level inflated by polished prose. - Mitigation: calibrate confidence from anchor density, source quality, and unresolved contradictions.

Variants and when to choose them

Variant A: Paper-page uncertainty block (single publication). - Use when curating type: paper. - Prioritize protocol transparency, parameter omissions, and in-study validation limits.

Variant B: Synthesis-page uncertainty block (cross-paper). - Use for concept/material/forcefield/debate pages. - Prioritize comparability constraints, heterogeneity of methods, and corpus coverage gaps.

Variant C: Entry-point page uncertainty block (decision support). - Use for query-first pages that route readers. - Prioritize decision risk, common misreads, and minimum verification before method selection.

Expected outputs after applying this protocol: - Updated uncertainty/limitation prose in the target page, evidence-anchored and scope-bounded. - Frontmatter confidence aligned with uncertainty burden. - Improved retrieval reliability for MAS workflows that rank by confidence and evidence density.

Recommended downstream maintenance: - If uncertainty themes recur across pages, open or update a relevant debate page under wiki/debates/. - If recurring method caveats are operational, link to or extend a companion protocol page. - Refresh section-level chunks after substantive edits using: - python3 scripts/build_chunks.py

Evidence anchors

Anchor hierarchy for this protocol: 1. Version-of-record paper PDF and SI (if locally available). 2. Curated paper page Methods/Findings sections tied to that publication. 3. source_refs on synthesis pages with explicit paper_id locators. 4. Extract-only fallback with explicit quality caveat.

Evidence anchor formatting guidance: - Prefer compact citations in prose that include paper_id and locator context. - Keep locator language stable enough that another curator can find it. - Never use an uncertainty claim without an anchor path or explicit "not reported" statement.

Minimum anchor density targets: - Paper pages: at least one anchor for each major limitation category. - Synthesis pages: at least one anchor per cross-paper uncertainty cluster.

MAS / retrieval

Stable id: methodprotocol:uncertainty-and-limitation-reporting. Query synonyms: "uncertainty section", "limitations section", "evidence-qualified claims", "confidence calibration". Refresh when corpus source quality changes (new PDFs/SI, corrected extractions, or updated paper-level Methods/Findings curation).