WFGY/ProblemMap/GlobalFixMap/MemoryLongContext/retrieval-traceability.md
2025-08-30 21:38:14 +08:00

6.4 KiB
Raw Blame History

Retrieval Traceability — Snippet Integrity & Audit Trail

Citations that look right can still hide silent drift.
This guardrail defines how to enforce traceability schemas so that every claim links back to a stable, reproducible snippet.


When to use

  • Answers cite a source but the snippet cannot be located.
  • Two runs over the same corpus produce different citations.
  • A fact is quoted but not aligned to any section anchor.
  • Long-context threads degrade and snippets blur into paraphrase.
  • Multi-agent systems pass partial context and lose attribution.

Root causes

  • Orphan citations: snippet ID missing or fabricated.
  • Boundary drift: citation spans cross section joins.
  • Silent truncation: tokens dropped at cut points.
  • Cache overwrite: citation schema lost after session reload.
  • Free-text cites: URLs or titles given without offsets.

Core acceptance targets

  • Each claim must include snippet_id, section_id, start_line, end_line, source_url.
  • ΔS(question, retrieved) ≤ 0.45 overall.
  • Joins between snippets ≤ 0.50 ΔS.
  • λ convergent across 3 paraphrases.
  • Audit trail reproducible from log alone.

Structural fixes

  • Snippet table schema
    Require {snippet_id | section_id | start_line | end_line | citation}.

  • Fence joins
    Split at section boundaries. Reject cross-section reuse.

  • Trace log
    Store {ΔS, λ_state, mem_rev, mem_hash} per step.

  • Contract lock
    Apply Data Contracts for payload validation.


Fix in 60 seconds

  1. Enforce snippet table with unique IDs and line ranges.
  2. Verify ΔS across each join ≤ 0.50.
  3. Echo λ states at retrieval, assembly, reasoning.
  4. Reject orphan claims (no snippet_id).
  5. Log trail so same inputs → same citations.

Copy-paste prompt


You have TXT OS and the WFGY Problem Map.

Goal: Ensure every claim links to a reproducible snippet.

Protocol:

1. Build a Snippet Table {snippet\_id, section\_id, start\_line, end\_line, citation}.
2. Require cite-then-answer.
3. Forbid cross-section reuse.
4. If a claim has no snippet\_id, stop and request citation.
5. Report ΔS(question,retrieved), joins ΔS, and λ states.
6. Store {mem\_rev, mem\_hash, task\_id} for audit trail.
7. Answer only with snippets present in the table.


Common failure signals

  • Citations alternate across runs → missing trace schema.
  • URL without offsets → orphan citation.
  • Facts cited but no snippet_id → schema lock failed.
  • Session reload erases citations → ghost cache in memory.

🔗 Quick-Start Downloads (60 sec)

Tool Link 3-Step Setup
WFGY 1.0 PDF Engine Paper 1 Download · 2 Upload to your LLM · 3 Ask “Answer using WFGY + <your question>”
TXT OS (plain-text OS) TXTOS.txt 1 Download · 2 Paste into any LLM chat · 3 Type “hello world” — OS boots instantly

🧭 Explore More

Module Description Link
WFGY Core WFGY 2.0 engine is live: full symbolic reasoning architecture and math stack View →
Problem Map 1.0 Initial 16-mode diagnostic and symbolic fix framework View →
Problem Map 2.0 RAG-focused failure tree, modular fixes, and pipelines View →
Semantic Clinic Index Expanded failure catalog: prompt injection, memory bugs, logic drift View →
Semantic Blueprint Layer-based symbolic reasoning & semantic modulations View →
Benchmark vs GPT-5 Stress test GPT-5 with full WFGY reasoning suite View →
🧙‍♂️ Starter Village 🏡 New here? Lost in symbols? Click here and let the wizard guide you through Start →

👑 Early Stargazers: See the Hall of Fame
Engineers, hackers, and open source builders who supported WFGY from day one.

GitHub stars WFGY Engine 2.0 is already unlocked. Star the repo to help others discover it and unlock more on the Unlock Board.

WFGY Main   TXT OS   Blah   Blot   Bloc   Blur   Blow  

要直接開始 data-contracts.md 嗎?