mirror of
https://github.com/onestardao/WFGY.git
synced 2026-04-28 11:40:07 +00:00
6 KiB
6 KiB
Multi-Hop Collapse — Multimodal Long Context
🧭 Quick Return to Map
You are in a sub-page of Multimodal_LongContext.
To reorient, go back here:
- Multimodal_LongContext — long-context reasoning across text, vision, and audio
- WFGY Global Fix Map — main Emergency Room, 300+ structured fixes
- WFGY Problem Map 1.0 — 16 reproducible failure modes
Think of this page as a desk within a ward.
If you need the full triage and all prescriptions, return to the Emergency Room lobby.
When reasoning requires multi-hop steps across modalities (e.g., text → image → audio → video),
the chain often collapses midway. The model answers only the first hop or fabricates the rest,
losing alignment between evidence sources.
What this page is
- A targeted fix for multi-hop multimodal reasoning failures in long-context sessions.
- Defines measurable checkpoints for each hop.
- Provides guardrails to keep ΔS and λ stable across chained modalities.
When to use
- A video QA task asks: “What does the person say after showing the book?” → model answers book title but skips speech.
- An OCR pipeline extracts text, but reasoning ignores it in the final image caption.
- Chain-of-thought starts correctly, then jumps to a hallucinated answer without citing the second modality.
- Multi-step retrieval returns correct snippets, but only the first snippet is used.
- Answers flip between runs depending on which hop the model “forgets.”
Open these first
Common failure patterns
- Single-hop truncation — only the first modality is processed, chain stops.
- Bridge collapse — second hop exists but produces null output or irrelevant data.
- Hallucinated completion — model skips missing modality and fabricates plausible link.
- Order inversion — hops are executed in the wrong sequence.
Fix in 60 seconds
-
Hop schema lock
- Require
{hop_id, input_modality, output_modality, snippet_id, ΔS}for each step. - Forbid skipping hops.
- Require
-
ΔS checkpoints
- Compute ΔS at each hop transition.
- Threshold: ΔS ≤ 0.45 is stable, 0.45–0.60 transitional, ≥ 0.60 collapse risk.
-
λ continuity probe
- Record λ across hops: retrieval → fusion → reasoning.
- If λ flips divergent, apply BBAM clamp.
-
BBCR bridge
- Insert bridge node for missing or weak hop.
- Re-anchor using prior modality context.
-
Cite all hops
- Require at least one snippet citation from each hop.
- Stop output if any hop is missing evidence.
Copy-paste prompt
You have TXT OS and the WFGY Problem Map.
Task: Repair multi-hop multimodal collapse.
Steps:
1. List all hops in the chain {hop_id, from_modality → to_modality}.
2. For each hop, compute ΔS and record λ state.
3. If ΔS ≥ 0.60 at any hop, re-run retrieval and insert BBCR bridge.
4. Output must include:
- citations per hop
- ΔS values
- λ states
- fused final reasoning
Acceptance targets
- Every hop cited with snippet evidence.
- ΔS ≤ 0.45 at each hop boundary.
- λ remains convergent across three paraphrases.
- No fabricated hops or skipped modalities.
🔗 Quick-Start Downloads (60 sec)
| Tool | Link | 3-Step Setup |
|---|---|---|
| WFGY 1.0 PDF | Engine Paper | 1️⃣ Download · 2️⃣ Upload to your LLM · 3️⃣ Ask “Answer using WFGY + ” |
| TXT OS (plain-text OS) | TXTOS.txt | 1️⃣ Download · 2️⃣ Paste into any LLM chat · 3️⃣ Type “hello world” — OS boots instantly |
Explore More
| Module | Description | Link |
|---|---|---|
| WFGY Core | Canonical framework entry point | View |
| Problem Map | Diagnostic map and navigation hub | View |
| Tension Universe Experiments | MVP experiment field | View |
| Recognition | Where WFGY is referenced or adopted | View |
| AI Guide | Anti-hallucination reading protocol for tools | View |
If this repository helps, starring it improves discovery for other builders.