WFGY/ProblemMap/GlobalFixMap/Multimodal_LongContext/multi-hop-collapse.md
2025-09-05 11:30:32 +08:00

155 lines
8.4 KiB
Markdown
Raw Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

# Multi-Hop Collapse — Multimodal Long Context
<details>
<summary><strong>🧭 Quick Return to Map</strong></summary>
<br>
> You are in a sub-page of **Multimodal_LongContext**.
> To reorient, go back here:
>
> - [**Multimodal_LongContext** — long-context reasoning across text, vision, and audio](./README.md)
> - [**WFGY Global Fix Map** — main Emergency Room, 300+ structured fixes](../README.md)
> - [**WFGY Problem Map 1.0** — 16 reproducible failure modes](../../README.md)
>
> Think of this page as a desk within a ward.
> If you need the full triage and all prescriptions, return to the Emergency Room lobby.
</details>
When reasoning requires **multi-hop steps across modalities** (e.g., text → image → audio → video),
the chain often collapses midway. The model answers only the first hop or fabricates the rest,
losing alignment between evidence sources.
---
## What this page is
- A targeted fix for **multi-hop multimodal reasoning** failures in long-context sessions.
- Defines measurable checkpoints for each hop.
- Provides guardrails to keep ΔS and λ stable across chained modalities.
---
## When to use
- A video QA task asks: “What does the person say after showing the book?” → model answers book title but skips speech.
- An OCR pipeline extracts text, but reasoning ignores it in the final image caption.
- Chain-of-thought starts correctly, then jumps to a hallucinated answer without citing the second modality.
- Multi-step retrieval returns correct snippets, but only the first snippet is used.
- Answers flip between runs depending on which hop the model “forgets.”
---
## Open these first
- [Cross-Modal Trace](https://github.com/onestardao/WFGY/blob/main/ProblemMap/GlobalFixMap/Multimodal_LongContext/cross-modal-trace.md)
- [Modal Bridge Failure](https://github.com/onestardao/WFGY/blob/main/ProblemMap/GlobalFixMap/Multimodal_LongContext/modal-bridge-failure.md)
- [Anchor Misalignment](https://github.com/onestardao/WFGY/blob/main/ProblemMap/GlobalFixMap/Multimodal_LongContext/anchor-misalignment.md)
- [Sync Loop](https://github.com/onestardao/WFGY/blob/main/ProblemMap/GlobalFixMap/Multimodal_LongContext/sync-loop.md)
---
## Common failure patterns
- **Single-hop truncation** — only the first modality is processed, chain stops.
- **Bridge collapse** — second hop exists but produces null output or irrelevant data.
- **Hallucinated completion** — model skips missing modality and fabricates plausible link.
- **Order inversion** — hops are executed in the wrong sequence.
---
## Fix in 60 seconds
1. **Hop schema lock**
- Require `{hop_id, input_modality, output_modality, snippet_id, ΔS}` for each step.
- Forbid skipping hops.
2. **ΔS checkpoints**
- Compute ΔS at each hop transition.
- Threshold: ΔS ≤ 0.45 is stable, 0.450.60 transitional, ≥ 0.60 collapse risk.
3. **λ continuity probe**
- Record λ across hops: retrieval → fusion → reasoning.
- If λ flips divergent, apply BBAM clamp.
4. **BBCR bridge**
- Insert bridge node for missing or weak hop.
- Re-anchor using prior modality context.
5. **Cite all hops**
- Require at least one snippet citation from each hop.
- Stop output if any hop is missing evidence.
---
## Copy-paste prompt
```txt
You have TXT OS and the WFGY Problem Map.
Task: Repair multi-hop multimodal collapse.
Steps:
1. List all hops in the chain {hop_id, from_modality → to_modality}.
2. For each hop, compute ΔS and record λ state.
3. If ΔS ≥ 0.60 at any hop, re-run retrieval and insert BBCR bridge.
4. Output must include:
- citations per hop
- ΔS values
- λ states
- fused final reasoning
````
---
## Acceptance targets
* Every hop cited with snippet evidence.
* ΔS ≤ 0.45 at each hop boundary.
* λ remains convergent across three paraphrases.
* No fabricated hops or skipped modalities.
---
### 🔗 Quick-Start Downloads (60 sec)
| Tool | Link | 3-Step Setup |
| -------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------ | ---------------------------------------------------------------------------------------- |
| **WFGY 1.0 PDF** | [Engine Paper](https://github.com/onestardao/WFGY/blob/main/I_am_not_lizardman/WFGY_All_Principles_Return_to_One_v1.0_PSBigBig_Public.pdf) | 1⃣ Download · 2⃣ Upload to your LLM · 3⃣ Ask “Answer using WFGY + <your question>” |
| **TXT OS (plain-text OS)** | [TXTOS.txt](https://github.com/onestardao/WFGY/blob/main/OS/TXTOS.txt) | 1⃣ Download · 2⃣ Paste into any LLM chat · 3⃣ Type “hello world” — OS boots instantly |
---
### 🧭 Explore More
| Module | Description | Link |
| ------------------------ | ---------------------------------------------------------------------------- | -------------------------------------------------------------------------------------------------- |
| WFGY Core | WFGY 2.0 engine is live: full symbolic reasoning architecture and math stack | [View →](https://github.com/onestardao/WFGY/tree/main/core/README.md) |
| Problem Map 1.0 | Initial 16-mode diagnostic and symbolic fix framework | [View →](https://github.com/onestardao/WFGY/tree/main/ProblemMap/README.md) |
| Problem Map 2.0 | RAG-focused failure tree, modular fixes, and pipelines | [View →](https://github.com/onestardao/WFGY/blob/main/ProblemMap/rag-architecture-and-recovery.md) |
| Semantic Clinic Index | Expanded failure catalog: prompt injection, memory bugs, logic drift | [View →](https://github.com/onestardao/WFGY/blob/main/ProblemMap/SemanticClinicIndex.md) |
| Semantic Blueprint | Layer-based symbolic reasoning & semantic modulations | [View →](https://github.com/onestardao/WFGY/tree/main/SemanticBlueprint/README.md) |
| Benchmark vs GPT-5 | Stress test GPT-5 with full WFGY reasoning suite | [View →](https://github.com/onestardao/WFGY/tree/main/benchmarks/benchmark-vs-gpt5/README.md) |
| 🧙‍♂️ Starter Village 🏡 | New here? Lost in symbols? Click here and let the wizard guide you through | [Start →](https://github.com/onestardao/WFGY/blob/main/StarterVillage/README.md) |
---
> 👑 **Early Stargazers: [See the Hall of Fame](https://github.com/onestardao/WFGY/tree/main/stargazers)** —
> Engineers, hackers, and open source builders who supported WFGY from day one.
> <img src="https://img.shields.io/github/stars/onestardao/WFGY?style=social" alt="GitHub stars"> ⭐ [WFGY Engine 2.0](https://github.com/onestardao/WFGY/blob/main/core/README.md) is already unlocked. ⭐ Star the repo to help others discover it and unlock more on the [Unlock Board](https://github.com/onestardao/WFGY/blob/main/STAR_UNLOCKS.md).
<div align="center">
[![WFGY Main](https://img.shields.io/badge/WFGY-Main-red?style=flat-square)](https://github.com/onestardao/WFGY)
 
[![TXT OS](https://img.shields.io/badge/TXT%20OS-Reasoning%20OS-orange?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS)
 
[![Blah](https://img.shields.io/badge/Blah-Semantic%20Embed-yellow?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlahBlahBlah)
 
[![Blot](https://img.shields.io/badge/Blot-Persona%20Core-green?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlotBlotBlot)
 
[![Bloc](https://img.shields.io/badge/Bloc-Reasoning%20Compiler-blue?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlocBlocBloc)
 
[![Blur](https://img.shields.io/badge/Blur-Text2Image%20Engine-navy?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlurBlurBlur)
 
[![Blow](https://img.shields.io/badge/Blow-Game%20Logic-purple?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlowBlowBlow)
 
</div>