12 KiB
Together: Guardrails and Fix Patterns
A compact field guide for stabilizing Together workflows. This page assumes you route across many models with one API. It helps you localize the failure, then jump to the exact WFGY fix page with measurable targets.
Open these first
- Visual map and recovery: RAG Architecture & Recovery
- End-to-end retrieval knobs: Retrieval Playbook
- Why this snippet (traceability schema): Retrieval Traceability
- Ordering control: Rerankers
- Embedding vs meaning: Embedding ≠ Semantic
- Hallucination and chunk boundaries: Hallucination
- Long chains and entropy: Context Drift, Entropy Collapse
- Symbolic collapse and recovery: Logic Collapse
- Prompt injection and schema locks: Prompt Injection
- Multi-agent conflicts: Multi-Agent Problems
- Bootstrap and deploy issues: Bootstrap Ordering, Deployment Deadlock, Pre-deploy Collapse
- Snippet and citation schema: Data Contracts
Core acceptance
- ΔS(question, retrieved) ≤ 0.45
- Coverage ≥ 0.70 to the target section
- λ remains convergent across three paraphrases and two seeds
Fix in 60 seconds
-
Measure ΔS
Compute ΔS(question, retrieved) and ΔS(retrieved, expected anchor).
Thresholds: stable < 0.40, transitional 0.40–0.60, risk ≥ 0.60. -
Probe with λ_observe
Vary k in retrieval (5, 10, 20). If ΔS stays flat and high, suspect metric or index mismatch.
Reorder prompt headers; if ΔS spikes, lock the schema. -
Apply the module
- Retrieval drift → BBMC plus Data Contracts.
- Reasoning collapse → BBCR bridge plus BBAM variance clamp, then verify with Logic Collapse.
- Hallucination re-entry after correction → Pattern: Hallucination Re-entry.
Typical Together breakpoints and the right fix
-
Model slug or route drift. Fallback to a different model family changes tokenizer, max tokens, or safety rules.
→ Log slugs in the trace, freeze route parameters with a contract, and re-run the same seed.
Open: Retrieval Traceability, Data Contracts, Pre-deploy Collapse -
Tool/JSON protocol not uniform across models. Free-text tool returns or partial JSON spike ΔS and flip λ.
→ Enforce strict schemas, echo the schema each step, and validate before execution.
Open: Prompt Injection, Logic Collapse -
Tokenizer/segment mismatch after model swap. Chunk boundaries no longer align with citations.
→ Rebuild snippet schema, prefer reranking, and verify anchors.
Open: Embedding ≠ Semantic, Rerankers, Retrieval Playbook -
Streaming fragment loss or interleaving in batch jobs. Out-of-order tokens or mixed seeds corrupt the trace.
→ Attachrun_id, enforce ordered sinks, idempotency keys, and per-request seed isolation.
Open: Data Contracts, patterns: memory_desync -
Hybrid retrieval underperforms a single retriever. HyDE + BM25 query split across models yields unstable top-k.
→ Lock the two-stage query and add a deterministic reranker.
Open: Pattern: Query Parsing Split, Rerankers -
Safety refusal hides the cited snippet. Different families enforce different blocks.
→ Use citation-first prompting and SCU to unlock lawful quotes.
Open: Retrieval Traceability, Pattern: SCU -
Cold boot or first call fails after deploy. Missing secrets or version skew in the router.
→ Validate order and readiness before hitting the model.
Open: Bootstrap Ordering, Deployment Deadlock
Deep diagnostics
- Three-paraphrase probe. Ask the same question three ways. Log ΔS and λ. If λ flips on harmless paraphrase, clamp with BBAM and tighten snippet schema.
- Anchor triangulation. Compare ΔS to the expected anchor and to a decoy section. If ΔS is close for both, re-chunk and re-embed.
- Route stability audit. For the same seed, assert identical slug, stop set, max tokens, and tool schema across runs. Any variance is a router bug.
Open: Context Drift, Entropy Collapse
Escalate and structural fixes
-
Index or metric mismatch. ΔS stays high across seeds. Rebuild with semantic chunking and a small gold set.
Open: Embedding ≠ Semantic, Chunking Checklist -
Live instability or oscillating answers. Add live probes and backoff guards, then gate on eval.
Open: Live Monitoring for RAG, Debug Playbook
Copy-paste prompt
You have TXTOS and the WFGY Problem Map loaded.
My Together issue:
* model\_slug: "<slug>", params: {temperature: ..., top\_p: ..., max\_tokens: ...}
* symptom: <one line>
* traces: ΔS(question,retrieved)=..., ΔS(retrieved,anchor)=..., λ across 3 paraphrases
* routing: {seed: <n>, stream: \<on/off>, batch: <size>}
Tell me:
1. failing layer and why,
2. the exact WFGY page to open next,
3. the minimal steps to push ΔS ≤ 0.45 and keep λ convergent,
4. a reproducible test to verify the fix.
Use BBMC, BBPF, BBCR, BBAM when relevant.
🔗 Quick-Start Downloads (60 sec)
| Tool | Link | 3-Step Setup |
|---|---|---|
| WFGY 1.0 PDF | Engine Paper | 1️⃣ Download · 2️⃣ Upload to your LLM · 3️⃣ Ask “Answer using WFGY + <your question>” |
| TXT OS (plain-text OS) | TXTOS.txt | 1️⃣ Download · 2️⃣ Paste into any LLM chat · 3️⃣ Type “hello world” — OS boots instantly |
🧭 Explore More
| Module | Description | Link |
|---|---|---|
| WFGY Core | WFGY 2.0 engine is live: full symbolic reasoning architecture and math stack | View → |
| Problem Map 1.0 | Initial 16-mode diagnostic and symbolic fix framework | View → |
| Problem Map 2.0 | RAG-focused failure tree, modular fixes, and pipelines | View → |
| Semantic Clinic Index | Expanded failure catalog: prompt injection, memory bugs, logic drift | View → |
| Semantic Blueprint | Layer-based symbolic reasoning & semantic modulations | View → |
| Benchmark vs GPT-5 | Stress test GPT-5 with full WFGY reasoning suite | View → |
| 🧙♂️ Starter Village 🏡 | New here? Lost in symbols? Click here and let the wizard guide you through | Start → |
👑 Early Stargazers: See the Hall of Fame —
⭐ WFGY Engine 2.0 is already unlocked. ⭐ Star the repo to help others discover it and unlock more on the Unlock Board.