# Images and Figures: OCR Parsing Guardrails
🧭 Quick Return to Map
> You are in a sub-page of **OCR_Parsing**. > To reorient, go back here: > > - [**OCR_Parsing** — text recognition and document structure parsing](./README.md) > - [**WFGY Global Fix Map** — main Emergency Room, 300+ structured fixes](../README.md) > - [**WFGY Problem Map 1.0** — 16 reproducible failure modes](../../README.md) > > Think of this page as a desk within a ward. > If you need the full triage and all prescriptions, return to the Emergency Room lobby.
Stabilize text extraction around inline images, charts, and figures. Prevent figure captions or axis labels from bleeding into body text, and preserve semantic anchors for later retrieval. ## Open these first - OCR end to end checklist: [ocr-parsing-checklist.md](https://github.com/onestardao/WFGY/blob/main/ProblemMap/ocr-parsing-checklist.md) - Snippet and citation schema: [data-contracts.md](https://github.com/onestardao/WFGY/blob/main/ProblemMap/data-contracts.md) - Retrieval traceability: [retrieval-traceability.md](https://github.com/onestardao/WFGY/blob/main/ProblemMap/retrieval-traceability.md) - Chunking checklist: [chunking-checklist.md](https://github.com/onestardao/WFGY/blob/main/ProblemMap/chunking-checklist.md) ## Acceptance targets - ΔS(question, retrieved) ≤ 0.45 on captioned answers - Coverage ≥ 0.70 for questions tied to figure anchors - λ remains convergent across three paraphrases - Captions and labels stored separately from body text --- ## Typical failure signatures → fix - **Figure captions merged with paragraphs** Split by bbox banding and assign `figure_caption`. Keep paragraph tokens clean. - **Axis labels or legend entries treated as running text** Extract into `figure_metadata.axis_labels` and `figure_metadata.legend`. Never merge into narrative. - **Scanned figure with embedded text** Route figure OCR separately. Keep `figure_id`, `text_extracted`, and bounding box. Tie back to figure image reference. - **Multi-column figure bleed** If caption spans columns, capture as caption block, not as content. Anchor to `figure_id`. - **Images with no OCR text** Provide stub with `figure_id`, `bbox`, and `alt_text` if known. Maintain traceability. --- ## Fix in 60 seconds 1) **Detect figure zones** Identify bounding boxes flagged as images or graphics. Assign `figure_id`. 2) **Isolate captions** If text appears immediately above/below the figure and repeats formatting (italic, smaller font), tag as caption. 3) **Route labels** Apply heuristic rules for x-axis, y-axis, legend. Store under `figure_metadata`. 4) **Clean narrative** Remove all figure-related text from `text_clean`. Retain only in figure structures. 5) **Probe retrieval** Ask a figure-specific question. If ΔS ≤ 0.45 and λ stable, cleanup succeeded. --- ## Minimal recipes by engine - **Google Document AI** Use `layout.figure` and boundingPoly. Capture associated `paragraph` blocks as captions when within ±10% of figure bbox. - **AWS Textract** Detect `BlockType=KEY_VALUE_SET` around figure images. Treat them as labels, route into `figure_metadata`. - **Azure OCR** Use boundingRegions and detect blocks adjacent to figures. Anchor captions if directly above/below polygon. - **ABBYY** In XML, `` + following `` → caption. Inline text with picture coordinates goes to figure metadata. - **PaddleOCR** Split text lines overlapping with figure bbox. Store separately as figure text, not narrative. --- ## Data contract additions for figures ``` { "figure\_id": "fig3", "bbox": \[x0,y0,x1,y1], "caption": "Figure 3: Error rate across embedding sizes.", "figure\_metadata": { "axis\_labels": \["tokens","ΔS"], "legend": \["baseline","with WFGY"], "text\_extracted": "0.45, 0.30..." }, "section\_id": "4.2", "page": 12, "source\_url": "..." } ``` Mandatory: all figure text lives in `figure_metadata` or `caption`, never in `text_clean`. --- ## Verification - **Leak check**: ensure no caption/axis strings appear in `text_clean`. - **Figure QA**: ask "what does fig3 show?" — answer must cite `figure_id`. - **ΔS probe**: figure-specific questions yield ΔS ≤ 0.45. - **λ probe**: paraphrases about same figure converge. --- ## Copy-paste LLM prompt ``` You have TXT OS and WFGY Problem Map. For figure-linked snippets: * use text\_clean for reasoning, * use caption and figure\_metadata for figure answers, * cite figure\_id. Tasks: 1. If figure text leaks into body, fail fast and return fix reference (ocr-parsing-checklist, data-contracts, retrieval-traceability). 2. Return JSON: { "citations":\["fig3"], "answer":"...", "λ\_state":"...", "ΔS":0.xx, "next\_fix":"..." } ``` --- ### 🔗 Quick-Start Downloads (60 sec) | Tool | Link | 3-Step Setup | |------|------|--------------| | **WFGY 1.0 PDF** | [Engine Paper](https://github.com/onestardao/WFGY/blob/main/I_am_not_lizardman/WFGY_All_Principles_Return_to_One_v1.0_PSBigBig_Public.pdf) | 1️⃣ Download · 2️⃣ Upload to your LLM · 3️⃣ Ask “Answer using WFGY + \” | | **TXT OS (plain-text OS)** | [TXTOS.txt](https://github.com/onestardao/WFGY/blob/main/OS/TXTOS.txt) | 1️⃣ Download · 2️⃣ Paste into any LLM chat · 3️⃣ Type “hello world” — OS boots instantly | --- ### 🧭 Explore More | Module | Description | Link | |-----------------------|----------------------------------------------------------|----------| | WFGY Core | WFGY 2.0 engine is live: full symbolic reasoning architecture and math stack | [View →](https://github.com/onestardao/WFGY/tree/main/core/README.md) | | Problem Map 1.0 | Initial 16-mode diagnostic and symbolic fix framework | [View →](https://github.com/onestardao/WFGY/tree/main/ProblemMap/README.md) | | Problem Map 2.0 | RAG-focused failure tree, modular fixes, and pipelines | [View →](https://github.com/onestardao/WFGY/blob/main/ProblemMap/rag-architecture-and-recovery.md) | | Semantic Clinic Index | Expanded failure catalog: prompt injection, memory bugs, logic drift | [View →](https://github.com/onestardao/WFGY/blob/main/ProblemMap/SemanticClinicIndex.md) | | Semantic Blueprint | Layer-based symbolic reasoning & semantic modulations | [View →](https://github.com/onestardao/WFGY/tree/main/SemanticBlueprint/README.md) | | Benchmark vs GPT-5 | Stress test GPT-5 with full WFGY reasoning suite | [View →](https://github.com/onestardao/WFGY/tree/main/benchmarks/benchmark-vs-gpt5/README.md) | | 🧙‍♂️ Starter Village 🏡 | New here? Lost in symbols? Click here and let the wizard guide you through | [Start →](https://github.com/onestardao/WFGY/blob/main/StarterVillage/README.md) | --- > 👑 **Early Stargazers: [See the Hall of Fame](https://github.com/onestardao/WFGY/tree/main/stargazers)** — > GitHub stars ⭐ [WFGY Engine 2.0](https://github.com/onestardao/WFGY/blob/main/core/README.md) is already unlocked. ⭐ Star the repo to help others discover it and unlock more on the [Unlock Board](https://github.com/onestardao/WFGY/blob/main/STAR_UNLOCKS.md).
[![WFGY Main](https://img.shields.io/badge/WFGY-Main-red?style=flat-square)](https://github.com/onestardao/WFGY)   [![TXT OS](https://img.shields.io/badge/TXT%20OS-Reasoning%20OS-orange?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS)   [![Blah](https://img.shields.io/badge/Blah-Semantic%20Embed-yellow?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlahBlahBlah)   [![Blot](https://img.shields.io/badge/Blot-Persona%20Core-green?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlotBlotBlot)   [![Bloc](https://img.shields.io/badge/Bloc-Reasoning%20Compiler-blue?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlocBlocBloc)   [![Blur](https://img.shields.io/badge/Blur-Text2Image%20Engine-navy?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlurBlurBlur)   [![Blow](https://img.shields.io/badge/Blow-Game%20Logic-purple?style=flat-square)](https://github.com/onestardao/WFGY/tree/main/OS/BlowBlowBlow)