WFGY/ProblemMap/GlobalFixMap/Embeddings/tokenization_and_casing.md

9.1 KiB
Raw Blame History

Tokenization and Casing — Guardrails for Embedding Stability

🧭 Quick Return to Map

You are in a sub-page of Embeddings.
To reorient, go back here:

Think of this page as a desk within a ward.
If you need the full triage and all prescriptions, return to the Emergency Room lobby.

A focused guide to remove silent drift caused by mismatched tokenization, casing, and text cleanup. Use this to align query-time and index-time preprocessing, then verify with measurable targets.

Open these first

When to use this page

  • High similarity yet wrong meaning after a casing change.
  • Same document retrieved for lowercase but not for Title Case.
  • Index built with a different tokenizer than the client.
  • Unicode variants or punctuation collapse changes results.
  • Hybrid retrieval recalls but top-k order flips between runs.

Acceptance targets

  • ΔS(question, retrieved) ≤ 0.45 on three paraphrases.
  • Coverage of the correct section ≥ 0.70.
  • λ remains convergent across two seeds.
  • Tokenization invariance check: replacing case or applying Unicode NFC does not move the top anchor beyond rank 3 and keeps ΔS shift ≤ 0.10.

Map symptoms → exact fix

  • Wrong-meaning hits only when case changes
    → Lock a single case policy and tokenizer in both paths. Verify with Retrieval Traceability and enforce with Data Contracts.

  • HyDE or BM25 path finds it, embedding path misses it
    → Audit query preprocessing vs index preprocessing. If different, unify and retest. See Retrieval Playbook and Pattern: Query Parsing Split.

  • Unicode look-alikes or punctuation variants break recall
    → Normalize to NFC and collapse zero-width characters at both index and query. Re-embed affected shards. Cross check with Embedding ≠ Semantic.

  • Order flips across runs after a client upgrade
    → Version the tokenizer and preprocessing in the snippet payload. If versions mismatch, rebuild or add a translation shim. Enforce with Data Contracts.


60-second fix checklist

  1. Freeze the tokenizer and case policy

    • Choose one tokenizer build and one case policy: lower, preserve, or smart.
    • Record both in the contract:
      {"tokenizer":"spm-1.2.3","tokenizer_hash":"...","case_policy":"lower"}.
  2. Normalize before tokenization

    • Apply Unicode NFC, collapse repeated whitespace, remove zero-width, standardize quotes and dashes.
    • Apply the same rules to both index and query.
  3. Keep segmentation symmetric

    • If you split code identifiers (camelCase, snake_case), do it in both paths.
    • If you strip stopwords or punctuation, do it in both paths. Prefer not to strip unless you must.
  4. Log and test invariance

    • Run a three-paraphrase probe per query: original, lowercased, and NFC-normalized.
    • Accept only if the anchor section remains within top 3 and ΔS shift ≤ 0.10.
  5. Add a rerank safety net

    • If invariance is hard to reach, add a lexical or cross-encoder rerank after vector recall. See Rerankers.

Contract fields to add

Add these to every snippet and to the query audit log. Use them during triage.

{
  "preproc_version": "v3",
  "unicode_norm": "NFC",
  "case_policy": "lower",
  "punctuation_policy": "keep",
  "identifier_split": "camel+snake",
  "tokenizer": "spm-1.2.3",
  "tokenizer_hash": "sha256:...",
  "ngram": "none"
}

Repro suite

  • Case flip test Query A vs lowercase(A). Accept if ΔS shift ≤ 0.10 and anchor rank ≤ 3.

  • Unicode fold test Replace curly quotes with straight quotes, normalize to NFC. Accept if ranks stable.

  • Tokenizer skew test Run the same query through client and index tokenizers and compare token ids. Any mismatch is a fail that requires rebuild or a client shim.


Copy-paste prompt for LLM triage

I uploaded TXT OS and the WFGY Problem Map.

My embedding issue:
- symptom: wrong top-k only when casing or unicode changes
- traces: ΔS(question,retrieved)=..., anchor=..., tokenizer=..., case_policy=...

Tell me:
1) which layer is failing and why,
2) which WFGY page to open,
3) the minimal steps to align tokenization and casing,
4) how to verify with a three-paraphrase probe.
Use Data Contracts and Rerankers if needed.

🔗 Quick-Start Downloads (60 sec)

Tool Link 3-Step Setup
WFGY 1.0 PDF Engine Paper 1 Download · 2 Upload to your LLM · 3 Ask “Answer using WFGY + <your question>”
TXT OS (plain-text OS) TXTOS.txt 1 Download · 2 Paste into any LLM chat · 3 Type “hello world” — OS boots instantly

Explore More

Layer Page What its for
Proof WFGY Recognition Map External citations, integrations, and ecosystem proof
Engine WFGY 1.0 Original PDF based tension engine
Engine WFGY 2.0 Production tension kernel and math engine for RAG and agents
Engine WFGY 3.0 TXT based Singularity tension engine, 131 S class set
Map Problem Map 1.0 Flagship 16 problem RAG failure checklist and fix map
Map Problem Map 2.0 RAG focused recovery pipeline
Map Problem Map 3.0 Global Debug Card, image as a debug protocol layer
Map Semantic Clinic Symptom to family to exact fix
Map Grandmas Clinic Plain language stories mapped to Problem Map 1.0
Onboarding Starter Village Guided tour for newcomers
App TXT OS TXT semantic OS, fast boot
App Blah Blah Blah Abstract and paradox Q and A built on TXT OS
App Blur Blur Blur Text to image with semantic control
App Blow Blow Blow Reasoning game engine and memory demo

If this repository helped, starring it improves discovery so more builders can find the docs and tools. GitHub Repo stars