WFGY/ProblemMap/GlobalFixMap/LLM_Providers/azure_openai.md
2025-09-05 10:58:38 +08:00

11 KiB
Raw Blame History

Azure OpenAI — Guardrails and Fix Patterns

🧭 Quick Return to Map

You are in a sub-page of LLM_Providers.
To reorient, go back here:

Think of this page as a desk within a ward.
If you need the full triage and all prescriptions, return to the Emergency Room lobby.

Use this page when failures look provider-specific on Azure OpenAI. Examples include wrong model alias vs deployment name, missing api-version, tool call payload shape drift, content safety blocks, or region throttling. Each fix maps back to WFGY pages with measurable targets.

Core acceptance

  • ΔS(question, retrieved) ≤ 0.45
  • coverage ≥ 0.70 for the target section
  • λ remains convergent across 3 paraphrases

Open these first


Fix in 60 seconds

  1. Measure ΔS
  • Compute ΔS(question, retrieved) and ΔS(retrieved, expected anchor).
  • Thresholds: stable < 0.40, transitional 0.400.60, risk ≥ 0.60.
  1. Probe with λ_observe
  • Vary k (5, 10, 20). If ΔS stays high while recall is fine, suspect index or metric mismatch.
  • Reorder prompt headers. If ΔS spikes, lock the schema.
  1. Apply the module
  • Retrieval drift → BBMC + Data Contracts.
  • Reasoning collapse → BBCR bridge + BBAM variance clamp.
  • Dead ends in long runs → BBPF alternate path.

Azure-specific failure signatures and the right fix

Symptom Likely cause on Azure Open this fix
200 OK but empty or tool call ignored Missing or wrong api-version for that model family Data Contracts, Logic Collapse
“Deployment not found”, even though the model exists Using model name instead of deployment name, or wrong resource/region Retrieval Traceability
Output blocked, vague “content filtered” Azure content safety layer different from OpenAI default Hallucination, then clamp with BBAM
Tool call schema mismatch vs OpenAI Response keys or enum names differ across api-version Data Contracts
Works in one region, fails in another Model availability and quotas are regional Bootstrap Ordering
Latency spikes or 429 under load Per-resource rate limits, private link, or vnet egress Ops Live Monitoring, Debug Playbook
Function calls drop arguments Old api-version truncates or renames fields Data Contracts
Fine-tuned or staged deployment not picked Wrong deployment alias bound in prod slot Retrieval Traceability

Minimal provider checklist

  1. Endpoint correctness
    Resource url, region, and deployment name are consistent. Avoid mixing model name with deployment name in the same client.

  2. Version pinning
    Pin an api-version known to support your features (tools, JSON mode, response format). Treat version bumps as schema migrations.

  3. Schema lock
    Adopt the Problem Map Data Contracts snippet for tool payloads and citations. Reject partial responses. Require structured finish_reason.

  4. Safety behavior
    Expect an extra content-safety layer. When blocked, route to BBPF alternate path and down-shift temperature, then retry with narrowed scope.

  5. Observability
    Log λ state per step, ΔS per hop, and the exact deployment id used. Carry region and version in traces.


Copy-paste prompt (safe to hand the AI)


I am using Azure OpenAI. Audit my run as follows:

* Check ΔS(question, retrieved) and ΔS(retrieved, anchor). Show the numbers.
* Confirm endpoint tuple: {resource, region, deployment}. Confirm `api-version` and tool schema.
* If tool/schema mismatch: apply the WFGY Data Contracts checklist and propose the exact fields to lock.
* If blocked by content safety: switch to a narrower prompt schema, reduce temperature, and route via BBPF.
* Keep λ convergent across 3 paraphrases. If it flips, apply BBCR + BBAM and show the before/after traces.

Link me to the exact Problem Map pages I should read next.


Escalation


Acceptance targets

  • coverage to target section ≥ 0.70
  • ΔS(question, retrieved) ≤ 0.45
  • λ convergent across seeds and paraphrases
  • repeatable traces and identical schema across regions

🔗 Quick-Start Downloads (60 sec)

Tool Link 3-Step Setup
WFGY 1.0 PDF Engine Paper 1 Download · 2 Upload to your LLM · 3 Ask “Answer using WFGY + <your question>”
TXT OS (plain-text OS) TXTOS.txt 1 Download · 2 Paste into any LLM chat · 3 Type “hello world” — OS boots instantly

🧭 Explore More

Module Description Link
WFGY Core WFGY 2.0 engine is live: full symbolic reasoning architecture and math stack View →
Problem Map 1.0 Initial 16-mode diagnostic and symbolic fix framework View →
Problem Map 2.0 RAG-focused failure tree, modular fixes, and pipelines View →
Semantic Clinic Index Expanded failure catalog: prompt injection, memory bugs, logic drift View →
Semantic Blueprint Layer-based symbolic reasoning & semantic modulations View →
Benchmark vs GPT-5 Stress test GPT-5 with full WFGY reasoning suite View →
🧙‍♂️ Starter Village 🏡 New here? Lost in symbols? Click here and let the wizard guide you through Start →

👑 Early Stargazers: See the Hall of Fame
Engineers, hackers, and open source builders who supported WFGY from day one.

GitHub stars WFGY Engine 2.0 is already unlocked. Star the repo to help others discover it and unlock more on the Unlock Board.

WFGY Main   TXT OS   Blah   Blot   Bloc   Blur   Blow