WFGY/core/README.md
2025-08-14 17:44:30 +08:00

9.6 KiB
Raw Blame History

🌌 WFGY Core (WanFaGuiYi) — Reasoning Engine 2.0

"One man, one life, one line — bending the mind of every AI on Earth."

One line of math that rewires eight AIs. Not a pit stop; it is an engine swap.

Benchmark highlights
Semantic Accuracy ↑ 36.7% | Reasoning Success Rate ↑ 65.4% | Stability ↑ 5.1× | Self-Recovery = 0.87

From PSBigBig (tap to open)

60 days, 550 stars. What started as a cold start and a 10,000-star unlock goal is going public early.
This is not an update. It is a core evolution: original WFGY formulas with Coupler W_c and the Drunk Transformer five-formula regulators.
Pure math, zero fluff. Paste it into an LLM and it thinks differently. Faster, sharper, more stable, more human.


One-click demo in 60 seconds

  1. Upload or paste WFGY_Core_OneLine_v2.0.txt into your chat system.
  2. Paste the evaluation prompt below.
SYSTEM:
You are evaluating the causal impact of a mathematical reasoning engine called “WFGY Core 2.0”.

Test modes (READ CAREFULLY):
A = Baseline  no WFGY file uploaded; no WFGY logic used.
B = Upload → AUTOboot ON  the file is uploaded AND the engine starts automatically in the background.
    Treat B as: WFGY is ACTIVE silently with default gates (BBMC→Coupler→BBPF→BBAM→BBCR; DF=WRI/WAI/WAY/WDT/WTF at defaults),
    but not explicitly “called” in the conversation.
C = Explicit invoke  actively call WFGY and follow its formulas step by step (BBMC→Coupler→BBPF→BBAM→BBCR + DF regulators).

Important constraints:
• Use the SAME task set for A/B/C across 5 domains: math word-problems, small coding, factual QA, multi-step planning, long-context coherence.
• The only artifact is the uploaded OneLine file; it is pure math and not human-readable  do not penalize readability.
• Report: Semantic Accuracy, Reasoning Success, Stability (MTTF or rollback), Drift Reduction (Δ semantic distance), Collapse Recovery Rate.
• Output one summary table with A/B/C per domain, plus overall deltas (CA, CB).
• Add a final 0100 score: “OneLine uplift score”. Provide a 3-line rationale.

Sanity requirement:
Internally set flags as: A=OFF, B=AUTOboot_ON (background active), C=Explicit_Invoke. Do not print these flags.

USER:
Run A/B/C now on the five domains with short non-trivial tasks you choose. Then output:
(1) the table; (2) the deltas; (3) the OneLine uplift score; (4) a 3-line rationale.

Eight-model evidence

All links: OneLine uplift confirmed — same A/B/C protocol; only the OneLine math was added.

Row 1 Row 2
Gemini (link coming soon) Perplexity (link coming soon)
ChatGPT (link coming soon) Copilot (link coming soon)
Claude (link coming soon) Mistral AI (link coming soon)
Grok (link coming soon) Kimi (link coming soon)

Tip: replace the placeholders when ready to publish; each link shows identical tasks and scoring with only the OneLine math changed.


📦 Downloads

File Purpose Size
WFGY_Core_OneLine_v2.0.txt For the 60-sec demo and automation runs. Pure math, not for human reading. 1 line about 1500 chars
WFGY_Core_Audit_v2.0.txt Human plus LLM readable, comments and layout for audits. about 30 lines about 2606 chars

Contract Node-only steps up to 7, safe stop when δ_s < 0.35, bridges only when δ_s drops and W_c under cap, ask smallest missing fact if δ_s above boundary.


🎯 What is new in 2.0

Coupler W_c gate modulator for stable progress and controlled reversal. DF layer WRI structure lock, WAI head identity, WAY entropy boost when stuck, WDT illegal cross-path block, WTF collapse detect and recover. Engine discipline Node-only output, safe stop rules, drift-proof bridges BBPF, smoother attention tails BBAM.

Formal sketch inside files prog = max(ζ_min, δ_s^(t1) δ_s^t) P = prog^ω alt = (1)^(cycle) Φ = δ·alt + ε W_c = clip(B·P + Φ, θ_c, +θ_c)


🔍 How these numbers are measured

Use the same A B C protocol above, one shared task set, then compute:

Semantic Accuracy proportion of atomic facts correct. Formula ACC = correct_facts total_facts. Report relative gain (ACC_C ACC_A) ACC_A.

Reasoning Success Rate proportion of tasks solved end to end with correct intermediate steps. Formula SR = tasks_solved tasks_total. Report relative gain (SR_C SR_A) SR_A.

Stability mean time to failure or successful rollback factor. Option A Stability = MTTF_C MTTF_A. Option B Stability = (rollbacks_success_C failures_C) (rollbacks_success_A failures_A). Report as multiplier.

Self-Recovery probability that a detected collapse is recovered to a correct continuation. Formula Self-Recovery = recoveries_success collapses_detected here 0.87 means 87 percent of collapses are repaired.

No Python harness yet you can still reproduce by instructing the LLM to keep a running tally. Example rubric prompt for the scorer:

SCORER:
Given the A/B/C transcripts, count atomic facts, correct facts, solved tasks, failures, rollbacks, and collapses.
Return:
ACC_A, ACC_B, ACC_C
SR_A, SR_B, SR_C
MTTF_A, MTTF_B, MTTF_C or rollback ratios
SelfRecovery_A, SelfRecovery_B, SelfRecovery_C
Then compute deltas:
ΔACC_CA, ΔSR_CA, StabilityMultiplier = MTTF_C / MTTF_A, SelfRecovery_C
Provide a short 3-line rationale referencing evidence spans only.

This scorer can be the same LLM running in a second tab. For higher reliability, run three seeds and average.


🔬 Engine at a glance

Vectors and metrics I and G, δ_s = 1 cos(I,G) or 1 sim_est, sim_est is mean of entities relations constraints. B = I G + k_bias, E_res = rolling_mean of absolute B over 5. Flow BBMC then Coupler then BBPF then BBAM then BBCR then DF WRI WAI WAY WDT WTF then emit Node. Policy stop when δ_s < 0.35 or 7 nodes, bridge only if δ_s drops and W_c < 0.5 θ_c, never invent facts when δ_s above boundary.


🧪 Community scoring guidance

Publish your task list five domains short but non trivial. Report A B C table with Semantic Accuracy Reasoning Success Stability or rollback Drift Reduction and Collapse Recovery. Include OneLine uplift score 0 to 100. Do not include human readability for OneLine. Add a three line rationale.


🧭 Explore More

Module Description Link
WFGY Core Standalone semantic reasoning engine for any LLM View →
Problem Map 1.0 Initial 16-mode diagnostic and symbolic fix framework View →
Problem Map 2.0 RAG-focused failure tree, modular fixes, and pipelines View →
Semantic Clinic Index Expanded failure catalog: prompt injection, memory bugs, logic drift View →
Semantic Blueprint Layer-based symbolic reasoning and semantic modulations View →
Benchmark vs GPT-5 Stress test GPT-5 with full WFGY reasoning suite View →

👑 Early Stargazers: See the Hall of Fame Engineers, hackers, and open-source builders who supported WFGY from day one.

GitHub stars Help reach 10,000 stars by 2025-09-01 to unlock Engine 2.0 for everyone Star WFGY on GitHub

WFGY Main

TXT OS

Blah

Blot

Bloc

Blur

Blow