Podracing III: Cubric Lite — 0.9362 BPB#782
Conversation
Per-order adaptive alpha scaling on legal score-first 7-gram backoff. Tracks per-order beat rate on already-scored tokens, suppresses noisy low orders (2-3 → 0.3x alpha), boosts accurate high orders (5-7 → 2.0x). Results (seeds 2045/43/300): Sliding BPB (no n-gram): 1.1198 mean Cubric n-gram BPB: 0.9362 mean (0.9357/0.9362/0.9365) Artifact: 15.59 MB (int6+zstd) 0.026 BPB improvement over Podracing II (openai#753, 0.9625). Original contribution: per-order adaptive alpha scaling. Co-Authored-By: Claude Opus 4.6 (1M context) <[email protected]>
Community Review — Podracing III: Cubric Lite — 0.9362 BPBBPB: 0.9362 | Compliance: FLAG — hashed n-gram cache with target-in-key (PR #779 family pattern) What I found in the code (head SHA The n-gram lookup key at line 1101 is constructed by XOR-ing the target token into the hash: The code default is This matches the Per Issue #1017 condition 1, Cluster context: this same structural pattern has been closed on 15+ PRs under the #779 ruling as of 2026-04-11 (#779 itself, #770, #798, #808, #825, #786, #797, #909, #940, #761, #776, #788, #774, #778, #715, #758, #702 upstream). CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.05s, dim=512, layers=11, vocab=1024, code=98717 B, SMOKE_TEST_PASS Verdict: COMPLIANCE FLAG — target-in-key hashed n-gram cache, same family as PR #779. N-gram cache confirmed active in submission logs (order=7). Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: CLOSE under the same ruling as the rest of the family-bug cluster. A context-only resubmission (drop the target from the lookup key and use a full-vocabulary reweighting from a single context row, per @valerio-oai's suggested legal path on #779) would be welcomed. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.05s, dim=512, layers=11, vocab=1024, code=98717 B, SMOKE_TEST_PASS. Classification via manual code review + submission log audit (classifier initially mis-tagged as PURE_NEURAL_CLEAN because |
Summary
What Changed vs Podracing II (#753)
One eval-time addition, no training changes:
Per-order adaptive alpha scaling ("Cubric Lite"): During n-gram eval, track how often each order's n-gram probability beats the model's probability on already-scored tokens. Every 32 batches, adjust per-order alpha multipliers. Converged multipliers: **UNDEREXPLORED
Key finding: bigrams and trigrams (orders 2-3) were actively harming BPB by injecting noisy predictions at the same alpha as high-order matches. Suppressing them to 30% of base alpha and boosting orders 5-7 to 200% = 0.026 BPB gain.
Compliance
Credits
Test plan
🤖 Generated with Claude Code