Record: SP4096 + Depth Recurrence + Parallel Residuals + MuonEq-R + QK-Gain 5.0 — val_bpb 1.0897 (3-seed mean)#1296
Conversation
…0940 (3-seed mean) 4096-vocab + MLP 4x + WD 0.090 + depth recurrence (layers 4,5) + MuonEq-R + full GPTQ int6 + brotli + selective pruning. 3-seed mean: 1.0940 BPB, beating merged SOTA (PR openai#1019, 1.1147 BPB) by 0.0208 BPB.
…d mean) LZMA self-extracting code wrapper (24KB vs 81KB) frees 57KB for model precision. No pruning needed. 3-seed mean improves from 1.0940 to 1.0926.
Added parallel residuals from layer 7+ (separate attn/MLP lanes). 3-seed mean improves from 1.0926 to 1.0904.
QK-Gain from 4.0 to 5.0 plus parallel residuals and depth recurrence. 3-seed mean: 1.0897 BPB (std 0.0003), delta -0.0250 vs merged SOTA.
Port depth recurrence from PR openai#1290 and parallel residuals from PR openai#1296. - Depth recurrence: layers 3,4 repeated in forward pass via virtual layer mapping - Parallel residuals: attn+mlp computed in parallel from layer 6 onward - Configurable via RECUR_LAYERS, RECUR_START_STEP, PARALLEL_START_LAYER env vars
Ports parallel residuals from PR openai#1296 to openai#1290 base: - Block.__init__ accepts parallel flag - Block.forward() computes attn+mlp in parallel when parallel=True - GPT.__init__ passes parallel_start_layer to Block constructors - Layers 7-10 run parallel, layers 0-6 sequential (default PARALLEL_START_LAYER=7) - Both base_model and eval_model wired up
- QK_GAIN_INIT: 1.5 -> 5.0 (matches openai#1296 proven config) - WARMDOWN_ITERS: already 4000 (matches openai#1290 run command) - MULTIRES_ENABLED: 1 -> 0 (multi-res failed: only 1.13x speedup) - BIGRAM: revert to 2048x128 (3072x112 exceeded 16MB artifact limit)
Decompressed PR openai#1296 codebase (SP4096 + depth recurrence + MuonEq-R + parallel residuals + QK5 + GPTQ + brotli, 1.0897 BPB). Plan: port Helix cross-injection onto their architecture, add loop-aware GPTQ. Their depth recurrence + our helix quant shielding = novel combo. Co-Authored-By: Claude Opus 4.6 (1M context) <[email protected]>
train_gpt_base.py — untouched PR openai#1296 decompressed source train_gpt_helix.py — same + Helix crawler block, cross-injection, merge gate 4 test arms: base, helix dim=64, helix dim=192, helix without recurrence. Tests whether Helix improves the field's SOTA recursion approach. Co-Authored-By: Claude Opus 4.6 (1M context) <[email protected]>
Community Review — Record: SP4096 + Depth Recurrence + Parallel Residuals + MuonEq-R + Full GPTQ — val_bpb 1.0904 (3-seed mean)BPB: 1.0904 | Compliance: LOOKS CLEAN — score-first-per-chunk TTT (legal #1416/#1423 pattern) What I found in the code (head SHA The TTT path at line 1521 implements the score-first-per-chunk pattern: each chunk is scored under Per Issue #402 and Issue #677, TTT is legal when each token is scored before the adapter updates on it, and that's what the code does here — chunk CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 5.18s, dim=512, layers=11, vocab=4096, code=24584 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending standard checks (3-seed validation, 16MB artifact cap, 10-min wallclock on 8×H100 SXM). The compliance picture matches the legal reference frontier and no flags were raised by the classification pass. Auto-classification caveat: this review was drafted by the AST-based classifier against a template derived from manually-reviewed cluster PRs (#1420, #1450, #1487, #1541, #1529, #1533, #1518). If I've misread a subtlety in your eval path — e.g., multi-epoch TTT that I mistook for single-pass, or a target-in-key lookup I missed in a helper function — please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 5.18s, dim=512, layers=11, vocab=4096, code=24584 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
Record: SP4096 + Depth Recurrence + Parallel Residuals + MuonEq-R + QK-Gain 5.0
val_bpb = 1.0897 (3-seed mean, std 0.0003) | ~15.99 MB | 8×H100 SXM
3-Seed Results
Merged SOTA (PR #1019): 1.1147 BPB. Delta: −0.0250 BPB.
Key Techniques
Compliance
Reproduction
Credits