Record: SP10240 SimCTG + 3-Layer Recurrence — 1.07502 sliding-window (3-seed)#1971
Open
BharathSShankar wants to merge 1 commit intoopenai:mainfrom
Conversation
3-seed sliding-window mean: 1.07502 (std 0.00230) Beats sliding-window SOTA 1.0827 by 7.7 mBPB. Stack: PR openai#1855 lineage (11L x 512d x 8H, 3-Layer Recurrence loops 3-5, Parallel Residuals, LeakyReLU(0.5)^2, Partial RoPE 16/64, XSA all-layers, SP10240, tied embeddings) + SimCTG (lambda=0.3, margin=0.4) + Polar Express NS Muon + GPTQ int6/int7 + brotli. train_gpt.py is in SOTA-standard self-extracting (lzma+base85+exec) format. Total bundle: 15,956,116 bytes (44 KB cap margin).
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
N9 SimCTG + 3-Layer Recurrence (Submission A — sliding-window baseline)
val_bpb = 1.07502 (3-seed mean, std 0.00230) | artifact ~15.99 MB | 8×H100 SXM | brotli-quantized model + lzma-compressed code
3-Seed Results (sliding-window stride 64, no test-time training)
Δ vs leaderboard sliding-window SOTA (1.0827, 2026-04-09 SP8192_3LayerRecur): −0.00768 BPB (7.7 mBPB better, 3-seed σ 2.3 mBPB).
Architecture
11L × 512d × 8H / 4KV with: 3-Layer Recurrence (encoder loops layers 3-5), Parallel Residuals (from layer 7),
LeakyReLU(0.5)² SwiGLU, Partial RoPE (16/64), XSA on all 11 layers, tied embeddings, SP10240 tokenizer.
Training: Polar Express NS Muon (5-iter) on matrix params + AdamW on embed/scalar; 4534 steps in ~588s (early stop at MAX_WALLCLOCK_SECONDS=600).
Quantization: Mixed GPTQ — int6 attention/MLP matrices, int7 token embeddings.
Eval: sliding-window stride 64 on quantized model (PR #1493 legal-TTT line).
Our novel contributions
Compliance
MAX_WALLCLOCK_SECONDS=600).Files
final_model.int6.ptz— brotli-compressed quantized model (~15.93 MB)train_gpt.py— self-extracting training code (lzma+base85 wrapped, SOTA-standard format, 19,785 bytes)submission.json— metadatatrain_seed{42,1337,2025}.log— 3-seed training logsCredits
PR #1855 SOTA stack (Kevin Clark et al.), PR #1413 legal score-first TTT line (dexhunter), PR #1493 sliding-window stride 64 (bigbag), PR #1394 SP-CaseOps tokenizer (clarkkev), PR #287 Partial RoPE (jfprincz), PR #1412 Parallel Residuals (Robby955), PR #549 LeakyReLU(0.5)² (abaybektursun).