Recursive Transformer 4B/7L + VE + QAT + TTT — val_bpb 1.1696 (3-seed mean)#927
Recursive Transformer 4B/7L + VE + QAT + TTT — val_bpb 1.1696 (3-seed mean)#927Tonyy1977 wants to merge 2 commits intoopenai:mainfrom
Conversation
… mean) True Universal Transformer: 4 shared blocks x 7 loops (7x weight reuse), dim=1024, int6 QAT from step 0, score-first TTT+sliding window eval. 3-seed mean: 1.1696 BPB, 15.85MB artifact, 600s training on 8xH100.
Required for zstd-22 compression of the int8 quantized model artifact. Without it, the script falls back to zlib which produces 17.5MB (over 16MB budget). Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Community Review — Recursive Transformer 4B/7L + VE + QAT + TTT — val_bpb 1.1696 (3-seed mean)BPB: 1.1696 | Compliance: LOOKS CLEAN — score-first-per-chunk TTT (legal #1416/#1423 pattern) What I found in the code (head SHA The TTT path at line 918 implements the score-first-per-chunk pattern: each chunk is scored under Per Issue #402 and Issue #677, TTT is legal when each token is scored before the adapter updates on it, and that's what the code does here — chunk CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.07s, dim=1024, layers=, vocab=1024, code=68750 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending standard checks (3-seed validation, 16MB artifact cap, 10-min wallclock on 8×H100 SXM). The compliance picture matches the legal reference frontier and no flags were raised by the classification pass. Auto-classification caveat: this review was drafted by the AST-based classifier against a template derived from manually-reviewed cluster PRs (#1420, #1450, #1487, #1541, #1529, #1533, #1518). If I've misread a subtlety in your eval path — e.g., multi-epoch TTT that I mistook for single-pass, or a target-in-key lookup I missed in a helper function — please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.07s, dim=1024, layers=, vocab=1024, code=68750 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
Summary
Recursive transformer: 4 shared blocks × 7 loops (7× weight reuse) at dim=1024, with ValueEmbedding, int6 QAT from step 0, and score-first TTT+sliding window eval.
3-seed mean: 1.1696 BPB | ~15.85MB artifact | 600s on 8xH100 SXM
Key novelty
Unlike other depth recurrence submissions that repeat 1-2 layers on top of 10-11 unique blocks (~1.2× reuse), this uses 4 shared blocks looped 7 times (7× reuse). This enables dim=1024 (2× wider than standard 512) while staying under 16MB.
Architecture highlights
See README.md in the submission folder for full details and negative results.