Non-Record: BPB 1.13872 — LeakyReLU(0.5)² + Per-Layer LR Legal TTT (3 seeds)#537
Conversation
- Replace ReLU² with LeakyReLU(0.5)² activation (-0.004 BPB pre-TTT) - Add per-layer LR groups: mlp.proj 3x, mlp.fc 0.5x for TTT - Add intra-chunk cosine LR schedule for TTT epochs - 3-seed validation: 1.13912, 1.14024, 1.13872 (mean 1.13936) - Score-first legal TTT with SGD momentum, 30 epochs, freeze-2 - Best seed (7): BPB 1.13872, artifact 15.36 MB
e15a447 to
3b619c7
Compare
Novel: TTT adapts ONLY scalar/control parameters (q_gain, attn_scale, mlp_scale, resid_mix, RMSNorm weights, skip_weights, skip_gates). Matrix weights (c_q/c_k/c_v/proj/MLP/tok_emb) stay frozen. This is mechanistically different from full-model TTT (openai#1413, openai#537): the model retunes its existing control knobs rather than learning new weight directions. Higher LR (0.01) since scalars need bigger steps. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Community Review — Non-Record: BPB 1.13872 — LeakyReLU(0.5)² + Per-Layer LR Legal TTT (3 seeds)BPB: 1.13872 | Compliance: LOOKS CLEAN — score-first-per-chunk TTT (legal #1416/#1423 pattern) What I found in the code (head SHA The TTT path at line 849 implements the score-first-per-chunk pattern: each chunk is scored under Per Issue #402 and Issue #677, TTT is legal when each token is scored before the adapter updates on it, and that's what the code does here — chunk CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.05s, dim=512, layers=11, vocab=1024, code=74210 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending standard checks (3-seed validation, 16MB artifact cap, 10-min wallclock on 8×H100 SXM). The compliance picture matches the legal reference frontier and no flags were raised by the classification pass. Auto-classification caveat: this review was drafted by the AST-based classifier against a template derived from manually-reviewed cluster PRs (#1420, #1450, #1487, #1541, #1529, #1533, #1518). If I've misread a subtlety in your eval path — e.g., multi-epoch TTT that I mistook for single-pass, or a target-in-key lookup I missed in a helper function — please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.05s, dim=512, layers=11, vocab=1024, code=74210 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
Non-Record: 11L LeakyReLU(0.5)² + Per-Layer LR Legal TTT (3 seeds)
val_bpb = 1.13872 (best seed) | Mean: 1.13936 ± 0.0008 | Pre-TTT mean: 1.1574 | Artifact: 15.36 MB
Non-record unlimited-compute submission (trained on 4×A100-40GB, ~42 min; eval ~3690s on 1×A100).
What Changed vs PR #526 (BPB 1.14252)
Note on TTT modifications: Per-layer LR and intra-chunk cosine were adopted from other PRs but showed no measurable TTT improvement in this data — TTT gain went from −0.0184 (PR #526) to −0.0182 (this PR), a slight regression. The entire final BPB improvement comes from the better pre-TTT model via LeakyReLU. These TTT modifications require further ablation.
3-Seed Validation
Architecture
TTT Protocol (Legal)
torch.inference_mode()during scoring, then trainCredits
This submission integrates work from many contributors: