Skip to content

Non-record: 5L MLP4x + SlidingWindow + SWA + QAT — val_bpb 1.33 (1xH100)#842

Open
JUSTSUJAY wants to merge 2 commits intoopenai:mainfrom
JUSTSUJAY:submission/5L-MLP4x-SlidingWindow-1.3827
Open

Non-record: 5L MLP4x + SlidingWindow + SWA + QAT — val_bpb 1.33 (1xH100)#842
JUSTSUJAY wants to merge 2 commits intoopenai:mainfrom
JUSTSUJAY:submission/5L-MLP4x-SlidingWindow-1.3827

Conversation

@JUSTSUJAY
Copy link
Copy Markdown

Autonomous AI-driven exploration of 16 experiments using autoresearch framework. Key discovery: 5L MLP4x significantly outperforms deeper narrower architectures on single-GPU compute budgets.

Techniques: BigramHash(4096), SmearGate, OrthoInit, QAT (int8 STE), SWA (18 checkpoints), sliding window eval (stride=64).

14.6MB artifact, 15.5M params, 4573 steps in 300s on 1xH100 80GB.

Autoresearch Agent and others added 2 commits March 26, 2026 11:35
…H100)

Autonomous AI-driven exploration of 16 experiments using autoresearch
framework. Key discovery: 5L MLP4x significantly outperforms deeper
narrower architectures on single-GPU compute budgets.

Techniques: BigramHash(4096), SmearGate, OrthoInit, QAT (int8 STE),
SWA (18 checkpoints), sliding window eval (stride=64).

14.6MB artifact, 15.5M params, 4573 steps in 300s on 1xH100 80GB.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Doubled training time to 10 minutes (challenge limit). 9353 steps
with 38 SWA checkpoints. val_bpb improved from 1.3827 to 1.3380.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
@JUSTSUJAY JUSTSUJAY changed the title Non-record: 5L MLP4x + SlidingWindow + SWA + QAT — val_bpb 1.3827 (1xH100) Non-record: 5L MLP4x + SlidingWindow + SWA + QAT — val_bpb 1.33 (1xH100) Mar 26, 2026
@MatoTeziTanka
Copy link
Copy Markdown

MatoTeziTanka commented Apr 11, 2026

Community Review — Non-record: 5L MLP4x + SlidingWindow + SWA + QAT — val_bpb 1.33 (1xH100)

Compliance: NEEDS AUTHOR ACTION — train_gpt.py fails to import on CT2038 (Python 3.10 / torch 2.10.0+cpu)

What I found: The CPU smoke test on CT2038 (proteus-engine, 128 GB RAM, Triton 3.6.0, flash_attn stub, cutlass_evt_fusion stub) failed at the import step with:

ModuleNotFoundError: No module named 'prepare_pgolf'

A few of the common patterns I've seen for this class of error in the 2026-04-11 sweep:

Recommendation: Could you run python3 -c "import py_compile; py_compile.compile('train_gpt.py')" on your records-folder train_gpt.py under Python 3.10 specifically? The eval image is Python 3.10 per Issue #17 / the README, so any parse error on 3.10 blocks the submission at import time before any of the scored-eval logic runs.

Once the parse/import issue is fixed, I'll re-run the compliance audit through the normal pipeline. No other flags identified yet because the audit halts at the import step.


Reviewed by @MatoTeziTankaThe Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): IMPORT_FAIL — ModuleNotFoundError: No module named 'prepare_pgolf'. Classification via classify_prs.py AST-based classifier; full compliance audit deferred until the import issue is resolved. Auto-drafted from a template and spot-checked before posting.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants