Non-record: VR + GA + Late QAT + Full GPTQ — 1.1418 BPB, 15.7 MB#601
Open
anantdgoel wants to merge 1 commit intoopenai:mainfrom
Open
Non-record: VR + GA + Late QAT + Full GPTQ — 1.1418 BPB, 15.7 MB#601anantdgoel wants to merge 1 commit intoopenai:mainfrom
anantdgoel wants to merge 1 commit intoopenai:mainfrom
Conversation
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
himanshudongre
added a commit
to himanshudongre/parameter-golf
that referenced
this pull request
Apr 4, 2026
Evidence from 4 independent configurations (PR openai#461, PR openai#601, PR openai#1326, and my own experiments) showing GPTQ's compensatory weight structure is destroyed by SGD-based test-time training. Key finding: SGD TTT gives -0.0165 BPB on simple int6 but provides negligible to negative improvement on GPTQ-quantized models (-0.0001 to +0.030 BPB). Includes complete SGD TTT implementation (sgd_ttt_eval.py) following PR openai#461 protocol, and LoRA TTT implementation (clark_ttt_eval.py). Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
|
I replicated this finding with LoRA TTT (rank-8 on Q,V projections) on a GPTQ int6 Clark 11L model: -0.0013 BPB, effectively zero improvement. PR #1326 (aryanbhosale) also independently confirmed with score-first SGD TTT on GPTQ: -0.0001 BPB. I've written a systematic analysis aggregating all 4 known TTT+GPTQ configurations in PR #1341, including root cause analysis (GPTQ's column-wise error compensation is destroyed by SGD) and proposed fix directions. |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
val_bpb: 1.1418 | 15.7 MB | 1x NVIDIA RTX A6000, ~14 hours
Summary
11-layer GPT combining the community meta-stack with novel techniques Value Residual (VR) and Gated Attention (GA), plus Late QAT during training and a Full GPTQ + Int5 MLP post-training quantization pipeline. Achieves 1.1418 BPB (stride=128) in a 15.7 MB artifact that fits under the 16 MB limit.
Update pending: BH10240 (bigram hash 10240 buckets) variant currently evaluating — expect improved results soon.
Novel Contributions
Ablation Results (stride=128)
Credits
Built on top of the excellent community meta-stack. Key techniques originated from:
Files
train_gpt.py— Full training + eval script with all techniquessubmission.json— MetadataREADME.md— Detailed writeup with ablations and reproducibility commands