Record: Int6 QAT + SmearGate + Muon WD (val_bpb=1.1669)#170
Open
baudrillardsgh0st wants to merge 2 commits intoopenai:mainfrom
Open
Record: Int6 QAT + SmearGate + Muon WD (val_bpb=1.1669)#170baudrillardsgh0st wants to merge 2 commits intoopenai:mainfrom
baudrillardsgh0st wants to merge 2 commits intoopenai:mainfrom
Conversation
9L 512dim int6 QAT with STE, SmearGate, Muon weight decay 0.01, int6-in-int8 zstd22 compression. 14.77MB artifact, 9706 steps @ 61.8ms/step. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
11-layer GPT with int6 QAT, SmearGate, and decoupled Muon weight decay 0.038. Artifact: 15.50MB (int6+zstd-22). Single seed, 7723 steps at 77ms/step. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
Key Techniques
Int6 QAT (Quantization-Aware Training): STE fake int6 quantization during forward pass with per-row symmetric scaling. Eliminates post-quant degradation without needing fp16 late-K layer passthrough.
Int6-in-Int8 zstd22 compression: Store int6 values (-32 to 31) in int8 containers — zstd-22 compresses the restricted value range ~35%. Achieves 14.77MB from 21.8M params. (Bit-packing int6 values destroys byte alignment and defeats compressors.)
SmearGate: ~513-param learned gate blending current + previous token embedding. Zero-initialized, very low LR. Provides cheap bigram context at the embedding layer.
Decoupled Muon weight decay (0.01): Applied in the Muon optimizer step for improved generalization and quantization robustness.
Sliding window evaluation (stride=64, batch=32 seqs): Full-context scoring at every token position.
FP16 tied embedding passthrough: Avoids compounding int6 errors through both input/output paths.
Results
Artifact Size
Test plan