Submission: DominationV2 + BOS-Reset Bigram Cache + TTT (val_bpb=1.1382, 3-seed mean)#958
Conversation
Community Review — Submission: DominationV2 + BOS-Reset Bigram Cache + TTT (val_bpb=1.1382, 3-seed mean)BPB: 1.1382 | Compliance: LOOKS CLEAN — pure-neural submission, no TTT/SLOT/n-gram-cache What I found in the code (head SHA Static code review found no TTT adaptation function, no SLOT optimization loop, no n-gram-cache class, and no pre-quant val-token fine-tune. The eval path uses the standard sliding-window stride-64 pattern. The submission is a pure-neural architecture iteration on the standard SP1024/SP4096/SP8192 baseline. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.03s, dim=512, layers=11, vocab=1024, code=55299 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending the usual record-track checks (3-seed validation, under-16MB artifact cap, ≤600s train + ≤600s eval on 8×H100 SXM). No compliance flags from the classification pass — this looks like a clean pure-neural iteration on the standard baseline. Auto-classification caveat: this review was drafted by the AST-based classifier. If there's a non-standard eval mechanism (logit postprocessing, hedge mixing, etc.) that I missed because it's factored into a helper file or a non-standard function name, please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.03s, dim=512, layers=11, vocab=1024, code=55299 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
DominationV2 + BOS-Reset Bigram Cache + TTT
val_bpb: 1.1382 (3-seed mean, std 0.0010) | ~15.5 MB | 8xH100 SXM
Results (8xH100 80GB SXM, PyTorch 2.9.1+cu128)
Timing Budget
BOS-Reset Bigram Cache
An eval-time bigram cache applied during sliding window evaluation, after quantization roundtrip and TTT.
For each scored token, the cache tracks bigram counts from already-scored tokens within the current document and blends with model probabilities:
Cache resets at every BOS token (document boundary). Updated only after each token is scored (score-first, same ordering as TTT in PR #549).
Architecture
DominationV2 stack:
Cache Settings
Run Command
python3 data/cached_challenge_fineweb.py --variant sp1024 pip install zstandard cd records/track_10min_16mb/2026-03-27_DominationV2_BigramCache_TTT DATA_PATH=../../data/datasets/fineweb10B_sp1024 \ TOKENIZER_PATH=../../data/tokenizers/fineweb_1024_bpe.model \ SEED=1337 \ torchrun --standalone --nproc_per_node=8 train_gpt.pyCredits