Non-record: 11L GEPA + 12k Steps + Pure Int6 + Legal TTT (val_bpb=1.1079)#612
Conversation
Non-record submission: 11L GEPA architecture trained for 12000 steps (7k peak-LR + 5k warmdown) on 4xA100-40GB with pure int6 per-row quantization using 15-candidate GPTQ-lite clip search and zstd-22 compression. Legal score-first TTT (SGD, 10 epochs, momentum 0.9) drives final BPB from 1.154 (int6 quant) to 1.1079. - Pre-TTT float base: 1.1268 (step 12000) - Post-quant pre-TTT: ~1.154 - Final with legal TTT: 1.10788263 - Artifact: 15.51 MB (14.72 MB model + 78 KB code) - 27M parameters, pure int6 quantization
Community Review — Non-record: 11L GEPA + 12k Steps + Pure Int6 + Legal TTT (val_bpb=1.1079)BPB: 1.1079 | Compliance: LOOKS CLEAN — score-first-per-chunk TTT (legal #1416/#1423 pattern) What I found in the code (head SHA The TTT path at line 399 implements the score-first-per-chunk pattern: each chunk is scored under Per Issue #402 and Issue #677, TTT is legal when each token is scored before the adapter updates on it, and that's what the code does here — chunk CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.03s, dim=512, layers=11, vocab=1024, code=78281 B, SMOKE_TEST_PASS Verdict: LOOKS CLEAN. Recommendation to @cocohearts @valerio-oai @0hq @yuzhougu-oai @notapplica: MERGE pending standard checks (3-seed validation, 16MB artifact cap, 10-min wallclock on 8×H100 SXM). The compliance picture matches the legal reference frontier and no flags were raised by the classification pass. Auto-classification caveat: this review was drafted by the AST-based classifier against a template derived from manually-reviewed cluster PRs (#1420, #1450, #1487, #1541, #1529, #1533, #1518). If I've misread a subtlety in your eval path — e.g., multi-epoch TTT that I mistook for single-pass, or a target-in-key lookup I missed in a helper function — please flag it and I'll re-run the audit manually. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): import OK in 0.03s, dim=512, layers=11, vocab=1024, code=78281 B, SMOKE_TEST_PASS. Classification via deterministic AST-based |
Non-Record Submission: 11L GEPA + 12k Steps + Pure Int6 + Legal TTT
val_bpb = 1.1079 (1.10788263 exact) | Pre-TTT float: 1.1268 | Int6 quant: ~1.154 | TTT gain: −0.046 | Artifact: 14.79 MB
Summary
GEPA architecture (11L, 27M params) trained for 12,000 steps (7k peak-LR + 5k warmdown) on 4×A100-40GB. Pure int6 per-row quantization with 15-candidate GPTQ-lite clip search + zstd-22 compression. Legal score-first TTT (SGD, momentum 0.9, lr=0.002, 10 epochs, freeze first 2 blocks).
Key Results
Novel Contributions
Track
track_non_record_16mb— unlimited compute, 16 MB artifact limit.Checklist