Non-record: 11L gated Krylov + AR GPTQ int6 + lzma, 1.09596 BPB#1446
Non-record: 11L gated Krylov + AR GPTQ int6 + lzma, 1.09596 BPB#1446LauraGomezjurado wants to merge 9 commits intoopenai:mainfrom
Conversation
Muon with gated Krylov correction on nonnormal square slices, AR self-gen Full-Hessian GPTQ int6 + lzma, selective ±1 pruning, sliding-window eval. 26.99M params, 15,925,099 bytes, 1xA100 80GB, 8h 52m. Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
…ai#1430 stalled, 2 new PRs validate deferred specs Patches 15/16/21 still uncontested in 150+ open + 10 closed PRs (5 audits in a row). Strong evidence of true novelty. PR openai#1430 still OPEN, 0 comments, no comp owner activity since creation. Increasingly likely to be reverted or outlawed. NEW PRs validate two of our deferred H100 escalation specs: - PR openai#1445 (1.0889): "Depth Recurrence + EMA 0.9965" → validates Patch 17 EMA spec - PR openai#1446 (1.0960): "int6 GPTQ + lzma" → validates Patch 23 INT6 GPTQ-Lite spec Combined with PR openai#1437/openai#1420 already validating Patch 23 N-gram Tilt, the 3-spec H100 escalation bundle (EMA + Tilt + INT6 GPTQ) is now triple- confirmed by independent comp PRs. Spend ~$3.00/$36 (8% utilization). Pod healthy at 6h uptime. Reminder: depth recurrence is back on the table — 5+ records use it now. LESSONS.md §29 needs another update from "stale" to "real direction". Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Community Review — Non-record: 11L gated Krylov + AR GPTQ int6 + lzma, 1.09596 BPBCompliance: NEEDS AUTHOR ACTION — What I found: The CPU smoke test on CT2038 (proteus-engine, 128 GB RAM, Triton 3.6.0, flash_attn stub, cutlass_evt_fusion stub) failed at the import step with: A few of the common patterns I've seen for this class of error in the 2026-04-11 sweep:
Recommendation: Could you run Once the parse/import issue is fixed, I'll re-run the compliance audit through the normal pipeline. No other flags identified yet because the audit halts at the import step. Reviewed by @MatoTeziTanka — The Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): IMPORT_FAIL — ModuleNotFoundError: No module named 'golf'. Classification via |
|
@MatoTeziTanka I have fixed the problem. Please let me know if any error persists. Thanks for your help! |
|
Re-audited at head SHA Fix confirmed. The The import-blocking issue from my original review is resolved. I haven't run the full compliance audit yet (TTT/SLOT/n-gram checks on the model architecture) — I'll queue that for the next sweep. No Thanks for the quick fix @g4nesh. Re-audit by @MatoTeziTanka. Verified import chain at lines 31-39, py_compile OK under Python 3.10. |
Submission
Approach
Standard SentencePiece GPT (11L, 512d, 26.99M params) with two main additions:
Gated Krylov correction on Muon: Estimates nonnormality of square weight slices via Hutchinson trace of
W^T W − W W^T. Slices exceeding a threshold get a small adaptive-rank Krylov residual correction blended in at α=0.05. Muon stays the base optimizer; Krylov fires selectively on nonnormal slices.AR self-generated Full-Hessian GPTQ int6 + lzma: Hessians calibrated on 64×2048 tokens from the model's own autoregressive output (temp=0.8), avoiding val/train data leakage. Percentile clipping across 5 levels, selective ±1 pruning to fit the 16MB cap.
Additional architecture: XSA across all 11 layers, BigramHash, SmearGate, VE128, partial RoPE, U-Net skips, LeakyReLU(0.5)^2, seq_len=2048, EMA (decay=0.997).
Results