Skip to content

Non-record: 11L gated Krylov + AR GPTQ int6 + lzma, 1.09596 BPB#1446

Open
LauraGomezjurado wants to merge 9 commits intoopenai:mainfrom
LauraGomezjurado:main
Open

Non-record: 11L gated Krylov + AR GPTQ int6 + lzma, 1.09596 BPB#1446
LauraGomezjurado wants to merge 9 commits intoopenai:mainfrom
LauraGomezjurado:main

Conversation

@LauraGomezjurado
Copy link
Copy Markdown

@LauraGomezjurado LauraGomezjurado commented Apr 7, 2026

Submission

  • Score: 1.09596320 BPB (sliding-window eval), 1.11953265 BPB exact roundtrip
  • Track: non-record-unlimited-compute-16mb
  • Artifact: 15,925,099 bytes (under 16MB cap)
  • Authors: Ganesh Talluri, Laura Gomezjurado, Hiroki Naganuma (@g4nesh, @LauraGomezjurado, @Hiroki11x)
  • Compute: 1×A100 80GB, 8h 52m, seed=1337

Approach

Standard SentencePiece GPT (11L, 512d, 26.99M params) with two main additions:

Gated Krylov correction on Muon: Estimates nonnormality of square weight slices via Hutchinson trace of W^T W − W W^T. Slices exceeding a threshold get a small adaptive-rank Krylov residual correction blended in at α=0.05. Muon stays the base optimizer; Krylov fires selectively on nonnormal slices.

AR self-generated Full-Hessian GPTQ int6 + lzma: Hessians calibrated on 64×2048 tokens from the model's own autoregressive output (temp=0.8), avoiding val/train data leakage. Percentile clipping across 5 levels, selective ±1 pruning to fit the 16MB cap.

Additional architecture: XSA across all 11 layers, BigramHash, SmearGate, VE128, partial RoPE, U-Net skips, LeakyReLU(0.5)^2, seq_len=2048, EMA (decay=0.997).

Results

Metric Value
Sliding BPB 1.09596320
Int6 roundtrip exact BPB 1.11953265
Artifact bytes 15,925,099
Training time 8h 52m (1×A100 80GB)

LauraGomezjurado and others added 3 commits April 7, 2026 10:30
Muon with gated Krylov correction on nonnormal square slices, AR self-gen
Full-Hessian GPTQ int6 + lzma, selective ±1 pruning, sliding-window eval.
26.99M params, 15,925,099 bytes, 1xA100 80GB, 8h 52m.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
taka6745 pushed a commit to taka6745/parameter-golf that referenced this pull request Apr 7, 2026
…ai#1430 stalled, 2 new PRs validate deferred specs

Patches 15/16/21 still uncontested in 150+ open + 10 closed PRs (5 audits
in a row). Strong evidence of true novelty.

PR openai#1430 still OPEN, 0 comments, no comp owner activity since creation.
Increasingly likely to be reverted or outlawed.

NEW PRs validate two of our deferred H100 escalation specs:
  - PR openai#1445 (1.0889): "Depth Recurrence + EMA 0.9965" → validates Patch 17 EMA spec
  - PR openai#1446 (1.0960): "int6 GPTQ + lzma" → validates Patch 23 INT6 GPTQ-Lite spec

Combined with PR openai#1437/openai#1420 already validating Patch 23 N-gram Tilt, the
3-spec H100 escalation bundle (EMA + Tilt + INT6 GPTQ) is now triple-
confirmed by independent comp PRs.

Spend ~$3.00/$36 (8% utilization). Pod healthy at 6h uptime.

Reminder: depth recurrence is back on the table — 5+ records use it now.
LESSONS.md §29 needs another update from "stale" to "real direction".

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
@MatoTeziTanka
Copy link
Copy Markdown

MatoTeziTanka commented Apr 11, 2026

Community Review — Non-record: 11L gated Krylov + AR GPTQ int6 + lzma, 1.09596 BPB

Compliance: NEEDS AUTHOR ACTION — train_gpt.py fails to import on CT2038 (Python 3.10 / torch 2.10.0+cpu)

What I found: The CPU smoke test on CT2038 (proteus-engine, 128 GB RAM, Triton 3.6.0, flash_attn stub, cutlass_evt_fusion stub) failed at the import step with:

ModuleNotFoundError: No module named 'golf'

A few of the common patterns I've seen for this class of error in the 2026-04-11 sweep:

Recommendation: Could you run python3 -c "import py_compile; py_compile.compile('train_gpt.py')" on your records-folder train_gpt.py under Python 3.10 specifically? The eval image is Python 3.10 per Issue #17 / the README, so any parse error on 3.10 blocks the submission at import time before any of the scored-eval logic runs.

Once the parse/import issue is fixed, I'll re-run the compliance audit through the normal pipeline. No other flags identified yet because the audit halts at the import step.


Reviewed by @MatoTeziTankaThe Agora. CPU smoke test (CT2038 proteus-engine, 2026-04-11): IMPORT_FAIL — ModuleNotFoundError: No module named 'golf'. Classification via classify_prs.py AST-based classifier; full compliance audit deferred until the import issue is resolved. Auto-drafted from a template and spot-checked before posting.

@g4nesh
Copy link
Copy Markdown

g4nesh commented Apr 11, 2026

@MatoTeziTanka I have fixed the problem. Please let me know if any error persists. Thanks for your help!

@MatoTeziTanka
Copy link
Copy Markdown

Re-audited at head SHA 4d4bb02.

Fix confirmed. The golf module import is now wrapped in a try/except chain (lines 31-39): first tries from hnet_tokenizer import HNetByteLM (sibling), then from golf.hnet_tokenizer import HNetByteLM (package), then falls back to HNetByteLM = None. File compiles clean under Python 3.10.

The import-blocking issue from my original review is resolved. I haven't run the full compliance audit yet (TTT/SLOT/n-gram checks on the model architecture) — I'll queue that for the next sweep. No hnet_tokenizer.py sibling file in the submission folder, so the HNet path is likely unused in this submission's active code path, but the graceful fallback to None means it won't block import regardless.

Thanks for the quick fix @g4nesh.


Re-audit by @MatoTeziTanka. Verified import chain at lines 31-39, py_compile OK under Python 3.10.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants