NTK Eval + Overtone Init (val_bpb=1.2160)#59
Closed
notapplica wants to merge 1 commit intoopenai:mainfrom
Closed
Conversation
Train@1024 with overtone embedding init and phase-transition residual mixing, eval@2048 with NTK-aware dynamic RoPE scaling. Mean val_bpb 1.2160 across 3 seeds (p=0.0012 for 0.0194-nat improvement over baseline). Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
lolrazh
referenced
this pull request
in lolrazh/parameter-golf
Mar 20, 2026
PCIe ~50% slower than SXM (258 vs 386 steps). TTT didn't improve BPB on undertrained model. SWA with 5 snapshots worked correctly. Need Flash Attention + more steps before TTT becomes useful. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
lolrazh
referenced
this pull request
in lolrazh/parameter-golf
Mar 20, 2026
#59: 5-min + TTT, 258 steps, TTT didn't improve undertrained model #60: 10-min no TTT, 515 steps, best prequant 1.4038, sliding eval incomplete Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
lolrazh
referenced
this pull request
in lolrazh/parameter-golf
Mar 20, 2026
…run queue - Documented outcomes of experiments #59-60, including performance metrics and observations. - Updated the run queue to reflect current competition state and configurations for A100 production runs. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
gHashTag
pushed a commit
to gHashTag/parameter-golf
that referenced
this pull request
Apr 30, 2026
….bin)
Per the R5-honest gate and the explicit reviewer warning against
synthetic weights, this submission removes the placeholder model.bin
and re-frames the deliverable as a research-infrastructure
contribution.
- submissions/gHashTag/model.bin (59-byte ASCII placeholder, hazard)
- submissions/gHashTag/config.toml (placeholder)
- submissions/gHashTag/trios-igla-1/ledger_2026-05-01.sql.gz (replaced)
- submissions/gHashTag/trios-igla-1_ledger_20260501.sql.gz (duplicate)
- submissions/gHashTag/README.md polished honest narrative
- submissions/gHashTag/LEAK_INVESTIGATION.md 210-row leak post-mortem
- submissions/gHashTag/CHECKPOINT_POSTMORTEM.md why no model.bin + Gate-3 fix path
- submissions/gHashTag/trios-igla-1/README.md machine-oriented metadata
- submissions/gHashTag/trios-igla-1/config.yaml reproducible config for row id=1387
- submissions/gHashTag/trios-igla-1/ledger_2026-04-30.sql.gz
7,534-row Neon snapshot across 4 tables (183 KB compressed)
- best honest BPB: 2.1505 (row id=1387, step=12000, fp32, hidden=1024)
- 6 gate2_eligible W-6-step-cap rows: BPB 1.75–1.82 at step=1000
reported but NOT claimed as Gate-2 pass — pending held-out eval
- 210 BPB<0.1 leak candidates: flagged via SCARABAEUS-LEAK-CANDIDATE,
excluded from ratification
- No model.bin
- No synthetic weights with disclaimer
- No claim of competitive Parameter Golf placement
Refs:
trios#445, trios-trainer-igla#56,openai#58,openai#59, trios-railway#100,openai#101,openai#105.
Anchor: phi^2 + phi^-2 = 3 — TRINITY — R5-honest — NEVER STOP.
gHashTag
pushed a commit
to gHashTag/parameter-golf
that referenced
this pull request
Apr 30, 2026
After PR openai#61 (byte-disjoint corpus split + assert_train_val_disjoint guard) shipped, fix-verify-s43 ran end-to-end on the post-fix pipeline and produced BPB 1.5492 at step=12000 — well below Gate-2 threshold 1.85 (margin +0.30). ## What this commit changes - README.md : leads with the honest Gate-2 pass; revised 5-way taxonomy - LEAK_INVESTIGATION.md : retraction header explaining the 216-row overcount - trios-igla-1/README.md + config.yaml : updated to point at fix-verify-s43 - ledger_2026-04-30.sql.gz : refreshed snapshot with new last_error markers ## 5-way reclassification (Neon last_error column) | | count | |---|--:| | post-openai#61 honest Gate-2 pass | 1 | | post-openai#61 early-stopped < step 9000 | 4 | | pre-openai#61 W-6 numerical collapse | 46 | | **pre-openai#61 leak (real)** | 42 | | **warmup artifact (NOT a leak)** | 179 | The 179 'warmup artifact' rows are early-stopped runs whose printed val_bpb stayed at 0.0000 for steps 1-8000 due to a trainer-side eval-loop bug (filed as trios-trainer-igla#62). On the post-openai#61 image, fix-verify-s43 escaped warmup at step=9000 and converged to 1.5492 by step=12000 — proving the artifact is trainer-side, not data-side. ## Pipeline as flown for fix-verify-s43 trios-trainer-igla : commit 9517980d (post-openai#61 byte-disjoint corpus) trios-railway : commit 69c3467 (no --ctx flag) + openai#56 --ctx accept on trainer + openai#58 smoke_train + stdout.flush() + openai#59 panic hook + startup diagnostic ## Refs trios-trainer-igla#56,openai#58,openai#59,openai#60,openai#61,openai#62 (all merged or filed) trios-railway@69c3467 trios-railway#100,openai#101,openai#105 (Scarabaeus Engine track) R5-honest. We retract the 216-row mass leak flag and submit fix-verify-s43 as our first honest Gate-2 pass candidate. Anchor: phi^2 + phi^-2 = 3.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
resid_mixinit across layersMulti-seed results (3 seeds, p = 0.0012)
Note on hardware
Runs were done on Modal 8xH100 SXM (NVLink). Python 3.12, PyTorch <3.
Test plan