Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
25 changes: 25 additions & 0 deletions 2026-03-21_ParGolfZero/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
# ParGolf-Zero v2 — 6-Layer Joint Optimization

## Status
Non-record submission. Awaiting H100 compute grant for final scored run.
Pipeline fully confirmed on Kaggle T4 — artifact 5.52MB under 16MB limit.

## Unique Contribution
Joint compression-aware training. Per-row weight range penalty minimizes
int8 quantization error during every gradient step. Nobody else is doing this.

## Layers
- L1 COMPAT: Auto-detects GPU platform
- L2 TRAIN: FP16 embed + Muon WD + warmdown + SWA
- L3 COMPRESS: Weight range penalty + QAT + zstd-22
- L4 EVAL: Sliding window stride=64, 960 token context
- L5 ADAPT: Low-rank Q/V regularization for TTT
- L6 BIGRAM: BigramHash(10240) learned bigram table

## Confirmed Results (T4 smoke test)
- Artifact size: 5.52MB ✅
- Pipeline: train → quantize → zstd → roundtrip ✅
- val_bpb (200 steps, tiny batch): 3.19 (not real score)

## Author
Sanjith G — github.com/sanjith3057
22 changes: 22 additions & 0 deletions 2026-03-21_ParGolfZero/submission.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,22 @@
{
"name": "Sanjith G",
"github_id": "sanjith3057",
"run_name": "ParGolf-Zero v2",
"track": "non_record_16mb",
"val_bpb": null,
"notes": "Non-record submission. 6-layer joint optimization system. Pipeline confirmed on Kaggle T4. Awaiting H100 compute grant for final scored run. Artifact confirmed at 5.52MB under 16MB limit.",
"approach": "Joint compression-aware training — per-row weight range penalty minimizes int8 quantization error during every gradient step. QAT activates in final 500 steps. BigramHash(10240) adds bigram lookup. SWA over last 40% of training. zstd-22 compression. TTT low-rank regularization on Q/V projections.",
"layers": {
"L1_COMPAT": "Auto-detects GPU platform, runs on T4/A100/H100/CPU",
"L2_TRAIN": "FP16 embed + Muon WD=0.02 + warmdown=20000 + SWA",
"L3_COMPRESS": "Weight range penalty + QAT final 500 steps + zstd-22",
"L4_EVAL": "Sliding window stride=64, 960 token context per token",
"L5_ADAPT": "Low-rank Q/V regularization for TTT LoRA readiness",
"L6_BIGRAM": "BigramHash(10240) learned bigram table on logits"
},
"artifact_size_bytes": 5524254,
"hardware_tested": "Kaggle T4 (smoke test, 200 steps)",
"hardware_target": "8xH100 SXM (pending compute grant)",
"val_bpb_smoke_test": 3.1899,
"roundtrip_val_bpb_smoke_test": 3.2311
}
25 changes: 25 additions & 0 deletions records/track_non_record_16mb/2026-03-23_ParGolfZero_v2/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
# ParGolf-Zero v2 — 6-Layer Joint Optimization

## Status
Non-record submission. Awaiting H100 compute grant for final scored run.
Pipeline confirmed on Kaggle T4 — artifact 5.52MB under 16MB limit.

## Unique Contribution
Joint compression-aware training. Per-row weight range penalty minimizes
int8 quantization error during every gradient step. Nobody else is doing this.

## Layers
- L1 COMPAT: Auto-detects GPU platform
- L2 TRAIN: FP16 embed + Muon WD + warmdown + SWA
- L3 COMPRESS: Weight range penalty + QAT + zstd-22
- L4 EVAL: Sliding window stride=64, 960 token context
- L5 ADAPT: Low-rank Q/V regularization for TTT
- L6 BIGRAM: BigramHash(10240) learned bigram table

## Results (T4 smoke test, 200 steps)
- Artifact: 5.52MB ✅
- val_bpb: 3.19 (smoke test only, not real score)
- roundtrip val_bpb: 3.23 ✅

## Author
Sanjith G — github.com/sanjith3057
Original file line number Diff line number Diff line change
@@ -0,0 +1,22 @@
{
"name": "Sanjith G",
"github_id": "sanjith3057",
"run_name": "ParGolf-Zero v2",
"track": "non_record_16mb",
"val_bpb": null,
"notes": "Non-record submission. Awaiting H100 compute grant for final scored run. Artifact confirmed 5.52MB under 16MB.",
"approach": "Joint compression-aware training. Per-row weight range penalty + QAT + BigramHash(10240) + SWA + zstd-22 + TTT low-rank regularization.",
"layers": {
"L1_COMPAT": "Auto-detects GPU platform",
"L2_TRAIN": "FP16 embed + Muon WD=0.02 + warmdown=20000 + SWA",
"L3_COMPRESS": "Weight range penalty + QAT final 500 steps + zstd-22",
"L4_EVAL": "Sliding window stride=64, 960 token context",
"L5_ADAPT": "Low-rank Q/V regularization for TTT LoRA",
"L6_BIGRAM": "BigramHash(10240) learned bigram table"
},
"artifact_size_bytes": 5524254,
"hardware_tested": "Kaggle T4 (200 steps smoke test)",
"hardware_target": "8xH100 SXM (pending compute grant)",
"val_bpb_smoke_test": 3.1899,
"roundtrip_val_bpb_smoke_test": 3.2311
}
Loading