Skip to content

Fix: classify.py reads artifact_size but fetch_prs.py writes artifact_bytes#5

Merged
MatoTeziTanka merged 1 commit intoMatoTeziTanka:gh-pagesfrom
fielding:fix/artifact-field-name
Mar 31, 2026
Merged

Fix: classify.py reads artifact_size but fetch_prs.py writes artifact_bytes#5
MatoTeziTanka merged 1 commit intoMatoTeziTanka:gh-pagesfrom
fielding:fix/artifact-field-name

Conversation

@fielding
Copy link
Copy Markdown

@fielding fielding commented Mar 31, 2026

Bug

fetch_prs.py writes artifact sizes to artifact_bytes (line 301), but classify.py reads artifact_size (line 315), which is never populated. Every open PR without banned-keyword flags gets classified as INCOMPLETE with artifact-unknown, even when artifact data exists.

Currently 0/641 INCOMPLETE PRs can become ALIVE... only merged PRs bypass the check.

Fix

# before
artifact_size = pr.get("artifact_size")

# after
artifact_size = pr.get("artifact_bytes", pr.get("artifact_size"))

Reads the field fetch_prs.py actually writes, with fallback for backwards compatibility.

See also: #4

@MatoTeziTanka MatoTeziTanka merged commit 2173904 into MatoTeziTanka:gh-pages Mar 31, 2026
MatoTeziTanka pushed a commit that referenced this pull request Apr 11, 2026
…cluster + CT2038 gauntlet provisioned

Reviewed all 20 highest-priority Tier 1 PRs from openai/parameter-golf.
Two cluster-level findings:

- N-gram family bug (10 PRs CLOSED + 1 already ruled): full_key = ((ctx_hash
  ^ (target * primes[k])) & mask) — target token hashed into the eval-cache
  lookup key, ruled illegal by valerio-oai on PR openai#779. Same verbatim pattern
  in openai#770/openai#798/openai#808/openai#825/openai#786/openai#797/openai#909/openai#940/openai#761 + openai#764 follow-up. Upstream
  parent: lukacf (openai#659/openai#702/openai#727 — task #5 audit queued).

- Standard SLOT cluster (4 HOLD pending openai#1336, 2 CLOSE): per-window
  delta+logit_bias optimized N steps against (per_token_nll * mask) where
  mask = scored positions [s:wlen]. PRs openai#1321/openai#1324/openai#1278/openai#1263 → HOLD;
  openai#1319/openai#1376 → CLOSE.

Clean MERGE-eligible: openai#1420 (token_hint-only post-fix) and openai#1450 (TMA
megakernel triple loop).

Eval-budget gate (openai#915/openai#889 anthony-maio pair): clean ngram code, ~14.9 min
ngram stage on 8xH100 SXM. One @0hq ruling on Issue openai#17 unblocks both PRs
plus ~30 ngram-cache PRs.

Infrastructure: provisioned CT2038 (proteus-engine, 128 GB RAM, 32 cores)
as the dedicated parameter-golf gauntlet host. Installed Triton 3.6.0,
deployed cpu_test.py + flash_attn_stub.py. Re-ran the 4 PRs originally
skipped due to FA3/Triton blockers — all PASS. Edited 4 GitHub comments
via gh api PATCH to add the rerun results. Coverage went from 9/20 to
14/20 fully gauntleted.

Side session handed off via SOW_HF_DATASET_REPUBLISH.md (Scylla 998→1254
fix + SP4096/SP8192/SP12288/SP16384 publish + Cloudflare R2 mirror).

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants