You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
fix: reduce Docker layers, add auto CI trigger, fix fake ops import (#363)
* fix: reduce Docker image layers to avoid overlay2 max depth limit
Aggressively merge RUN instructions in the Dockerfile to reduce total
layer count from ~126 to ~119. The inference image was hitting the
overlay2 128-layer limit ("failed to register layer: max depth
exceeded") on CI nodes.
devel stage: 8 RUN + 1 COPY -> 4 RUN + 1 COPY (-4 layers)
build stage: 4 RUN + 1 COPY -> 1 RUN + 1 COPY (-3 layers)
FBGEMM and TorchRec kept as separate layers for build cache efficiency.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
* ci: add pull_request_target trigger for auto CI on PR open/sync
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
* Fix imports for fake ops wrapper used in expor
* fix: remove invalid import of hstu.hstu_ops_gpu
The module hstu.hstu_ops_gpu does not exist as a Python module.
The C++ source hstu_ops_gpu.cpp compiles into hstu/fbgemm_gpu_experimental_hstu.so,
not a separate hstu_ops_gpu submodule. This import was incorrectly added in PR #327
and causes ModuleNotFoundError in CI.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
* fix: update FBGEMM submodule to include hstu_ops_gpu.py fake impl
Update from 04df536 to 65bad42 which adds fake tensor implementations
for torch.export (hstu_ops_gpu.py). This was missing since PR #340
accidentally reverted the submodule pointer.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
* ci: allow /build with flags by matching prefix instead of exact string
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
* ci: remove pull_request_target trigger, keep only /build comment
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
---------
Co-authored-by: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-authored-by: Junyi Qiu <junyiq@nvidia.com>
0 commit comments