We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 0ee663d commit 4eface0Copy full SHA for 4eface0
test/test_gpu/skip_tests_h100_pytorch.yaml
@@ -5,8 +5,6 @@
5
# op-name: to skip an entire operator
6
# op-name:\n\t- impl-name to skip an impl
7
flash_attention:
8
- # thunderkittens cannot handle the default input shapes
9
- - tk
10
# tma API changed in upstream
11
- triton_tutorial_flash_v2_tma
12
# triton_tutorial_*_ws kernels require triton-main
test/test_gpu/skip_tests_h100_triton_main.yaml
# _ws kernels require Triton with warp specialization
- triton_tutorial_flash_v2_ws
- triton_tutorial_flash_v2_tma_ws
0 commit comments