-
Notifications
You must be signed in to change notification settings - Fork 54
Expand file tree
/
Copy path__init__.py
More file actions
76 lines (71 loc) · 2.1 KB
/
__init__.py
File metadata and controls
76 lines (71 loc) · 2.1 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
# SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# SPDX-License-Identifier: MIT
"""cuTile backend implementations for all TileGym operations"""
from tilegym.backend import is_backend_available
# Only import if cutile backend is available
if is_backend_available("cutile"):
# Activation functions
# Non-DL operations
# Linear algebra operations
# NN operations
from . import activation
from . import attention
from . import dropout
from . import flash_decode
from . import group_gemm
from . import matmul
from . import mla
from . import mla_decoding
from . import mla_decoding_split_kv
from . import moe
from . import moe_align_block
from . import rms_norm
from . import rope
from . import silu_and_mul
from . import softmax
from . import splitk_reduce
from . import swiglu
# Import specific functions for direct access
from .flash_decode import fmha_decode
from .moe import fused_moe_kernel as invoke_fused_moe_kernel
from .moe_align_block import moe_align_block_size
from .rms_norm import get_rms_norm_module
from .rms_norm import rms_norm
from .rope import apply_rope_base
from .rope import get_apply_rope_func
from .silu_and_mul import silu_and_mul
from .softmax import softmax
from .splitk_reduce import splitk_reduce
from .swiglu import get_swiglu
from .swiglu import get_swiglu_module
__all__ = [
# NN operations
"fmha_decode",
"flash_decode",
"splitk_reduce",
"invoke_fused_moe_kernel",
"moe_align_block_size",
"attention",
"mla",
"mla_decoding",
"get_swiglu_module",
"get_swiglu",
"get_apply_rope_func",
"get_rms_norm_module",
"rms_norm",
"silu_and_mul",
"dropout",
"softmax",
"mla_decoding_split_kv",
"moe",
"moe_align_block",
"rope",
"swiglu",
"apply_rope_base",
# Linalg operations
"matmul",
"group_gemm",
]
else:
__all__ = []