drbh
commited on
Commit
·
f475609
1
Parent(s):
d76b04d
feat: rebuild outputs
Browse files- build/torch25-cxx11-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch25-cxx11-cu118-x86_64-linux/flash_mla/{_flash_mla_y6bdeh54o26h6.abi3.so → _flash_mla_xw2uadieawdvi.abi3.so} +2 -2
- build/torch25-cxx11-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch25-cxx11-cu121-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch25-cxx11-cu121-x86_64-linux/flash_mla/{_flash_mla_mytbuokq46mgm.abi3.so → _flash_mla_liludkqplkha6.abi3.so} +2 -2
- build/torch25-cxx11-cu121-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch25-cxx11-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch25-cxx11-cu124-x86_64-linux/flash_mla/{_flash_mla_iohry4qbuggqa.abi3.so → _flash_mla_rqg4eobrk7hgm.abi3.so} +2 -2
- build/torch25-cxx11-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch25-cxx98-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch25-cxx98-cu118-x86_64-linux/flash_mla/{_flash_mla_ks7izefym4ha2.abi3.so → _flash_mla_xc7pmvlhanmng.abi3.so} +2 -2
- build/torch25-cxx98-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch25-cxx98-cu121-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_cd6mvrbov7aye.abi3.so +0 -3
- build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_ooalrhhwrrl42.abi3.so +3 -0
- build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch25-cxx98-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_muvqsop7ydtdg.abi3.so +0 -3
- build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_r4jls6fpvnpkq.abi3.so +3 -0
- build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch26-cxx11-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_4gt6haj2eaeve.abi3.so +0 -3
- build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_eae4qwas6rh2a.abi3.so +3 -0
- build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch26-cxx11-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_defqzbdfoejoc.abi3.so +3 -0
- build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_fqlv2wdc4vwcs.abi3.so +0 -3
- build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch26-cxx11-cu126-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_b4xwzd3vzzbdy.abi3.so +0 -3
- build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_ozsllcybel7au.abi3.so +3 -0
- build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch26-cxx98-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_hxgeqtwgn6oxm.abi3.so +3 -0
- build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_iyfleyrgf34la.abi3.so +0 -3
- build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch26-cxx98-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_fywd4tw2yn6ew.abi3.so +0 -3
- build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_oza65fumibsj4.abi3.so +3 -0
- build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
- build/torch26-cxx98-cu126-x86_64-linux/flash_mla/__init__.py +1 -4
- build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_duoktlzcb6q5w.abi3.so +0 -3
- build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_zcupqkuarp5xg.abi3.so +3 -0
- build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_ops.py +3 -3
build/torch25-cxx11-cu118-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch25-cxx11-cu118-x86_64-linux/flash_mla/{_flash_mla_y6bdeh54o26h6.abi3.so → _flash_mla_xw2uadieawdvi.abi3.so}
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dda33043e05beab8819043b9aac8d73b71ca27b49a443618d8c828bb286eac93
|
3 |
+
size 2558664
|
build/torch25-cxx11-cu118-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_xw2uadieawdvi
|
3 |
+
ops = torch.ops._flash_mla_xw2uadieawdvi
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_xw2uadieawdvi::{op_name}"
|
build/torch25-cxx11-cu121-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch25-cxx11-cu121-x86_64-linux/flash_mla/{_flash_mla_mytbuokq46mgm.abi3.so → _flash_mla_liludkqplkha6.abi3.so}
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f5e56f6e252f51ebc7e7bbfdb4724c9f29441b00dd9d17f557272b6f423df74
|
3 |
+
size 2589688
|
build/torch25-cxx11-cu121-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_liludkqplkha6
|
3 |
+
ops = torch.ops._flash_mla_liludkqplkha6
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_liludkqplkha6::{op_name}"
|
build/torch25-cxx11-cu124-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch25-cxx11-cu124-x86_64-linux/flash_mla/{_flash_mla_iohry4qbuggqa.abi3.so → _flash_mla_rqg4eobrk7hgm.abi3.so}
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59123e3564bf077d73ad747c0f9e7d53e1314d75f13439c7b8bbb35c30074821
|
3 |
+
size 2574760
|
build/torch25-cxx11-cu124-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_rqg4eobrk7hgm
|
3 |
+
ops = torch.ops._flash_mla_rqg4eobrk7hgm
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_rqg4eobrk7hgm::{op_name}"
|
build/torch25-cxx98-cu118-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch25-cxx98-cu118-x86_64-linux/flash_mla/{_flash_mla_ks7izefym4ha2.abi3.so → _flash_mla_xc7pmvlhanmng.abi3.so}
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0304190c33988637aa9df9fca1e857ddfca5b944765012eaf498f708001852a4
|
3 |
+
size 2551312
|
build/torch25-cxx98-cu118-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_xc7pmvlhanmng
|
3 |
+
ops = torch.ops._flash_mla_xc7pmvlhanmng
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_xc7pmvlhanmng::{op_name}"
|
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_cd6mvrbov7aye.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:1dc0f7b58d315d7f713b383732f4866df52131833afeb02fb8c8c0290b867ec9
|
3 |
-
size 2590688
|
|
|
|
|
|
|
|
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_ooalrhhwrrl42.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:426ad2a900df3aaaa0b9e2e4f29927fe4c8f25c97b561693a140856e5b35008a
|
3 |
+
size 2586520
|
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_ooalrhhwrrl42
|
3 |
+
ops = torch.ops._flash_mla_ooalrhhwrrl42
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_ooalrhhwrrl42::{op_name}"
|
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_muvqsop7ydtdg.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:10bbc09993e05062a2e257ea388e3aeaee1c1c54960ac352857bb921973f5dc8
|
3 |
-
size 2571656
|
|
|
|
|
|
|
|
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_r4jls6fpvnpkq.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21727612adfc1c93c412eb639687311f5e848d704921cd623fc3ef4d9be5eb77
|
3 |
+
size 2571592
|
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_r4jls6fpvnpkq
|
3 |
+
ops = torch.ops._flash_mla_r4jls6fpvnpkq
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_r4jls6fpvnpkq::{op_name}"
|
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_4gt6haj2eaeve.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:b7b1010f156fb14c68680155d12b6415ebaf818eb8f8a162b8b1cd0f89f085a4
|
3 |
-
size 2563176
|
|
|
|
|
|
|
|
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_eae4qwas6rh2a.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d8b84d1bbfc70c3cf8ccd654df16a4691aa42bd6963a715d77a19ebd89c2a77
|
3 |
+
size 2559008
|
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_eae4qwas6rh2a
|
3 |
+
ops = torch.ops._flash_mla_eae4qwas6rh2a
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_eae4qwas6rh2a::{op_name}"
|
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_defqzbdfoejoc.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a49b08acd8e9ac8b8da3cdc2c448acc1476d03a2afe64843bf16f593f59c858
|
3 |
+
size 2575104
|
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_fqlv2wdc4vwcs.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:13d1bfb8d6fa769168838204b8bcc62bc4c48ed81275e8a75d18f467dab352a0
|
3 |
-
size 2575176
|
|
|
|
|
|
|
|
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_defqzbdfoejoc
|
3 |
+
ops = torch.ops._flash_mla_defqzbdfoejoc
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_defqzbdfoejoc::{op_name}"
|
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_b4xwzd3vzzbdy.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:8c6cfeb75b5f9c8b88624f573fe80afc3f4d9f08828630c64c84e8100afe1101
|
3 |
-
size 2555136
|
|
|
|
|
|
|
|
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_ozsllcybel7au.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a63288318d7a76761ea5c89a34d9648132010929f4778e81418776e9f14ade49
|
3 |
+
size 2555072
|
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_ozsllcybel7au
|
3 |
+
ops = torch.ops._flash_mla_ozsllcybel7au
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_ozsllcybel7au::{op_name}"
|
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_hxgeqtwgn6oxm.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92823f6df4d684d0d6fe96b20ff9580917a5f8f5f97a5e751a37686adfaa9ef3
|
3 |
+
size 2547560
|
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_iyfleyrgf34la.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:25a5dc0c023db2b535161f494470ebbf0cc0c261028906259db65cfa396fd6a5
|
3 |
-
size 2551728
|
|
|
|
|
|
|
|
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_hxgeqtwgn6oxm
|
3 |
+
ops = torch.ops._flash_mla_hxgeqtwgn6oxm
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_hxgeqtwgn6oxm::{op_name}"
|
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_fywd4tw2yn6ew.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:59ea97594e0550f64aa3b5ad0358e68306f284344acef250e0293ccf50e87273
|
3 |
-
size 2572008
|
|
|
|
|
|
|
|
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_oza65fumibsj4.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8afed8db3008d659356aeba76d163cde31d725005af54039a1145648a0a9710
|
3 |
+
size 2571936
|
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_oza65fumibsj4
|
3 |
+
ops = torch.ops._flash_mla_oza65fumibsj4
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_oza65fumibsj4::{op_name}"
|
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/__init__.py
CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
22 |
-
# TODO: remove when resolved
|
23 |
-
unknown_param = 0
|
24 |
return ops.mha_fwd_kvcache_mla(
|
25 |
q,
|
26 |
kcache,
|
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
|
|
31 |
softmax_scale,
|
32 |
is_causal_,
|
33 |
tile_scheduler_metadata,
|
34 |
-
num_splits
|
35 |
-
unknown_param,
|
36 |
)
|
|
|
19 |
tile_scheduler_metadata: torch.Tensor,
|
20 |
num_splits: torch.Tensor,
|
21 |
) -> torch.Tensor:
|
|
|
|
|
22 |
return ops.mha_fwd_kvcache_mla(
|
23 |
q,
|
24 |
kcache,
|
|
|
29 |
softmax_scale,
|
30 |
is_causal_,
|
31 |
tile_scheduler_metadata,
|
32 |
+
num_splits
|
|
|
33 |
)
|
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_duoktlzcb6q5w.abi3.so
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:e85fa15e0103badddb1cf6f894e06cf42ce1300431066af499feae6e5df4c66b
|
3 |
-
size 2551968
|
|
|
|
|
|
|
|
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_zcupqkuarp5xg.abi3.so
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b106cca9c05188c533fbffdfbf0c6237285d41926e8869f2a4dfe6a32c52626a
|
3 |
+
size 2551896
|
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_ops.py
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
import torch
|
2 |
-
from . import
|
3 |
-
ops = torch.ops.
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
-
return f"
|
|
|
1 |
import torch
|
2 |
+
from . import _flash_mla_zcupqkuarp5xg
|
3 |
+
ops = torch.ops._flash_mla_zcupqkuarp5xg
|
4 |
|
5 |
def add_op_namespace_prefix(op_name: str):
|
6 |
"""
|
7 |
Prefix op by namespace.
|
8 |
"""
|
9 |
+
return f"_flash_mla_zcupqkuarp5xg::{op_name}"
|