drbh commited on
Commit
f475609
·
1 Parent(s): d76b04d

feat: rebuild outputs

Browse files
Files changed (44) hide show
  1. build/torch25-cxx11-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
  2. build/torch25-cxx11-cu118-x86_64-linux/flash_mla/{_flash_mla_y6bdeh54o26h6.abi3.so → _flash_mla_xw2uadieawdvi.abi3.so} +2 -2
  3. build/torch25-cxx11-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
  4. build/torch25-cxx11-cu121-x86_64-linux/flash_mla/__init__.py +1 -4
  5. build/torch25-cxx11-cu121-x86_64-linux/flash_mla/{_flash_mla_mytbuokq46mgm.abi3.so → _flash_mla_liludkqplkha6.abi3.so} +2 -2
  6. build/torch25-cxx11-cu121-x86_64-linux/flash_mla/_ops.py +3 -3
  7. build/torch25-cxx11-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
  8. build/torch25-cxx11-cu124-x86_64-linux/flash_mla/{_flash_mla_iohry4qbuggqa.abi3.so → _flash_mla_rqg4eobrk7hgm.abi3.so} +2 -2
  9. build/torch25-cxx11-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
  10. build/torch25-cxx98-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
  11. build/torch25-cxx98-cu118-x86_64-linux/flash_mla/{_flash_mla_ks7izefym4ha2.abi3.so → _flash_mla_xc7pmvlhanmng.abi3.so} +2 -2
  12. build/torch25-cxx98-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
  13. build/torch25-cxx98-cu121-x86_64-linux/flash_mla/__init__.py +1 -4
  14. build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_cd6mvrbov7aye.abi3.so +0 -3
  15. build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_ooalrhhwrrl42.abi3.so +3 -0
  16. build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_ops.py +3 -3
  17. build/torch25-cxx98-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
  18. build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_muvqsop7ydtdg.abi3.so +0 -3
  19. build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_r4jls6fpvnpkq.abi3.so +3 -0
  20. build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
  21. build/torch26-cxx11-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
  22. build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_4gt6haj2eaeve.abi3.so +0 -3
  23. build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_eae4qwas6rh2a.abi3.so +3 -0
  24. build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
  25. build/torch26-cxx11-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
  26. build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_defqzbdfoejoc.abi3.so +3 -0
  27. build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_fqlv2wdc4vwcs.abi3.so +0 -3
  28. build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
  29. build/torch26-cxx11-cu126-x86_64-linux/flash_mla/__init__.py +1 -4
  30. build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_b4xwzd3vzzbdy.abi3.so +0 -3
  31. build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_ozsllcybel7au.abi3.so +3 -0
  32. build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_ops.py +3 -3
  33. build/torch26-cxx98-cu118-x86_64-linux/flash_mla/__init__.py +1 -4
  34. build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_hxgeqtwgn6oxm.abi3.so +3 -0
  35. build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_iyfleyrgf34la.abi3.so +0 -3
  36. build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_ops.py +3 -3
  37. build/torch26-cxx98-cu124-x86_64-linux/flash_mla/__init__.py +1 -4
  38. build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_fywd4tw2yn6ew.abi3.so +0 -3
  39. build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_oza65fumibsj4.abi3.so +3 -0
  40. build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_ops.py +3 -3
  41. build/torch26-cxx98-cu126-x86_64-linux/flash_mla/__init__.py +1 -4
  42. build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_duoktlzcb6q5w.abi3.so +0 -3
  43. build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_zcupqkuarp5xg.abi3.so +3 -0
  44. build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_ops.py +3 -3
build/torch25-cxx11-cu118-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch25-cxx11-cu118-x86_64-linux/flash_mla/{_flash_mla_y6bdeh54o26h6.abi3.so → _flash_mla_xw2uadieawdvi.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c90743918a3a5a7e3f21c3de54c524af2d04d5bf112b9b638ac8b05c5009538
3
- size 2558728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dda33043e05beab8819043b9aac8d73b71ca27b49a443618d8c828bb286eac93
3
+ size 2558664
build/torch25-cxx11-cu118-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_y6bdeh54o26h6
3
- ops = torch.ops._flash_mla_y6bdeh54o26h6
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_y6bdeh54o26h6::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_xw2uadieawdvi
3
+ ops = torch.ops._flash_mla_xw2uadieawdvi
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_xw2uadieawdvi::{op_name}"
build/torch25-cxx11-cu121-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch25-cxx11-cu121-x86_64-linux/flash_mla/{_flash_mla_mytbuokq46mgm.abi3.so → _flash_mla_liludkqplkha6.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d20c322601136e544cd1995cee8b21098f6cbc4279b7ebd8e007c59150d5df9
3
- size 2593856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f5e56f6e252f51ebc7e7bbfdb4724c9f29441b00dd9d17f557272b6f423df74
3
+ size 2589688
build/torch25-cxx11-cu121-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_mytbuokq46mgm
3
- ops = torch.ops._flash_mla_mytbuokq46mgm
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_mytbuokq46mgm::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_liludkqplkha6
3
+ ops = torch.ops._flash_mla_liludkqplkha6
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_liludkqplkha6::{op_name}"
build/torch25-cxx11-cu124-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch25-cxx11-cu124-x86_64-linux/flash_mla/{_flash_mla_iohry4qbuggqa.abi3.so → _flash_mla_rqg4eobrk7hgm.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccb2bc37dac026a6d0aff1fa3069d07fb3e90d3a0d88bb985508549918e1b454
3
- size 2574832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59123e3564bf077d73ad747c0f9e7d53e1314d75f13439c7b8bbb35c30074821
3
+ size 2574760
build/torch25-cxx11-cu124-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_iohry4qbuggqa
3
- ops = torch.ops._flash_mla_iohry4qbuggqa
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_iohry4qbuggqa::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_rqg4eobrk7hgm
3
+ ops = torch.ops._flash_mla_rqg4eobrk7hgm
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_rqg4eobrk7hgm::{op_name}"
build/torch25-cxx98-cu118-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch25-cxx98-cu118-x86_64-linux/flash_mla/{_flash_mla_ks7izefym4ha2.abi3.so → _flash_mla_xc7pmvlhanmng.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:770a95aee0d760d66c11cee67284473199f49ed0a37264c9e98911e520653e5d
3
- size 2551384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0304190c33988637aa9df9fca1e857ddfca5b944765012eaf498f708001852a4
3
+ size 2551312
build/torch25-cxx98-cu118-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_ks7izefym4ha2
3
- ops = torch.ops._flash_mla_ks7izefym4ha2
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_ks7izefym4ha2::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_xc7pmvlhanmng
3
+ ops = torch.ops._flash_mla_xc7pmvlhanmng
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_xc7pmvlhanmng::{op_name}"
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_cd6mvrbov7aye.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1dc0f7b58d315d7f713b383732f4866df52131833afeb02fb8c8c0290b867ec9
3
- size 2590688
 
 
 
 
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_flash_mla_ooalrhhwrrl42.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:426ad2a900df3aaaa0b9e2e4f29927fe4c8f25c97b561693a140856e5b35008a
3
+ size 2586520
build/torch25-cxx98-cu121-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_cd6mvrbov7aye
3
- ops = torch.ops._flash_mla_cd6mvrbov7aye
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_cd6mvrbov7aye::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_ooalrhhwrrl42
3
+ ops = torch.ops._flash_mla_ooalrhhwrrl42
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_ooalrhhwrrl42::{op_name}"
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_muvqsop7ydtdg.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:10bbc09993e05062a2e257ea388e3aeaee1c1c54960ac352857bb921973f5dc8
3
- size 2571656
 
 
 
 
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_r4jls6fpvnpkq.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21727612adfc1c93c412eb639687311f5e848d704921cd623fc3ef4d9be5eb77
3
+ size 2571592
build/torch25-cxx98-cu124-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_muvqsop7ydtdg
3
- ops = torch.ops._flash_mla_muvqsop7ydtdg
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_muvqsop7ydtdg::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_r4jls6fpvnpkq
3
+ ops = torch.ops._flash_mla_r4jls6fpvnpkq
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_r4jls6fpvnpkq::{op_name}"
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_4gt6haj2eaeve.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7b1010f156fb14c68680155d12b6415ebaf818eb8f8a162b8b1cd0f89f085a4
3
- size 2563176
 
 
 
 
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_flash_mla_eae4qwas6rh2a.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d8b84d1bbfc70c3cf8ccd654df16a4691aa42bd6963a715d77a19ebd89c2a77
3
+ size 2559008
build/torch26-cxx11-cu118-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_4gt6haj2eaeve
3
- ops = torch.ops._flash_mla_4gt6haj2eaeve
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_4gt6haj2eaeve::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_eae4qwas6rh2a
3
+ ops = torch.ops._flash_mla_eae4qwas6rh2a
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_eae4qwas6rh2a::{op_name}"
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_defqzbdfoejoc.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a49b08acd8e9ac8b8da3cdc2c448acc1476d03a2afe64843bf16f593f59c858
3
+ size 2575104
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_flash_mla_fqlv2wdc4vwcs.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:13d1bfb8d6fa769168838204b8bcc62bc4c48ed81275e8a75d18f467dab352a0
3
- size 2575176
 
 
 
 
build/torch26-cxx11-cu124-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_fqlv2wdc4vwcs
3
- ops = torch.ops._flash_mla_fqlv2wdc4vwcs
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_fqlv2wdc4vwcs::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_defqzbdfoejoc
3
+ ops = torch.ops._flash_mla_defqzbdfoejoc
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_defqzbdfoejoc::{op_name}"
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_b4xwzd3vzzbdy.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c6cfeb75b5f9c8b88624f573fe80afc3f4d9f08828630c64c84e8100afe1101
3
- size 2555136
 
 
 
 
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_flash_mla_ozsllcybel7au.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a63288318d7a76761ea5c89a34d9648132010929f4778e81418776e9f14ade49
3
+ size 2555072
build/torch26-cxx11-cu126-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_b4xwzd3vzzbdy
3
- ops = torch.ops._flash_mla_b4xwzd3vzzbdy
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_b4xwzd3vzzbdy::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_ozsllcybel7au
3
+ ops = torch.ops._flash_mla_ozsllcybel7au
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_ozsllcybel7au::{op_name}"
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_hxgeqtwgn6oxm.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92823f6df4d684d0d6fe96b20ff9580917a5f8f5f97a5e751a37686adfaa9ef3
3
+ size 2547560
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_flash_mla_iyfleyrgf34la.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:25a5dc0c023db2b535161f494470ebbf0cc0c261028906259db65cfa396fd6a5
3
- size 2551728
 
 
 
 
build/torch26-cxx98-cu118-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_iyfleyrgf34la
3
- ops = torch.ops._flash_mla_iyfleyrgf34la
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_iyfleyrgf34la::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_hxgeqtwgn6oxm
3
+ ops = torch.ops._flash_mla_hxgeqtwgn6oxm
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_hxgeqtwgn6oxm::{op_name}"
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_fywd4tw2yn6ew.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:59ea97594e0550f64aa3b5ad0358e68306f284344acef250e0293ccf50e87273
3
- size 2572008
 
 
 
 
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_flash_mla_oza65fumibsj4.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8afed8db3008d659356aeba76d163cde31d725005af54039a1145648a0a9710
3
+ size 2571936
build/torch26-cxx98-cu124-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_fywd4tw2yn6ew
3
- ops = torch.ops._flash_mla_fywd4tw2yn6ew
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_fywd4tw2yn6ew::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_oza65fumibsj4
3
+ ops = torch.ops._flash_mla_oza65fumibsj4
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_oza65fumibsj4::{op_name}"
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/__init__.py CHANGED
@@ -19,8 +19,6 @@ def mha_fwd_kvcache_mla(
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
22
- # TODO: remove when resolved
23
- unknown_param = 0
24
  return ops.mha_fwd_kvcache_mla(
25
  q,
26
  kcache,
@@ -31,6 +29,5 @@ def mha_fwd_kvcache_mla(
31
  softmax_scale,
32
  is_causal_,
33
  tile_scheduler_metadata,
34
- num_splits,
35
- unknown_param,
36
  )
 
19
  tile_scheduler_metadata: torch.Tensor,
20
  num_splits: torch.Tensor,
21
  ) -> torch.Tensor:
 
 
22
  return ops.mha_fwd_kvcache_mla(
23
  q,
24
  kcache,
 
29
  softmax_scale,
30
  is_causal_,
31
  tile_scheduler_metadata,
32
+ num_splits
 
33
  )
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_duoktlzcb6q5w.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e85fa15e0103badddb1cf6f894e06cf42ce1300431066af499feae6e5df4c66b
3
- size 2551968
 
 
 
 
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_flash_mla_zcupqkuarp5xg.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b106cca9c05188c533fbffdfbf0c6237285d41926e8869f2a4dfe6a32c52626a
3
+ size 2551896
build/torch26-cxx98-cu126-x86_64-linux/flash_mla/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _flash_mla_duoktlzcb6q5w
3
- ops = torch.ops._flash_mla_duoktlzcb6q5w
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_flash_mla_duoktlzcb6q5w::{op_name}"
 
1
  import torch
2
+ from . import _flash_mla_zcupqkuarp5xg
3
+ ops = torch.ops._flash_mla_zcupqkuarp5xg
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_flash_mla_zcupqkuarp5xg::{op_name}"