kernels-bot commited on
Commit
284c729
·
verified ·
1 Parent(s): 0a66e91

Uploaded using `kernel-builder`.

Browse files
.gitattributes CHANGED
@@ -108,3 +108,8 @@ build/torch210-cxx11-cu130-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so fil
108
  build/torch211-cxx11-cu128-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
109
  build/torch211-cxx11-cu130-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
110
  build/torch29-cxx11-cu129-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
108
  build/torch211-cxx11-cu128-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
109
  build/torch211-cxx11-cu130-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
110
  build/torch29-cxx11-cu129-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
111
+ build/torch210-cxx11-cu128-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
112
+ build/torch210-cxx11-cu130-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
113
+ build/torch211-cxx11-cu128-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
114
+ build/torch211-cxx11-cu130-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
115
+ build/torch29-cxx11-cu129-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
build/torch210-cxx11-cu128-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _sage_attention_cuda_5818884
3
- ops = torch.ops._sage_attention_cuda_5818884
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_sage_attention_cuda_5818884::{op_name}"
 
1
  import torch
2
+ from . import _sage_attention_cuda_6b6ce0f
3
+ ops = torch.ops._sage_attention_cuda_6b6ce0f
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_sage_attention_cuda_6b6ce0f::{op_name}"
build/torch210-cxx11-cu128-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ba419376bb9bfed00cadf4480c7d66c746c46b2b61105e7139403dffec136f7
3
  size 33330136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:553c3154f69cbe04ea53be416f630ceaa80cba65f02d9b53d1ae9f716913bf6b
3
  size 33330136
build/torch210-cxx11-cu130-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _sage_attention_cuda_5818884
3
- ops = torch.ops._sage_attention_cuda_5818884
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_sage_attention_cuda_5818884::{op_name}"
 
1
  import torch
2
+ from . import _sage_attention_cuda_6b6ce0f
3
+ ops = torch.ops._sage_attention_cuda_6b6ce0f
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_sage_attention_cuda_6b6ce0f::{op_name}"
build/torch210-cxx11-cu130-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e207e98fa54466498e9d8eb376ecf85df6240b020a49106b74e3f02e6417323b
3
  size 33810512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e1c89a1e6fb0108792fdf2706c6a181d3298ec1c3c45d9413fa2c7f7c116b5e
3
  size 33810512
build/torch211-cxx11-cu128-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _sage_attention_cuda_5818884
3
- ops = torch.ops._sage_attention_cuda_5818884
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_sage_attention_cuda_5818884::{op_name}"
 
1
  import torch
2
+ from . import _sage_attention_cuda_6b6ce0f
3
+ ops = torch.ops._sage_attention_cuda_6b6ce0f
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_sage_attention_cuda_6b6ce0f::{op_name}"
build/torch211-cxx11-cu128-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c275014c7585e628308c4d737d9589d26266f10c6178957ff241da4c59489e6
3
  size 33326264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0abf9c6ea86fe2a7655b6258291fe16bf6f7f5bf8625f93f6fe8e725488f981d
3
  size 33326264
build/torch211-cxx11-cu130-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _sage_attention_cuda_5818884
3
- ops = torch.ops._sage_attention_cuda_5818884
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_sage_attention_cuda_5818884::{op_name}"
 
1
  import torch
2
+ from . import _sage_attention_cuda_6b6ce0f
3
+ ops = torch.ops._sage_attention_cuda_6b6ce0f
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_sage_attention_cuda_6b6ce0f::{op_name}"
build/torch211-cxx11-cu130-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d41e9393dd1f29c28461831c1aeb4598cea7dbd51f0c171ea1f668ed812eed67
3
  size 33806648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3501aa9e3dfd9846a2b23c43f7210e4066c2d05b2289360d4cc4f38c75e03e58
3
  size 33806648
build/torch29-cxx11-cu129-aarch64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _sage_attention_cuda_5818884
3
- ops = torch.ops._sage_attention_cuda_5818884
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_sage_attention_cuda_5818884::{op_name}"
 
1
  import torch
2
+ from . import _sage_attention_cuda_6b6ce0f
3
+ ops = torch.ops._sage_attention_cuda_6b6ce0f
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_sage_attention_cuda_6b6ce0f::{op_name}"
build/torch29-cxx11-cu129-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcd0b4dc30126b840c1a40b05954f412d82ba0c011071d2b337cc84da69d003d
3
  size 33657728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f2e6fc2ac41520e92a88e1c724d6a5182bdabea7e482e32bcbc77889a516eaa
3
  size 33657728