Uploaded using `kernel-builder`.
Browse files- .gitattributes +5 -0
- build/torch210-cxx11-cu128-aarch64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu128-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} +1 -1
- build/torch210-cxx11-cu130-aarch64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu130-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} +1 -1
- build/torch211-cxx11-cu128-aarch64-linux/_ops.py +3 -3
- build/torch211-cxx11-cu128-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} +1 -1
- build/torch211-cxx11-cu130-aarch64-linux/_ops.py +3 -3
- build/torch211-cxx11-cu130-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} +1 -1
- build/torch29-cxx11-cu129-aarch64-linux/_ops.py +3 -3
- build/torch29-cxx11-cu129-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so} +1 -1
.gitattributes
CHANGED
|
@@ -108,3 +108,8 @@ build/torch210-cxx11-cu130-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so fil
|
|
| 108 |
build/torch211-cxx11-cu128-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 109 |
build/torch211-cxx11-cu130-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 110 |
build/torch29-cxx11-cu129-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 108 |
build/torch211-cxx11-cu128-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 109 |
build/torch211-cxx11-cu130-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 110 |
build/torch29-cxx11-cu129-x86_64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 111 |
+
build/torch210-cxx11-cu128-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 112 |
+
build/torch210-cxx11-cu130-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 113 |
+
build/torch211-cxx11-cu128-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 114 |
+
build/torch211-cxx11-cu130-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 115 |
+
build/torch29-cxx11-cu129-aarch64-linux/_sage_attention_cuda_6b6ce0f.abi3.so filter=lfs diff=lfs merge=lfs -text
|
build/torch210-cxx11-cu128-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _sage_attention_cuda_6b6ce0f
|
| 3 |
+
ops = torch.ops._sage_attention_cuda_6b6ce0f
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_sage_attention_cuda_6b6ce0f::{op_name}"
|
build/torch210-cxx11-cu128-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 33330136
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:553c3154f69cbe04ea53be416f630ceaa80cba65f02d9b53d1ae9f716913bf6b
|
| 3 |
size 33330136
|
build/torch210-cxx11-cu130-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _sage_attention_cuda_6b6ce0f
|
| 3 |
+
ops = torch.ops._sage_attention_cuda_6b6ce0f
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_sage_attention_cuda_6b6ce0f::{op_name}"
|
build/torch210-cxx11-cu130-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 33810512
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e1c89a1e6fb0108792fdf2706c6a181d3298ec1c3c45d9413fa2c7f7c116b5e
|
| 3 |
size 33810512
|
build/torch211-cxx11-cu128-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _sage_attention_cuda_6b6ce0f
|
| 3 |
+
ops = torch.ops._sage_attention_cuda_6b6ce0f
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_sage_attention_cuda_6b6ce0f::{op_name}"
|
build/torch211-cxx11-cu128-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 33326264
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0abf9c6ea86fe2a7655b6258291fe16bf6f7f5bf8625f93f6fe8e725488f981d
|
| 3 |
size 33326264
|
build/torch211-cxx11-cu130-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _sage_attention_cuda_6b6ce0f
|
| 3 |
+
ops = torch.ops._sage_attention_cuda_6b6ce0f
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_sage_attention_cuda_6b6ce0f::{op_name}"
|
build/torch211-cxx11-cu130-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 33806648
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3501aa9e3dfd9846a2b23c43f7210e4066c2d05b2289360d4cc4f38c75e03e58
|
| 3 |
size 33806648
|
build/torch29-cxx11-cu129-aarch64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _sage_attention_cuda_6b6ce0f
|
| 3 |
+
ops = torch.ops._sage_attention_cuda_6b6ce0f
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_sage_attention_cuda_6b6ce0f::{op_name}"
|
build/torch29-cxx11-cu129-aarch64-linux/{_sage_attention_cuda_5818884.abi3.so → _sage_attention_cuda_6b6ce0f.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 33657728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5f2e6fc2ac41520e92a88e1c724d6a5182bdabea7e482e32bcbc77889a516eaa
|
| 3 |
size 33657728
|