Build uploaded using `kernels`.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +14 -0
- build/torch210-cxx11-cpu-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-cpu-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-cpu-x86_64-linux/{_relu_cpu_smcctie2ikqu4.abi3.so → _relu_cpu_6261c06.abi3.so} +2 -2
- build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-cpu-x86_64-linux/metadata.json +1 -0
- build/torch210-cxx11-cu126-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-cu126-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu126-x86_64-linux/{_relu_cuda_byb5bdzb6z5zm.abi3.so → _relu_cuda_6261c06.abi3.so} +2 -2
- build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-cu126-x86_64-linux/metadata.json +16 -2
- build/torch210-cxx11-cu128-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-cu128-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu128-x86_64-linux/{_relu_cuda_ohbj3nuuoamci.abi3.so → _relu_cuda_6261c06.abi3.so} +2 -2
- build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-cu128-x86_64-linux/metadata.json +19 -2
- build/torch210-cxx11-cu130-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-cu130-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-cu130-x86_64-linux/{_relu_cuda_oheoc6gkfjkwm.abi3.so → _relu_cuda_6261c06.abi3.so} +2 -2
- build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-cu130-x86_64-linux/metadata.json +17 -2
- build/torch210-cxx11-rocm70-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-rocm70-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so +3 -0
- build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so +0 -3
- build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-rocm70-x86_64-linux/metadata.json +15 -2
- build/torch210-cxx11-rocm71-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-rocm71-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so +3 -0
- build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so +0 -3
- build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-rocm71-x86_64-linux/metadata.json +15 -2
- build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py +5 -1
- build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py +3 -3
- build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so +3 -0
- build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so +0 -3
- build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py +11 -0
- build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json +1 -0
- build/torch29-cxx11-cpu-x86_64-linux/__init__.py +5 -1
- build/torch29-cxx11-cpu-x86_64-linux/_ops.py +3 -3
- build/torch29-cxx11-cpu-x86_64-linux/{_relu_cpu_liz2pcyxsjqzq.abi3.so → _relu_cpu_6261c06.abi3.so} +2 -2
- build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py +11 -0
- build/torch29-cxx11-cpu-x86_64-linux/metadata.json +1 -0
- build/torch29-cxx11-cu126-x86_64-linux/__init__.py +5 -1
- build/torch29-cxx11-cu126-x86_64-linux/_ops.py +3 -3
- build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so +3 -0
- build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so +0 -3
- build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py +11 -0
- build/torch29-cxx11-cu126-x86_64-linux/metadata.json +16 -2
.gitattributes
CHANGED
|
@@ -259,3 +259,17 @@ build/torch29-cxx11-cu126-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs di
|
|
| 259 |
build/torch29-cxx11-cu128-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 260 |
build/torch29-cxx11-cu130-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 261 |
build/torch210-cu128-x86_64-windows/_relu_cuda_657300a.pyd filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 259 |
build/torch29-cxx11-cu128-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 260 |
build/torch29-cxx11-cu130-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 261 |
build/torch210-cu128-x86_64-windows/_relu_cuda_657300a.pyd filter=lfs diff=lfs merge=lfs -text
|
| 262 |
+
build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 263 |
+
build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 264 |
+
build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 265 |
+
build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 266 |
+
build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 267 |
+
build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 268 |
+
build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 269 |
+
build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 270 |
+
build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 271 |
+
build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 272 |
+
build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 273 |
+
build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 274 |
+
build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
| 275 |
+
build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
|
build/torch210-cxx11-cpu-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-cpu-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_cpu_6261c06
|
| 3 |
+
ops = torch.ops._relu_cpu_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_cpu_6261c06::{op_name}"
|
build/torch210-cxx11-cpu-x86_64-linux/{_relu_cpu_smcctie2ikqu4.abi3.so → _relu_cpu_6261c06.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9bc51be4e362c70be9c87e4e3e8464904547c487496ce8a0ebe552c5fa6effab
|
| 3 |
+
size 1778080
|
build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-cpu-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
|
|
|
| 3 |
"python-depends": []
|
| 4 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
"python-depends": []
|
| 5 |
}
|
build/torch210-cxx11-cu126-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-cu126-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_cuda_6261c06
|
| 3 |
+
ops = torch.ops._relu_cuda_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_cuda_6261c06::{op_name}"
|
build/torch210-cxx11-cu126-x86_64-linux/{_relu_cuda_byb5bdzb6z5zm.abi3.so → _relu_cuda_6261c06.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3aaf0a90799a5102aba64409350132a8bba560a8816e00be51a8b5f41f9475d
|
| 3 |
+
size 2019880
|
build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-cu126-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,18 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
-
"
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
+
"python-depends": [],
|
| 5 |
+
"backend": {
|
| 6 |
+
"type": "cuda",
|
| 7 |
+
"archs": [
|
| 8 |
+
"7.0",
|
| 9 |
+
"7.2",
|
| 10 |
+
"7.5",
|
| 11 |
+
"8.0",
|
| 12 |
+
"8.6",
|
| 13 |
+
"8.7",
|
| 14 |
+
"8.9",
|
| 15 |
+
"9.0+PTX"
|
| 16 |
+
]
|
| 17 |
+
}
|
| 18 |
+
}
|
build/torch210-cxx11-cu128-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-cu128-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_cuda_6261c06
|
| 3 |
+
ops = torch.ops._relu_cuda_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_cuda_6261c06::{op_name}"
|
build/torch210-cxx11-cu128-x86_64-linux/{_relu_cuda_ohbj3nuuoamci.abi3.so → _relu_cuda_6261c06.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e433d2b9bf2ec78c5567fce9b71cef81fc0fe3e5ce5c68ed35774fb4bfc3c33
|
| 3 |
+
size 2111840
|
build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-cu128-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,21 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
-
"
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
+
"python-depends": [],
|
| 5 |
+
"backend": {
|
| 6 |
+
"type": "cuda",
|
| 7 |
+
"archs": [
|
| 8 |
+
"10.0",
|
| 9 |
+
"10.1",
|
| 10 |
+
"12.0+PTX",
|
| 11 |
+
"7.0",
|
| 12 |
+
"7.2",
|
| 13 |
+
"7.5",
|
| 14 |
+
"8.0",
|
| 15 |
+
"8.6",
|
| 16 |
+
"8.7",
|
| 17 |
+
"8.9",
|
| 18 |
+
"9.0"
|
| 19 |
+
]
|
| 20 |
+
}
|
| 21 |
+
}
|
build/torch210-cxx11-cu130-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-cu130-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_cuda_6261c06
|
| 3 |
+
ops = torch.ops._relu_cuda_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_cuda_6261c06::{op_name}"
|
build/torch210-cxx11-cu130-x86_64-linux/{_relu_cuda_oheoc6gkfjkwm.abi3.so → _relu_cuda_6261c06.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7d44340251ab3a892a7c74890b06230810cca22518d174d620562ae216bc60ae
|
| 3 |
+
size 2129656
|
build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-cu130-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,19 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
-
"
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
+
"python-depends": [],
|
| 5 |
+
"backend": {
|
| 6 |
+
"type": "cuda",
|
| 7 |
+
"archs": [
|
| 8 |
+
"10.0",
|
| 9 |
+
"11.0",
|
| 10 |
+
"12.0+PTX",
|
| 11 |
+
"7.5",
|
| 12 |
+
"8.0",
|
| 13 |
+
"8.6",
|
| 14 |
+
"8.7",
|
| 15 |
+
"8.9",
|
| 16 |
+
"9.0"
|
| 17 |
+
]
|
| 18 |
+
}
|
| 19 |
+
}
|
build/torch210-cxx11-rocm70-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-rocm70-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_rocm_6261c06
|
| 3 |
+
ops = torch.ops._relu_rocm_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_rocm_6261c06::{op_name}"
|
build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:111ce943531a26f724e8c241b86959b2c65e247c1c3d89b0b5a8f7f910ed8d3f
|
| 3 |
+
size 1990368
|
build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:b3a0a5621e558e37c57d96dc9dd467d1a48d142e07acfe6fe7339098ef7d88fa
|
| 3 |
-
size 1990392
|
|
|
|
|
|
|
|
|
|
|
|
build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-rocm70-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,17 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
-
"
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
+
"python-depends": [],
|
| 5 |
+
"backend": {
|
| 6 |
+
"type": "rocm",
|
| 7 |
+
"archs": [
|
| 8 |
+
"gfx1030",
|
| 9 |
+
"gfx1100",
|
| 10 |
+
"gfx1101",
|
| 11 |
+
"gfx906",
|
| 12 |
+
"gfx908",
|
| 13 |
+
"gfx90a",
|
| 14 |
+
"gfx942"
|
| 15 |
+
]
|
| 16 |
+
}
|
| 17 |
+
}
|
build/torch210-cxx11-rocm71-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-rocm71-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_rocm_6261c06
|
| 3 |
+
ops = torch.ops._relu_rocm_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_rocm_6261c06::{op_name}"
|
build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c4c743ab2b2fc7cbf616579558ef2acbb30f9610f2ea076c63ddfc9e50c0c69b
|
| 3 |
+
size 1990272
|
build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:12860f472b8ba796983a3693696b35280ae6a749233e4693d88572be022e962e
|
| 3 |
-
size 1990296
|
|
|
|
|
|
|
|
|
|
|
|
build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-rocm71-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,17 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
-
"
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
+
"python-depends": [],
|
| 5 |
+
"backend": {
|
| 6 |
+
"type": "rocm",
|
| 7 |
+
"archs": [
|
| 8 |
+
"gfx1030",
|
| 9 |
+
"gfx1100",
|
| 10 |
+
"gfx1101",
|
| 11 |
+
"gfx906",
|
| 12 |
+
"gfx908",
|
| 13 |
+
"gfx90a",
|
| 14 |
+
"gfx942"
|
| 15 |
+
]
|
| 16 |
+
}
|
| 17 |
+
}
|
build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_xpu_6261c06
|
| 3 |
+
ops = torch.ops._relu_xpu_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_xpu_6261c06::{op_name}"
|
build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ccc014010a351d00431933062b5f385e6bf809b2eb5f6324bcd38ef7ab894a4
|
| 3 |
+
size 2050448
|
build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:7420cb6337e94e6131081a2b2eefd548990da151c007283ed0ba5f8003095687
|
| 3 |
-
size 2050464
|
|
|
|
|
|
|
|
|
|
|
|
build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
|
|
|
| 3 |
"python-depends": []
|
| 4 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
"python-depends": []
|
| 5 |
}
|
build/torch29-cxx11-cpu-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch29-cxx11-cpu-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_cpu_6261c06
|
| 3 |
+
ops = torch.ops._relu_cpu_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_cpu_6261c06::{op_name}"
|
build/torch29-cxx11-cpu-x86_64-linux/{_relu_cpu_liz2pcyxsjqzq.abi3.so → _relu_cpu_6261c06.abi3.so}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e5f3b27de92bba3fdbb9bef157f6e15bcb9119581699420e1dca69f0c7cacce
|
| 3 |
+
size 311616
|
build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch29-cxx11-cpu-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
|
|
|
| 3 |
"python-depends": []
|
| 4 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
"python-depends": []
|
| 5 |
}
|
build/torch29-cxx11-cu126-x86_64-linux/__init__.py
CHANGED
|
@@ -4,9 +4,13 @@ import torch
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
|
|
|
|
|
|
| 7 |
|
| 8 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 9 |
if out is None:
|
| 10 |
out = torch.empty_like(x)
|
| 11 |
ops.relu(out, x)
|
| 12 |
-
return out
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
from ._ops import ops
|
| 6 |
|
| 7 |
+
from . import layers
|
| 8 |
+
|
| 9 |
|
| 10 |
def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
|
| 11 |
if out is None:
|
| 12 |
out = torch.empty_like(x)
|
| 13 |
ops.relu(out, x)
|
| 14 |
+
return out
|
| 15 |
+
|
| 16 |
+
__all__ = ["relu", "layers"]
|
build/torch29-cxx11-cu126-x86_64-linux/_ops.py
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
-
from . import
|
| 3 |
-
ops = torch.ops.
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
-
return f"
|
|
|
|
| 1 |
import torch
|
| 2 |
+
from . import _relu_cuda_6261c06
|
| 3 |
+
ops = torch.ops._relu_cuda_6261c06
|
| 4 |
|
| 5 |
def add_op_namespace_prefix(op_name: str):
|
| 6 |
"""
|
| 7 |
Prefix op by namespace.
|
| 8 |
"""
|
| 9 |
+
return f"_relu_cuda_6261c06::{op_name}"
|
build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ccb1d2b06fd0f7636eaa740b43937e557c1f07b47479b412b095089f49bc3d2
|
| 3 |
+
size 2014200
|
build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:7de615584690456b6ed2b89ce95611a9ea06e6368a81cddaa35f5cc00678b271
|
| 3 |
-
size 2014224
|
|
|
|
|
|
|
|
|
|
|
|
build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from .._ops import ops
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
class ReLU(nn.Module):
|
| 8 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
| 9 |
+
out = torch.empty_like(x)
|
| 10 |
+
ops.relu(out, x)
|
| 11 |
+
return out
|
build/torch29-cxx11-cu126-x86_64-linux/metadata.json
CHANGED
|
@@ -1,4 +1,18 @@
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
-
"
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
{
|
| 2 |
"version": 1,
|
| 3 |
+
"license": "Apache-2.0",
|
| 4 |
+
"python-depends": [],
|
| 5 |
+
"backend": {
|
| 6 |
+
"type": "cuda",
|
| 7 |
+
"archs": [
|
| 8 |
+
"7.0",
|
| 9 |
+
"7.2",
|
| 10 |
+
"7.5",
|
| 11 |
+
"8.0",
|
| 12 |
+
"8.6",
|
| 13 |
+
"8.7",
|
| 14 |
+
"8.9",
|
| 15 |
+
"9.0+PTX"
|
| 16 |
+
]
|
| 17 |
+
}
|
| 18 |
+
}
|