diff --git a/.gitattributes b/.gitattributes index 06f972ef85f96053023a86820eea90dcac7778cf..0fb83c61089ea83fe9fa748aadf790ad1c064404 100644 --- a/.gitattributes +++ b/.gitattributes @@ -259,3 +259,17 @@ build/torch29-cxx11-cu126-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs di build/torch29-cxx11-cu128-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text build/torch29-cxx11-cu130-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text build/torch210-cu128-x86_64-windows/_relu_cuda_657300a.pyd filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text +build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text diff --git a/build/torch210-cxx11-cpu-x86_64-linux/__init__.py b/build/torch210-cxx11-cpu-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-cpu-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-cpu-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-cpu-x86_64-linux/_ops.py b/build/torch210-cxx11-cpu-x86_64-linux/_ops.py index 0cef0261d994be3bdc1647852633379f3a80dde2..ff8b9666569cab35ae0a56dc9d3363e5970dfe59 100644 --- a/build/torch210-cxx11-cpu-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-cpu-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cpu_smcctie2ikqu4 -ops = torch.ops._relu_cpu_smcctie2ikqu4 +from . import _relu_cpu_6261c06 +ops = torch.ops._relu_cpu_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cpu_smcctie2ikqu4::{op_name}" + return f"_relu_cpu_6261c06::{op_name}" diff --git a/build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so b/build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..c7277c34a37c54a802f56518c7b4136de5710606 --- /dev/null +++ b/build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bc51be4e362c70be9c87e4e3e8464904547c487496ce8a0ebe552c5fa6effab +size 1778080 diff --git a/build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_smcctie2ikqu4.abi3.so b/build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_smcctie2ikqu4.abi3.so deleted file mode 100644 index 1c71c9cbc5c19fdc69e640dc93ae28fd89e3b8dd..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_smcctie2ikqu4.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:15d0e318702082b3d2aa29dcdd5e64cd8f3b3a15c1bd8b9310f2ee2e43a5482e -size 1778104 diff --git a/build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-cpu-x86_64-linux/metadata.json b/build/torch210-cxx11-cpu-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..a5381dd80836f863378b9f33a559815688de9287 100644 --- a/build/torch210-cxx11-cpu-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-cpu-x86_64-linux/metadata.json @@ -1,4 +1,5 @@ { "version": 1, + "license": "Apache-2.0", "python-depends": [] } \ No newline at end of file diff --git a/build/torch210-cxx11-cu126-x86_64-linux/__init__.py b/build/torch210-cxx11-cu126-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-cu126-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-cu126-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-cu126-x86_64-linux/_ops.py b/build/torch210-cxx11-cu126-x86_64-linux/_ops.py index 221939ed9d33e3e3d08220a3a7c0ffcc78968c7f..54798b7a318e36d9aea831cec37d64d7c39f74a8 100644 --- a/build/torch210-cxx11-cu126-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-cu126-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cuda_byb5bdzb6z5zm -ops = torch.ops._relu_cuda_byb5bdzb6z5zm +from . import _relu_cuda_6261c06 +ops = torch.ops._relu_cuda_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cuda_byb5bdzb6z5zm::{op_name}" + return f"_relu_cuda_6261c06::{op_name}" diff --git a/build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so b/build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..fb346f2c8a9146b284c5d83390b909996fce3d45 --- /dev/null +++ b/build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3aaf0a90799a5102aba64409350132a8bba560a8816e00be51a8b5f41f9475d +size 2019880 diff --git a/build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_byb5bdzb6z5zm.abi3.so b/build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_byb5bdzb6z5zm.abi3.so deleted file mode 100644 index 029345deb243e7bf98568673ee808aece239101a..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_byb5bdzb6z5zm.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:628fca28e54413192316cd014179733076bf926d757ec04eac79650dc2c562d3 -size 2019904 diff --git a/build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-cu126-x86_64-linux/metadata.json b/build/torch210-cxx11-cu126-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..f5902b55ab0b2b561c0cf97567c9806c60839c7f 100644 --- a/build/torch210-cxx11-cu126-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-cu126-x86_64-linux/metadata.json @@ -1,4 +1,18 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "cuda", + "archs": [ + "7.0", + "7.2", + "7.5", + "8.0", + "8.6", + "8.7", + "8.9", + "9.0+PTX" + ] + } +} diff --git a/build/torch210-cxx11-cu128-x86_64-linux/__init__.py b/build/torch210-cxx11-cu128-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-cu128-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-cu128-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-cu128-x86_64-linux/_ops.py b/build/torch210-cxx11-cu128-x86_64-linux/_ops.py index 6f6e60ba613be0912d3fe3abbc7de30494857504..54798b7a318e36d9aea831cec37d64d7c39f74a8 100644 --- a/build/torch210-cxx11-cu128-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-cu128-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cuda_ohbj3nuuoamci -ops = torch.ops._relu_cuda_ohbj3nuuoamci +from . import _relu_cuda_6261c06 +ops = torch.ops._relu_cuda_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cuda_ohbj3nuuoamci::{op_name}" + return f"_relu_cuda_6261c06::{op_name}" diff --git a/build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so b/build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..d793f14c07ed0d973ca90b657caaafe928ff6fdc --- /dev/null +++ b/build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e433d2b9bf2ec78c5567fce9b71cef81fc0fe3e5ce5c68ed35774fb4bfc3c33 +size 2111840 diff --git a/build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_ohbj3nuuoamci.abi3.so b/build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_ohbj3nuuoamci.abi3.so deleted file mode 100644 index a55bb3cea6651ec6c8453cc57f38d62c444fc41b..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_ohbj3nuuoamci.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8eec7269c3936d34e2b5208a9dbe3b3b234c2a0aa97f6bb22ba0df83da511f4e -size 2111856 diff --git a/build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-cu128-x86_64-linux/metadata.json b/build/torch210-cxx11-cu128-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..8b796af185fbbd8594fcd846949aa5fadc0ccdda 100644 --- a/build/torch210-cxx11-cu128-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-cu128-x86_64-linux/metadata.json @@ -1,4 +1,21 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "cuda", + "archs": [ + "10.0", + "10.1", + "12.0+PTX", + "7.0", + "7.2", + "7.5", + "8.0", + "8.6", + "8.7", + "8.9", + "9.0" + ] + } +} diff --git a/build/torch210-cxx11-cu130-x86_64-linux/__init__.py b/build/torch210-cxx11-cu130-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-cu130-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-cu130-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-cu130-x86_64-linux/_ops.py b/build/torch210-cxx11-cu130-x86_64-linux/_ops.py index 44955e49ec4794f32e3e1cbfb830f62aa18e3e3e..54798b7a318e36d9aea831cec37d64d7c39f74a8 100644 --- a/build/torch210-cxx11-cu130-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-cu130-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cuda_oheoc6gkfjkwm -ops = torch.ops._relu_cuda_oheoc6gkfjkwm +from . import _relu_cuda_6261c06 +ops = torch.ops._relu_cuda_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cuda_oheoc6gkfjkwm::{op_name}" + return f"_relu_cuda_6261c06::{op_name}" diff --git a/build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so b/build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..552feeceb03720a3ae735a91320e2c122b257668 --- /dev/null +++ b/build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d44340251ab3a892a7c74890b06230810cca22518d174d620562ae216bc60ae +size 2129656 diff --git a/build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_oheoc6gkfjkwm.abi3.so b/build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_oheoc6gkfjkwm.abi3.so deleted file mode 100644 index 137274b70b75d7ad0602854469218cf158978e60..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_oheoc6gkfjkwm.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:01d13efe21c6ca11829cf5e1fc0cc48e2b77a365ebacfa3d37c1460471979f0b -size 2129672 diff --git a/build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-cu130-x86_64-linux/metadata.json b/build/torch210-cxx11-cu130-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..66651b7d3f95ac9e5ce5fc2a641b6f0f50788f87 100644 --- a/build/torch210-cxx11-cu130-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-cu130-x86_64-linux/metadata.json @@ -1,4 +1,19 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "cuda", + "archs": [ + "10.0", + "11.0", + "12.0+PTX", + "7.5", + "8.0", + "8.6", + "8.7", + "8.9", + "9.0" + ] + } +} diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py b/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-rocm70-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py b/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py index 28043827c057d81616781fdfc9978b95c25f8c65..6522586762fd33860e634154c091fa05059ddfe1 100644 --- a/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-rocm70-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_rocm_g773e4bf7332k -ops = torch.ops._relu_rocm_g773e4bf7332k +from . import _relu_rocm_6261c06 +ops = torch.ops._relu_rocm_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_rocm_g773e4bf7332k::{op_name}" + return f"_relu_rocm_6261c06::{op_name}" diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so b/build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..2f5173fd715aa33763d20b8f0ed7c9d04c800df7 --- /dev/null +++ b/build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:111ce943531a26f724e8c241b86959b2c65e247c1c3d89b0b5a8f7f910ed8d3f +size 1990368 diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so b/build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so deleted file mode 100644 index 0258176f09707da99f7f280fc3ee3735fa7746ad..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b3a0a5621e558e37c57d96dc9dd467d1a48d142e07acfe6fe7339098ef7d88fa -size 1990392 diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json b/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..3e8d811f1dc42febd33121b2627f809447622baf 100644 --- a/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-rocm70-x86_64-linux/metadata.json @@ -1,4 +1,17 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "rocm", + "archs": [ + "gfx1030", + "gfx1100", + "gfx1101", + "gfx906", + "gfx908", + "gfx90a", + "gfx942" + ] + } +} diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py b/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-rocm71-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py b/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py index 013f73d70c5109dd140c0916236e8b0313e40e53..6522586762fd33860e634154c091fa05059ddfe1 100644 --- a/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-rocm71-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_rocm_ijhrndbhfiwtg -ops = torch.ops._relu_rocm_ijhrndbhfiwtg +from . import _relu_rocm_6261c06 +ops = torch.ops._relu_rocm_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_rocm_ijhrndbhfiwtg::{op_name}" + return f"_relu_rocm_6261c06::{op_name}" diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so b/build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..58d07bc923af985cce574847e499e8839db1b612 --- /dev/null +++ b/build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4c743ab2b2fc7cbf616579558ef2acbb30f9610f2ea076c63ddfc9e50c0c69b +size 1990272 diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so b/build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so deleted file mode 100644 index be146a0c8492b8f165c36529b9a8fb8441e4cd88..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:12860f472b8ba796983a3693696b35280ae6a749233e4693d88572be022e962e -size 1990296 diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json b/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..3e8d811f1dc42febd33121b2627f809447622baf 100644 --- a/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-rocm71-x86_64-linux/metadata.json @@ -1,4 +1,17 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "rocm", + "archs": [ + "gfx1030", + "gfx1100", + "gfx1101", + "gfx906", + "gfx908", + "gfx90a", + "gfx942" + ] + } +} diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py b/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py +++ b/build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py b/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py index 50ca861d780398ac524511329b24876d519245a9..a045efb682ea904f9f35ac056d7a5de18ae067ce 100644 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py +++ b/build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_xpu_gnle2ptjfkifg -ops = torch.ops._relu_xpu_gnle2ptjfkifg +from . import _relu_xpu_6261c06 +ops = torch.ops._relu_xpu_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_xpu_gnle2ptjfkifg::{op_name}" + return f"_relu_xpu_6261c06::{op_name}" diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so b/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..2aedff592a21d355e45aa25bf66b532d3c9fe558 --- /dev/null +++ b/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ccc014010a351d00431933062b5f385e6bf809b2eb5f6324bcd38ef7ab894a4 +size 2050448 diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so b/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so deleted file mode 100644 index bed624079755bb0eb3e26be0ef64bd966ee24552..0000000000000000000000000000000000000000 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7420cb6337e94e6131081a2b2eefd548990da151c007283ed0ba5f8003095687 -size 2050464 diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py b/build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json b/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..a5381dd80836f863378b9f33a559815688de9287 100644 --- a/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json +++ b/build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json @@ -1,4 +1,5 @@ { "version": 1, + "license": "Apache-2.0", "python-depends": [] } \ No newline at end of file diff --git a/build/torch29-cxx11-cpu-x86_64-linux/__init__.py b/build/torch29-cxx11-cpu-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-cpu-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-cpu-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-cpu-x86_64-linux/_ops.py b/build/torch29-cxx11-cpu-x86_64-linux/_ops.py index 7df1ae20db495cf2f2e5e31c2e0075077194ef1b..ff8b9666569cab35ae0a56dc9d3363e5970dfe59 100644 --- a/build/torch29-cxx11-cpu-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-cpu-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cpu_liz2pcyxsjqzq -ops = torch.ops._relu_cpu_liz2pcyxsjqzq +from . import _relu_cpu_6261c06 +ops = torch.ops._relu_cpu_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cpu_liz2pcyxsjqzq::{op_name}" + return f"_relu_cpu_6261c06::{op_name}" diff --git a/build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so b/build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..b416f4fa6480e576708770ea40f9042a309cb974 --- /dev/null +++ b/build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e5f3b27de92bba3fdbb9bef157f6e15bcb9119581699420e1dca69f0c7cacce +size 311616 diff --git a/build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_liz2pcyxsjqzq.abi3.so b/build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_liz2pcyxsjqzq.abi3.so deleted file mode 100644 index e79dcbe4b6d3bb029d46278aff23133e3a57f653..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_liz2pcyxsjqzq.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a2c59f561d7ff702ada015764f08651af54296dca20dafa731ede8c16bedaa07 -size 311648 diff --git a/build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-cpu-x86_64-linux/metadata.json b/build/torch29-cxx11-cpu-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..a5381dd80836f863378b9f33a559815688de9287 100644 --- a/build/torch29-cxx11-cpu-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-cpu-x86_64-linux/metadata.json @@ -1,4 +1,5 @@ { "version": 1, + "license": "Apache-2.0", "python-depends": [] } \ No newline at end of file diff --git a/build/torch29-cxx11-cu126-x86_64-linux/__init__.py b/build/torch29-cxx11-cu126-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-cu126-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-cu126-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-cu126-x86_64-linux/_ops.py b/build/torch29-cxx11-cu126-x86_64-linux/_ops.py index 2cbd2d6b480928a0f9f70395a4b8da02ab6f0bd0..54798b7a318e36d9aea831cec37d64d7c39f74a8 100644 --- a/build/torch29-cxx11-cu126-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-cu126-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cuda_udp2pzwonis42 -ops = torch.ops._relu_cuda_udp2pzwonis42 +from . import _relu_cuda_6261c06 +ops = torch.ops._relu_cuda_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cuda_udp2pzwonis42::{op_name}" + return f"_relu_cuda_6261c06::{op_name}" diff --git a/build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so b/build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..d964226d994f4b599766b25a7d96ae1a42ad402b --- /dev/null +++ b/build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ccb1d2b06fd0f7636eaa740b43937e557c1f07b47479b412b095089f49bc3d2 +size 2014200 diff --git a/build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so b/build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so deleted file mode 100644 index 9ff132b8b2c2bc5516dcbfaf379193593ab54b69..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7de615584690456b6ed2b89ce95611a9ea06e6368a81cddaa35f5cc00678b271 -size 2014224 diff --git a/build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-cu126-x86_64-linux/metadata.json b/build/torch29-cxx11-cu126-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..f5902b55ab0b2b561c0cf97567c9806c60839c7f 100644 --- a/build/torch29-cxx11-cu126-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-cu126-x86_64-linux/metadata.json @@ -1,4 +1,18 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "cuda", + "archs": [ + "7.0", + "7.2", + "7.5", + "8.0", + "8.6", + "8.7", + "8.9", + "9.0+PTX" + ] + } +} diff --git a/build/torch29-cxx11-cu128-x86_64-linux/__init__.py b/build/torch29-cxx11-cu128-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-cu128-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-cu128-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-cu128-x86_64-linux/_ops.py b/build/torch29-cxx11-cu128-x86_64-linux/_ops.py index 01e81b91613adafb9efc66e2c9dccbf29d1cb4bc..54798b7a318e36d9aea831cec37d64d7c39f74a8 100644 --- a/build/torch29-cxx11-cu128-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-cu128-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cuda_ijqh67vz4tgr4 -ops = torch.ops._relu_cuda_ijqh67vz4tgr4 +from . import _relu_cuda_6261c06 +ops = torch.ops._relu_cuda_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cuda_ijqh67vz4tgr4::{op_name}" + return f"_relu_cuda_6261c06::{op_name}" diff --git a/build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so b/build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..94e3d5947233ec1b2a2e0ac8a71e0e06e9e65475 --- /dev/null +++ b/build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82802833c8c59b5e4136deaaf4b783aa0cd13288aedbccbcf278fb4e40d673f4 +size 2101928 diff --git a/build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_ijqh67vz4tgr4.abi3.so b/build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_ijqh67vz4tgr4.abi3.so deleted file mode 100644 index ef8f95cb901d1d852b6b83acfe478b50a4819794..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_ijqh67vz4tgr4.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e93b1d3439fbcb62c491cf15b4b79cad992b80042c62944874361e54efeb35b4 -size 2101960 diff --git a/build/torch29-cxx11-cu128-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-cu128-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-cu128-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-cu128-x86_64-linux/metadata.json b/build/torch29-cxx11-cu128-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..8b796af185fbbd8594fcd846949aa5fadc0ccdda 100644 --- a/build/torch29-cxx11-cu128-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-cu128-x86_64-linux/metadata.json @@ -1,4 +1,21 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "cuda", + "archs": [ + "10.0", + "10.1", + "12.0+PTX", + "7.0", + "7.2", + "7.5", + "8.0", + "8.6", + "8.7", + "8.9", + "9.0" + ] + } +} diff --git a/build/torch29-cxx11-cu130-x86_64-linux/__init__.py b/build/torch29-cxx11-cu130-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-cu130-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-cu130-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-cu130-x86_64-linux/_ops.py b/build/torch29-cxx11-cu130-x86_64-linux/_ops.py index aadd3dfad65a7eb99944de3ae51d2ac9daf94a34..54798b7a318e36d9aea831cec37d64d7c39f74a8 100644 --- a/build/torch29-cxx11-cu130-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-cu130-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_cuda_7mqpgn6tykjbc -ops = torch.ops._relu_cuda_7mqpgn6tykjbc +from . import _relu_cuda_6261c06 +ops = torch.ops._relu_cuda_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_cuda_7mqpgn6tykjbc::{op_name}" + return f"_relu_cuda_6261c06::{op_name}" diff --git a/build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so b/build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..995a6052cc1ed17481c96b3ee07d0ab4a4ff8c87 --- /dev/null +++ b/build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fcbd057483865c6527156408ecd11a1b7dca95e70ae0b2f88b22857633624349 +size 2127936 diff --git a/build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_7mqpgn6tykjbc.abi3.so b/build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_7mqpgn6tykjbc.abi3.so deleted file mode 100644 index 89737f870b7bea40fca7fc4fb70a94f147a6007c..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_7mqpgn6tykjbc.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:65b1af2fbd28380e40962aef4468d83cfb4cae2e49bacc81ab7420cb2f770e4d -size 2127968 diff --git a/build/torch29-cxx11-cu130-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-cu130-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-cu130-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-cu130-x86_64-linux/metadata.json b/build/torch29-cxx11-cu130-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..66651b7d3f95ac9e5ce5fc2a641b6f0f50788f87 100644 --- a/build/torch29-cxx11-cu130-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-cu130-x86_64-linux/metadata.json @@ -1,4 +1,19 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "cuda", + "archs": [ + "10.0", + "11.0", + "12.0+PTX", + "7.5", + "8.0", + "8.6", + "8.7", + "8.9", + "9.0" + ] + } +} diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py b/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-rocm63-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py b/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py index d0c6bd8d5949523786b2c23012328591b7d30fd8..6522586762fd33860e634154c091fa05059ddfe1 100644 --- a/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-rocm63-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_rocm_vox4kgutt6m6i -ops = torch.ops._relu_rocm_vox4kgutt6m6i +from . import _relu_rocm_6261c06 +ops = torch.ops._relu_rocm_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_rocm_vox4kgutt6m6i::{op_name}" + return f"_relu_rocm_6261c06::{op_name}" diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_6261c06.abi3.so b/build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..10d97307ff909883bed62a2eacc2f61d3a0efffc --- /dev/null +++ b/build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:592555b16e93327db8d0e75a3288de3c1f9e380128d39f8ef0a965d7e44e0c9a +size 1970472 diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_vox4kgutt6m6i.abi3.so b/build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_vox4kgutt6m6i.abi3.so deleted file mode 100644 index 41dc2f416e040ca083bafb67d97871031c5366a2..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_vox4kgutt6m6i.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ba40972743ed267fb8c29c2986f75b6a09cdd61ceb07f8ee1a64074cd5d0a060 -size 1970504 diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-rocm63-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-rocm63-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json b/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..3e8d811f1dc42febd33121b2627f809447622baf 100644 --- a/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-rocm63-x86_64-linux/metadata.json @@ -1,4 +1,17 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "rocm", + "archs": [ + "gfx1030", + "gfx1100", + "gfx1101", + "gfx906", + "gfx908", + "gfx90a", + "gfx942" + ] + } +} diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py b/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-rocm64-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py b/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py index 30b12340b9dd5254c7a27eadad68887dac3dd0d1..6522586762fd33860e634154c091fa05059ddfe1 100644 --- a/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-rocm64-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_rocm_dd3ars5bdez2c -ops = torch.ops._relu_rocm_dd3ars5bdez2c +from . import _relu_rocm_6261c06 +ops = torch.ops._relu_rocm_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_rocm_dd3ars5bdez2c::{op_name}" + return f"_relu_rocm_6261c06::{op_name}" diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_6261c06.abi3.so b/build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..fe783f5b7c98f0f03f37f21ffbd7dc7c3b6b25de --- /dev/null +++ b/build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a26f035ba427cb3a8c8b20210931ae6c26980a5295d45b4c6b92c0be7eef382 +size 1989248 diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_dd3ars5bdez2c.abi3.so b/build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_dd3ars5bdez2c.abi3.so deleted file mode 100644 index fd202b5dea6057a48d375cde99d89efb702d6c17..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_dd3ars5bdez2c.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:da786a1b6e769a91c1d86486b04efa804d9f610f289126702324d7e6ed68c0f0 -size 1989280 diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-rocm64-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-rocm64-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json b/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..3e8d811f1dc42febd33121b2627f809447622baf 100644 --- a/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-rocm64-x86_64-linux/metadata.json @@ -1,4 +1,17 @@ { "version": 1, - "python-depends": [] -} \ No newline at end of file + "license": "Apache-2.0", + "python-depends": [], + "backend": { + "type": "rocm", + "archs": [ + "gfx1030", + "gfx1100", + "gfx1101", + "gfx906", + "gfx908", + "gfx90a", + "gfx942" + ] + } +} diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py b/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py index 8050dfd765d1a50720e99d17870c2c854d9e2574..1034928e8aa6d9cf8c85742c404d42d77ab38514 100644 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py +++ b/build/torch29-cxx11-xpu20252-x86_64-linux/__init__.py @@ -4,9 +4,13 @@ import torch from ._ops import ops +from . import layers + def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor: if out is None: out = torch.empty_like(x) ops.relu(out, x) - return out \ No newline at end of file + return out + +__all__ = ["relu", "layers"] diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py b/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py index 75b0822bbf0c1119af785eef623959c210c9ba40..a045efb682ea904f9f35ac056d7a5de18ae067ce 100644 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py +++ b/build/torch29-cxx11-xpu20252-x86_64-linux/_ops.py @@ -1,9 +1,9 @@ import torch -from . import _relu_xpu_wcav6tli7hepe -ops = torch.ops._relu_xpu_wcav6tli7hepe +from . import _relu_xpu_6261c06 +ops = torch.ops._relu_xpu_6261c06 def add_op_namespace_prefix(op_name: str): """ Prefix op by namespace. """ - return f"_relu_xpu_wcav6tli7hepe::{op_name}" + return f"_relu_xpu_6261c06::{op_name}" diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_6261c06.abi3.so b/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_6261c06.abi3.so new file mode 100644 index 0000000000000000000000000000000000000000..1ce8be1882919c61119fc0b61920ff8f9fea4bb2 --- /dev/null +++ b/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_6261c06.abi3.so @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d12874e554a6955647bacb35738e09cd9a157f6203273163e48f19ab90449a9 +size 2048328 diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_wcav6tli7hepe.abi3.so b/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_wcav6tli7hepe.abi3.so deleted file mode 100644 index a81ddcce6eff0dc4485aa1068b1c3fe2516262b1..0000000000000000000000000000000000000000 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_wcav6tli7hepe.abi3.so +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cbcf4a55e88b789b750c1fc35cabaf571999dd31f67751ec161b50b2183111ff -size 2048352 diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/layers/__init__.py b/build/torch29-cxx11-xpu20252-x86_64-linux/layers/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..6105a19101b9f01a5ed9f061f94cc92f3c3ab157 --- /dev/null +++ b/build/torch29-cxx11-xpu20252-x86_64-linux/layers/__init__.py @@ -0,0 +1,11 @@ +import torch +import torch.nn as nn + +from .._ops import ops + + +class ReLU(nn.Module): + def forward(self, x: torch.Tensor) -> torch.Tensor: + out = torch.empty_like(x) + ops.relu(out, x) + return out diff --git a/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json b/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json index 9cf5deed9898dce769f4cc73913d3530b92a0bd8..a5381dd80836f863378b9f33a559815688de9287 100644 --- a/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json +++ b/build/torch29-cxx11-xpu20252-x86_64-linux/metadata.json @@ -1,4 +1,5 @@ { "version": 1, + "license": "Apache-2.0", "python-depends": [] } \ No newline at end of file