danieldk HF Staff commited on
Commit
c8cdee3
·
verified ·
1 Parent(s): fd833df

Build uploaded using `kernels`.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +14 -0
  2. build/torch210-cxx11-cpu-x86_64-linux/__init__.py +5 -1
  3. build/torch210-cxx11-cpu-x86_64-linux/_ops.py +3 -3
  4. build/torch210-cxx11-cpu-x86_64-linux/{_relu_cpu_smcctie2ikqu4.abi3.so → _relu_cpu_6261c06.abi3.so} +2 -2
  5. build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py +11 -0
  6. build/torch210-cxx11-cpu-x86_64-linux/metadata.json +1 -0
  7. build/torch210-cxx11-cu126-x86_64-linux/__init__.py +5 -1
  8. build/torch210-cxx11-cu126-x86_64-linux/_ops.py +3 -3
  9. build/torch210-cxx11-cu126-x86_64-linux/{_relu_cuda_byb5bdzb6z5zm.abi3.so → _relu_cuda_6261c06.abi3.so} +2 -2
  10. build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py +11 -0
  11. build/torch210-cxx11-cu126-x86_64-linux/metadata.json +16 -2
  12. build/torch210-cxx11-cu128-x86_64-linux/__init__.py +5 -1
  13. build/torch210-cxx11-cu128-x86_64-linux/_ops.py +3 -3
  14. build/torch210-cxx11-cu128-x86_64-linux/{_relu_cuda_ohbj3nuuoamci.abi3.so → _relu_cuda_6261c06.abi3.so} +2 -2
  15. build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py +11 -0
  16. build/torch210-cxx11-cu128-x86_64-linux/metadata.json +19 -2
  17. build/torch210-cxx11-cu130-x86_64-linux/__init__.py +5 -1
  18. build/torch210-cxx11-cu130-x86_64-linux/_ops.py +3 -3
  19. build/torch210-cxx11-cu130-x86_64-linux/{_relu_cuda_oheoc6gkfjkwm.abi3.so → _relu_cuda_6261c06.abi3.so} +2 -2
  20. build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py +11 -0
  21. build/torch210-cxx11-cu130-x86_64-linux/metadata.json +17 -2
  22. build/torch210-cxx11-rocm70-x86_64-linux/__init__.py +5 -1
  23. build/torch210-cxx11-rocm70-x86_64-linux/_ops.py +3 -3
  24. build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so +3 -0
  25. build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so +0 -3
  26. build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py +11 -0
  27. build/torch210-cxx11-rocm70-x86_64-linux/metadata.json +15 -2
  28. build/torch210-cxx11-rocm71-x86_64-linux/__init__.py +5 -1
  29. build/torch210-cxx11-rocm71-x86_64-linux/_ops.py +3 -3
  30. build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so +3 -0
  31. build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so +0 -3
  32. build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py +11 -0
  33. build/torch210-cxx11-rocm71-x86_64-linux/metadata.json +15 -2
  34. build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py +5 -1
  35. build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py +3 -3
  36. build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so +3 -0
  37. build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so +0 -3
  38. build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py +11 -0
  39. build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json +1 -0
  40. build/torch29-cxx11-cpu-x86_64-linux/__init__.py +5 -1
  41. build/torch29-cxx11-cpu-x86_64-linux/_ops.py +3 -3
  42. build/torch29-cxx11-cpu-x86_64-linux/{_relu_cpu_liz2pcyxsjqzq.abi3.so → _relu_cpu_6261c06.abi3.so} +2 -2
  43. build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py +11 -0
  44. build/torch29-cxx11-cpu-x86_64-linux/metadata.json +1 -0
  45. build/torch29-cxx11-cu126-x86_64-linux/__init__.py +5 -1
  46. build/torch29-cxx11-cu126-x86_64-linux/_ops.py +3 -3
  47. build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so +3 -0
  48. build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so +0 -3
  49. build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py +11 -0
  50. build/torch29-cxx11-cu126-x86_64-linux/metadata.json +16 -2
.gitattributes CHANGED
@@ -259,3 +259,17 @@ build/torch29-cxx11-cu126-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs di
259
  build/torch29-cxx11-cu128-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
260
  build/torch29-cxx11-cu130-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
261
  build/torch210-cu128-x86_64-windows/_relu_cuda_657300a.pyd filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
259
  build/torch29-cxx11-cu128-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
260
  build/torch29-cxx11-cu130-aarch64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
261
  build/torch210-cu128-x86_64-windows/_relu_cuda_657300a.pyd filter=lfs diff=lfs merge=lfs -text
262
+ build/torch210-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
263
+ build/torch210-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
264
+ build/torch210-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
265
+ build/torch210-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
266
+ build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
267
+ build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
268
+ build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
269
+ build/torch29-cxx11-cpu-x86_64-linux/_relu_cpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
270
+ build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
271
+ build/torch29-cxx11-cu128-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
272
+ build/torch29-cxx11-cu130-x86_64-linux/_relu_cuda_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
273
+ build/torch29-cxx11-rocm63-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
274
+ build/torch29-cxx11-rocm64-x86_64-linux/_relu_rocm_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
275
+ build/torch29-cxx11-xpu20252-x86_64-linux/_relu_xpu_6261c06.abi3.so filter=lfs diff=lfs merge=lfs -text
build/torch210-cxx11-cpu-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-cpu-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_cpu_smcctie2ikqu4
3
- ops = torch.ops._relu_cpu_smcctie2ikqu4
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_cpu_smcctie2ikqu4::{op_name}"
 
1
  import torch
2
+ from . import _relu_cpu_6261c06
3
+ ops = torch.ops._relu_cpu_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_cpu_6261c06::{op_name}"
build/torch210-cxx11-cpu-x86_64-linux/{_relu_cpu_smcctie2ikqu4.abi3.so → _relu_cpu_6261c06.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:15d0e318702082b3d2aa29dcdd5e64cd8f3b3a15c1bd8b9310f2ee2e43a5482e
3
- size 1778104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bc51be4e362c70be9c87e4e3e8464904547c487496ce8a0ebe552c5fa6effab
3
+ size 1778080
build/torch210-cxx11-cpu-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-cpu-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,5 @@
1
  {
2
  "version": 1,
 
3
  "python-depends": []
4
  }
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
  "python-depends": []
5
  }
build/torch210-cxx11-cu126-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-cu126-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_cuda_byb5bdzb6z5zm
3
- ops = torch.ops._relu_cuda_byb5bdzb6z5zm
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_cuda_byb5bdzb6z5zm::{op_name}"
 
1
  import torch
2
+ from . import _relu_cuda_6261c06
3
+ ops = torch.ops._relu_cuda_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_cuda_6261c06::{op_name}"
build/torch210-cxx11-cu126-x86_64-linux/{_relu_cuda_byb5bdzb6z5zm.abi3.so → _relu_cuda_6261c06.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:628fca28e54413192316cd014179733076bf926d757ec04eac79650dc2c562d3
3
- size 2019904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3aaf0a90799a5102aba64409350132a8bba560a8816e00be51a8b5f41f9475d
3
+ size 2019880
build/torch210-cxx11-cu126-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-cu126-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,18 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "7.0",
9
+ "7.2",
10
+ "7.5",
11
+ "8.0",
12
+ "8.6",
13
+ "8.7",
14
+ "8.9",
15
+ "9.0+PTX"
16
+ ]
17
+ }
18
+ }
build/torch210-cxx11-cu128-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-cu128-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_cuda_ohbj3nuuoamci
3
- ops = torch.ops._relu_cuda_ohbj3nuuoamci
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_cuda_ohbj3nuuoamci::{op_name}"
 
1
  import torch
2
+ from . import _relu_cuda_6261c06
3
+ ops = torch.ops._relu_cuda_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_cuda_6261c06::{op_name}"
build/torch210-cxx11-cu128-x86_64-linux/{_relu_cuda_ohbj3nuuoamci.abi3.so → _relu_cuda_6261c06.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eec7269c3936d34e2b5208a9dbe3b3b234c2a0aa97f6bb22ba0df83da511f4e
3
- size 2111856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e433d2b9bf2ec78c5567fce9b71cef81fc0fe3e5ce5c68ed35774fb4bfc3c33
3
+ size 2111840
build/torch210-cxx11-cu128-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-cu128-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,21 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0",
9
+ "10.1",
10
+ "12.0+PTX",
11
+ "7.0",
12
+ "7.2",
13
+ "7.5",
14
+ "8.0",
15
+ "8.6",
16
+ "8.7",
17
+ "8.9",
18
+ "9.0"
19
+ ]
20
+ }
21
+ }
build/torch210-cxx11-cu130-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-cu130-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_cuda_oheoc6gkfjkwm
3
- ops = torch.ops._relu_cuda_oheoc6gkfjkwm
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_cuda_oheoc6gkfjkwm::{op_name}"
 
1
  import torch
2
+ from . import _relu_cuda_6261c06
3
+ ops = torch.ops._relu_cuda_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_cuda_6261c06::{op_name}"
build/torch210-cxx11-cu130-x86_64-linux/{_relu_cuda_oheoc6gkfjkwm.abi3.so → _relu_cuda_6261c06.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01d13efe21c6ca11829cf5e1fc0cc48e2b77a365ebacfa3d37c1460471979f0b
3
- size 2129672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d44340251ab3a892a7c74890b06230810cca22518d174d620562ae216bc60ae
3
+ size 2129656
build/torch210-cxx11-cu130-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-cu130-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,19 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "10.0",
9
+ "11.0",
10
+ "12.0+PTX",
11
+ "7.5",
12
+ "8.0",
13
+ "8.6",
14
+ "8.7",
15
+ "8.9",
16
+ "9.0"
17
+ ]
18
+ }
19
+ }
build/torch210-cxx11-rocm70-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-rocm70-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_rocm_g773e4bf7332k
3
- ops = torch.ops._relu_rocm_g773e4bf7332k
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_rocm_g773e4bf7332k::{op_name}"
 
1
  import torch
2
+ from . import _relu_rocm_6261c06
3
+ ops = torch.ops._relu_rocm_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_rocm_6261c06::{op_name}"
build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_6261c06.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:111ce943531a26f724e8c241b86959b2c65e247c1c3d89b0b5a8f7f910ed8d3f
3
+ size 1990368
build/torch210-cxx11-rocm70-x86_64-linux/_relu_rocm_g773e4bf7332k.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3a0a5621e558e37c57d96dc9dd467d1a48d142e07acfe6fe7339098ef7d88fa
3
- size 1990392
 
 
 
 
build/torch210-cxx11-rocm70-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-rocm70-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,17 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "rocm",
7
+ "archs": [
8
+ "gfx1030",
9
+ "gfx1100",
10
+ "gfx1101",
11
+ "gfx906",
12
+ "gfx908",
13
+ "gfx90a",
14
+ "gfx942"
15
+ ]
16
+ }
17
+ }
build/torch210-cxx11-rocm71-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-rocm71-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_rocm_ijhrndbhfiwtg
3
- ops = torch.ops._relu_rocm_ijhrndbhfiwtg
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_rocm_ijhrndbhfiwtg::{op_name}"
 
1
  import torch
2
+ from . import _relu_rocm_6261c06
3
+ ops = torch.ops._relu_rocm_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_rocm_6261c06::{op_name}"
build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_6261c06.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4c743ab2b2fc7cbf616579558ef2acbb30f9610f2ea076c63ddfc9e50c0c69b
3
+ size 1990272
build/torch210-cxx11-rocm71-x86_64-linux/_relu_rocm_ijhrndbhfiwtg.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:12860f472b8ba796983a3693696b35280ae6a749233e4693d88572be022e962e
3
- size 1990296
 
 
 
 
build/torch210-cxx11-rocm71-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-rocm71-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,17 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "rocm",
7
+ "archs": [
8
+ "gfx1030",
9
+ "gfx1100",
10
+ "gfx1101",
11
+ "gfx906",
12
+ "gfx908",
13
+ "gfx90a",
14
+ "gfx942"
15
+ ]
16
+ }
17
+ }
build/torch210-cxx11-xpu20253-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch210-cxx11-xpu20253-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_xpu_gnle2ptjfkifg
3
- ops = torch.ops._relu_xpu_gnle2ptjfkifg
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_xpu_gnle2ptjfkifg::{op_name}"
 
1
  import torch
2
+ from . import _relu_xpu_6261c06
3
+ ops = torch.ops._relu_xpu_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_xpu_6261c06::{op_name}"
build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_6261c06.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ccc014010a351d00431933062b5f385e6bf809b2eb5f6324bcd38ef7ab894a4
3
+ size 2050448
build/torch210-cxx11-xpu20253-x86_64-linux/_relu_xpu_gnle2ptjfkifg.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7420cb6337e94e6131081a2b2eefd548990da151c007283ed0ba5f8003095687
3
- size 2050464
 
 
 
 
build/torch210-cxx11-xpu20253-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch210-cxx11-xpu20253-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,5 @@
1
  {
2
  "version": 1,
 
3
  "python-depends": []
4
  }
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
  "python-depends": []
5
  }
build/torch29-cxx11-cpu-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch29-cxx11-cpu-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_cpu_liz2pcyxsjqzq
3
- ops = torch.ops._relu_cpu_liz2pcyxsjqzq
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_cpu_liz2pcyxsjqzq::{op_name}"
 
1
  import torch
2
+ from . import _relu_cpu_6261c06
3
+ ops = torch.ops._relu_cpu_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_cpu_6261c06::{op_name}"
build/torch29-cxx11-cpu-x86_64-linux/{_relu_cpu_liz2pcyxsjqzq.abi3.so → _relu_cpu_6261c06.abi3.so} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2c59f561d7ff702ada015764f08651af54296dca20dafa731ede8c16bedaa07
3
- size 311648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e5f3b27de92bba3fdbb9bef157f6e15bcb9119581699420e1dca69f0c7cacce
3
+ size 311616
build/torch29-cxx11-cpu-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch29-cxx11-cpu-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,5 @@
1
  {
2
  "version": 1,
 
3
  "python-depends": []
4
  }
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
  "python-depends": []
5
  }
build/torch29-cxx11-cu126-x86_64-linux/__init__.py CHANGED
@@ -4,9 +4,13 @@ import torch
4
 
5
  from ._ops import ops
6
 
 
 
7
 
8
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
9
  if out is None:
10
  out = torch.empty_like(x)
11
  ops.relu(out, x)
12
- return out
 
 
 
4
 
5
  from ._ops import ops
6
 
7
+ from . import layers
8
+
9
 
10
  def relu(x: torch.Tensor, out: Optional[torch.Tensor] = None) -> torch.Tensor:
11
  if out is None:
12
  out = torch.empty_like(x)
13
  ops.relu(out, x)
14
+ return out
15
+
16
+ __all__ = ["relu", "layers"]
build/torch29-cxx11-cu126-x86_64-linux/_ops.py CHANGED
@@ -1,9 +1,9 @@
1
  import torch
2
- from . import _relu_cuda_udp2pzwonis42
3
- ops = torch.ops._relu_cuda_udp2pzwonis42
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
- return f"_relu_cuda_udp2pzwonis42::{op_name}"
 
1
  import torch
2
+ from . import _relu_cuda_6261c06
3
+ ops = torch.ops._relu_cuda_6261c06
4
 
5
  def add_op_namespace_prefix(op_name: str):
6
  """
7
  Prefix op by namespace.
8
  """
9
+ return f"_relu_cuda_6261c06::{op_name}"
build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_6261c06.abi3.so ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ccb1d2b06fd0f7636eaa740b43937e557c1f07b47479b412b095089f49bc3d2
3
+ size 2014200
build/torch29-cxx11-cu126-x86_64-linux/_relu_cuda_udp2pzwonis42.abi3.so DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7de615584690456b6ed2b89ce95611a9ea06e6368a81cddaa35f5cc00678b271
3
- size 2014224
 
 
 
 
build/torch29-cxx11-cu126-x86_64-linux/layers/__init__.py ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from .._ops import ops
5
+
6
+
7
+ class ReLU(nn.Module):
8
+ def forward(self, x: torch.Tensor) -> torch.Tensor:
9
+ out = torch.empty_like(x)
10
+ ops.relu(out, x)
11
+ return out
build/torch29-cxx11-cu126-x86_64-linux/metadata.json CHANGED
@@ -1,4 +1,18 @@
1
  {
2
  "version": 1,
3
- "python-depends": []
4
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  {
2
  "version": 1,
3
+ "license": "Apache-2.0",
4
+ "python-depends": [],
5
+ "backend": {
6
+ "type": "cuda",
7
+ "archs": [
8
+ "7.0",
9
+ "7.2",
10
+ "7.5",
11
+ "8.0",
12
+ "8.6",
13
+ "8.7",
14
+ "8.9",
15
+ "9.0+PTX"
16
+ ]
17
+ }
18
+ }