Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .cache/pip/http-v2/0/1/f/2/0/01f2082df50502ba9492d64e69db99d1fdb5730707a16c6264b355b8.body +3 -0
- .cache/pip/http-v2/0/2/d/4/2/02d4221e858694abc22129c65515f1df2c4c326330eb1a34ceb0b382.body +3 -0
- .cache/pip/http-v2/4/0/2/3/b/4023be7b5b37a7a4144c804ce69828082d4fb2a124d9d8aabc855da8.body +3 -0
- .cache/pip/http-v2/4/f/d/2/5/4fd254dbd56deb4021e55d22c4b489f6c776c69c316eb7345bc91691.body +3 -0
- .cache/pip/http-v2/9/6/e/8/3/96e83221dd149da9a3d38feebc955beb2034effd910108971c5b167b.body +3 -0
- .cache/pip/http-v2/9/e/8/c/8/9e8c8c0496d6d3384d616902379ed05e07b6b1dba9673d70b5fef231.body +3 -0
- .cache/pip/http-v2/a/e/7/a/2/ae7a241673cf118ca18eca030dc29d2715b1980127dd0e2949514433.body +3 -0
- .cache/pip/http-v2/d/3/3/a/b/d33abf9ad709d023fff05902f39da682c1afb233bcd9f2c479487586.body +3 -0
- .cache/pip/http-v2/d/b/1/f/6/db1f6b45c0850c8e2ce7d8b47148edeca6e8115413af41f4ecc8ce32.body +3 -0
- .cache/pip/http-v2/f/5/2/7/6/f52769e4b4d00542e1e056baf2db3e5ad8f277bff67f2636cace711d.body +3 -0
- .cache/pip/wheels/7e/e3/c3/89c7a2f3c4adc07cd1c675f8bb7b9ad4d18f64a72bccdfe826/flash_attn-2.6.3-cp310-cp310-linux_x86_64.whl +3 -0
- .gitattributes +20 -0
- .launchpadlib/api.launchpad.net/cache/api.launchpad.net,devel,-application,vnd.sun.wadl+xml,2f09acb494bdefdbf8ef0d1396a05e86 +0 -0
- .local/share/jupyter/nbextensions/go_to_current_running_cell/auto_focus.gif +3 -0
- .local/share/jupyter/nbextensions/nbTranslate/demo1.gif +3 -0
- .local/share/jupyter/nbextensions/nbTranslate/demo2.gif +3 -0
- .local/share/jupyter/nbextensions/scratchpad/demo.gif +3 -0
- .local/share/jupyter/nbextensions/toc2/demo.gif +3 -0
- .local/share/jupyter/nbextensions/zenmode/images/back11.jpg +0 -0
- .local/share/jupyter/nbextensions/zenmode/images/back21.jpg +0 -0
- .local/share/jupyter/nbextensions/zenmode/images/ipynblogo1.png +0 -0
- .triton/dump/11759acf26ac56366b171628132485d6/triton_.llir +347 -0
- .triton/dump/11759acf26ac56366b171628132485d6/triton_.ttgir +78 -0
- .triton/dump/11759acf26ac56366b171628132485d6/triton_.ttir +76 -0
- .triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ptx +988 -0
- .triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ttir +104 -0
- .triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ptx +782 -0
- .triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ttgir +88 -0
- .triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.llir +980 -0
- .triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ptx +1654 -0
- .triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ttir +104 -0
- .triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ptx +465 -0
- .triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttgir +39 -0
- .triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttir +38 -0
- .triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.cubin +0 -0
- .triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.llir +550 -0
- .triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.cubin +0 -0
- .triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.llir +85 -0
- .triton/dump/510522bb05917b836ed253751364fcad/triton_.cubin +0 -0
- .triton/dump/510522bb05917b836ed253751364fcad/triton_.ptx +1810 -0
- .triton/dump/510522bb05917b836ed253751364fcad/triton_.ttgir +153 -0
- .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.cubin +0 -0
- .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.llir +1360 -0
- .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ptx +2004 -0
- .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttgir +164 -0
- .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttir +151 -0
- .triton/dump/76fb48b96c75cb8e388c291a18ef9b02/triton_.ttgir +169 -0
- .triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttgir +110 -0
- .triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttir +91 -0
- .triton/dump/ab89f85e55d5ddd9676325b49df9419f/triton_.ttgir +68 -0
.cache/pip/http-v2/0/1/f/2/0/01f2082df50502ba9492d64e69db99d1fdb5730707a16c6264b355b8.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d4511c52caacf3c4981d1ae2df85908bd31853f33d30b345c8b6830763f769c
|
3 |
+
size 1080866
|
.cache/pip/http-v2/0/2/d/4/2/02d4221e858694abc22129c65515f1df2c4c326330eb1a34ceb0b382.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5bfae9500ad8e7d2937ebccb4906f3bc464d1bf66eedd0e4adabd520811c7b52
|
3 |
+
size 2631958
|
.cache/pip/http-v2/4/0/2/3/b/4023be7b5b37a7a4144c804ce69828082d4fb2a124d9d8aabc855da8.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd4c97d69242efd604c1a2077c8b56341e236cfaca78c40f59dcef9b95464fdc
|
3 |
+
size 9663908
|
.cache/pip/http-v2/4/f/d/2/5/4fd254dbd56deb4021e55d22c4b489f6c776c69c316eb7345bc91691.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f99e4769b4750076cd4235c044b61232110733322384a94a63791d2e7beacc66
|
3 |
+
size 9995162
|
.cache/pip/http-v2/9/6/e/8/3/96e83221dd149da9a3d38feebc955beb2034effd910108971c5b167b.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e32dced201274bf96899e6491d9ba3e9a5f6b336708656466ad0522d8528f69
|
3 |
+
size 41178528
|
.cache/pip/http-v2/9/e/8/c/8/9e8c8c0496d6d3384d616902379ed05e07b6b1dba9673d70b5fef231.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3d2665c5df629eb2f981dab244c01bfa6cdc185f4ffa026639286c4d56fafb54
|
3 |
+
size 1221827
|
.cache/pip/http-v2/a/e/7/a/2/ae7a241673cf118ca18eca030dc29d2715b1980127dd0e2949514433.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9dca7c3956b03b7663fac4d150f5e6d4f6f38b2462c1e9afd83bcf7019f17913
|
3 |
+
size 1080679
|
.cache/pip/http-v2/d/3/3/a/b/d33abf9ad709d023fff05902f39da682c1afb233bcd9f2c479487586.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba0d021a166865d2265246961bec0152ff124de910c5cc39f1156ce3fa7c69dc
|
3 |
+
size 2110226
|
.cache/pip/http-v2/d/b/1/f/6/db1f6b45c0850c8e2ce7d8b47148edeca6e8115413af41f4ecc8ce32.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7ae2de664e0b158d1607699a16a488de3d008ba99b3a7aa5de1cbc13574d047
|
3 |
+
size 39855626
|
.cache/pip/http-v2/f/5/2/7/6/f52769e4b4d00542e1e056baf2db3e5ad8f277bff67f2636cace711d.body
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86976a1c5b25ae3f8ccae3a5306e443569ee3c3faf444dfd0f41cda24667ad57
|
3 |
+
size 13064210
|
.cache/pip/wheels/7e/e3/c3/89c7a2f3c4adc07cd1c675f8bb7b9ad4d18f64a72bccdfe826/flash_attn-2.6.3-cp310-cp310-linux_x86_64.whl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8991eedb5038a1ee6fc9904f99c12b40213d66753ed91e261a43d085f5aeab8f
|
3 |
+
size 187219571
|
.gitattributes
CHANGED
@@ -33,3 +33,23 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
.cache/pip/http-v2/a/e/7/a/2/ae7a241673cf118ca18eca030dc29d2715b1980127dd0e2949514433.body filter=lfs diff=lfs merge=lfs -text
|
37 |
+
.cache/pip/http-v2/0/2/d/4/2/02d4221e858694abc22129c65515f1df2c4c326330eb1a34ceb0b382.body filter=lfs diff=lfs merge=lfs -text
|
38 |
+
.local/share/jupyter/nbextensions/nbTranslate/demo2.gif filter=lfs diff=lfs merge=lfs -text
|
39 |
+
.cache/pip/http-v2/4/f/d/2/5/4fd254dbd56deb4021e55d22c4b489f6c776c69c316eb7345bc91691.body filter=lfs diff=lfs merge=lfs -text
|
40 |
+
.cache/pip/http-v2/0/1/f/2/0/01f2082df50502ba9492d64e69db99d1fdb5730707a16c6264b355b8.body filter=lfs diff=lfs merge=lfs -text
|
41 |
+
.cache/pip/http-v2/4/0/2/3/b/4023be7b5b37a7a4144c804ce69828082d4fb2a124d9d8aabc855da8.body filter=lfs diff=lfs merge=lfs -text
|
42 |
+
.local/share/jupyter/nbextensions/scratchpad/demo.gif filter=lfs diff=lfs merge=lfs -text
|
43 |
+
.cache/pip/http-v2/f/5/2/7/6/f52769e4b4d00542e1e056baf2db3e5ad8f277bff67f2636cace711d.body filter=lfs diff=lfs merge=lfs -text
|
44 |
+
.local/share/jupyter/nbextensions/go_to_current_running_cell/auto_focus.gif filter=lfs diff=lfs merge=lfs -text
|
45 |
+
.cache/pip/http-v2/9/e/8/c/8/9e8c8c0496d6d3384d616902379ed05e07b6b1dba9673d70b5fef231.body filter=lfs diff=lfs merge=lfs -text
|
46 |
+
.local/share/jupyter/nbextensions/nbTranslate/demo1.gif filter=lfs diff=lfs merge=lfs -text
|
47 |
+
.local/share/jupyter/nbextensions/toc2/demo.gif filter=lfs diff=lfs merge=lfs -text
|
48 |
+
.cache/pip/http-v2/d/3/3/a/b/d33abf9ad709d023fff05902f39da682c1afb233bcd9f2c479487586.body filter=lfs diff=lfs merge=lfs -text
|
49 |
+
wandb/run-20240927_074011-gzu8f7wl/run-gzu8f7wl.wandb filter=lfs diff=lfs merge=lfs -text
|
50 |
+
.cache/pip/http-v2/9/6/e/8/3/96e83221dd149da9a3d38feebc955beb2034effd910108971c5b167b.body filter=lfs diff=lfs merge=lfs -text
|
51 |
+
wandb/run-20240927_005424-60260ulk/run-60260ulk.wandb filter=lfs diff=lfs merge=lfs -text
|
52 |
+
wandb/run-20240926_180814-1klxtkie/run-1klxtkie.wandb filter=lfs diff=lfs merge=lfs -text
|
53 |
+
wandb/run-20240926_111257-7yvor5gh/run-7yvor5gh.wandb filter=lfs diff=lfs merge=lfs -text
|
54 |
+
.cache/pip/http-v2/d/b/1/f/6/db1f6b45c0850c8e2ce7d8b47148edeca6e8115413af41f4ecc8ce32.body filter=lfs diff=lfs merge=lfs -text
|
55 |
+
.cache/pip/wheels/7e/e3/c3/89c7a2f3c4adc07cd1c675f8bb7b9ad4d18f64a72bccdfe826/flash_attn-2.6.3-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
|
.launchpadlib/api.launchpad.net/cache/api.launchpad.net,devel,-application,vnd.sun.wadl+xml,2f09acb494bdefdbf8ef0d1396a05e86
ADDED
The diff for this file is too large to render.
See raw diff
|
|
.local/share/jupyter/nbextensions/go_to_current_running_cell/auto_focus.gif
ADDED
![]() |
Git LFS Details
|
.local/share/jupyter/nbextensions/nbTranslate/demo1.gif
ADDED
![]() |
Git LFS Details
|
.local/share/jupyter/nbextensions/nbTranslate/demo2.gif
ADDED
![]() |
Git LFS Details
|
.local/share/jupyter/nbextensions/scratchpad/demo.gif
ADDED
![]() |
Git LFS Details
|
.local/share/jupyter/nbextensions/toc2/demo.gif
ADDED
![]() |
Git LFS Details
|
.local/share/jupyter/nbextensions/zenmode/images/back11.jpg
ADDED
![]() |
.local/share/jupyter/nbextensions/zenmode/images/back21.jpg
ADDED
![]() |
.local/share/jupyter/nbextensions/zenmode/images/ipynblogo1.png
ADDED
![]() |
.triton/dump/11759acf26ac56366b171628132485d6/triton_.llir
ADDED
@@ -0,0 +1,347 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
; ModuleID = 'LLVMDialectModule'
|
2 |
+
source_filename = "LLVMDialectModule"
|
3 |
+
|
4 |
+
@global_smem = external addrspace(3) global [0 x i8]
|
5 |
+
@.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
|
6 |
+
|
7 |
+
define void @triton__0d1d2d3d4d5d6d7d8de9de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, ptr addrspace(1) %6, ptr addrspace(1) %7, i32 %8, i32 %9) local_unnamed_addr !dbg !7 {
|
8 |
+
%11 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
|
9 |
+
%12 = and i32 %11, 31, !dbg !10
|
10 |
+
%13 = lshr i32 %11, 5, !dbg !10
|
11 |
+
%14 = and i32 %13, 1, !dbg !10
|
12 |
+
%urem = and i32 %11, 63, !dbg !10
|
13 |
+
%15 = shl nuw nsw i32 %urem, 2, !dbg !10
|
14 |
+
%16 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !11
|
15 |
+
%17 = shl i32 %16, 8, !dbg !12
|
16 |
+
%18 = or i32 %17, %15, !dbg !13
|
17 |
+
%19 = sext i32 %18 to i64, !dbg !14
|
18 |
+
%20 = getelementptr float, ptr addrspace(1) %1, i64 %19, !dbg !14
|
19 |
+
%21 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %20, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !15
|
20 |
+
%22 = extractvalue { i32, i32, i32, i32 } %21, 0, !dbg !15
|
21 |
+
%23 = extractvalue { i32, i32, i32, i32 } %21, 1, !dbg !15
|
22 |
+
%24 = extractvalue { i32, i32, i32, i32 } %21, 2, !dbg !15
|
23 |
+
%25 = extractvalue { i32, i32, i32, i32 } %21, 3, !dbg !15
|
24 |
+
%26 = bitcast i32 %24 to float, !dbg !15
|
25 |
+
%27 = bitcast i32 %25 to float, !dbg !15
|
26 |
+
%28 = getelementptr i16, ptr addrspace(1) %2, i64 %19, !dbg !16
|
27 |
+
%29 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %28, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !17
|
28 |
+
%30 = extractvalue { i32, i32 } %29, 0, !dbg !17
|
29 |
+
%31 = extractvalue { i32, i32 } %29, 1, !dbg !17
|
30 |
+
%32 = trunc i32 %30 to i16, !dbg !17
|
31 |
+
%extelt.offset = lshr i32 %30, 16, !dbg !17
|
32 |
+
%33 = trunc i32 %extelt.offset to i16, !dbg !17
|
33 |
+
%34 = trunc i32 %31 to i16, !dbg !17
|
34 |
+
%extelt.offset1 = lshr i32 %31, 16, !dbg !17
|
35 |
+
%35 = trunc i32 %extelt.offset1 to i16, !dbg !17
|
36 |
+
%36 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %32) #6, !dbg !18
|
37 |
+
%37 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %33) #6, !dbg !18
|
38 |
+
%38 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %34) #6, !dbg !18
|
39 |
+
%39 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %35) #6, !dbg !18
|
40 |
+
%40 = getelementptr i16, ptr addrspace(1) %3, i64 %19, !dbg !19
|
41 |
+
%41 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %40, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !20
|
42 |
+
%42 = extractvalue { i32, i32 } %41, 0, !dbg !20
|
43 |
+
%43 = extractvalue { i32, i32 } %41, 1, !dbg !20
|
44 |
+
%44 = trunc i32 %42 to i16, !dbg !20
|
45 |
+
%extelt.offset2 = lshr i32 %42, 16, !dbg !20
|
46 |
+
%45 = trunc i32 %extelt.offset2 to i16, !dbg !20
|
47 |
+
%46 = trunc i32 %43 to i16, !dbg !20
|
48 |
+
%extelt.offset3 = lshr i32 %43, 16, !dbg !20
|
49 |
+
%47 = trunc i32 %extelt.offset3 to i16, !dbg !20
|
50 |
+
%48 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %44) #6, !dbg !21
|
51 |
+
%49 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %45) #6, !dbg !21
|
52 |
+
%50 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %46) #6, !dbg !21
|
53 |
+
%51 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %47) #6, !dbg !21
|
54 |
+
%52 = getelementptr i16, ptr addrspace(1) %4, i64 %19, !dbg !22
|
55 |
+
%53 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %52, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !23
|
56 |
+
%54 = extractvalue { i32, i32 } %53, 0, !dbg !23
|
57 |
+
%55 = extractvalue { i32, i32 } %53, 1, !dbg !23
|
58 |
+
%56 = trunc i32 %54 to i16, !dbg !23
|
59 |
+
%extelt.offset4 = lshr i32 %54, 16, !dbg !23
|
60 |
+
%57 = trunc i32 %extelt.offset4 to i16, !dbg !23
|
61 |
+
%58 = trunc i32 %55 to i16, !dbg !23
|
62 |
+
%extelt.offset5 = lshr i32 %55, 16, !dbg !23
|
63 |
+
%59 = trunc i32 %extelt.offset5 to i16, !dbg !23
|
64 |
+
%60 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %56) #6, !dbg !24
|
65 |
+
%61 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %57) #6, !dbg !24
|
66 |
+
%62 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %58) #6, !dbg !24
|
67 |
+
%63 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %59) #6, !dbg !24
|
68 |
+
%64 = zext nneg i32 %15 to i64, !dbg !25
|
69 |
+
%65 = getelementptr float, ptr addrspace(1) %5, i64 %64, !dbg !25
|
70 |
+
%66 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %65, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !26
|
71 |
+
%67 = fadd float %38, %26, !dbg !27
|
72 |
+
%68 = fadd float %39, %27, !dbg !27
|
73 |
+
%69 = fadd float %67, %50, !dbg !28
|
74 |
+
%70 = fadd float %68, %51, !dbg !28
|
75 |
+
%71 = insertelement <2 x i32> poison, i32 %22, i64 0, !dbg !15
|
76 |
+
%72 = insertelement <2 x i32> %71, i32 %23, i64 1, !dbg !15
|
77 |
+
%73 = bitcast <2 x i32> %72 to <2 x float>, !dbg !15
|
78 |
+
%74 = insertelement <2 x float> poison, float %36, i64 0, !dbg !27
|
79 |
+
%75 = insertelement <2 x float> %74, float %37, i64 1, !dbg !27
|
80 |
+
%76 = fadd <2 x float> %75, %73, !dbg !27
|
81 |
+
%77 = insertelement <2 x float> poison, float %48, i64 0, !dbg !28
|
82 |
+
%78 = insertelement <2 x float> %77, float %49, i64 1, !dbg !28
|
83 |
+
%79 = fadd <2 x float> %76, %78, !dbg !28
|
84 |
+
%80 = insertelement <2 x float> poison, float %60, i64 0, !dbg !29
|
85 |
+
%81 = insertelement <2 x float> %80, float %61, i64 1, !dbg !29
|
86 |
+
%82 = fadd <2 x float> %79, %81, !dbg !29
|
87 |
+
%83 = fadd float %69, %62, !dbg !29
|
88 |
+
%84 = fadd float %70, %63, !dbg !29
|
89 |
+
%85 = extractelement <2 x float> %82, i64 0, !dbg !30
|
90 |
+
%86 = extractelement <2 x float> %82, i64 1, !dbg !30
|
91 |
+
%87 = fadd float %85, %86, !dbg !30
|
92 |
+
%88 = fadd float %87, %83, !dbg !30
|
93 |
+
%89 = fadd float %88, %84, !dbg !30
|
94 |
+
%90 = bitcast float %89 to i32, !dbg !36
|
95 |
+
%91 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %90, i32 16, i32 31), !dbg !36
|
96 |
+
%92 = bitcast i32 %91 to float, !dbg !36
|
97 |
+
%93 = fadd float %89, %92, !dbg !30
|
98 |
+
%94 = bitcast float %93 to i32, !dbg !36
|
99 |
+
%95 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %94, i32 8, i32 31), !dbg !36
|
100 |
+
%96 = bitcast i32 %95 to float, !dbg !36
|
101 |
+
%97 = fadd float %93, %96, !dbg !30
|
102 |
+
%98 = bitcast float %97 to i32, !dbg !36
|
103 |
+
%99 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %98, i32 4, i32 31), !dbg !36
|
104 |
+
%100 = bitcast i32 %99 to float, !dbg !36
|
105 |
+
%101 = fadd float %97, %100, !dbg !30
|
106 |
+
%102 = bitcast float %101 to i32, !dbg !36
|
107 |
+
%103 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %102, i32 2, i32 31), !dbg !36
|
108 |
+
%104 = bitcast i32 %103 to float, !dbg !36
|
109 |
+
%105 = fadd float %101, %104, !dbg !30
|
110 |
+
%106 = bitcast float %105 to i32, !dbg !36
|
111 |
+
%107 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %106, i32 1, i32 31), !dbg !36
|
112 |
+
%108 = bitcast i32 %107 to float, !dbg !36
|
113 |
+
%109 = fadd float %105, %108, !dbg !30
|
114 |
+
%110 = icmp eq i32 %12, 0, !dbg !36
|
115 |
+
%111 = zext nneg i32 %14 to i64, !dbg !36
|
116 |
+
%112 = getelementptr float, ptr addrspace(3) @global_smem, i64 %111, !dbg !36
|
117 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %112, float %109, i1 %110) #6, !dbg !36
|
118 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !36
|
119 |
+
%113 = icmp slt i32 %11, 2, !dbg !36
|
120 |
+
%114 = sext i32 %11 to i64, !dbg !36
|
121 |
+
%115 = getelementptr float, ptr addrspace(3) @global_smem, i64 %114, !dbg !36
|
122 |
+
%116 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %115, i1 %113) #6, !dbg !36
|
123 |
+
%117 = bitcast float %116 to i32, !dbg !36
|
124 |
+
%118 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %117, i32 1, i32 31), !dbg !36
|
125 |
+
%119 = bitcast i32 %118 to float, !dbg !36
|
126 |
+
%120 = fadd float %116, %119, !dbg !30
|
127 |
+
%121 = and i32 %11, 1, !dbg !36
|
128 |
+
%122 = icmp eq i32 %121, 0, !dbg !36
|
129 |
+
%123 = and i1 %113, %122, !dbg !36
|
130 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %115, float %120, i1 %123) #6, !dbg !36
|
131 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !36
|
132 |
+
%124 = load float, ptr addrspace(3) @global_smem, align 4, !dbg !36
|
133 |
+
%125 = fadd float %124, 0.000000e+00, !dbg !38
|
134 |
+
%126 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %125, float 2.560000e+02) #6, !dbg !42
|
135 |
+
%127 = fsub float %85, %126, !dbg !43
|
136 |
+
%128 = fsub float %86, %126, !dbg !43
|
137 |
+
%129 = fsub float %83, %126, !dbg !43
|
138 |
+
%130 = fsub float %84, %126, !dbg !43
|
139 |
+
%131 = fmul float %127, %127, !dbg !44
|
140 |
+
%132 = fmul float %128, %128, !dbg !44
|
141 |
+
%133 = fmul float %129, %129, !dbg !44
|
142 |
+
%134 = fmul float %130, %130, !dbg !44
|
143 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !45
|
144 |
+
%135 = fadd float %131, %132, !dbg !47
|
145 |
+
%136 = fadd float %133, %135, !dbg !47
|
146 |
+
%137 = fadd float %134, %136, !dbg !47
|
147 |
+
%138 = bitcast float %137 to i32, !dbg !45
|
148 |
+
%139 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %138, i32 16, i32 31), !dbg !45
|
149 |
+
%140 = bitcast i32 %139 to float, !dbg !45
|
150 |
+
%141 = fadd float %137, %140, !dbg !47
|
151 |
+
%142 = bitcast float %141 to i32, !dbg !45
|
152 |
+
%143 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %142, i32 8, i32 31), !dbg !45
|
153 |
+
%144 = bitcast i32 %143 to float, !dbg !45
|
154 |
+
%145 = fadd float %141, %144, !dbg !47
|
155 |
+
%146 = bitcast float %145 to i32, !dbg !45
|
156 |
+
%147 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %146, i32 4, i32 31), !dbg !45
|
157 |
+
%148 = bitcast i32 %147 to float, !dbg !45
|
158 |
+
%149 = fadd float %145, %148, !dbg !47
|
159 |
+
%150 = bitcast float %149 to i32, !dbg !45
|
160 |
+
%151 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %150, i32 2, i32 31), !dbg !45
|
161 |
+
%152 = bitcast i32 %151 to float, !dbg !45
|
162 |
+
%153 = fadd float %149, %152, !dbg !47
|
163 |
+
%154 = bitcast float %153 to i32, !dbg !45
|
164 |
+
%155 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %154, i32 1, i32 31), !dbg !45
|
165 |
+
%156 = bitcast i32 %155 to float, !dbg !45
|
166 |
+
%157 = fadd float %153, %156, !dbg !47
|
167 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %112, float %157, i1 %110) #6, !dbg !45
|
168 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !45
|
169 |
+
%158 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %115, i1 %113) #6, !dbg !45
|
170 |
+
%159 = bitcast float %158 to i32, !dbg !45
|
171 |
+
%160 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %159, i32 1, i32 31), !dbg !45
|
172 |
+
%161 = bitcast i32 %160 to float, !dbg !45
|
173 |
+
%162 = fadd float %158, %161, !dbg !47
|
174 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %115, float %162, i1 %123) #6, !dbg !45
|
175 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !45
|
176 |
+
%163 = load float, ptr addrspace(3) @global_smem, align 4, !dbg !45
|
177 |
+
%164 = fadd float %163, 0.000000e+00, !dbg !50
|
178 |
+
%165 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %164, float 2.560000e+02) #6, !dbg !52
|
179 |
+
%166 = fadd float %165, 0x3EE4F8B580000000, !dbg !53
|
180 |
+
%167 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !54
|
181 |
+
%.not.i = icmp eq i32 %167, 0, !dbg !54
|
182 |
+
br i1 %.not.i, label %170, label %168, !dbg !54
|
183 |
+
|
184 |
+
168: ; preds = %10
|
185 |
+
%169 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %166), !dbg !54
|
186 |
+
br label %__nv_rsqrtf.exit, !dbg !54
|
187 |
+
|
188 |
+
170: ; preds = %10
|
189 |
+
%171 = tail call float @llvm.nvvm.rsqrt.approx.f(float %166), !dbg !54
|
190 |
+
br label %__nv_rsqrtf.exit, !dbg !54
|
191 |
+
|
192 |
+
__nv_rsqrtf.exit: ; preds = %168, %170
|
193 |
+
%.0.i = phi float [ %169, %168 ], [ %171, %170 ], !dbg !54
|
194 |
+
%172 = extractvalue { i32, i32, i32, i32 } %66, 3, !dbg !26
|
195 |
+
%173 = bitcast i32 %172 to float, !dbg !26
|
196 |
+
%174 = extractvalue { i32, i32, i32, i32 } %66, 2, !dbg !26
|
197 |
+
%175 = bitcast i32 %174 to float, !dbg !26
|
198 |
+
%176 = extractvalue { i32, i32, i32, i32 } %66, 1, !dbg !26
|
199 |
+
%177 = bitcast i32 %176 to float, !dbg !26
|
200 |
+
%178 = extractvalue { i32, i32, i32, i32 } %66, 0, !dbg !26
|
201 |
+
%179 = bitcast i32 %178 to float, !dbg !26
|
202 |
+
%180 = fmul float %127, %.0.i, !dbg !55
|
203 |
+
%181 = fmul float %128, %.0.i, !dbg !55
|
204 |
+
%182 = fmul float %129, %.0.i, !dbg !55
|
205 |
+
%183 = fmul float %130, %.0.i, !dbg !55
|
206 |
+
%184 = fmul float %180, %179, !dbg !56
|
207 |
+
%185 = fmul float %181, %177, !dbg !56
|
208 |
+
%186 = fmul float %182, %175, !dbg !56
|
209 |
+
%187 = fmul float %183, %173, !dbg !56
|
210 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !57
|
211 |
+
%188 = sext i32 %16 to i64, !dbg !58
|
212 |
+
%189 = getelementptr float, ptr addrspace(1) %0, i64 %188, !dbg !58
|
213 |
+
%190 = icmp eq i32 %urem, 0, !dbg !59
|
214 |
+
%191 = bitcast float %.0.i to i32, !dbg !59
|
215 |
+
tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %191, ptr addrspace(1) %189, i1 %190) #6, !dbg !59
|
216 |
+
%192 = getelementptr i16, ptr addrspace(1) %7, i64 %19, !dbg !60
|
217 |
+
%193 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %184) #6, !dbg !61
|
218 |
+
%194 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %185) #6, !dbg !61
|
219 |
+
%195 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %186) #6, !dbg !61
|
220 |
+
%196 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %187) #6, !dbg !61
|
221 |
+
%197 = insertelement <2 x i16> undef, i16 %193, i64 0, !dbg !61
|
222 |
+
%198 = insertelement <2 x i16> %197, i16 %194, i64 1, !dbg !61
|
223 |
+
%199 = bitcast <2 x i16> %198 to i32, !dbg !61
|
224 |
+
%200 = insertelement <2 x i16> undef, i16 %195, i64 0, !dbg !61
|
225 |
+
%201 = insertelement <2 x i16> %200, i16 %196, i64 1, !dbg !61
|
226 |
+
%202 = bitcast <2 x i16> %201 to i32, !dbg !61
|
227 |
+
tail call void asm sideeffect "@$3 st.global.v2.b32 [ $2 + 0 ], { $0, $1 };", "r,r,l,b"(i32 %199, i32 %202, ptr addrspace(1) %192, i1 true) #6, !dbg !61
|
228 |
+
%203 = getelementptr float, ptr addrspace(1) %6, i64 %188, !dbg !62
|
229 |
+
%204 = bitcast float %126 to i32, !dbg !63
|
230 |
+
tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %204, ptr addrspace(1) %203, i1 %190) #6, !dbg !63
|
231 |
+
ret void, !dbg !64
|
232 |
+
}
|
233 |
+
|
234 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
|
235 |
+
declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
|
236 |
+
|
237 |
+
; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
|
238 |
+
declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
|
239 |
+
|
240 |
+
; Function Attrs: convergent nocallback nounwind
|
241 |
+
declare void @llvm.nvvm.barrier0() #2
|
242 |
+
|
243 |
+
; Function Attrs: alwaysinline nounwind
|
244 |
+
define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
|
245 |
+
%1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
|
246 |
+
%.not = icmp eq i32 %1, 0
|
247 |
+
br i1 %.not, label %4, label %2
|
248 |
+
|
249 |
+
2: ; preds = %0
|
250 |
+
%3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
|
251 |
+
br label %6
|
252 |
+
|
253 |
+
4: ; preds = %0
|
254 |
+
%5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
|
255 |
+
br label %6
|
256 |
+
|
257 |
+
6: ; preds = %4, %2
|
258 |
+
%.0 = phi float [ %3, %2 ], [ %5, %4 ]
|
259 |
+
ret float %.0
|
260 |
+
}
|
261 |
+
|
262 |
+
declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
|
263 |
+
|
264 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
265 |
+
declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
|
266 |
+
|
267 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
268 |
+
declare float @llvm.nvvm.rsqrt.approx.f(float) #5
|
269 |
+
|
270 |
+
attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
|
271 |
+
attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
|
272 |
+
attributes #2 = { convergent nocallback nounwind }
|
273 |
+
attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
274 |
+
attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
275 |
+
attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
|
276 |
+
attributes #6 = { nounwind }
|
277 |
+
|
278 |
+
!llvm.module.flags = !{!0, !1}
|
279 |
+
!llvm.dbg.cu = !{!2}
|
280 |
+
!nvvm.annotations = !{!4, !5, !5, !4}
|
281 |
+
!llvm.ident = !{!6}
|
282 |
+
|
283 |
+
!0 = !{i32 2, !"Debug Info Version", i32 3}
|
284 |
+
!1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
|
285 |
+
!2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
|
286 |
+
!3 = !DIFile(filename: "cdxa5yqgsimvskocpuiz4ajfrjfcwys3opyrdv53xfphj4576qx7.py", directory: "/tmp/torchinductor_root/dx")
|
287 |
+
!4 = !{ptr @triton__0d1d2d3d4d5d6d7d8de9de, !"kernel", i32 1}
|
288 |
+
!5 = !{ptr @triton__0d1d2d3d4d5d6d7d8de9de, !"maxntidx", i32 64}
|
289 |
+
!6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
|
290 |
+
!7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5d6d7d8de9de", linkageName: "triton__0d1d2d3d4d5d6d7d8de9de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
|
291 |
+
!8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
|
292 |
+
!9 = !{}
|
293 |
+
!10 = !DILocation(line: 26, column: 26, scope: !7)
|
294 |
+
!11 = !DILocation(line: 23, column: 28, scope: !7)
|
295 |
+
!12 = !DILocation(line: 30, column: 40, scope: !7)
|
296 |
+
!13 = !DILocation(line: 30, column: 36, scope: !7)
|
297 |
+
!14 = !DILocation(line: 30, column: 30, scope: !7)
|
298 |
+
!15 = !DILocation(line: 30, column: 46, scope: !7)
|
299 |
+
!16 = !DILocation(line: 31, column: 30, scope: !7)
|
300 |
+
!17 = !DILocation(line: 31, column: 46, scope: !7)
|
301 |
+
!18 = !DILocation(line: 31, column: 67, scope: !7)
|
302 |
+
!19 = !DILocation(line: 32, column: 30, scope: !7)
|
303 |
+
!20 = !DILocation(line: 32, column: 46, scope: !7)
|
304 |
+
!21 = !DILocation(line: 32, column: 67, scope: !7)
|
305 |
+
!22 = !DILocation(line: 33, column: 30, scope: !7)
|
306 |
+
!23 = !DILocation(line: 33, column: 46, scope: !7)
|
307 |
+
!24 = !DILocation(line: 33, column: 67, scope: !7)
|
308 |
+
!25 = !DILocation(line: 34, column: 31, scope: !7)
|
309 |
+
!26 = !DILocation(line: 34, column: 36, scope: !7)
|
310 |
+
!27 = !DILocation(line: 36, column: 18, scope: !7)
|
311 |
+
!28 = !DILocation(line: 38, column: 18, scope: !7)
|
312 |
+
!29 = !DILocation(line: 40, column: 18, scope: !7)
|
313 |
+
!30 = !DILocation(line: 233, column: 15, scope: !31, inlinedAt: !34)
|
314 |
+
!31 = distinct !DILexicalBlockFile(scope: !33, file: !32, discriminator: 0)
|
315 |
+
!32 = !DIFile(filename: "standard.py", directory: "/usr/local/lib/python3.10/dist-packages/triton/language")
|
316 |
+
!33 = distinct !DILexicalBlockFile(scope: !7, file: !32, discriminator: 0)
|
317 |
+
!34 = !DILocation(line: 243, column: 36, scope: !31, inlinedAt: !35)
|
318 |
+
!35 = !DILocation(line: 45, column: 59, scope: !31)
|
319 |
+
!36 = !DILocation(line: 243, column: 36, scope: !33, inlinedAt: !37)
|
320 |
+
!37 = !DILocation(line: 45, column: 59, scope: !33)
|
321 |
+
!38 = !DILocation(line: 8, column: 15, scope: !39, inlinedAt: !41)
|
322 |
+
!39 = distinct !DILexicalBlockFile(scope: !7, file: !40, discriminator: 0)
|
323 |
+
!40 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
|
324 |
+
!41 = !DILocation(line: 45, column: 45, scope: !39)
|
325 |
+
!42 = !DILocation(line: 48, column: 20, scope: !7)
|
326 |
+
!43 = !DILocation(line: 49, column: 20, scope: !7)
|
327 |
+
!44 = !DILocation(line: 50, column: 20, scope: !7)
|
328 |
+
!45 = !DILocation(line: 243, column: 36, scope: !33, inlinedAt: !46)
|
329 |
+
!46 = !DILocation(line: 53, column: 59, scope: !33)
|
330 |
+
!47 = !DILocation(line: 233, column: 15, scope: !31, inlinedAt: !48)
|
331 |
+
!48 = !DILocation(line: 243, column: 36, scope: !31, inlinedAt: !49)
|
332 |
+
!49 = !DILocation(line: 53, column: 59, scope: !31)
|
333 |
+
!50 = !DILocation(line: 8, column: 15, scope: !39, inlinedAt: !51)
|
334 |
+
!51 = !DILocation(line: 53, column: 45, scope: !39)
|
335 |
+
!52 = !DILocation(line: 55, column: 20, scope: !7)
|
336 |
+
!53 = !DILocation(line: 57, column: 20, scope: !7)
|
337 |
+
!54 = !DILocation(line: 58, column: 26, scope: !7)
|
338 |
+
!55 = !DILocation(line: 60, column: 20, scope: !7)
|
339 |
+
!56 = !DILocation(line: 61, column: 20, scope: !7)
|
340 |
+
!57 = !DILocation(line: 63, column: 4, scope: !7)
|
341 |
+
!58 = !DILocation(line: 64, column: 28, scope: !7)
|
342 |
+
!59 = !DILocation(line: 64, column: 40, scope: !7)
|
343 |
+
!60 = !DILocation(line: 65, column: 25, scope: !7)
|
344 |
+
!61 = !DILocation(line: 65, column: 48, scope: !7)
|
345 |
+
!62 = !DILocation(line: 66, column: 25, scope: !7)
|
346 |
+
!63 = !DILocation(line: 66, column: 37, scope: !7)
|
347 |
+
!64 = !DILocation(line: 66, column: 4, scope: !7)
|
.triton/dump/11759acf26ac56366b171628132485d6/triton_.ttgir
ADDED
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [4], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
|
2 |
+
#blocked1 = #triton_gpu.blocked<{sizePerThread = [1], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
|
3 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
4 |
+
tt.func public @triton__0d1d2d3d4d5d6d7d8de9de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg6: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg7: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg8: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg9: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
5 |
+
%cst = arith.constant dense<256> : tensor<256xi32, #blocked>
|
6 |
+
%cst_0 = arith.constant 9.99999974E-6 : f32
|
7 |
+
%cst_1 = arith.constant 2.560000e+02 : f32
|
8 |
+
%cst_2 = arith.constant 0.000000e+00 : f32
|
9 |
+
%c256_i32 = arith.constant 256 : i32
|
10 |
+
%cst_3 = arith.constant dense<0.000000e+00> : tensor<256xf32, #blocked>
|
11 |
+
%cst_4 = arith.constant dense<0.000000e+00> : tensor<256xbf16, #blocked>
|
12 |
+
%0 = tt.get_program_id x : i32
|
13 |
+
%1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32, #blocked>
|
14 |
+
%2 = arith.cmpi slt, %1, %cst : tensor<256xi32, #blocked>
|
15 |
+
%3 = arith.muli %0, %c256_i32 : i32
|
16 |
+
%4 = tt.splat %3 : (i32) -> tensor<256xi32, #blocked>
|
17 |
+
%5 = arith.addi %1, %4 : tensor<256xi32, #blocked>
|
18 |
+
%6 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
19 |
+
%7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
20 |
+
%8 = tt.load %7, %2, %cst_3 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
21 |
+
%9 = tt.splat %arg2 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
22 |
+
%10 = tt.addptr %9, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
23 |
+
%11 = tt.load %10, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
|
24 |
+
%12 = arith.extf %11 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
|
25 |
+
%13 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
26 |
+
%14 = tt.addptr %13, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
27 |
+
%15 = tt.load %14, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
|
28 |
+
%16 = arith.extf %15 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
|
29 |
+
%17 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
30 |
+
%18 = tt.addptr %17, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
31 |
+
%19 = tt.load %18, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
|
32 |
+
%20 = arith.extf %19 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
|
33 |
+
%21 = tt.splat %arg5 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
34 |
+
%22 = tt.addptr %21, %1 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
35 |
+
%23 = tt.load %22, %2, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
36 |
+
%24 = arith.addf %8, %12 : tensor<256xf32, #blocked>
|
37 |
+
%25 = arith.addf %24, %16 : tensor<256xf32, #blocked>
|
38 |
+
%26 = arith.addf %25, %20 : tensor<256xf32, #blocked>
|
39 |
+
%27 = arith.select %2, %26, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
40 |
+
%28 = "tt.reduce"(%27) <{axis = 0 : i32}> ({
|
41 |
+
^bb0(%arg10: f32, %arg11: f32):
|
42 |
+
%52 = arith.addf %arg10, %arg11 : f32
|
43 |
+
tt.reduce.return %52 : f32
|
44 |
+
}) : (tensor<256xf32, #blocked>) -> f32
|
45 |
+
%29 = arith.addf %28, %cst_2 : f32
|
46 |
+
%30 = arith.divf %29, %cst_1 : f32
|
47 |
+
%31 = tt.splat %30 : (f32) -> tensor<1xf32, #blocked1>
|
48 |
+
%32 = tt.splat %30 : (f32) -> tensor<256xf32, #blocked>
|
49 |
+
%33 = arith.subf %26, %32 : tensor<256xf32, #blocked>
|
50 |
+
%34 = arith.mulf %33, %33 : tensor<256xf32, #blocked>
|
51 |
+
%35 = arith.select %2, %34, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
52 |
+
%36 = "tt.reduce"(%35) <{axis = 0 : i32}> ({
|
53 |
+
^bb0(%arg10: f32, %arg11: f32):
|
54 |
+
%52 = arith.addf %arg10, %arg11 : f32
|
55 |
+
tt.reduce.return %52 : f32
|
56 |
+
}) : (tensor<256xf32, #blocked>) -> f32
|
57 |
+
%37 = arith.addf %36, %cst_2 : f32
|
58 |
+
%38 = arith.divf %37, %cst_1 : f32
|
59 |
+
%39 = arith.addf %38, %cst_0 : f32
|
60 |
+
%40 = tt.extern_elementwise %39 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (f32) -> f32
|
61 |
+
%41 = tt.splat %40 : (f32) -> tensor<1xf32, #blocked1>
|
62 |
+
%42 = tt.splat %40 : (f32) -> tensor<256xf32, #blocked>
|
63 |
+
%43 = arith.mulf %33, %42 : tensor<256xf32, #blocked>
|
64 |
+
%44 = arith.mulf %43, %23 : tensor<256xf32, #blocked>
|
65 |
+
gpu.barrier
|
66 |
+
%45 = tt.addptr %arg0, %0 : !tt.ptr<f32, 1>, i32
|
67 |
+
%46 = tt.splat %45 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
|
68 |
+
tt.store %46, %41 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
|
69 |
+
%47 = tt.splat %arg7 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
70 |
+
%48 = tt.addptr %47, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
71 |
+
%49 = arith.truncf %44 : tensor<256xf32, #blocked> to tensor<256xbf16, #blocked>
|
72 |
+
tt.store %48, %49, %2 {cache = 1 : i32, evict = 1 : i32} : tensor<256xbf16, #blocked>
|
73 |
+
%50 = tt.addptr %arg6, %0 : !tt.ptr<f32, 1>, i32
|
74 |
+
%51 = tt.splat %50 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
|
75 |
+
tt.store %51, %31 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
|
76 |
+
tt.return
|
77 |
+
}
|
78 |
+
}
|
.triton/dump/11759acf26ac56366b171628132485d6/triton_.ttir
ADDED
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
module {
|
2 |
+
tt.func public @triton__0d1d2d3d4d5d6d7d8de9de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg6: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg7: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg8: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg9: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
3 |
+
%c256_i32 = arith.constant 256 : i32
|
4 |
+
%cst = arith.constant dense<0.000000e+00> : tensor<256xbf16>
|
5 |
+
%cst_0 = arith.constant 0.000000e+00 : f32
|
6 |
+
%cst_1 = arith.constant 2.560000e+02 : f32
|
7 |
+
%cst_2 = arith.constant 9.99999974E-6 : f32
|
8 |
+
%cst_3 = arith.constant dense<0.000000e+00> : tensor<256xf32>
|
9 |
+
%cst_4 = arith.constant dense<256> : tensor<256xi32>
|
10 |
+
%0 = tt.get_program_id x : i32
|
11 |
+
%1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32>
|
12 |
+
%2 = arith.cmpi slt, %1, %cst_4 : tensor<256xi32>
|
13 |
+
%3 = arith.muli %0, %c256_i32 : i32
|
14 |
+
%4 = tt.splat %3 : (i32) -> tensor<256xi32>
|
15 |
+
%5 = arith.addi %1, %4 : tensor<256xi32>
|
16 |
+
%6 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>>
|
17 |
+
%7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<f32, 1>>, tensor<256xi32>
|
18 |
+
%8 = tt.load %7, %2, %cst_3 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32>
|
19 |
+
%9 = tt.splat %arg2 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
|
20 |
+
%10 = tt.addptr %9, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
|
21 |
+
%11 = tt.load %10, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16>
|
22 |
+
%12 = arith.extf %11 : tensor<256xbf16> to tensor<256xf32>
|
23 |
+
%13 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
|
24 |
+
%14 = tt.addptr %13, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
|
25 |
+
%15 = tt.load %14, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16>
|
26 |
+
%16 = arith.extf %15 : tensor<256xbf16> to tensor<256xf32>
|
27 |
+
%17 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
|
28 |
+
%18 = tt.addptr %17, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
|
29 |
+
%19 = tt.load %18, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16>
|
30 |
+
%20 = arith.extf %19 : tensor<256xbf16> to tensor<256xf32>
|
31 |
+
%21 = tt.splat %arg5 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>>
|
32 |
+
%22 = tt.addptr %21, %1 : tensor<256x!tt.ptr<f32, 1>>, tensor<256xi32>
|
33 |
+
%23 = tt.load %22, %2, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32>
|
34 |
+
%24 = arith.addf %8, %12 : tensor<256xf32>
|
35 |
+
%25 = arith.addf %24, %16 : tensor<256xf32>
|
36 |
+
%26 = arith.addf %25, %20 : tensor<256xf32>
|
37 |
+
%27 = arith.select %2, %26, %cst_3 : tensor<256xi1>, tensor<256xf32>
|
38 |
+
%28 = "tt.reduce"(%27) <{axis = 0 : i32}> ({
|
39 |
+
^bb0(%arg10: f32, %arg11: f32):
|
40 |
+
%52 = arith.addf %arg10, %arg11 : f32
|
41 |
+
tt.reduce.return %52 : f32
|
42 |
+
}) : (tensor<256xf32>) -> f32
|
43 |
+
%29 = arith.addf %28, %cst_0 : f32
|
44 |
+
%30 = arith.divf %29, %cst_1 : f32
|
45 |
+
%31 = tt.splat %30 : (f32) -> tensor<1xf32>
|
46 |
+
%32 = tt.splat %30 : (f32) -> tensor<256xf32>
|
47 |
+
%33 = arith.subf %26, %32 : tensor<256xf32>
|
48 |
+
%34 = arith.mulf %33, %33 : tensor<256xf32>
|
49 |
+
%35 = arith.select %2, %34, %cst_3 : tensor<256xi1>, tensor<256xf32>
|
50 |
+
%36 = "tt.reduce"(%35) <{axis = 0 : i32}> ({
|
51 |
+
^bb0(%arg10: f32, %arg11: f32):
|
52 |
+
%52 = arith.addf %arg10, %arg11 : f32
|
53 |
+
tt.reduce.return %52 : f32
|
54 |
+
}) : (tensor<256xf32>) -> f32
|
55 |
+
%37 = arith.addf %36, %cst_0 : f32
|
56 |
+
%38 = arith.divf %37, %cst_1 : f32
|
57 |
+
%39 = arith.addf %38, %cst_2 : f32
|
58 |
+
%40 = tt.extern_elementwise %39 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (f32) -> f32
|
59 |
+
%41 = tt.splat %40 : (f32) -> tensor<1xf32>
|
60 |
+
%42 = tt.splat %40 : (f32) -> tensor<256xf32>
|
61 |
+
%43 = arith.mulf %33, %42 : tensor<256xf32>
|
62 |
+
%44 = arith.mulf %43, %23 : tensor<256xf32>
|
63 |
+
gpu.barrier
|
64 |
+
%45 = tt.addptr %arg0, %0 : !tt.ptr<f32, 1>, i32
|
65 |
+
%46 = tt.splat %45 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>>
|
66 |
+
tt.store %46, %41 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32>
|
67 |
+
%47 = tt.splat %arg7 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
|
68 |
+
%48 = tt.addptr %47, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
|
69 |
+
%49 = arith.truncf %44 : tensor<256xf32> to tensor<256xbf16>
|
70 |
+
tt.store %48, %49, %2 {cache = 1 : i32, evict = 1 : i32} : tensor<256xbf16>
|
71 |
+
%50 = tt.addptr %arg6, %0 : !tt.ptr<f32, 1>, i32
|
72 |
+
%51 = tt.splat %50 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>>
|
73 |
+
tt.store %51, %31 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32>
|
74 |
+
tt.return
|
75 |
+
}
|
76 |
+
}
|
.triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ptx
ADDED
@@ -0,0 +1,988 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
//
|
2 |
+
// Generated by LLVM NVPTX Back-End
|
3 |
+
//
|
4 |
+
|
5 |
+
.version 8.2
|
6 |
+
.target sm_89
|
7 |
+
.address_size 64
|
8 |
+
|
9 |
+
// .globl triton__0d1d2d3d4d5de6de
|
10 |
+
.extern .func __assertfail
|
11 |
+
(
|
12 |
+
.param .b64 __assertfail_param_0,
|
13 |
+
.param .b64 __assertfail_param_1,
|
14 |
+
.param .b32 __assertfail_param_2,
|
15 |
+
.param .b64 __assertfail_param_3,
|
16 |
+
.param .b64 __assertfail_param_4
|
17 |
+
)
|
18 |
+
;
|
19 |
+
.global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
20 |
+
.global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
21 |
+
.global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
22 |
+
.global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
23 |
+
.global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
24 |
+
.global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
25 |
+
.extern .shared .align 1 .b8 global_smem[];
|
26 |
+
.global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
|
27 |
+
|
28 |
+
.visible .entry triton__0d1d2d3d4d5de6de(
|
29 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_0,
|
30 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_1,
|
31 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_2,
|
32 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_3,
|
33 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_4,
|
34 |
+
.param .u32 triton__0d1d2d3d4d5de6de_param_5,
|
35 |
+
.param .u32 triton__0d1d2d3d4d5de6de_param_6
|
36 |
+
)
|
37 |
+
.maxntid 128, 1, 1
|
38 |
+
{
|
39 |
+
.reg .pred %p<50>;
|
40 |
+
.reg .b16 %rs<5>;
|
41 |
+
.reg .b32 %r<169>;
|
42 |
+
.reg .f32 %f<153>;
|
43 |
+
.reg .b64 %rd<53>;
|
44 |
+
.loc 1 18 0
|
45 |
+
$L__func_begin0:
|
46 |
+
.loc 1 18 0
|
47 |
+
|
48 |
+
ld.param.u64 %rd6, [triton__0d1d2d3d4d5de6de_param_3];
|
49 |
+
ld.param.u64 %rd5, [triton__0d1d2d3d4d5de6de_param_1];
|
50 |
+
ld.param.u64 %rd19, [triton__0d1d2d3d4d5de6de_param_0];
|
51 |
+
$L__tmp0:
|
52 |
+
.loc 1 22 44
|
53 |
+
mov.u32 %r1, %tid.x;
|
54 |
+
and.b32 %r2, %r1, 31;
|
55 |
+
ld.param.u64 %rd20, [triton__0d1d2d3d4d5de6de_param_2];
|
56 |
+
bfe.u32 %r3, %r1, 6, 1;
|
57 |
+
and.b32 %r4, %r1, 1;
|
58 |
+
.loc 1 24 33
|
59 |
+
bfe.u32 %r5, %r1, 5, 1;
|
60 |
+
shl.b32 %r24, %r1, 2;
|
61 |
+
and.b32 %r6, %r24, 252;
|
62 |
+
shl.b32 %r25, %r1, 1;
|
63 |
+
and.b32 %r7, %r25, 254;
|
64 |
+
.loc 1 21 28
|
65 |
+
mov.u32 %r15, %ctaid.x;
|
66 |
+
.loc 1 21 33
|
67 |
+
shl.b32 %r26, %r15, 1;
|
68 |
+
.loc 1 22 23
|
69 |
+
or.b32 %r8, %r26, %r3;
|
70 |
+
or.b32 %r27, %r26, %r4;
|
71 |
+
.loc 1 26 30
|
72 |
+
mul.wide.s32 %rd21, %r8, 8;
|
73 |
+
add.s64 %rd9, %rd19, %rd21;
|
74 |
+
mul.wide.s32 %rd22, %r27, 8;
|
75 |
+
add.s64 %rd17, %rd19, %rd22;
|
76 |
+
mov.pred %p44, -1;
|
77 |
+
.loc 1 26 35
|
78 |
+
mov.u64 %rd8, 0x0;
|
79 |
+
@%p44 ld.global.L1::evict_last.b64 { %rd8 }, [ %rd9 + 0 ];
|
80 |
+
mov.u64 %rd10, 0x0;
|
81 |
+
@%p44 ld.global.L1::evict_last.b64 { %rd10 }, [ %rd9 + 0 ];
|
82 |
+
mov.u64 %rd12, 0x0;
|
83 |
+
@%p44 ld.global.L1::evict_last.b64 { %rd12 }, [ %rd9 + 0 ];
|
84 |
+
mov.u64 %rd14, 0x0;
|
85 |
+
@%p44 ld.global.L1::evict_last.b64 { %rd14 }, [ %rd9 + 0 ];
|
86 |
+
mov.u64 %rd16, 0x0;
|
87 |
+
@%p44 ld.global.L1::evict_last.b64 { %rd16 }, [ %rd17 + 0 ];
|
88 |
+
.loc 1 27 18
|
89 |
+
bfe.s32 %r28, %r15, 30, 1;
|
90 |
+
shr.u32 %r29, %r28, 23;
|
91 |
+
add.s32 %r30, %r8, %r29;
|
92 |
+
and.b32 %r31, %r30, 16776704;
|
93 |
+
sub.s32 %r32, %r8, %r31;
|
94 |
+
.loc 1 35 44
|
95 |
+
shl.b32 %r33, %r32, 8;
|
96 |
+
.loc 1 35 40
|
97 |
+
or.b32 %r34, %r33, %r6;
|
98 |
+
.loc 1 35 34
|
99 |
+
mul.wide.s32 %rd23, %r34, 4;
|
100 |
+
add.s64 %rd33, %rd20, %rd23;
|
101 |
+
mov.b32 %r137, 0;
|
102 |
+
.loc 1 35 50
|
103 |
+
mov.u32 %r16, 0x0;
|
104 |
+
mov.u32 %r17, 0x0;
|
105 |
+
mov.u32 %r18, 0x0;
|
106 |
+
mov.u32 %r19, 0x0;
|
107 |
+
@%p44 ld.global.L1::evict_last.v4.b32 { %r16, %r17, %r18, %r19 }, [ %rd33 + 0 ];
|
108 |
+
@!%p44 mov.u32 %r16, %r137;
|
109 |
+
@!%p44 mov.u32 %r17, %r137;
|
110 |
+
@!%p44 mov.u32 %r18, %r137;
|
111 |
+
@!%p44 mov.u32 %r19, %r137;
|
112 |
+
mov.b32 %f1, %r16;
|
113 |
+
mov.b32 %f2, %r17;
|
114 |
+
mov.b32 %f3, %r18;
|
115 |
+
mov.b32 %f4, %r19;
|
116 |
+
.loc 1 36 22
|
117 |
+
add.s64 %rd24, %rd16, 50257;
|
118 |
+
.loc 1 37 22
|
119 |
+
setp.lt.s64 %p11, %rd16, 0;
|
120 |
+
.loc 1 38 36
|
121 |
+
selp.b64 %rd3, %rd24, %rd16, %p11;
|
122 |
+
.loc 1 39 40
|
123 |
+
setp.lt.u64 %p12, %rd3, 50257;
|
124 |
+
mov.b32 %r168, 883;
|
125 |
+
mov.u64 %rd52, 1;
|
126 |
+
.loc 1 39 55
|
127 |
+
@%p12 bra $L__BB0_2;
|
128 |
+
mov.u64 %rd25, assertMessage_0;
|
129 |
+
cvta.global.u64 %rd26, %rd25;
|
130 |
+
mov.u64 %rd27, assertFile_0;
|
131 |
+
cvta.global.u64 %rd28, %rd27;
|
132 |
+
mov.u64 %rd29, assertFunc_0;
|
133 |
+
cvta.global.u64 %rd30, %rd29;
|
134 |
+
{ // callseq 4, 0
|
135 |
+
.reg .b32 temp_param_reg;
|
136 |
+
.param .b64 param0;
|
137 |
+
st.param.b64 [param0+0], %rd26;
|
138 |
+
.param .b64 param1;
|
139 |
+
st.param.b64 [param1+0], %rd28;
|
140 |
+
.param .b32 param2;
|
141 |
+
st.param.b32 [param2+0], %r168;
|
142 |
+
.param .b64 param3;
|
143 |
+
st.param.b64 [param3+0], %rd30;
|
144 |
+
.param .b64 param4;
|
145 |
+
st.param.b64 [param4+0], %rd52;
|
146 |
+
call.uni
|
147 |
+
__assertfail,
|
148 |
+
(
|
149 |
+
param0,
|
150 |
+
param1,
|
151 |
+
param2,
|
152 |
+
param3,
|
153 |
+
param4
|
154 |
+
);
|
155 |
+
} // callseq 4
|
156 |
+
$L__BB0_2:
|
157 |
+
.loc 1 0 55
|
158 |
+
ld.param.u64 %rd7, [triton__0d1d2d3d4d5de6de_param_4];
|
159 |
+
.loc 1 37 22
|
160 |
+
setp.lt.s64 %p36, %rd8, 0;
|
161 |
+
.loc 1 40 44
|
162 |
+
shl.b64 %rd35, %rd8, 8;
|
163 |
+
add.s64 %rd36, %rd35, 12865792;
|
164 |
+
selp.b64 %rd37, %rd36, %rd35, %p36;
|
165 |
+
cvt.u64.u32 %rd38, %r6;
|
166 |
+
.loc 1 40 40
|
167 |
+
or.b64 %rd39, %rd37, %rd38;
|
168 |
+
.loc 1 40 34
|
169 |
+
shl.b64 %rd40, %rd39, 2;
|
170 |
+
add.s64 %rd49, %rd5, %rd40;
|
171 |
+
.loc 1 40 52
|
172 |
+
mov.u32 %r36, 0x0;
|
173 |
+
mov.u32 %r37, 0x0;
|
174 |
+
mov.u32 %r38, 0x0;
|
175 |
+
mov.u32 %r39, 0x0;
|
176 |
+
@%p44 ld.global.L1::evict_last.v4.b32 { %r36, %r37, %r38, %r39 }, [ %rd49 + 0 ];
|
177 |
+
@!%p44 mov.u32 %r36, %r137;
|
178 |
+
@!%p44 mov.u32 %r37, %r137;
|
179 |
+
@!%p44 mov.u32 %r38, %r137;
|
180 |
+
@!%p44 mov.u32 %r39, %r137;
|
181 |
+
mov.b32 %f7, %r36;
|
182 |
+
mov.b32 %f8, %r37;
|
183 |
+
mov.b32 %f9, %r38;
|
184 |
+
mov.b32 %f10, %r39;
|
185 |
+
.loc 1 41 22
|
186 |
+
add.f32 %f11, %f1, %f7;
|
187 |
+
add.f32 %f12, %f2, %f8;
|
188 |
+
add.f32 %f13, %f3, %f9;
|
189 |
+
add.f32 %f14, %f4, %f10;
|
190 |
+
$L__tmp1:
|
191 |
+
.loc 2 98 22
|
192 |
+
add.f32 %f15, %f11, 0f00000000;
|
193 |
+
add.f32 %f16, %f12, 0f00000000;
|
194 |
+
add.f32 %f17, %f13, 0f00000000;
|
195 |
+
add.f32 %f18, %f14, 0f00000000;
|
196 |
+
.loc 2 101 30
|
197 |
+
sub.f32 %f19, %f11, %f15;
|
198 |
+
sub.f32 %f20, %f12, %f16;
|
199 |
+
sub.f32 %f21, %f13, %f17;
|
200 |
+
sub.f32 %f22, %f14, %f18;
|
201 |
+
.loc 2 101 13
|
202 |
+
fma.rn.f32 %f23, %f11, %f19, 0f00000000;
|
203 |
+
fma.rn.f32 %f24, %f12, %f20, 0f00000000;
|
204 |
+
fma.rn.f32 %f25, %f13, %f21, 0f00000000;
|
205 |
+
fma.rn.f32 %f26, %f14, %f22, 0f00000000;
|
206 |
+
$L__tmp2:
|
207 |
+
.loc 2 108 21
|
208 |
+
sub.f32 %f27, %f16, %f15;
|
209 |
+
mov.b32 %r45, 1065353216;
|
210 |
+
mov.b32 %r46, 1073741824;
|
211 |
+
.loc 2 110 60
|
212 |
+
div.full.f32 %r44, %r45, %r46;
|
213 |
+
mov.b32 %f28, %r44;
|
214 |
+
.loc 2 112 17
|
215 |
+
fma.rn.f32 %f29, %f28, %f27, %f15;
|
216 |
+
.loc 2 113 15
|
217 |
+
add.f32 %f30, %f23, %f24;
|
218 |
+
.loc 2 113 30
|
219 |
+
mul.f32 %f31, %f27, %f27;
|
220 |
+
.loc 2 113 22
|
221 |
+
fma.rn.f32 %f32, %f28, %f31, %f30;
|
222 |
+
.loc 2 108 21
|
223 |
+
sub.f32 %f33, %f17, %f29;
|
224 |
+
mov.b32 %r49, 1077936128;
|
225 |
+
.loc 2 110 60
|
226 |
+
div.full.f32 %r47, %r45, %r49;
|
227 |
+
mov.b32 %f34, %r47;
|
228 |
+
.loc 2 112 17
|
229 |
+
fma.rn.f32 %f35, %f34, %f33, %f29;
|
230 |
+
.loc 2 113 15
|
231 |
+
add.f32 %f36, %f25, %f32;
|
232 |
+
.loc 2 113 30
|
233 |
+
mul.f32 %f37, %f33, %f33;
|
234 |
+
.loc 2 113 38
|
235 |
+
fma.rn.f32 %f38, %f33, %f33, %f37;
|
236 |
+
.loc 2 113 22
|
237 |
+
fma.rn.f32 %f39, %f34, %f38, %f36;
|
238 |
+
.loc 2 108 21
|
239 |
+
sub.f32 %f40, %f18, %f35;
|
240 |
+
mov.b32 %r52, 1082130432;
|
241 |
+
.loc 2 110 60
|
242 |
+
div.full.f32 %r50, %r45, %r52;
|
243 |
+
mov.b32 %f41, %r50;
|
244 |
+
.loc 2 112 17
|
245 |
+
fma.rn.f32 %f42, %f41, %f40, %f35;
|
246 |
+
.loc 2 113 15
|
247 |
+
add.f32 %f43, %f26, %f39;
|
248 |
+
.loc 2 113 30
|
249 |
+
mul.f32 %f44, %f40, %f40;
|
250 |
+
.loc 2 113 38
|
251 |
+
mul.f32 %f45, %f44, 0f40400000;
|
252 |
+
.loc 2 113 22
|
253 |
+
fma.rn.f32 %f46, %f41, %f45, %f43;
|
254 |
+
$L__tmp3:
|
255 |
+
.loc 2 120 46
|
256 |
+
mov.b32 %r101, %f42;
|
257 |
+
shfl.sync.bfly.b32 %r102, %r101, 16, 31, -1;
|
258 |
+
mov.b32 %f47, %r102;
|
259 |
+
mov.b32 %r103, %f46;
|
260 |
+
shfl.sync.bfly.b32 %r104, %r103, 16, 31, -1;
|
261 |
+
mov.b32 %f48, %r104;
|
262 |
+
shfl.sync.bfly.b32 %r54, %r52, 16, 31, -1;
|
263 |
+
mov.b32 %f49, %r54;
|
264 |
+
$L__tmp4:
|
265 |
+
.loc 2 108 21
|
266 |
+
sub.f32 %f50, %f47, %f42;
|
267 |
+
.loc 2 109 28
|
268 |
+
add.f32 %f51, %f49, 0f40800000;
|
269 |
+
.loc 2 110 39
|
270 |
+
setp.eq.f32 %p37, %f51, 0f00000000;
|
271 |
+
.loc 2 110 60
|
272 |
+
mov.b32 %r55, %f51;
|
273 |
+
div.full.f32 %r53, %r54, %r55;
|
274 |
+
mov.b32 %f52, %r53;
|
275 |
+
.loc 2 110 49
|
276 |
+
selp.f32 %f53, 0f00000000, %f52, %p37;
|
277 |
+
.loc 2 112 17
|
278 |
+
fma.rn.f32 %f54, %f53, %f50, %f42;
|
279 |
+
.loc 2 113 15
|
280 |
+
add.f32 %f55, %f46, %f48;
|
281 |
+
.loc 2 113 30
|
282 |
+
mul.f32 %f56, %f50, %f50;
|
283 |
+
.loc 2 113 38
|
284 |
+
mul.f32 %f57, %f56, 0f40800000;
|
285 |
+
.loc 2 113 22
|
286 |
+
fma.rn.f32 %f58, %f53, %f57, %f55;
|
287 |
+
$L__tmp5:
|
288 |
+
.loc 2 120 46
|
289 |
+
mov.b32 %r105, %f54;
|
290 |
+
shfl.sync.bfly.b32 %r106, %r105, 8, 31, -1;
|
291 |
+
mov.b32 %f59, %r106;
|
292 |
+
mov.b32 %r107, %f58;
|
293 |
+
shfl.sync.bfly.b32 %r108, %r107, 8, 31, -1;
|
294 |
+
mov.b32 %f60, %r108;
|
295 |
+
shfl.sync.bfly.b32 %r57, %r55, 8, 31, -1;
|
296 |
+
mov.b32 %f61, %r57;
|
297 |
+
$L__tmp6:
|
298 |
+
.loc 2 108 21
|
299 |
+
sub.f32 %f62, %f59, %f54;
|
300 |
+
.loc 2 109 28
|
301 |
+
add.f32 %f63, %f51, %f61;
|
302 |
+
.loc 2 110 39
|
303 |
+
setp.eq.f32 %p38, %f63, 0f00000000;
|
304 |
+
.loc 2 110 60
|
305 |
+
mov.b32 %r58, %f63;
|
306 |
+
div.full.f32 %r56, %r57, %r58;
|
307 |
+
mov.b32 %f64, %r56;
|
308 |
+
.loc 2 110 49
|
309 |
+
selp.f32 %f65, 0f00000000, %f64, %p38;
|
310 |
+
.loc 2 112 17
|
311 |
+
fma.rn.f32 %f66, %f65, %f62, %f54;
|
312 |
+
.loc 2 113 15
|
313 |
+
add.f32 %f67, %f58, %f60;
|
314 |
+
.loc 2 113 30
|
315 |
+
mul.f32 %f68, %f62, %f62;
|
316 |
+
.loc 2 113 38
|
317 |
+
mul.f32 %f69, %f51, %f68;
|
318 |
+
.loc 2 113 22
|
319 |
+
fma.rn.f32 %f70, %f65, %f69, %f67;
|
320 |
+
$L__tmp7:
|
321 |
+
.loc 2 120 46
|
322 |
+
mov.b32 %r109, %f66;
|
323 |
+
shfl.sync.bfly.b32 %r110, %r109, 4, 31, -1;
|
324 |
+
mov.b32 %f71, %r110;
|
325 |
+
mov.b32 %r111, %f70;
|
326 |
+
shfl.sync.bfly.b32 %r112, %r111, 4, 31, -1;
|
327 |
+
mov.b32 %f72, %r112;
|
328 |
+
shfl.sync.bfly.b32 %r60, %r58, 4, 31, -1;
|
329 |
+
mov.b32 %f73, %r60;
|
330 |
+
$L__tmp8:
|
331 |
+
.loc 2 108 21
|
332 |
+
sub.f32 %f74, %f71, %f66;
|
333 |
+
.loc 2 109 28
|
334 |
+
add.f32 %f75, %f63, %f73;
|
335 |
+
.loc 2 110 39
|
336 |
+
setp.eq.f32 %p39, %f75, 0f00000000;
|
337 |
+
.loc 2 110 60
|
338 |
+
mov.b32 %r61, %f75;
|
339 |
+
div.full.f32 %r59, %r60, %r61;
|
340 |
+
mov.b32 %f76, %r59;
|
341 |
+
.loc 2 110 49
|
342 |
+
selp.f32 %f77, 0f00000000, %f76, %p39;
|
343 |
+
.loc 2 112 17
|
344 |
+
fma.rn.f32 %f78, %f77, %f74, %f66;
|
345 |
+
.loc 2 113 15
|
346 |
+
add.f32 %f79, %f70, %f72;
|
347 |
+
.loc 2 113 30
|
348 |
+
mul.f32 %f80, %f74, %f74;
|
349 |
+
.loc 2 113 38
|
350 |
+
mul.f32 %f81, %f63, %f80;
|
351 |
+
.loc 2 113 22
|
352 |
+
fma.rn.f32 %f82, %f77, %f81, %f79;
|
353 |
+
$L__tmp9:
|
354 |
+
.loc 2 120 46
|
355 |
+
mov.b32 %r113, %f78;
|
356 |
+
shfl.sync.bfly.b32 %r114, %r113, 2, 31, -1;
|
357 |
+
mov.b32 %f83, %r114;
|
358 |
+
mov.b32 %r115, %f82;
|
359 |
+
shfl.sync.bfly.b32 %r116, %r115, 2, 31, -1;
|
360 |
+
mov.b32 %f84, %r116;
|
361 |
+
shfl.sync.bfly.b32 %r63, %r61, 2, 31, -1;
|
362 |
+
mov.b32 %f85, %r63;
|
363 |
+
$L__tmp10:
|
364 |
+
.loc 2 108 21
|
365 |
+
sub.f32 %f86, %f83, %f78;
|
366 |
+
.loc 2 109 28
|
367 |
+
add.f32 %f87, %f75, %f85;
|
368 |
+
.loc 2 110 39
|
369 |
+
setp.eq.f32 %p40, %f87, 0f00000000;
|
370 |
+
.loc 2 110 60
|
371 |
+
mov.b32 %r64, %f87;
|
372 |
+
div.full.f32 %r62, %r63, %r64;
|
373 |
+
mov.b32 %f88, %r62;
|
374 |
+
.loc 2 110 49
|
375 |
+
selp.f32 %f89, 0f00000000, %f88, %p40;
|
376 |
+
.loc 2 112 17
|
377 |
+
fma.rn.f32 %f90, %f89, %f86, %f78;
|
378 |
+
.loc 2 113 15
|
379 |
+
add.f32 %f91, %f82, %f84;
|
380 |
+
.loc 2 113 30
|
381 |
+
mul.f32 %f92, %f86, %f86;
|
382 |
+
.loc 2 113 38
|
383 |
+
mul.f32 %f93, %f75, %f92;
|
384 |
+
.loc 2 113 22
|
385 |
+
fma.rn.f32 %f94, %f89, %f93, %f91;
|
386 |
+
$L__tmp11:
|
387 |
+
.loc 2 120 46
|
388 |
+
mov.b32 %r117, %f90;
|
389 |
+
shfl.sync.bfly.b32 %r118, %r117, 1, 31, -1;
|
390 |
+
mov.b32 %f95, %r118;
|
391 |
+
mov.b32 %r119, %f94;
|
392 |
+
shfl.sync.bfly.b32 %r120, %r119, 1, 31, -1;
|
393 |
+
mov.b32 %f96, %r120;
|
394 |
+
shfl.sync.bfly.b32 %r66, %r64, 1, 31, -1;
|
395 |
+
mov.b32 %f97, %r66;
|
396 |
+
$L__tmp12:
|
397 |
+
.loc 2 108 21
|
398 |
+
sub.f32 %f98, %f95, %f90;
|
399 |
+
.loc 2 109 28
|
400 |
+
add.f32 %f99, %f87, %f97;
|
401 |
+
.loc 2 110 39
|
402 |
+
setp.eq.f32 %p41, %f99, 0f00000000;
|
403 |
+
.loc 2 110 60
|
404 |
+
mov.b32 %r67, %f99;
|
405 |
+
div.full.f32 %r65, %r66, %r67;
|
406 |
+
mov.b32 %f100, %r65;
|
407 |
+
.loc 2 110 49
|
408 |
+
selp.f32 %f101, 0f00000000, %f100, %p41;
|
409 |
+
.loc 2 112 17
|
410 |
+
fma.rn.f32 %f102, %f98, %f101, %f90;
|
411 |
+
.loc 2 113 15
|
412 |
+
add.f32 %f103, %f94, %f96;
|
413 |
+
.loc 2 113 30
|
414 |
+
mul.f32 %f104, %f98, %f98;
|
415 |
+
.loc 2 113 38
|
416 |
+
mul.f32 %f105, %f87, %f104;
|
417 |
+
.loc 2 113 22
|
418 |
+
fma.rn.f32 %f106, %f101, %f105, %f103;
|
419 |
+
$L__tmp13:
|
420 |
+
.loc 2 120 46
|
421 |
+
setp.eq.s32 %p18, %r2, 0;
|
422 |
+
shl.b32 %r121, %r5, 2;
|
423 |
+
shl.b32 %r122, %r3, 3;
|
424 |
+
or.b32 %r123, %r122, %r121;
|
425 |
+
mov.u32 %r124, global_smem;
|
426 |
+
add.s32 %r68, %r124, %r123;
|
427 |
+
mov.b32 %r69, %f102;
|
428 |
+
@%p18 st.shared.b32 [ %r68 + 0 ], %r69;
|
429 |
+
add.s32 %r125, %r124, 16;
|
430 |
+
add.s32 %r70, %r125, %r123;
|
431 |
+
mov.b32 %r71, %f106;
|
432 |
+
@%p18 st.shared.b32 [ %r70 + 0 ], %r71;
|
433 |
+
add.s32 %r126, %r124, 32;
|
434 |
+
add.s32 %r72, %r126, %r123;
|
435 |
+
@%p18 st.shared.b32 [ %r72 + 0 ], %r67;
|
436 |
+
bar.sync 0;
|
437 |
+
setp.lt.s32 %p21, %r1, 4;
|
438 |
+
add.s32 %r75, %r124, %r24;
|
439 |
+
@%p21 ld.shared.b32 %r74, [ %r75 + 0 ];
|
440 |
+
mov.b32 %f107, %r74;
|
441 |
+
add.s32 %r77, %r125, %r24;
|
442 |
+
@%p21 ld.shared.b32 %r76, [ %r77 + 0 ];
|
443 |
+
mov.b32 %f108, %r76;
|
444 |
+
add.s32 %r79, %r126, %r24;
|
445 |
+
@%p21 ld.shared.b32 %r78, [ %r79 + 0 ];
|
446 |
+
mov.b32 %f109, %r78;
|
447 |
+
shfl.sync.bfly.b32 %r128, %r74, 1, 31, -1;
|
448 |
+
mov.b32 %f110, %r128;
|
449 |
+
shfl.sync.bfly.b32 %r129, %r76, 1, 31, -1;
|
450 |
+
mov.b32 %f111, %r129;
|
451 |
+
shfl.sync.bfly.b32 %r81, %r78, 1, 31, -1;
|
452 |
+
mov.b32 %f112, %r81;
|
453 |
+
$L__tmp14:
|
454 |
+
.loc 2 108 21
|
455 |
+
sub.f32 %f113, %f110, %f107;
|
456 |
+
.loc 2 109 28
|
457 |
+
add.f32 %f114, %f109, %f112;
|
458 |
+
.loc 2 110 39
|
459 |
+
setp.eq.f32 %p42, %f114, 0f00000000;
|
460 |
+
.loc 2 110 60
|
461 |
+
mov.b32 %r82, %f114;
|
462 |
+
div.full.f32 %r80, %r81, %r82;
|
463 |
+
mov.b32 %f115, %r80;
|
464 |
+
.loc 2 110 49
|
465 |
+
selp.f32 %f116, 0f00000000, %f115, %p42;
|
466 |
+
.loc 2 112 17
|
467 |
+
fma.rn.f32 %f117, %f113, %f116, %f107;
|
468 |
+
.loc 2 113 15
|
469 |
+
add.f32 %f118, %f108, %f111;
|
470 |
+
.loc 2 113 30
|
471 |
+
mul.f32 %f119, %f113, %f113;
|
472 |
+
.loc 2 113 38
|
473 |
+
mul.f32 %f120, %f109, %f119;
|
474 |
+
.loc 2 113 22
|
475 |
+
fma.rn.f32 %f121, %f120, %f116, %f118;
|
476 |
+
$L__tmp15:
|
477 |
+
.loc 2 120 46
|
478 |
+
setp.eq.s32 %p43, %r4, 0;
|
479 |
+
and.pred %p24, %p21, %p43;
|
480 |
+
mov.b32 %r84, %f117;
|
481 |
+
@%p24 st.shared.b32 [ %r75 + 0 ], %r84;
|
482 |
+
mov.b32 %r86, %f121;
|
483 |
+
@%p24 st.shared.b32 [ %r77 + 0 ], %r86;
|
484 |
+
@%p24 st.shared.b32 [ %r79 + 0 ], %r82;
|
485 |
+
bar.sync 0;
|
486 |
+
add.s32 %r130, %r124, %r122;
|
487 |
+
ld.shared.f32 %f5, [%r130];
|
488 |
+
add.s32 %r131, %r125, %r122;
|
489 |
+
ld.shared.f32 %f6, [%r131];
|
490 |
+
$L__tmp16:
|
491 |
+
.loc 1 59 51
|
492 |
+
mov.u32 %r89, 0x0;
|
493 |
+
mov.u32 %r90, 0x0;
|
494 |
+
mov.u32 %r91, 0x0;
|
495 |
+
mov.u32 %r92, 0x0;
|
496 |
+
@%p44 ld.global.L1::evict_last.v4.b32 { %r89, %r90, %r91, %r92 }, [ %rd33 + 0 ];
|
497 |
+
@!%p44 mov.u32 %r89, %r137;
|
498 |
+
@!%p44 mov.u32 %r90, %r137;
|
499 |
+
@!%p44 mov.u32 %r91, %r137;
|
500 |
+
@!%p44 mov.u32 %r92, %r137;
|
501 |
+
.loc 1 60 35
|
502 |
+
mul.wide.u32 %rd41, %r7, 4;
|
503 |
+
add.s64 %rd34, %rd6, %rd41;
|
504 |
+
.loc 1 60 40
|
505 |
+
mov.u32 %r97, 0x0;
|
506 |
+
mov.u32 %r98, 0x0;
|
507 |
+
@%p44 ld.global.L1::evict_last.v2.b32 { %r97, %r98 }, [ %rd34 + 0 ];
|
508 |
+
@!%p44 mov.u32 %r97, %r137;
|
509 |
+
@!%p44 mov.u32 %r98, %r137;
|
510 |
+
.loc 1 64 57
|
511 |
+
@%p12 bra $L__BB0_4;
|
512 |
+
mov.u64 %rd42, assertMessage_1;
|
513 |
+
cvta.global.u64 %rd43, %rd42;
|
514 |
+
mov.u64 %rd44, assertFile_1;
|
515 |
+
cvta.global.u64 %rd45, %rd44;
|
516 |
+
mov.u64 %rd46, assertFunc_1;
|
517 |
+
cvta.global.u64 %rd47, %rd46;
|
518 |
+
{ // callseq 5, 0
|
519 |
+
.reg .b32 temp_param_reg;
|
520 |
+
.param .b64 param0;
|
521 |
+
st.param.b64 [param0+0], %rd43;
|
522 |
+
.param .b64 param1;
|
523 |
+
st.param.b64 [param1+0], %rd45;
|
524 |
+
.param .b32 param2;
|
525 |
+
st.param.b32 [param2+0], %r168;
|
526 |
+
.param .b64 param3;
|
527 |
+
st.param.b64 [param3+0], %rd47;
|
528 |
+
.param .b64 param4;
|
529 |
+
st.param.b64 [param4+0], %rd52;
|
530 |
+
call.uni
|
531 |
+
__assertfail,
|
532 |
+
(
|
533 |
+
param0,
|
534 |
+
param1,
|
535 |
+
param2,
|
536 |
+
param3,
|
537 |
+
param4
|
538 |
+
);
|
539 |
+
} // callseq 5
|
540 |
+
$L__BB0_4:
|
541 |
+
.loc 1 65 54
|
542 |
+
mov.u32 %r133, 0x0;
|
543 |
+
mov.u32 %r134, 0x0;
|
544 |
+
mov.u32 %r135, 0x0;
|
545 |
+
mov.u32 %r136, 0x0;
|
546 |
+
@%p44 ld.global.L1::evict_first.v4.b32 { %r133, %r134, %r135, %r136 }, [ %rd49 + 0 ];
|
547 |
+
@!%p44 mov.u32 %r133, %r137;
|
548 |
+
@!%p44 mov.u32 %r134, %r137;
|
549 |
+
@!%p44 mov.u32 %r135, %r137;
|
550 |
+
@!%p44 mov.u32 %r136, %r137;
|
551 |
+
.loc 1 69 23
|
552 |
+
mov.b32 %r142, %f6;
|
553 |
+
mov.b32 %r143, 1132462080;
|
554 |
+
div.full.f32 %r141, %r142, %r143;
|
555 |
+
mov.b32 %f122, %r141;
|
556 |
+
.loc 1 71 24
|
557 |
+
add.f32 %f123, %f122, 0f3727C5AC;
|
558 |
+
.loc 1 72 30
|
559 |
+
rsqrt.approx.ftz.f32 %f124, %f123;
|
560 |
+
.loc 1 65 54
|
561 |
+
mov.b32 %f125, %r136;
|
562 |
+
.loc 1 59 51
|
563 |
+
mov.b32 %f126, %r92;
|
564 |
+
.loc 1 66 24
|
565 |
+
add.f32 %f127, %f126, %f125;
|
566 |
+
.loc 1 67 24
|
567 |
+
sub.f32 %f128, %f127, %f5;
|
568 |
+
.loc 1 65 54
|
569 |
+
mov.b32 %f129, %r135;
|
570 |
+
.loc 1 59 51
|
571 |
+
mov.b32 %f130, %r91;
|
572 |
+
.loc 1 66 24
|
573 |
+
add.f32 %f131, %f130, %f129;
|
574 |
+
.loc 1 67 24
|
575 |
+
sub.f32 %f132, %f131, %f5;
|
576 |
+
.loc 1 65 54
|
577 |
+
mov.b32 %f133, %r134;
|
578 |
+
.loc 1 59 51
|
579 |
+
mov.b32 %f134, %r90;
|
580 |
+
.loc 1 66 24
|
581 |
+
add.f32 %f135, %f134, %f133;
|
582 |
+
.loc 1 67 24
|
583 |
+
sub.f32 %f136, %f135, %f5;
|
584 |
+
.loc 1 65 54
|
585 |
+
mov.b32 %f137, %r133;
|
586 |
+
.loc 1 59 51
|
587 |
+
mov.b32 %f138, %r89;
|
588 |
+
.loc 1 66 24
|
589 |
+
add.f32 %f139, %f138, %f137;
|
590 |
+
.loc 1 67 24
|
591 |
+
sub.f32 %f140, %f139, %f5;
|
592 |
+
.loc 1 73 24
|
593 |
+
mul.f32 %f141, %f140, %f124;
|
594 |
+
mul.f32 %f142, %f136, %f124;
|
595 |
+
mul.f32 %f143, %f132, %f124;
|
596 |
+
mul.f32 %f144, %f128, %f124;
|
597 |
+
.loc 1 74 24
|
598 |
+
bar.sync 0;
|
599 |
+
shl.b32 %r159, %r7, 2;
|
600 |
+
add.s32 %r161, %r124, %r159;
|
601 |
+
st.shared.v2.u32 [%r161], {%r97, %r98};
|
602 |
+
bar.sync 0;
|
603 |
+
shl.b32 %r162, %r6, 2;
|
604 |
+
add.s32 %r163, %r124, %r162;
|
605 |
+
ld.shared.v4.f32 {%f145, %f146, %f147, %f148}, [%r163];
|
606 |
+
mul.f32 %f149, %f141, %f145;
|
607 |
+
mul.f32 %f150, %f142, %f146;
|
608 |
+
mul.f32 %f151, %f143, %f147;
|
609 |
+
mul.f32 %f152, %f144, %f148;
|
610 |
+
.loc 1 76 39
|
611 |
+
shl.b32 %r164, %r8, 8;
|
612 |
+
.loc 1 76 35
|
613 |
+
or.b32 %r165, %r164, %r6;
|
614 |
+
.loc 1 76 29
|
615 |
+
mul.wide.s32 %rd51, %r165, 2;
|
616 |
+
add.s64 %rd50, %rd7, %rd51;
|
617 |
+
.loc 1 76 52
|
618 |
+
mov.b32 %r153, %f149;
|
619 |
+
cvt.rn.bf16.f32 %rs1, %r153;
|
620 |
+
mov.b32 %r154, %f150;
|
621 |
+
cvt.rn.bf16.f32 %rs2, %r154;
|
622 |
+
mov.b32 %r155, %f151;
|
623 |
+
cvt.rn.bf16.f32 %rs3, %r155;
|
624 |
+
mov.b32 %r156, %f152;
|
625 |
+
cvt.rn.bf16.f32 %rs4, %r156;
|
626 |
+
mov.b32 %r166, {%rs1, %rs2};
|
627 |
+
mov.b32 %r167, {%rs3, %rs4};
|
628 |
+
@%p44 st.global.v2.b32 [ %rd50 + 0 ], { %r166, %r167 };
|
629 |
+
.loc 1 55 4
|
630 |
+
ret;
|
631 |
+
$L__tmp17:
|
632 |
+
$L__func_end0:
|
633 |
+
|
634 |
+
}
|
635 |
+
// .globl __nv_rsqrtf
|
636 |
+
.visible .func (.param .b32 func_retval0) __nv_rsqrtf(
|
637 |
+
.param .b32 __nv_rsqrtf_param_0
|
638 |
+
)
|
639 |
+
{
|
640 |
+
.reg .f32 %f<3>;
|
641 |
+
$L__func_begin1:
|
642 |
+
|
643 |
+
ld.param.f32 %f1, [__nv_rsqrtf_param_0];
|
644 |
+
rsqrt.approx.ftz.f32 %f2, %f1;
|
645 |
+
st.param.f32 [func_retval0+0], %f2;
|
646 |
+
ret;
|
647 |
+
$L__func_end1:
|
648 |
+
|
649 |
+
}
|
650 |
+
.file 1 "/tmp/torchinductor_root/lh/clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py"
|
651 |
+
.file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
|
652 |
+
.section .debug_abbrev
|
653 |
+
{
|
654 |
+
.b8 1
|
655 |
+
.b8 17
|
656 |
+
.b8 1
|
657 |
+
.b8 37
|
658 |
+
.b8 8
|
659 |
+
.b8 19
|
660 |
+
.b8 5
|
661 |
+
.b8 3
|
662 |
+
.b8 8
|
663 |
+
.b8 16
|
664 |
+
.b8 6
|
665 |
+
.b8 27
|
666 |
+
.b8 8
|
667 |
+
.b8 180
|
668 |
+
.b8 66
|
669 |
+
.b8 12
|
670 |
+
.b8 17
|
671 |
+
.b8 1
|
672 |
+
.b8 18
|
673 |
+
.b8 1
|
674 |
+
.b8 0
|
675 |
+
.b8 0
|
676 |
+
.b8 2
|
677 |
+
.b8 46
|
678 |
+
.b8 0
|
679 |
+
.b8 135
|
680 |
+
.b8 64
|
681 |
+
.b8 8
|
682 |
+
.b8 3
|
683 |
+
.b8 8
|
684 |
+
.b8 58
|
685 |
+
.b8 11
|
686 |
+
.b8 59
|
687 |
+
.b8 11
|
688 |
+
.b8 63
|
689 |
+
.b8 12
|
690 |
+
.b8 32
|
691 |
+
.b8 11
|
692 |
+
.b8 0
|
693 |
+
.b8 0
|
694 |
+
.b8 3
|
695 |
+
.b8 46
|
696 |
+
.b8 1
|
697 |
+
.b8 17
|
698 |
+
.b8 1
|
699 |
+
.b8 18
|
700 |
+
.b8 1
|
701 |
+
.b8 64
|
702 |
+
.b8 10
|
703 |
+
.b8 49
|
704 |
+
.b8 19
|
705 |
+
.b8 0
|
706 |
+
.b8 0
|
707 |
+
.b8 4
|
708 |
+
.b8 29
|
709 |
+
.b8 0
|
710 |
+
.b8 49
|
711 |
+
.b8 19
|
712 |
+
.b8 17
|
713 |
+
.b8 1
|
714 |
+
.b8 18
|
715 |
+
.b8 1
|
716 |
+
.b8 88
|
717 |
+
.b8 11
|
718 |
+
.b8 89
|
719 |
+
.b8 11
|
720 |
+
.b8 87
|
721 |
+
.b8 11
|
722 |
+
.b8 0
|
723 |
+
.b8 0
|
724 |
+
.b8 5
|
725 |
+
.b8 29
|
726 |
+
.b8 1
|
727 |
+
.b8 49
|
728 |
+
.b8 19
|
729 |
+
.b8 17
|
730 |
+
.b8 1
|
731 |
+
.b8 18
|
732 |
+
.b8 1
|
733 |
+
.b8 88
|
734 |
+
.b8 11
|
735 |
+
.b8 89
|
736 |
+
.b8 11
|
737 |
+
.b8 87
|
738 |
+
.b8 11
|
739 |
+
.b8 0
|
740 |
+
.b8 0
|
741 |
+
.b8 0
|
742 |
+
}
|
743 |
+
.section .debug_info
|
744 |
+
{
|
745 |
+
.b32 298
|
746 |
+
.b8 2
|
747 |
+
.b8 0
|
748 |
+
.b32 .debug_abbrev
|
749 |
+
.b8 8
|
750 |
+
.b8 1
|
751 |
+
.b8 116
|
752 |
+
.b8 114
|
753 |
+
.b8 105
|
754 |
+
.b8 116
|
755 |
+
.b8 111
|
756 |
+
.b8 110
|
757 |
+
.b8 0
|
758 |
+
.b8 2
|
759 |
+
.b8 0
|
760 |
+
.b8 99
|
761 |
+
.b8 108
|
762 |
+
.b8 104
|
763 |
+
.b8 101
|
764 |
+
.b8 52
|
765 |
+
.b8 97
|
766 |
+
.b8 51
|
767 |
+
.b8 115
|
768 |
+
.b8 116
|
769 |
+
.b8 118
|
770 |
+
.b8 117
|
771 |
+
.b8 102
|
772 |
+
.b8 120
|
773 |
+
.b8 97
|
774 |
+
.b8 102
|
775 |
+
.b8 109
|
776 |
+
.b8 113
|
777 |
+
.b8 51
|
778 |
+
.b8 107
|
779 |
+
.b8 107
|
780 |
+
.b8 53
|
781 |
+
.b8 104
|
782 |
+
.b8 111
|
783 |
+
.b8 100
|
784 |
+
.b8 97
|
785 |
+
.b8 122
|
786 |
+
.b8 122
|
787 |
+
.b8 50
|
788 |
+
.b8 101
|
789 |
+
.b8 102
|
790 |
+
.b8 99
|
791 |
+
.b8 116
|
792 |
+
.b8 102
|
793 |
+
.b8 102
|
794 |
+
.b8 116
|
795 |
+
.b8 101
|
796 |
+
.b8 54
|
797 |
+
.b8 52
|
798 |
+
.b8 54
|
799 |
+
.b8 122
|
800 |
+
.b8 110
|
801 |
+
.b8 106
|
802 |
+
.b8 100
|
803 |
+
.b8 110
|
804 |
+
.b8 118
|
805 |
+
.b8 51
|
806 |
+
.b8 108
|
807 |
+
.b8 113
|
808 |
+
.b8 105
|
809 |
+
.b8 53
|
810 |
+
.b8 111
|
811 |
+
.b8 97
|
812 |
+
.b8 46
|
813 |
+
.b8 112
|
814 |
+
.b8 121
|
815 |
+
.b8 0
|
816 |
+
.b32 .debug_line
|
817 |
+
.b8 47
|
818 |
+
.b8 116
|
819 |
+
.b8 109
|
820 |
+
.b8 112
|
821 |
+
.b8 47
|
822 |
+
.b8 116
|
823 |
+
.b8 111
|
824 |
+
.b8 114
|
825 |
+
.b8 99
|
826 |
+
.b8 104
|
827 |
+
.b8 105
|
828 |
+
.b8 110
|
829 |
+
.b8 100
|
830 |
+
.b8 117
|
831 |
+
.b8 99
|
832 |
+
.b8 116
|
833 |
+
.b8 111
|
834 |
+
.b8 114
|
835 |
+
.b8 95
|
836 |
+
.b8 114
|
837 |
+
.b8 111
|
838 |
+
.b8 111
|
839 |
+
.b8 116
|
840 |
+
.b8 47
|
841 |
+
.b8 108
|
842 |
+
.b8 104
|
843 |
+
.b8 0
|
844 |
+
.b8 1
|
845 |
+
.b64 $L__func_begin0
|
846 |
+
.b64 $L__func_end0
|
847 |
+
.b8 2
|
848 |
+
.b8 116
|
849 |
+
.b8 114
|
850 |
+
.b8 105
|
851 |
+
.b8 116
|
852 |
+
.b8 111
|
853 |
+
.b8 110
|
854 |
+
.b8 95
|
855 |
+
.b8 95
|
856 |
+
.b8 48
|
857 |
+
.b8 100
|
858 |
+
.b8 49
|
859 |
+
.b8 100
|
860 |
+
.b8 50
|
861 |
+
.b8 100
|
862 |
+
.b8 51
|
863 |
+
.b8 100
|
864 |
+
.b8 52
|
865 |
+
.b8 100
|
866 |
+
.b8 53
|
867 |
+
.b8 100
|
868 |
+
.b8 101
|
869 |
+
.b8 54
|
870 |
+
.b8 100
|
871 |
+
.b8 101
|
872 |
+
.b8 0
|
873 |
+
.b8 116
|
874 |
+
.b8 114
|
875 |
+
.b8 105
|
876 |
+
.b8 116
|
877 |
+
.b8 111
|
878 |
+
.b8 110
|
879 |
+
.b8 95
|
880 |
+
.b8 95
|
881 |
+
.b8 48
|
882 |
+
.b8 100
|
883 |
+
.b8 49
|
884 |
+
.b8 100
|
885 |
+
.b8 50
|
886 |
+
.b8 100
|
887 |
+
.b8 51
|
888 |
+
.b8 100
|
889 |
+
.b8 52
|
890 |
+
.b8 100
|
891 |
+
.b8 53
|
892 |
+
.b8 100
|
893 |
+
.b8 101
|
894 |
+
.b8 54
|
895 |
+
.b8 100
|
896 |
+
.b8 101
|
897 |
+
.b8 0
|
898 |
+
.b8 1
|
899 |
+
.b8 18
|
900 |
+
.b8 1
|
901 |
+
.b8 1
|
902 |
+
.b8 3
|
903 |
+
.b64 $L__func_begin0
|
904 |
+
.b64 $L__func_end0
|
905 |
+
.b8 1
|
906 |
+
.b8 156
|
907 |
+
.b32 125
|
908 |
+
.b8 4
|
909 |
+
.b32 125
|
910 |
+
.b64 $L__tmp1
|
911 |
+
.b64 $L__tmp2
|
912 |
+
.b8 2
|
913 |
+
.b8 44
|
914 |
+
.b8 38
|
915 |
+
.b8 5
|
916 |
+
.b32 125
|
917 |
+
.b64 $L__tmp2
|
918 |
+
.b64 $L__tmp15
|
919 |
+
.b8 2
|
920 |
+
.b8 50
|
921 |
+
.b8 41
|
922 |
+
.b8 4
|
923 |
+
.b32 125
|
924 |
+
.b64 $L__tmp2
|
925 |
+
.b64 $L__tmp15
|
926 |
+
.b8 2
|
927 |
+
.b8 120
|
928 |
+
.b8 46
|
929 |
+
.b8 0
|
930 |
+
.b8 4
|
931 |
+
.b32 125
|
932 |
+
.b64 $L__tmp3
|
933 |
+
.b64 $L__tmp16
|
934 |
+
.b8 2
|
935 |
+
.b8 50
|
936 |
+
.b8 41
|
937 |
+
.b8 0
|
938 |
+
.b8 0
|
939 |
+
}
|
940 |
+
.section .debug_pubnames
|
941 |
+
{
|
942 |
+
.b32 $L__pubNames_end0-$L__pubNames_start0
|
943 |
+
$L__pubNames_start0:
|
944 |
+
.b8 2
|
945 |
+
.b8 0
|
946 |
+
.b32 .debug_info
|
947 |
+
.b32 302
|
948 |
+
.b32 125
|
949 |
+
.b8 116
|
950 |
+
.b8 114
|
951 |
+
.b8 105
|
952 |
+
.b8 116
|
953 |
+
.b8 111
|
954 |
+
.b8 110
|
955 |
+
.b8 95
|
956 |
+
.b8 95
|
957 |
+
.b8 48
|
958 |
+
.b8 100
|
959 |
+
.b8 49
|
960 |
+
.b8 100
|
961 |
+
.b8 50
|
962 |
+
.b8 100
|
963 |
+
.b8 51
|
964 |
+
.b8 100
|
965 |
+
.b8 52
|
966 |
+
.b8 100
|
967 |
+
.b8 53
|
968 |
+
.b8 100
|
969 |
+
.b8 101
|
970 |
+
.b8 54
|
971 |
+
.b8 100
|
972 |
+
.b8 101
|
973 |
+
.b8 0
|
974 |
+
.b32 0
|
975 |
+
$L__pubNames_end0:
|
976 |
+
}
|
977 |
+
.section .debug_pubtypes
|
978 |
+
{
|
979 |
+
.b32 $L__pubTypes_end0-$L__pubTypes_start0
|
980 |
+
$L__pubTypes_start0:
|
981 |
+
.b8 2
|
982 |
+
.b8 0
|
983 |
+
.b32 .debug_info
|
984 |
+
.b32 302
|
985 |
+
.b32 0
|
986 |
+
$L__pubTypes_end0:
|
987 |
+
}
|
988 |
+
.section .debug_loc { }
|
.triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ttir
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
module {
|
2 |
+
tt.func public @triton__0d1d2d3d4d5de6de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
3 |
+
%cst = arith.constant dense<1.000000e+00> : tensor<1x256xf32>
|
4 |
+
%cst_0 = arith.constant dense<0.000000e+00> : tensor<1x256xf32>
|
5 |
+
%cst_1 = arith.constant 0.000000e+00 : f32
|
6 |
+
%cst_2 = arith.constant dense<256> : tensor<2x1xi64>
|
7 |
+
%cst_3 = arith.constant dense<50257> : tensor<2x1xi64>
|
8 |
+
%cst_4 = arith.constant dense<0> : tensor<2x1xi64>
|
9 |
+
%cst_5 = arith.constant dense<9.99999974E-6> : tensor<2x1xf32>
|
10 |
+
%cst_6 = arith.constant dense<2.560000e+02> : tensor<2x1xf32>
|
11 |
+
%cst_7 = arith.constant dense<0.000000e+00> : tensor<2x256xf32>
|
12 |
+
%cst_8 = arith.constant dense<256> : tensor<2x1xi32>
|
13 |
+
%cst_9 = arith.constant dense<256> : tensor<1x256xi32>
|
14 |
+
%cst_10 = arith.constant dense<512> : tensor<2x1xi32>
|
15 |
+
%c2_i32 = arith.constant 2 : i32
|
16 |
+
%0 = tt.get_program_id x : i32
|
17 |
+
%1 = arith.muli %0, %c2_i32 : i32
|
18 |
+
%2 = tt.make_range {end = 2 : i32, start = 0 : i32} : tensor<2xi32>
|
19 |
+
%3 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<2xi32>) -> tensor<2x1xi32>
|
20 |
+
%4 = tt.splat %1 : (i32) -> tensor<2x1xi32>
|
21 |
+
%5 = arith.addi %4, %3 : tensor<2x1xi32>
|
22 |
+
%6 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32>
|
23 |
+
%7 = tt.expand_dims %6 {axis = 0 : i32} : (tensor<256xi32>) -> tensor<1x256xi32>
|
24 |
+
%8 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<2x1x!tt.ptr<i64, 1>>
|
25 |
+
%9 = tt.addptr %8, %5 : tensor<2x1x!tt.ptr<i64, 1>>, tensor<2x1xi32>
|
26 |
+
%10 = tt.load %9 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x1xi64>
|
27 |
+
%11 = arith.remsi %5, %cst_10 : tensor<2x1xi32>
|
28 |
+
%12 = arith.cmpi slt, %7, %cst_9 : tensor<1x256xi32>
|
29 |
+
%13 = arith.muli %11, %cst_8 : tensor<2x1xi32>
|
30 |
+
%14 = tt.broadcast %7 : (tensor<1x256xi32>) -> tensor<2x256xi32>
|
31 |
+
%15 = tt.broadcast %13 : (tensor<2x1xi32>) -> tensor<2x256xi32>
|
32 |
+
%16 = arith.addi %14, %15 : tensor<2x256xi32>
|
33 |
+
%17 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<2x256x!tt.ptr<f32, 1>>
|
34 |
+
%18 = tt.addptr %17, %16 : tensor<2x256x!tt.ptr<f32, 1>>, tensor<2x256xi32>
|
35 |
+
%19 = tt.broadcast %12 : (tensor<1x256xi1>) -> tensor<2x256xi1>
|
36 |
+
%20 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x256xf32>
|
37 |
+
%21 = arith.addi %10, %cst_3 : tensor<2x1xi64>
|
38 |
+
%22 = arith.cmpi slt, %10, %cst_4 : tensor<2x1xi64>
|
39 |
+
%23 = arith.select %22, %21, %10 : tensor<2x1xi1>, tensor<2x1xi64>
|
40 |
+
%24 = arith.cmpi sge, %23, %cst_4 : tensor<2x1xi64>
|
41 |
+
%25 = arith.cmpi slt, %23, %cst_3 : tensor<2x1xi64>
|
42 |
+
%26 = arith.andi %24, %25 : tensor<2x1xi1>
|
43 |
+
tt.assert %26, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<2x1xi1>
|
44 |
+
%27 = arith.muli %23, %cst_2 : tensor<2x1xi64>
|
45 |
+
%28 = tt.broadcast %27 : (tensor<2x1xi64>) -> tensor<2x256xi64>
|
46 |
+
%29 = arith.extsi %7 : tensor<1x256xi32> to tensor<1x256xi64>
|
47 |
+
%30 = tt.broadcast %29 : (tensor<1x256xi64>) -> tensor<2x256xi64>
|
48 |
+
%31 = arith.addi %30, %28 : tensor<2x256xi64>
|
49 |
+
%32 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<2x256x!tt.ptr<f32, 1>>
|
50 |
+
%33 = tt.addptr %32, %31 : tensor<2x256x!tt.ptr<f32, 1>>, tensor<2x256xi64>
|
51 |
+
%34 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x256xf32>
|
52 |
+
%35 = arith.addf %34, %20 : tensor<2x256xf32>
|
53 |
+
%36 = arith.addf %35, %cst_7 : tensor<2x256xf32>
|
54 |
+
%37 = arith.subf %35, %36 : tensor<2x256xf32>
|
55 |
+
%38 = arith.mulf %35, %37 : tensor<2x256xf32>
|
56 |
+
%39 = arith.addf %38, %cst_7 : tensor<2x256xf32>
|
57 |
+
%40 = arith.select %19, %36, %cst_7 : tensor<2x256xi1>, tensor<2x256xf32>
|
58 |
+
%41 = arith.select %19, %39, %cst_7 : tensor<2x256xi1>, tensor<2x256xf32>
|
59 |
+
%42 = arith.select %12, %cst, %cst_0 : tensor<1x256xi1>, tensor<1x256xf32>
|
60 |
+
%43 = tt.broadcast %42 : (tensor<1x256xf32>) -> tensor<2x256xf32>
|
61 |
+
%44:3 = "tt.reduce"(%40, %41, %43) <{axis = 1 : i32}> ({
|
62 |
+
^bb0(%arg7: f32, %arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32):
|
63 |
+
%68 = arith.subf %arg10, %arg7 : f32
|
64 |
+
%69 = arith.addf %arg9, %arg12 : f32
|
65 |
+
%70 = arith.cmpf oeq, %69, %cst_1 : f32
|
66 |
+
%71 = arith.divf %arg12, %69 : f32
|
67 |
+
%72 = arith.select %70, %cst_1, %71 : f32
|
68 |
+
%73 = arith.mulf %68, %72 : f32
|
69 |
+
%74 = arith.addf %arg7, %73 : f32
|
70 |
+
%75 = arith.addf %arg8, %arg11 : f32
|
71 |
+
%76 = arith.mulf %68, %68 : f32
|
72 |
+
%77 = arith.mulf %76, %arg9 : f32
|
73 |
+
%78 = arith.mulf %77, %72 : f32
|
74 |
+
%79 = arith.addf %75, %78 : f32
|
75 |
+
tt.reduce.return %74, %79, %69 : f32, f32, f32
|
76 |
+
}) : (tensor<2x256xf32>, tensor<2x256xf32>, tensor<2x256xf32>) -> (tensor<2xf32>, tensor<2xf32>, tensor<2xf32>)
|
77 |
+
%45 = tt.expand_dims %44#0 {axis = 1 : i32} : (tensor<2xf32>) -> tensor<2x1xf32>
|
78 |
+
%46 = tt.expand_dims %44#1 {axis = 1 : i32} : (tensor<2xf32>) -> tensor<2x1xf32>
|
79 |
+
%47 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x256xf32>
|
80 |
+
%48 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x256x!tt.ptr<f32, 1>>
|
81 |
+
%49 = tt.addptr %48, %7 : tensor<1x256x!tt.ptr<f32, 1>>, tensor<1x256xi32>
|
82 |
+
%50 = tt.load %49, %12, %cst_0 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x256xf32>
|
83 |
+
tt.assert %26, "index out of bounds: 0 <= tmp13 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<2x1xi1>
|
84 |
+
%51 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<2x256xf32>
|
85 |
+
%52 = arith.addf %51, %47 : tensor<2x256xf32>
|
86 |
+
%53 = tt.broadcast %45 : (tensor<2x1xf32>) -> tensor<2x256xf32>
|
87 |
+
%54 = arith.subf %52, %53 : tensor<2x256xf32>
|
88 |
+
%55 = arith.divf %46, %cst_6 : tensor<2x1xf32>
|
89 |
+
%56 = arith.addf %55, %cst_5 : tensor<2x1xf32>
|
90 |
+
%57 = tt.extern_elementwise %56 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<2x1xf32>) -> tensor<2x1xf32>
|
91 |
+
%58 = tt.broadcast %57 : (tensor<2x1xf32>) -> tensor<2x256xf32>
|
92 |
+
%59 = arith.mulf %54, %58 : tensor<2x256xf32>
|
93 |
+
%60 = tt.broadcast %50 : (tensor<1x256xf32>) -> tensor<2x256xf32>
|
94 |
+
%61 = arith.mulf %59, %60 : tensor<2x256xf32>
|
95 |
+
%62 = arith.muli %5, %cst_8 : tensor<2x1xi32>
|
96 |
+
%63 = tt.broadcast %62 : (tensor<2x1xi32>) -> tensor<2x256xi32>
|
97 |
+
%64 = arith.addi %14, %63 : tensor<2x256xi32>
|
98 |
+
%65 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<2x256x!tt.ptr<bf16, 1>>
|
99 |
+
%66 = tt.addptr %65, %64 : tensor<2x256x!tt.ptr<bf16, 1>>, tensor<2x256xi32>
|
100 |
+
%67 = arith.truncf %61 : tensor<2x256xf32> to tensor<2x256xbf16>
|
101 |
+
tt.store %66, %67, %19 {cache = 1 : i32, evict = 1 : i32} : tensor<2x256xbf16>
|
102 |
+
tt.return
|
103 |
+
}
|
104 |
+
}
|
.triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ptx
ADDED
@@ -0,0 +1,782 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
//
|
2 |
+
// Generated by LLVM NVPTX Back-End
|
3 |
+
//
|
4 |
+
|
5 |
+
.version 8.2
|
6 |
+
.target sm_89
|
7 |
+
.address_size 64
|
8 |
+
|
9 |
+
// .globl triton__0d1d2d3d4d5d6d7d8de9de
|
10 |
+
.extern .shared .align 1 .b8 global_smem[];
|
11 |
+
|
12 |
+
.visible .entry triton__0d1d2d3d4d5d6d7d8de9de(
|
13 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_0,
|
14 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_1,
|
15 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_2,
|
16 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_3,
|
17 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_4,
|
18 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_5,
|
19 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_6,
|
20 |
+
.param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_7,
|
21 |
+
.param .u32 triton__0d1d2d3d4d5d6d7d8de9de_param_8,
|
22 |
+
.param .u32 triton__0d1d2d3d4d5d6d7d8de9de_param_9
|
23 |
+
)
|
24 |
+
.maxntid 64, 1, 1
|
25 |
+
{
|
26 |
+
.reg .pred %p<45>;
|
27 |
+
.reg .b16 %rs<5>;
|
28 |
+
.reg .b32 %r<106>;
|
29 |
+
.reg .f32 %f<90>;
|
30 |
+
.reg .b64 %rd<44>;
|
31 |
+
.loc 1 18 0
|
32 |
+
$L__func_begin0:
|
33 |
+
.loc 1 18 0
|
34 |
+
|
35 |
+
ld.param.u64 %rd25, [triton__0d1d2d3d4d5d6d7d8de9de_param_0];
|
36 |
+
ld.param.u64 %rd26, [triton__0d1d2d3d4d5d6d7d8de9de_param_1];
|
37 |
+
$L__tmp0:
|
38 |
+
.loc 1 26 26
|
39 |
+
mov.u32 %r74, %tid.x;
|
40 |
+
and.b32 %r75, %r74, 31;
|
41 |
+
ld.param.u64 %rd27, [triton__0d1d2d3d4d5d6d7d8de9de_param_2];
|
42 |
+
ld.param.u64 %rd28, [triton__0d1d2d3d4d5d6d7d8de9de_param_3];
|
43 |
+
ld.param.u64 %rd29, [triton__0d1d2d3d4d5d6d7d8de9de_param_4];
|
44 |
+
shl.b32 %r76, %r74, 2;
|
45 |
+
ld.param.u64 %rd30, [triton__0d1d2d3d4d5d6d7d8de9de_param_5];
|
46 |
+
and.b32 %r77, %r76, 252;
|
47 |
+
ld.param.u64 %rd31, [triton__0d1d2d3d4d5d6d7d8de9de_param_6];
|
48 |
+
ld.param.u64 %rd32, [triton__0d1d2d3d4d5d6d7d8de9de_param_7];
|
49 |
+
.loc 1 23 28
|
50 |
+
mov.u32 %r1, %ctaid.x;
|
51 |
+
.loc 1 30 40
|
52 |
+
shl.b32 %r78, %r1, 8;
|
53 |
+
.loc 1 30 36
|
54 |
+
or.b32 %r79, %r78, %r77;
|
55 |
+
.loc 1 30 30
|
56 |
+
mul.wide.s32 %rd33, %r79, 2;
|
57 |
+
add.s64 %rd1, %rd26, %rd33;
|
58 |
+
mov.b32 %r4, 0;
|
59 |
+
mov.pred %p1, -1;
|
60 |
+
.loc 1 30 46
|
61 |
+
mov.u32 %r2, 0x0;
|
62 |
+
mov.u32 %r3, 0x0;
|
63 |
+
@%p1 ld.global.v2.b32 { %r2, %r3 }, [ %rd1 + 0 ];
|
64 |
+
@!%p1 mov.u32 %r2, %r4;
|
65 |
+
@!%p1 mov.u32 %r3, %r4;
|
66 |
+
cvt.u16.u32 %rs1, %r2;
|
67 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs2}, %r2; }
|
68 |
+
cvt.u16.u32 %rs3, %r3;
|
69 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs4}, %r3; }
|
70 |
+
.loc 1 30 67
|
71 |
+
cvt.f32.bf16 %r6, %rs1;
|
72 |
+
mov.b32 %f1, %r6;
|
73 |
+
cvt.f32.bf16 %r7, %rs2;
|
74 |
+
mov.b32 %f2, %r7;
|
75 |
+
cvt.f32.bf16 %r8, %rs3;
|
76 |
+
mov.b32 %f3, %r8;
|
77 |
+
cvt.f32.bf16 %r9, %rs4;
|
78 |
+
mov.b32 %f4, %r9;
|
79 |
+
.loc 1 31 30
|
80 |
+
cvt.u64.u32 %rd34, %r77;
|
81 |
+
mul.wide.u32 %rd35, %r77, 4;
|
82 |
+
add.s64 %rd2, %rd27, %rd35;
|
83 |
+
.loc 1 31 35
|
84 |
+
mov.u32 %r10, 0x0;
|
85 |
+
mov.u32 %r11, 0x0;
|
86 |
+
mov.u32 %r12, 0x0;
|
87 |
+
mov.u32 %r13, 0x0;
|
88 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r10, %r11, %r12, %r13 }, [ %rd2 + 0 ];
|
89 |
+
@!%p1 mov.u32 %r10, %r4;
|
90 |
+
@!%p1 mov.u32 %r11, %r4;
|
91 |
+
@!%p1 mov.u32 %r12, %r4;
|
92 |
+
@!%p1 mov.u32 %r13, %r4;
|
93 |
+
mov.b32 %f5, %r10;
|
94 |
+
mov.b32 %f6, %r11;
|
95 |
+
mov.b32 %f7, %r12;
|
96 |
+
mov.b32 %f8, %r13;
|
97 |
+
.loc 1 32 30
|
98 |
+
mul.wide.s32 %rd36, %r79, 4;
|
99 |
+
add.s64 %rd3, %rd28, %rd36;
|
100 |
+
.loc 1 32 46
|
101 |
+
mov.u32 %r18, 0x0;
|
102 |
+
mov.u32 %r19, 0x0;
|
103 |
+
mov.u32 %r20, 0x0;
|
104 |
+
mov.u32 %r21, 0x0;
|
105 |
+
@%p1 ld.global.v4.b32 { %r18, %r19, %r20, %r21 }, [ %rd3 + 0 ];
|
106 |
+
@!%p1 mov.u32 %r18, %r4;
|
107 |
+
@!%p1 mov.u32 %r19, %r4;
|
108 |
+
@!%p1 mov.u32 %r20, %r4;
|
109 |
+
@!%p1 mov.u32 %r21, %r4;
|
110 |
+
mov.b32 %f9, %r18;
|
111 |
+
mov.b32 %f10, %r19;
|
112 |
+
mov.b32 %f11, %r20;
|
113 |
+
mov.b32 %f12, %r21;
|
114 |
+
.loc 1 33 30
|
115 |
+
mul.wide.s32 %rd37, %r1, 4;
|
116 |
+
add.s64 %rd4, %rd29, %rd37;
|
117 |
+
.loc 1 33 35
|
118 |
+
mov.u32 %r26, 0x0;
|
119 |
+
@%p1 ld.global.L1::evict_last.b32 { %r26 }, [ %rd4 + 0 ];
|
120 |
+
mov.b32 %f13, %r26;
|
121 |
+
mov.u32 %r27, 0x0;
|
122 |
+
@%p1 ld.global.L1::evict_last.b32 { %r27 }, [ %rd4 + 0 ];
|
123 |
+
mov.u32 %r28, 0x0;
|
124 |
+
@%p1 ld.global.L1::evict_last.b32 { %r28 }, [ %rd4 + 0 ];
|
125 |
+
mov.u32 %r29, 0x0;
|
126 |
+
@%p1 ld.global.L1::evict_last.b32 { %r29 }, [ %rd4 + 0 ];
|
127 |
+
.loc 1 34 31
|
128 |
+
add.s64 %rd8, %rd30, %rd37;
|
129 |
+
.loc 1 34 36
|
130 |
+
mov.u32 %r55, 0x0;
|
131 |
+
@%p1 ld.global.L1::evict_last.b32 { %r55 }, [ %rd8 + 0 ];
|
132 |
+
mov.b32 %f14, %r55;
|
133 |
+
mov.u32 %r31, 0x0;
|
134 |
+
@%p1 ld.global.L1::evict_last.b32 { %r31 }, [ %rd8 + 0 ];
|
135 |
+
mov.u32 %r32, 0x0;
|
136 |
+
@%p1 ld.global.L1::evict_last.b32 { %r32 }, [ %rd8 + 0 ];
|
137 |
+
mov.u32 %r33, 0x0;
|
138 |
+
@%p1 ld.global.L1::evict_last.b32 { %r33 }, [ %rd8 + 0 ];
|
139 |
+
.loc 1 35 31
|
140 |
+
mul.wide.s32 %rd38, %r1, 8;
|
141 |
+
add.s64 %rd13, %rd31, %rd38;
|
142 |
+
.loc 1 35 36
|
143 |
+
mov.u64 %rd12, 0x0;
|
144 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd12 }, [ %rd13 + 0 ];
|
145 |
+
mov.u64 %rd14, 0x0;
|
146 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd14 }, [ %rd13 + 0 ];
|
147 |
+
mov.u64 %rd16, 0x0;
|
148 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd16 }, [ %rd13 + 0 ];
|
149 |
+
mov.u64 %rd18, 0x0;
|
150 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd18 }, [ %rd13 + 0 ];
|
151 |
+
.loc 1 36 35
|
152 |
+
add.s64 %rd20, %rd25, %rd36;
|
153 |
+
.loc 1 36 51
|
154 |
+
mov.u32 %r34, 0x0;
|
155 |
+
mov.u32 %r35, 0x0;
|
156 |
+
mov.u32 %r36, 0x0;
|
157 |
+
mov.u32 %r37, 0x0;
|
158 |
+
@%p1 ld.global.v4.b32 { %r34, %r35, %r36, %r37 }, [ %rd20 + 0 ];
|
159 |
+
@!%p1 mov.u32 %r34, %r4;
|
160 |
+
@!%p1 mov.u32 %r35, %r4;
|
161 |
+
@!%p1 mov.u32 %r36, %r4;
|
162 |
+
@!%p1 mov.u32 %r37, %r4;
|
163 |
+
mov.b32 %f15, %r34;
|
164 |
+
mov.b32 %f16, %r35;
|
165 |
+
mov.b32 %f17, %r36;
|
166 |
+
mov.b32 %f18, %r37;
|
167 |
+
.loc 1 38 18
|
168 |
+
mul.f32 %f19, %f1, %f5;
|
169 |
+
mul.f32 %f20, %f2, %f6;
|
170 |
+
mul.f32 %f21, %f3, %f7;
|
171 |
+
mul.f32 %f22, %f4, %f8;
|
172 |
+
$L__tmp1:
|
173 |
+
.loc 2 233 15
|
174 |
+
fma.rn.f32 %f23, %f1, %f5, %f20;
|
175 |
+
fma.rn.f32 %f24, %f3, %f7, %f23;
|
176 |
+
fma.rn.f32 %f25, %f4, %f8, %f24;
|
177 |
+
$L__tmp2:
|
178 |
+
.loc 2 243 36
|
179 |
+
mov.b32 %r80, %f25;
|
180 |
+
shfl.sync.bfly.b32 %r81, %r80, 16, 31, -1;
|
181 |
+
mov.b32 %f26, %r81;
|
182 |
+
$L__tmp3:
|
183 |
+
.loc 2 233 15
|
184 |
+
add.f32 %f27, %f25, %f26;
|
185 |
+
$L__tmp4:
|
186 |
+
.loc 2 243 36
|
187 |
+
mov.b32 %r82, %f27;
|
188 |
+
shfl.sync.bfly.b32 %r83, %r82, 8, 31, -1;
|
189 |
+
mov.b32 %f28, %r83;
|
190 |
+
$L__tmp5:
|
191 |
+
.loc 2 233 15
|
192 |
+
add.f32 %f29, %f27, %f28;
|
193 |
+
$L__tmp6:
|
194 |
+
.loc 2 243 36
|
195 |
+
mov.b32 %r84, %f29;
|
196 |
+
shfl.sync.bfly.b32 %r85, %r84, 4, 31, -1;
|
197 |
+
mov.b32 %f30, %r85;
|
198 |
+
$L__tmp7:
|
199 |
+
.loc 2 233 15
|
200 |
+
add.f32 %f31, %f29, %f30;
|
201 |
+
$L__tmp8:
|
202 |
+
.loc 2 243 36
|
203 |
+
mov.b32 %r86, %f31;
|
204 |
+
shfl.sync.bfly.b32 %r87, %r86, 2, 31, -1;
|
205 |
+
mov.b32 %f32, %r87;
|
206 |
+
$L__tmp9:
|
207 |
+
.loc 2 233 15
|
208 |
+
add.f32 %f33, %f31, %f32;
|
209 |
+
$L__tmp10:
|
210 |
+
.loc 2 243 36
|
211 |
+
mov.b32 %r88, %f33;
|
212 |
+
shfl.sync.bfly.b32 %r89, %r88, 1, 31, -1;
|
213 |
+
mov.b32 %f34, %r89;
|
214 |
+
$L__tmp11:
|
215 |
+
.loc 2 233 15
|
216 |
+
add.f32 %f35, %f33, %f34;
|
217 |
+
$L__tmp12:
|
218 |
+
.loc 2 243 36
|
219 |
+
setp.eq.s32 %p31, %r75, 0;
|
220 |
+
shr.u32 %r90, %r74, 3;
|
221 |
+
and.b32 %r91, %r90, 4;
|
222 |
+
mov.u32 %r92, global_smem;
|
223 |
+
add.s32 %r42, %r92, %r91;
|
224 |
+
mov.b32 %r43, %f35;
|
225 |
+
@%p31 st.shared.b32 [ %r42 + 0 ], %r43;
|
226 |
+
bar.sync 0;
|
227 |
+
setp.lt.s32 %p32, %r74, 2;
|
228 |
+
add.s32 %r45, %r92, %r76;
|
229 |
+
@%p32 ld.shared.b32 %r44, [ %r45 + 0 ];
|
230 |
+
mov.b32 %f36, %r44;
|
231 |
+
shfl.sync.bfly.b32 %r93, %r44, 1, 31, -1;
|
232 |
+
mov.b32 %f37, %r93;
|
233 |
+
$L__tmp13:
|
234 |
+
.loc 2 233 15
|
235 |
+
add.f32 %f38, %f36, %f37;
|
236 |
+
$L__tmp14:
|
237 |
+
.loc 2 243 36
|
238 |
+
and.b32 %r94, %r74, 1;
|
239 |
+
setp.eq.b32 %p41, %r94, 1;
|
240 |
+
not.pred %p42, %p41;
|
241 |
+
and.pred %p33, %p32, %p42;
|
242 |
+
mov.b32 %r47, %f38;
|
243 |
+
@%p33 st.shared.b32 [ %r45 + 0 ], %r47;
|
244 |
+
bar.sync 0;
|
245 |
+
ld.shared.f32 %f39, [global_smem];
|
246 |
+
$L__tmp15:
|
247 |
+
.loc 3 8 15
|
248 |
+
add.f32 %f40, %f39, 0f00000000;
|
249 |
+
$L__tmp16:
|
250 |
+
.loc 1 42 19
|
251 |
+
sub.f32 %f41, %f9, %f13;
|
252 |
+
sub.f32 %f42, %f10, %f13;
|
253 |
+
sub.f32 %f43, %f11, %f13;
|
254 |
+
sub.f32 %f44, %f12, %f13;
|
255 |
+
.loc 1 43 20
|
256 |
+
mul.f32 %f45, %f41, %f14;
|
257 |
+
mul.f32 %f46, %f42, %f14;
|
258 |
+
mul.f32 %f47, %f43, %f14;
|
259 |
+
mul.f32 %f48, %f44, %f14;
|
260 |
+
.loc 1 44 19
|
261 |
+
mul.f32 %f49, %f20, %f46;
|
262 |
+
$L__tmp17:
|
263 |
+
.loc 2 243 36
|
264 |
+
bar.sync 0;
|
265 |
+
$L__tmp18:
|
266 |
+
.loc 2 233 15
|
267 |
+
fma.rn.f32 %f50, %f19, %f45, %f49;
|
268 |
+
fma.rn.f32 %f51, %f21, %f47, %f50;
|
269 |
+
fma.rn.f32 %f52, %f22, %f48, %f51;
|
270 |
+
$L__tmp19:
|
271 |
+
.loc 2 243 36
|
272 |
+
mov.b32 %r95, %f52;
|
273 |
+
shfl.sync.bfly.b32 %r96, %r95, 16, 31, -1;
|
274 |
+
mov.b32 %f53, %r96;
|
275 |
+
$L__tmp20:
|
276 |
+
.loc 2 233 15
|
277 |
+
add.f32 %f54, %f52, %f53;
|
278 |
+
$L__tmp21:
|
279 |
+
.loc 2 243 36
|
280 |
+
mov.b32 %r97, %f54;
|
281 |
+
shfl.sync.bfly.b32 %r98, %r97, 8, 31, -1;
|
282 |
+
mov.b32 %f55, %r98;
|
283 |
+
$L__tmp22:
|
284 |
+
.loc 2 233 15
|
285 |
+
add.f32 %f56, %f54, %f55;
|
286 |
+
$L__tmp23:
|
287 |
+
.loc 2 243 36
|
288 |
+
mov.b32 %r99, %f56;
|
289 |
+
shfl.sync.bfly.b32 %r100, %r99, 4, 31, -1;
|
290 |
+
mov.b32 %f57, %r100;
|
291 |
+
$L__tmp24:
|
292 |
+
.loc 2 233 15
|
293 |
+
add.f32 %f58, %f56, %f57;
|
294 |
+
$L__tmp25:
|
295 |
+
.loc 2 243 36
|
296 |
+
mov.b32 %r101, %f58;
|
297 |
+
shfl.sync.bfly.b32 %r102, %r101, 2, 31, -1;
|
298 |
+
mov.b32 %f59, %r102;
|
299 |
+
$L__tmp26:
|
300 |
+
.loc 2 233 15
|
301 |
+
add.f32 %f60, %f58, %f59;
|
302 |
+
$L__tmp27:
|
303 |
+
.loc 2 243 36
|
304 |
+
mov.b32 %r103, %f60;
|
305 |
+
shfl.sync.bfly.b32 %r104, %r103, 1, 31, -1;
|
306 |
+
mov.b32 %f61, %r104;
|
307 |
+
$L__tmp28:
|
308 |
+
.loc 2 233 15
|
309 |
+
add.f32 %f62, %f60, %f61;
|
310 |
+
$L__tmp29:
|
311 |
+
.loc 2 243 36
|
312 |
+
mov.b32 %r49, %f62;
|
313 |
+
@%p31 st.shared.b32 [ %r42 + 0 ], %r49;
|
314 |
+
bar.sync 0;
|
315 |
+
@%p32 ld.shared.b32 %r50, [ %r45 + 0 ];
|
316 |
+
mov.b32 %f63, %r50;
|
317 |
+
shfl.sync.bfly.b32 %r105, %r50, 1, 31, -1;
|
318 |
+
mov.b32 %f64, %r105;
|
319 |
+
$L__tmp30:
|
320 |
+
.loc 2 233 15
|
321 |
+
add.f32 %f65, %f63, %f64;
|
322 |
+
$L__tmp31:
|
323 |
+
.loc 2 243 36
|
324 |
+
mov.b32 %r53, %f65;
|
325 |
+
@%p33 st.shared.b32 [ %r45 + 0 ], %r53;
|
326 |
+
bar.sync 0;
|
327 |
+
ld.shared.f32 %f66, [global_smem];
|
328 |
+
$L__tmp32:
|
329 |
+
.loc 3 8 15
|
330 |
+
add.f32 %f67, %f66, 0f00000000;
|
331 |
+
$L__tmp33:
|
332 |
+
.loc 1 49 21
|
333 |
+
setp.eq.s64 %p43, %rd12, -1;
|
334 |
+
mov.b32 %r56, 1132462080;
|
335 |
+
.loc 1 51 20
|
336 |
+
div.full.f32 %r54, %r55, %r56;
|
337 |
+
mov.b32 %f68, %r54;
|
338 |
+
.loc 1 53 20
|
339 |
+
neg.f32 %f69, %f40;
|
340 |
+
fma.rn.f32 %f70, %f19, 0f43800000, %f69;
|
341 |
+
fma.rn.f32 %f71, %f20, 0f43800000, %f69;
|
342 |
+
fma.rn.f32 %f72, %f21, 0f43800000, %f69;
|
343 |
+
fma.rn.f32 %f73, %f22, 0f43800000, %f69;
|
344 |
+
.loc 1 55 20
|
345 |
+
neg.f32 %f74, %f45;
|
346 |
+
fma.rn.f32 %f75, %f74, %f67, %f70;
|
347 |
+
neg.f32 %f76, %f46;
|
348 |
+
fma.rn.f32 %f77, %f76, %f67, %f71;
|
349 |
+
neg.f32 %f78, %f47;
|
350 |
+
fma.rn.f32 %f79, %f78, %f67, %f72;
|
351 |
+
neg.f32 %f80, %f48;
|
352 |
+
fma.rn.f32 %f81, %f80, %f67, %f73;
|
353 |
+
.loc 1 57 20
|
354 |
+
fma.rn.f32 %f82, %f68, %f75, %f15;
|
355 |
+
fma.rn.f32 %f83, %f68, %f77, %f16;
|
356 |
+
fma.rn.f32 %f84, %f68, %f79, %f17;
|
357 |
+
fma.rn.f32 %f85, %f68, %f81, %f18;
|
358 |
+
.loc 1 59 35
|
359 |
+
selp.f32 %f86, 0f00000000, %f82, %p43;
|
360 |
+
selp.f32 %f87, 0f00000000, %f83, %p43;
|
361 |
+
selp.f32 %f88, 0f00000000, %f84, %p43;
|
362 |
+
selp.f32 %f89, 0f00000000, %f85, %p43;
|
363 |
+
.loc 1 61 20
|
364 |
+
setp.lt.s64 %p44, %rd12, 0;
|
365 |
+
.loc 1 63 56
|
366 |
+
shl.b64 %rd39, %rd12, 8;
|
367 |
+
add.s64 %rd40, %rd39, 12865792;
|
368 |
+
selp.b64 %rd41, %rd40, %rd39, %p44;
|
369 |
+
.loc 1 63 52
|
370 |
+
or.b64 %rd42, %rd41, %rd34;
|
371 |
+
.loc 1 63 30
|
372 |
+
shl.b64 %rd43, %rd42, 2;
|
373 |
+
add.s64 %rd21, %rd32, %rd43;
|
374 |
+
add.s64 %rd22, %rd21, 4;
|
375 |
+
add.s64 %rd23, %rd21, 8;
|
376 |
+
add.s64 %rd24, %rd21, 12;
|
377 |
+
.loc 1 63 83
|
378 |
+
mov.b32 %r67, %f86;
|
379 |
+
mov.u32 %r66, 0x0;
|
380 |
+
@%p1 atom.global.gpu.acq_rel.add.f32 %r66, [ %rd21 + 0 ], %r67;
|
381 |
+
mov.b32 %r69, %f87;
|
382 |
+
mov.u32 %r68, 0x0;
|
383 |
+
@%p1 atom.global.gpu.acq_rel.add.f32 %r68, [ %rd22 + 0 ], %r69;
|
384 |
+
mov.b32 %r71, %f88;
|
385 |
+
mov.u32 %r70, 0x0;
|
386 |
+
@%p1 atom.global.gpu.acq_rel.add.f32 %r70, [ %rd23 + 0 ], %r71;
|
387 |
+
mov.b32 %r73, %f89;
|
388 |
+
mov.u32 %r72, 0x0;
|
389 |
+
@%p1 atom.global.gpu.acq_rel.add.f32 %r72, [ %rd24 + 0 ], %r73;
|
390 |
+
.loc 1 63 4
|
391 |
+
ret;
|
392 |
+
$L__tmp34:
|
393 |
+
$L__func_end0:
|
394 |
+
|
395 |
+
}
|
396 |
+
.file 1 "/tmp/torchinductor_root/qr/cqryxm46jcxyr3qdktqirn53eap7h3pjjqiqavyqqyvflabjpvmd.py"
|
397 |
+
.file 2 "/usr/local/lib/python3.10/dist-packages/triton/language/standard.py"
|
398 |
+
.file 3 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
|
399 |
+
.section .debug_abbrev
|
400 |
+
{
|
401 |
+
.b8 1
|
402 |
+
.b8 17
|
403 |
+
.b8 1
|
404 |
+
.b8 37
|
405 |
+
.b8 8
|
406 |
+
.b8 19
|
407 |
+
.b8 5
|
408 |
+
.b8 3
|
409 |
+
.b8 8
|
410 |
+
.b8 16
|
411 |
+
.b8 6
|
412 |
+
.b8 27
|
413 |
+
.b8 8
|
414 |
+
.b8 180
|
415 |
+
.b8 66
|
416 |
+
.b8 12
|
417 |
+
.b8 17
|
418 |
+
.b8 1
|
419 |
+
.b8 18
|
420 |
+
.b8 1
|
421 |
+
.b8 0
|
422 |
+
.b8 0
|
423 |
+
.b8 2
|
424 |
+
.b8 46
|
425 |
+
.b8 0
|
426 |
+
.b8 135
|
427 |
+
.b8 64
|
428 |
+
.b8 8
|
429 |
+
.b8 3
|
430 |
+
.b8 8
|
431 |
+
.b8 58
|
432 |
+
.b8 11
|
433 |
+
.b8 59
|
434 |
+
.b8 11
|
435 |
+
.b8 63
|
436 |
+
.b8 12
|
437 |
+
.b8 32
|
438 |
+
.b8 11
|
439 |
+
.b8 0
|
440 |
+
.b8 0
|
441 |
+
.b8 3
|
442 |
+
.b8 46
|
443 |
+
.b8 1
|
444 |
+
.b8 17
|
445 |
+
.b8 1
|
446 |
+
.b8 18
|
447 |
+
.b8 1
|
448 |
+
.b8 64
|
449 |
+
.b8 10
|
450 |
+
.b8 49
|
451 |
+
.b8 19
|
452 |
+
.b8 0
|
453 |
+
.b8 0
|
454 |
+
.b8 4
|
455 |
+
.b8 29
|
456 |
+
.b8 1
|
457 |
+
.b8 49
|
458 |
+
.b8 19
|
459 |
+
.b8 17
|
460 |
+
.b8 1
|
461 |
+
.b8 18
|
462 |
+
.b8 1
|
463 |
+
.b8 88
|
464 |
+
.b8 11
|
465 |
+
.b8 89
|
466 |
+
.b8 11
|
467 |
+
.b8 87
|
468 |
+
.b8 11
|
469 |
+
.b8 0
|
470 |
+
.b8 0
|
471 |
+
.b8 5
|
472 |
+
.b8 29
|
473 |
+
.b8 0
|
474 |
+
.b8 49
|
475 |
+
.b8 19
|
476 |
+
.b8 17
|
477 |
+
.b8 1
|
478 |
+
.b8 18
|
479 |
+
.b8 1
|
480 |
+
.b8 88
|
481 |
+
.b8 11
|
482 |
+
.b8 89
|
483 |
+
.b8 11
|
484 |
+
.b8 87
|
485 |
+
.b8 11
|
486 |
+
.b8 0
|
487 |
+
.b8 0
|
488 |
+
.b8 0
|
489 |
+
}
|
490 |
+
.section .debug_info
|
491 |
+
{
|
492 |
+
.b32 407
|
493 |
+
.b8 2
|
494 |
+
.b8 0
|
495 |
+
.b32 .debug_abbrev
|
496 |
+
.b8 8
|
497 |
+
.b8 1
|
498 |
+
.b8 116
|
499 |
+
.b8 114
|
500 |
+
.b8 105
|
501 |
+
.b8 116
|
502 |
+
.b8 111
|
503 |
+
.b8 110
|
504 |
+
.b8 0
|
505 |
+
.b8 2
|
506 |
+
.b8 0
|
507 |
+
.b8 99
|
508 |
+
.b8 113
|
509 |
+
.b8 114
|
510 |
+
.b8 121
|
511 |
+
.b8 120
|
512 |
+
.b8 109
|
513 |
+
.b8 52
|
514 |
+
.b8 54
|
515 |
+
.b8 106
|
516 |
+
.b8 99
|
517 |
+
.b8 120
|
518 |
+
.b8 121
|
519 |
+
.b8 114
|
520 |
+
.b8 51
|
521 |
+
.b8 113
|
522 |
+
.b8 100
|
523 |
+
.b8 107
|
524 |
+
.b8 116
|
525 |
+
.b8 113
|
526 |
+
.b8 105
|
527 |
+
.b8 114
|
528 |
+
.b8 110
|
529 |
+
.b8 53
|
530 |
+
.b8 51
|
531 |
+
.b8 101
|
532 |
+
.b8 97
|
533 |
+
.b8 112
|
534 |
+
.b8 55
|
535 |
+
.b8 104
|
536 |
+
.b8 51
|
537 |
+
.b8 112
|
538 |
+
.b8 106
|
539 |
+
.b8 106
|
540 |
+
.b8 113
|
541 |
+
.b8 105
|
542 |
+
.b8 113
|
543 |
+
.b8 97
|
544 |
+
.b8 118
|
545 |
+
.b8 121
|
546 |
+
.b8 113
|
547 |
+
.b8 113
|
548 |
+
.b8 121
|
549 |
+
.b8 118
|
550 |
+
.b8 102
|
551 |
+
.b8 108
|
552 |
+
.b8 97
|
553 |
+
.b8 98
|
554 |
+
.b8 106
|
555 |
+
.b8 112
|
556 |
+
.b8 118
|
557 |
+
.b8 109
|
558 |
+
.b8 100
|
559 |
+
.b8 46
|
560 |
+
.b8 112
|
561 |
+
.b8 121
|
562 |
+
.b8 0
|
563 |
+
.b32 .debug_line
|
564 |
+
.b8 47
|
565 |
+
.b8 116
|
566 |
+
.b8 109
|
567 |
+
.b8 112
|
568 |
+
.b8 47
|
569 |
+
.b8 116
|
570 |
+
.b8 111
|
571 |
+
.b8 114
|
572 |
+
.b8 99
|
573 |
+
.b8 104
|
574 |
+
.b8 105
|
575 |
+
.b8 110
|
576 |
+
.b8 100
|
577 |
+
.b8 117
|
578 |
+
.b8 99
|
579 |
+
.b8 116
|
580 |
+
.b8 111
|
581 |
+
.b8 114
|
582 |
+
.b8 95
|
583 |
+
.b8 114
|
584 |
+
.b8 111
|
585 |
+
.b8 111
|
586 |
+
.b8 116
|
587 |
+
.b8 47
|
588 |
+
.b8 113
|
589 |
+
.b8 114
|
590 |
+
.b8 0
|
591 |
+
.b8 1
|
592 |
+
.b64 $L__func_begin0
|
593 |
+
.b64 $L__func_end0
|
594 |
+
.b8 2
|
595 |
+
.b8 116
|
596 |
+
.b8 114
|
597 |
+
.b8 105
|
598 |
+
.b8 116
|
599 |
+
.b8 111
|
600 |
+
.b8 110
|
601 |
+
.b8 95
|
602 |
+
.b8 95
|
603 |
+
.b8 48
|
604 |
+
.b8 100
|
605 |
+
.b8 49
|
606 |
+
.b8 100
|
607 |
+
.b8 50
|
608 |
+
.b8 100
|
609 |
+
.b8 51
|
610 |
+
.b8 100
|
611 |
+
.b8 52
|
612 |
+
.b8 100
|
613 |
+
.b8 53
|
614 |
+
.b8 100
|
615 |
+
.b8 54
|
616 |
+
.b8 100
|
617 |
+
.b8 55
|
618 |
+
.b8 100
|
619 |
+
.b8 56
|
620 |
+
.b8 100
|
621 |
+
.b8 101
|
622 |
+
.b8 57
|
623 |
+
.b8 100
|
624 |
+
.b8 101
|
625 |
+
.b8 0
|
626 |
+
.b8 116
|
627 |
+
.b8 114
|
628 |
+
.b8 105
|
629 |
+
.b8 116
|
630 |
+
.b8 111
|
631 |
+
.b8 110
|
632 |
+
.b8 95
|
633 |
+
.b8 95
|
634 |
+
.b8 48
|
635 |
+
.b8 100
|
636 |
+
.b8 49
|
637 |
+
.b8 100
|
638 |
+
.b8 50
|
639 |
+
.b8 100
|
640 |
+
.b8 51
|
641 |
+
.b8 100
|
642 |
+
.b8 52
|
643 |
+
.b8 100
|
644 |
+
.b8 53
|
645 |
+
.b8 100
|
646 |
+
.b8 54
|
647 |
+
.b8 100
|
648 |
+
.b8 55
|
649 |
+
.b8 100
|
650 |
+
.b8 56
|
651 |
+
.b8 100
|
652 |
+
.b8 101
|
653 |
+
.b8 57
|
654 |
+
.b8 100
|
655 |
+
.b8 101
|
656 |
+
.b8 0
|
657 |
+
.b8 1
|
658 |
+
.b8 18
|
659 |
+
.b8 1
|
660 |
+
.b8 1
|
661 |
+
.b8 3
|
662 |
+
.b64 $L__func_begin0
|
663 |
+
.b64 $L__func_end0
|
664 |
+
.b8 1
|
665 |
+
.b8 156
|
666 |
+
.b32 125
|
667 |
+
.b8 4
|
668 |
+
.b32 125
|
669 |
+
.b64 $L__tmp1
|
670 |
+
.b64 $L__tmp14
|
671 |
+
.b8 2
|
672 |
+
.b8 41
|
673 |
+
.b8 57
|
674 |
+
.b8 5
|
675 |
+
.b32 125
|
676 |
+
.b64 $L__tmp1
|
677 |
+
.b64 $L__tmp14
|
678 |
+
.b8 2
|
679 |
+
.b8 243
|
680 |
+
.b8 36
|
681 |
+
.b8 0
|
682 |
+
.b8 5
|
683 |
+
.b32 125
|
684 |
+
.b64 $L__tmp2
|
685 |
+
.b64 $L__tmp15
|
686 |
+
.b8 2
|
687 |
+
.b8 41
|
688 |
+
.b8 57
|
689 |
+
.b8 5
|
690 |
+
.b32 125
|
691 |
+
.b64 $L__tmp15
|
692 |
+
.b64 $L__tmp16
|
693 |
+
.b8 3
|
694 |
+
.b8 41
|
695 |
+
.b8 44
|
696 |
+
.b8 5
|
697 |
+
.b32 125
|
698 |
+
.b64 $L__tmp17
|
699 |
+
.b64 $L__tmp32
|
700 |
+
.b8 2
|
701 |
+
.b8 47
|
702 |
+
.b8 59
|
703 |
+
.b8 4
|
704 |
+
.b32 125
|
705 |
+
.b64 $L__tmp18
|
706 |
+
.b64 $L__tmp31
|
707 |
+
.b8 2
|
708 |
+
.b8 47
|
709 |
+
.b8 59
|
710 |
+
.b8 5
|
711 |
+
.b32 125
|
712 |
+
.b64 $L__tmp18
|
713 |
+
.b64 $L__tmp31
|
714 |
+
.b8 2
|
715 |
+
.b8 243
|
716 |
+
.b8 36
|
717 |
+
.b8 0
|
718 |
+
.b8 5
|
719 |
+
.b32 125
|
720 |
+
.b64 $L__tmp32
|
721 |
+
.b64 $L__tmp33
|
722 |
+
.b8 3
|
723 |
+
.b8 47
|
724 |
+
.b8 45
|
725 |
+
.b8 0
|
726 |
+
.b8 0
|
727 |
+
}
|
728 |
+
.section .debug_pubnames
|
729 |
+
{
|
730 |
+
.b32 $L__pubNames_end0-$L__pubNames_start0
|
731 |
+
$L__pubNames_start0:
|
732 |
+
.b8 2
|
733 |
+
.b8 0
|
734 |
+
.b32 .debug_info
|
735 |
+
.b32 411
|
736 |
+
.b32 125
|
737 |
+
.b8 116
|
738 |
+
.b8 114
|
739 |
+
.b8 105
|
740 |
+
.b8 116
|
741 |
+
.b8 111
|
742 |
+
.b8 110
|
743 |
+
.b8 95
|
744 |
+
.b8 95
|
745 |
+
.b8 48
|
746 |
+
.b8 100
|
747 |
+
.b8 49
|
748 |
+
.b8 100
|
749 |
+
.b8 50
|
750 |
+
.b8 100
|
751 |
+
.b8 51
|
752 |
+
.b8 100
|
753 |
+
.b8 52
|
754 |
+
.b8 100
|
755 |
+
.b8 53
|
756 |
+
.b8 100
|
757 |
+
.b8 54
|
758 |
+
.b8 100
|
759 |
+
.b8 55
|
760 |
+
.b8 100
|
761 |
+
.b8 56
|
762 |
+
.b8 100
|
763 |
+
.b8 101
|
764 |
+
.b8 57
|
765 |
+
.b8 100
|
766 |
+
.b8 101
|
767 |
+
.b8 0
|
768 |
+
.b32 0
|
769 |
+
$L__pubNames_end0:
|
770 |
+
}
|
771 |
+
.section .debug_pubtypes
|
772 |
+
{
|
773 |
+
.b32 $L__pubTypes_end0-$L__pubTypes_start0
|
774 |
+
$L__pubTypes_start0:
|
775 |
+
.b8 2
|
776 |
+
.b8 0
|
777 |
+
.b32 .debug_info
|
778 |
+
.b32 411
|
779 |
+
.b32 0
|
780 |
+
$L__pubTypes_end0:
|
781 |
+
}
|
782 |
+
.section .debug_loc { }
|
.triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ttgir
ADDED
@@ -0,0 +1,88 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [4], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
|
2 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
3 |
+
tt.func public @triton__0d1d2d3d4d5d6d7d8de9de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg6: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg7: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg8: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg9: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
4 |
+
%cst = arith.constant dense<256> : tensor<256xi32, #blocked>
|
5 |
+
%cst_0 = arith.constant dense<-1> : tensor<1xi64, #blocked>
|
6 |
+
%cst_1 = arith.constant dense<2.560000e+02> : tensor<1xf32, #blocked>
|
7 |
+
%cst_2 = arith.constant dense<256> : tensor<1xi64, #blocked>
|
8 |
+
%cst_3 = arith.constant dense<0> : tensor<1xi64, #blocked>
|
9 |
+
%cst_4 = arith.constant dense<50257> : tensor<1xi64, #blocked>
|
10 |
+
%cst_5 = arith.constant 0.000000e+00 : f32
|
11 |
+
%c256_i32 = arith.constant 256 : i32
|
12 |
+
%cst_6 = arith.constant dense<0.000000e+00> : tensor<256xf32, #blocked>
|
13 |
+
%cst_7 = arith.constant dense<2.560000e+02> : tensor<256xf32, #blocked>
|
14 |
+
%cst_8 = arith.constant dense<0.000000e+00> : tensor<256xbf16, #blocked>
|
15 |
+
%0 = tt.get_program_id x : i32
|
16 |
+
%1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32, #blocked>
|
17 |
+
%2 = arith.cmpi slt, %1, %cst : tensor<256xi32, #blocked>
|
18 |
+
%3 = arith.muli %0, %c256_i32 : i32
|
19 |
+
%4 = tt.splat %3 : (i32) -> tensor<256xi32, #blocked>
|
20 |
+
%5 = arith.addi %1, %4 : tensor<256xi32, #blocked>
|
21 |
+
%6 = tt.splat %arg1 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
22 |
+
%7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
23 |
+
%8 = tt.load %7, %2, %cst_8 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
|
24 |
+
%9 = arith.extf %8 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
|
25 |
+
%10 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
26 |
+
%11 = tt.addptr %10, %1 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
27 |
+
%12 = tt.load %11, %2, %cst_6 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
28 |
+
%13 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
29 |
+
%14 = tt.addptr %13, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
30 |
+
%15 = tt.load %14, %2, %cst_6 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
31 |
+
%16 = tt.addptr %arg4, %0 : !tt.ptr<f32, 1>, i32
|
32 |
+
%17 = tt.splat %16 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked>
|
33 |
+
%18 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1xf32, #blocked>
|
34 |
+
%19 = tt.addptr %arg5, %0 : !tt.ptr<f32, 1>, i32
|
35 |
+
%20 = tt.splat %19 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked>
|
36 |
+
%21 = tt.load %20 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1xf32, #blocked>
|
37 |
+
%22 = tt.addptr %arg6, %0 : !tt.ptr<i64, 1>, i32
|
38 |
+
%23 = tt.splat %22 : (!tt.ptr<i64, 1>) -> tensor<1x!tt.ptr<i64, 1>, #blocked>
|
39 |
+
%24 = tt.load %23 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1xi64, #blocked>
|
40 |
+
%25 = tt.splat %arg0 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
41 |
+
%26 = tt.addptr %25, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
42 |
+
%27 = tt.load %26, %2, %cst_6 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
43 |
+
%28 = arith.mulf %9, %12 : tensor<256xf32, #blocked>
|
44 |
+
%29 = arith.select %2, %28, %cst_6 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
45 |
+
%30 = "tt.reduce"(%29) <{axis = 0 : i32}> ({
|
46 |
+
^bb0(%arg10: f32, %arg11: f32):
|
47 |
+
%63 = arith.addf %arg10, %arg11 : f32
|
48 |
+
tt.reduce.return %63 : f32
|
49 |
+
}) : (tensor<256xf32, #blocked>) -> f32
|
50 |
+
%31 = arith.addf %30, %cst_5 : f32
|
51 |
+
%32 = tt.broadcast %18 : (tensor<1xf32, #blocked>) -> tensor<256xf32, #blocked>
|
52 |
+
%33 = arith.subf %15, %32 : tensor<256xf32, #blocked>
|
53 |
+
%34 = tt.broadcast %21 : (tensor<1xf32, #blocked>) -> tensor<256xf32, #blocked>
|
54 |
+
%35 = arith.mulf %33, %34 : tensor<256xf32, #blocked>
|
55 |
+
%36 = arith.mulf %28, %35 : tensor<256xf32, #blocked>
|
56 |
+
%37 = arith.select %2, %36, %cst_6 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
57 |
+
%38 = "tt.reduce"(%37) <{axis = 0 : i32}> ({
|
58 |
+
^bb0(%arg10: f32, %arg11: f32):
|
59 |
+
%63 = arith.addf %arg10, %arg11 : f32
|
60 |
+
tt.reduce.return %63 : f32
|
61 |
+
}) : (tensor<256xf32, #blocked>) -> f32
|
62 |
+
%39 = arith.addf %38, %cst_5 : f32
|
63 |
+
%40 = arith.cmpi eq, %24, %cst_0 : tensor<1xi64, #blocked>
|
64 |
+
%41 = arith.divf %21, %cst_1 : tensor<1xf32, #blocked>
|
65 |
+
%42 = arith.mulf %28, %cst_7 : tensor<256xf32, #blocked>
|
66 |
+
%43 = tt.splat %31 : (f32) -> tensor<256xf32, #blocked>
|
67 |
+
%44 = arith.subf %42, %43 : tensor<256xf32, #blocked>
|
68 |
+
%45 = tt.splat %39 : (f32) -> tensor<256xf32, #blocked>
|
69 |
+
%46 = arith.mulf %35, %45 : tensor<256xf32, #blocked>
|
70 |
+
%47 = arith.subf %44, %46 : tensor<256xf32, #blocked>
|
71 |
+
%48 = tt.broadcast %41 : (tensor<1xf32, #blocked>) -> tensor<256xf32, #blocked>
|
72 |
+
%49 = arith.mulf %48, %47 : tensor<256xf32, #blocked>
|
73 |
+
%50 = arith.addf %27, %49 : tensor<256xf32, #blocked>
|
74 |
+
%51 = tt.broadcast %40 : (tensor<1xi1, #blocked>) -> tensor<256xi1, #blocked>
|
75 |
+
%52 = arith.select %51, %cst_6, %50 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
76 |
+
%53 = arith.addi %24, %cst_4 : tensor<1xi64, #blocked>
|
77 |
+
%54 = arith.cmpi slt, %24, %cst_3 : tensor<1xi64, #blocked>
|
78 |
+
%55 = arith.select %54, %53, %24 : tensor<1xi1, #blocked>, tensor<1xi64, #blocked>
|
79 |
+
%56 = arith.muli %55, %cst_2 : tensor<1xi64, #blocked>
|
80 |
+
%57 = tt.broadcast %56 : (tensor<1xi64, #blocked>) -> tensor<256xi64, #blocked>
|
81 |
+
%58 = arith.extsi %1 : tensor<256xi32, #blocked> to tensor<256xi64, #blocked>
|
82 |
+
%59 = arith.addi %58, %57 : tensor<256xi64, #blocked>
|
83 |
+
%60 = tt.splat %arg7 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
84 |
+
%61 = tt.addptr %60, %59 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi64, #blocked>
|
85 |
+
%62 = "tt.atomic_rmw"(%61, %52, %2) <{atomic_rmw_op = 5 : i32, scope = 1 : i32, sem = 4 : i32}> : (tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xf32, #blocked>, tensor<256xi1, #blocked>) -> tensor<256xf32, #blocked>
|
86 |
+
tt.return
|
87 |
+
}
|
88 |
+
}
|
.triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.llir
ADDED
@@ -0,0 +1,980 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
; ModuleID = 'LLVMDialectModule'
|
2 |
+
source_filename = "LLVMDialectModule"
|
3 |
+
|
4 |
+
@assertFunc_1 = internal constant [25 x i8] c"_call_with_frames_removed"
|
5 |
+
@assertFile_1 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
|
6 |
+
@assertMessage_1 = internal constant [39 x i8] c"index out of bounds: 0 <= tmp13 < 50257"
|
7 |
+
@assertFunc_0 = internal constant [25 x i8] c"_call_with_frames_removed"
|
8 |
+
@assertFile_0 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
|
9 |
+
@assertMessage_0 = internal constant [38 x i8] c"index out of bounds: 0 <= tmp3 < 50257"
|
10 |
+
@global_smem = external local_unnamed_addr addrspace(3) global [0 x i8]
|
11 |
+
@.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
|
12 |
+
|
13 |
+
declare void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr
|
14 |
+
|
15 |
+
define void @triton__0d1d2d3d4d5de6de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, i32 %5, i32 %6) local_unnamed_addr !dbg !7 {
|
16 |
+
%8 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
|
17 |
+
%9 = lshr i32 %8, 5, !dbg !10
|
18 |
+
%10 = and i32 %9, 7, !dbg !10
|
19 |
+
%11 = and i32 %8, 15, !dbg !10
|
20 |
+
%12 = shl i32 %8, 3, !dbg !11
|
21 |
+
%13 = and i32 %12, 248, !dbg !11
|
22 |
+
%14 = or i32 %13, 4, !dbg !11
|
23 |
+
%urem = and i32 %8, 255, !dbg !11
|
24 |
+
%15 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !12
|
25 |
+
%16 = shl i32 %15, 4, !dbg !13
|
26 |
+
%17 = or i32 %16, %10, !dbg !14
|
27 |
+
%18 = or i32 %17, 8, !dbg !14
|
28 |
+
%19 = or i32 %16, %11, !dbg !14
|
29 |
+
%20 = sext i32 %17 to i64, !dbg !15
|
30 |
+
%21 = getelementptr i64, ptr addrspace(1) %0, i64 %20, !dbg !15
|
31 |
+
%22 = sext i32 %18 to i64, !dbg !15
|
32 |
+
%23 = getelementptr i64, ptr addrspace(1) %0, i64 %22, !dbg !15
|
33 |
+
%24 = sext i32 %19 to i64, !dbg !15
|
34 |
+
%25 = getelementptr i64, ptr addrspace(1) %0, i64 %24, !dbg !15
|
35 |
+
%26 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
36 |
+
%27 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
37 |
+
%28 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
38 |
+
%29 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
39 |
+
%30 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
40 |
+
%31 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
41 |
+
%32 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
42 |
+
%33 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
|
43 |
+
%34 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
44 |
+
%35 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
45 |
+
%36 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
46 |
+
%37 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
47 |
+
%38 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
48 |
+
%39 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
49 |
+
%40 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
50 |
+
%41 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
|
51 |
+
%42 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !16
|
52 |
+
%43 = srem i32 %17, 512, !dbg !17
|
53 |
+
%44 = srem i32 %18, 512, !dbg !17
|
54 |
+
%45 = shl nsw i32 %43, 8, !dbg !18
|
55 |
+
%46 = shl nsw i32 %44, 8, !dbg !18
|
56 |
+
%47 = or i32 %45, %13, !dbg !19
|
57 |
+
%48 = or i32 %45, %14, !dbg !19
|
58 |
+
%49 = or i32 %46, %13, !dbg !19
|
59 |
+
%50 = or i32 %46, %14, !dbg !19
|
60 |
+
%51 = sext i32 %47 to i64, !dbg !20
|
61 |
+
%52 = getelementptr float, ptr addrspace(1) %2, i64 %51, !dbg !20
|
62 |
+
%53 = sext i32 %48 to i64, !dbg !20
|
63 |
+
%54 = getelementptr float, ptr addrspace(1) %2, i64 %53, !dbg !20
|
64 |
+
%55 = sext i32 %49 to i64, !dbg !20
|
65 |
+
%56 = getelementptr float, ptr addrspace(1) %2, i64 %55, !dbg !20
|
66 |
+
%57 = sext i32 %50 to i64, !dbg !20
|
67 |
+
%58 = getelementptr float, ptr addrspace(1) %2, i64 %57, !dbg !20
|
68 |
+
%59 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %52, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
|
69 |
+
%60 = extractvalue { i32, i32, i32, i32 } %59, 0, !dbg !21
|
70 |
+
%61 = extractvalue { i32, i32, i32, i32 } %59, 1, !dbg !21
|
71 |
+
%62 = extractvalue { i32, i32, i32, i32 } %59, 2, !dbg !21
|
72 |
+
%63 = extractvalue { i32, i32, i32, i32 } %59, 3, !dbg !21
|
73 |
+
%64 = bitcast i32 %60 to float, !dbg !21
|
74 |
+
%65 = bitcast i32 %61 to float, !dbg !21
|
75 |
+
%66 = bitcast i32 %62 to float, !dbg !21
|
76 |
+
%67 = bitcast i32 %63 to float, !dbg !21
|
77 |
+
%68 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %54, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
|
78 |
+
%69 = extractvalue { i32, i32, i32, i32 } %68, 0, !dbg !21
|
79 |
+
%70 = extractvalue { i32, i32, i32, i32 } %68, 1, !dbg !21
|
80 |
+
%71 = extractvalue { i32, i32, i32, i32 } %68, 2, !dbg !21
|
81 |
+
%72 = extractvalue { i32, i32, i32, i32 } %68, 3, !dbg !21
|
82 |
+
%73 = bitcast i32 %69 to float, !dbg !21
|
83 |
+
%74 = bitcast i32 %70 to float, !dbg !21
|
84 |
+
%75 = bitcast i32 %71 to float, !dbg !21
|
85 |
+
%76 = bitcast i32 %72 to float, !dbg !21
|
86 |
+
%77 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %56, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
|
87 |
+
%78 = extractvalue { i32, i32, i32, i32 } %77, 0, !dbg !21
|
88 |
+
%79 = extractvalue { i32, i32, i32, i32 } %77, 1, !dbg !21
|
89 |
+
%80 = extractvalue { i32, i32, i32, i32 } %77, 2, !dbg !21
|
90 |
+
%81 = extractvalue { i32, i32, i32, i32 } %77, 3, !dbg !21
|
91 |
+
%82 = bitcast i32 %78 to float, !dbg !21
|
92 |
+
%83 = bitcast i32 %79 to float, !dbg !21
|
93 |
+
%84 = bitcast i32 %80 to float, !dbg !21
|
94 |
+
%85 = bitcast i32 %81 to float, !dbg !21
|
95 |
+
%86 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %58, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
|
96 |
+
%87 = extractvalue { i32, i32, i32, i32 } %86, 0, !dbg !21
|
97 |
+
%88 = extractvalue { i32, i32, i32, i32 } %86, 1, !dbg !21
|
98 |
+
%89 = extractvalue { i32, i32, i32, i32 } %86, 2, !dbg !21
|
99 |
+
%90 = extractvalue { i32, i32, i32, i32 } %86, 3, !dbg !21
|
100 |
+
%91 = bitcast i32 %87 to float, !dbg !21
|
101 |
+
%92 = bitcast i32 %88 to float, !dbg !21
|
102 |
+
%93 = bitcast i32 %89 to float, !dbg !21
|
103 |
+
%94 = bitcast i32 %90 to float, !dbg !21
|
104 |
+
%95 = add i64 %42, 50257, !dbg !22
|
105 |
+
%96 = icmp slt i64 %26, 0, !dbg !23
|
106 |
+
%97 = icmp slt i64 %34, 0, !dbg !23
|
107 |
+
%98 = icmp slt i64 %42, 0, !dbg !23
|
108 |
+
%99 = select i1 %98, i64 %95, i64 %42, !dbg !24
|
109 |
+
%100 = icmp ugt i64 %99, 50256, !dbg !25
|
110 |
+
br i1 %100, label %101, label %102, !dbg !26
|
111 |
+
|
112 |
+
101: ; preds = %7
|
113 |
+
tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 883, ptr nonnull @assertFunc_0, i64 1), !dbg !26
|
114 |
+
br label %102, !dbg !26
|
115 |
+
|
116 |
+
102: ; preds = %101, %7
|
117 |
+
%103 = shl i64 %26, 8, !dbg !27
|
118 |
+
%104 = add i64 %103, 12865792, !dbg !27
|
119 |
+
%105 = select i1 %96, i64 %104, i64 %103, !dbg !27
|
120 |
+
%106 = shl i64 %34, 8, !dbg !27
|
121 |
+
%107 = add i64 %106, 12865792, !dbg !27
|
122 |
+
%108 = select i1 %97, i64 %107, i64 %106, !dbg !27
|
123 |
+
%109 = zext nneg i32 %13 to i64
|
124 |
+
%110 = zext nneg i32 %14 to i64
|
125 |
+
%111 = or i64 %105, %109, !dbg !28
|
126 |
+
%112 = or i64 %105, %110, !dbg !28
|
127 |
+
%113 = or i64 %108, %109, !dbg !28
|
128 |
+
%114 = or i64 %108, %110, !dbg !28
|
129 |
+
%115 = getelementptr float, ptr addrspace(1) %1, i64 %111, !dbg !29
|
130 |
+
%116 = getelementptr float, ptr addrspace(1) %1, i64 %112, !dbg !29
|
131 |
+
%117 = getelementptr float, ptr addrspace(1) %1, i64 %113, !dbg !29
|
132 |
+
%118 = getelementptr float, ptr addrspace(1) %1, i64 %114, !dbg !29
|
133 |
+
%119 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %115, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
|
134 |
+
%120 = extractvalue { i32, i32, i32, i32 } %119, 0, !dbg !30
|
135 |
+
%121 = extractvalue { i32, i32, i32, i32 } %119, 1, !dbg !30
|
136 |
+
%122 = extractvalue { i32, i32, i32, i32 } %119, 2, !dbg !30
|
137 |
+
%123 = extractvalue { i32, i32, i32, i32 } %119, 3, !dbg !30
|
138 |
+
%124 = bitcast i32 %120 to float, !dbg !30
|
139 |
+
%125 = bitcast i32 %121 to float, !dbg !30
|
140 |
+
%126 = bitcast i32 %122 to float, !dbg !30
|
141 |
+
%127 = bitcast i32 %123 to float, !dbg !30
|
142 |
+
%128 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %116, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
|
143 |
+
%129 = extractvalue { i32, i32, i32, i32 } %128, 0, !dbg !30
|
144 |
+
%130 = extractvalue { i32, i32, i32, i32 } %128, 1, !dbg !30
|
145 |
+
%131 = extractvalue { i32, i32, i32, i32 } %128, 2, !dbg !30
|
146 |
+
%132 = extractvalue { i32, i32, i32, i32 } %128, 3, !dbg !30
|
147 |
+
%133 = bitcast i32 %129 to float, !dbg !30
|
148 |
+
%134 = bitcast i32 %130 to float, !dbg !30
|
149 |
+
%135 = bitcast i32 %131 to float, !dbg !30
|
150 |
+
%136 = bitcast i32 %132 to float, !dbg !30
|
151 |
+
%137 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %117, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
|
152 |
+
%138 = extractvalue { i32, i32, i32, i32 } %137, 0, !dbg !30
|
153 |
+
%139 = extractvalue { i32, i32, i32, i32 } %137, 1, !dbg !30
|
154 |
+
%140 = extractvalue { i32, i32, i32, i32 } %137, 2, !dbg !30
|
155 |
+
%141 = extractvalue { i32, i32, i32, i32 } %137, 3, !dbg !30
|
156 |
+
%142 = bitcast i32 %138 to float, !dbg !30
|
157 |
+
%143 = bitcast i32 %139 to float, !dbg !30
|
158 |
+
%144 = bitcast i32 %140 to float, !dbg !30
|
159 |
+
%145 = bitcast i32 %141 to float, !dbg !30
|
160 |
+
%146 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %118, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
|
161 |
+
%147 = extractvalue { i32, i32, i32, i32 } %146, 0, !dbg !30
|
162 |
+
%148 = extractvalue { i32, i32, i32, i32 } %146, 1, !dbg !30
|
163 |
+
%149 = extractvalue { i32, i32, i32, i32 } %146, 2, !dbg !30
|
164 |
+
%150 = extractvalue { i32, i32, i32, i32 } %146, 3, !dbg !30
|
165 |
+
%151 = bitcast i32 %147 to float, !dbg !30
|
166 |
+
%152 = bitcast i32 %148 to float, !dbg !30
|
167 |
+
%153 = bitcast i32 %149 to float, !dbg !30
|
168 |
+
%154 = bitcast i32 %150 to float, !dbg !30
|
169 |
+
%155 = fadd float %64, %124, !dbg !31
|
170 |
+
%156 = fadd float %65, %125, !dbg !31
|
171 |
+
%157 = fadd float %66, %126, !dbg !31
|
172 |
+
%158 = fadd float %67, %127, !dbg !31
|
173 |
+
%159 = fadd float %73, %133, !dbg !31
|
174 |
+
%160 = fadd float %74, %134, !dbg !31
|
175 |
+
%161 = fadd float %75, %135, !dbg !31
|
176 |
+
%162 = fadd float %76, %136, !dbg !31
|
177 |
+
%163 = fadd float %82, %142, !dbg !31
|
178 |
+
%164 = fadd float %83, %143, !dbg !31
|
179 |
+
%165 = fadd float %84, %144, !dbg !31
|
180 |
+
%166 = fadd float %85, %145, !dbg !31
|
181 |
+
%167 = fadd float %91, %151, !dbg !31
|
182 |
+
%168 = fadd float %92, %152, !dbg !31
|
183 |
+
%169 = fadd float %93, %153, !dbg !31
|
184 |
+
%170 = fadd float %94, %154, !dbg !31
|
185 |
+
%171 = fadd float %155, 0.000000e+00, !dbg !32
|
186 |
+
%172 = fadd float %156, 0.000000e+00, !dbg !32
|
187 |
+
%173 = fadd float %157, 0.000000e+00, !dbg !32
|
188 |
+
%174 = fadd float %158, 0.000000e+00, !dbg !32
|
189 |
+
%175 = fadd float %159, 0.000000e+00, !dbg !32
|
190 |
+
%176 = fadd float %160, 0.000000e+00, !dbg !32
|
191 |
+
%177 = fadd float %161, 0.000000e+00, !dbg !32
|
192 |
+
%178 = fadd float %162, 0.000000e+00, !dbg !32
|
193 |
+
%179 = fadd float %163, 0.000000e+00, !dbg !32
|
194 |
+
%180 = fadd float %164, 0.000000e+00, !dbg !32
|
195 |
+
%181 = fadd float %165, 0.000000e+00, !dbg !32
|
196 |
+
%182 = fadd float %166, 0.000000e+00, !dbg !32
|
197 |
+
%183 = fadd float %167, 0.000000e+00, !dbg !32
|
198 |
+
%184 = fadd float %168, 0.000000e+00, !dbg !32
|
199 |
+
%185 = fadd float %169, 0.000000e+00, !dbg !32
|
200 |
+
%186 = fadd float %170, 0.000000e+00, !dbg !32
|
201 |
+
%187 = fsub float %155, %171, !dbg !36
|
202 |
+
%188 = fsub float %156, %172, !dbg !36
|
203 |
+
%189 = fsub float %157, %173, !dbg !36
|
204 |
+
%190 = fsub float %158, %174, !dbg !36
|
205 |
+
%191 = fsub float %159, %175, !dbg !36
|
206 |
+
%192 = fsub float %160, %176, !dbg !36
|
207 |
+
%193 = fsub float %161, %177, !dbg !36
|
208 |
+
%194 = fsub float %162, %178, !dbg !36
|
209 |
+
%195 = fsub float %163, %179, !dbg !36
|
210 |
+
%196 = fsub float %164, %180, !dbg !36
|
211 |
+
%197 = fsub float %165, %181, !dbg !36
|
212 |
+
%198 = fsub float %166, %182, !dbg !36
|
213 |
+
%199 = fsub float %167, %183, !dbg !36
|
214 |
+
%200 = fsub float %168, %184, !dbg !36
|
215 |
+
%201 = fsub float %169, %185, !dbg !36
|
216 |
+
%202 = fsub float %170, %186, !dbg !36
|
217 |
+
%203 = fmul float %155, %187, !dbg !37
|
218 |
+
%204 = fmul float %156, %188, !dbg !37
|
219 |
+
%205 = fmul float %157, %189, !dbg !37
|
220 |
+
%206 = fmul float %158, %190, !dbg !37
|
221 |
+
%207 = fmul float %159, %191, !dbg !37
|
222 |
+
%208 = fmul float %160, %192, !dbg !37
|
223 |
+
%209 = fmul float %161, %193, !dbg !37
|
224 |
+
%210 = fmul float %162, %194, !dbg !37
|
225 |
+
%211 = fmul float %163, %195, !dbg !37
|
226 |
+
%212 = fmul float %164, %196, !dbg !37
|
227 |
+
%213 = fmul float %165, %197, !dbg !37
|
228 |
+
%214 = fmul float %166, %198, !dbg !37
|
229 |
+
%215 = fmul float %167, %199, !dbg !37
|
230 |
+
%216 = fmul float %168, %200, !dbg !37
|
231 |
+
%217 = fmul float %169, %201, !dbg !37
|
232 |
+
%218 = fmul float %170, %202, !dbg !37
|
233 |
+
%219 = fadd float %203, 0.000000e+00, !dbg !38
|
234 |
+
%220 = fadd float %204, 0.000000e+00, !dbg !38
|
235 |
+
%221 = fadd float %205, 0.000000e+00, !dbg !38
|
236 |
+
%222 = fadd float %206, 0.000000e+00, !dbg !38
|
237 |
+
%223 = fadd float %207, 0.000000e+00, !dbg !38
|
238 |
+
%224 = fadd float %208, 0.000000e+00, !dbg !38
|
239 |
+
%225 = fadd float %209, 0.000000e+00, !dbg !38
|
240 |
+
%226 = fadd float %210, 0.000000e+00, !dbg !38
|
241 |
+
%227 = fadd float %211, 0.000000e+00, !dbg !38
|
242 |
+
%228 = fadd float %212, 0.000000e+00, !dbg !38
|
243 |
+
%229 = fadd float %213, 0.000000e+00, !dbg !38
|
244 |
+
%230 = fadd float %214, 0.000000e+00, !dbg !38
|
245 |
+
%231 = fadd float %215, 0.000000e+00, !dbg !38
|
246 |
+
%232 = fadd float %216, 0.000000e+00, !dbg !38
|
247 |
+
%233 = fadd float %217, 0.000000e+00, !dbg !38
|
248 |
+
%234 = fadd float %218, 0.000000e+00, !dbg !38
|
249 |
+
%235 = fsub float %172, %171, !dbg !39
|
250 |
+
%236 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 2.000000e+00) #6, !dbg !43
|
251 |
+
%237 = fmul float %236, %235, !dbg !44
|
252 |
+
%238 = fadd float %171, %237, !dbg !45
|
253 |
+
%239 = fadd float %219, %220, !dbg !46
|
254 |
+
%240 = fmul float %235, %235, !dbg !47
|
255 |
+
%241 = fmul float %236, %240, !dbg !48
|
256 |
+
%242 = fadd float %241, %239, !dbg !49
|
257 |
+
%243 = fsub float %173, %238, !dbg !39
|
258 |
+
%244 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 3.000000e+00) #6, !dbg !43
|
259 |
+
%245 = fmul float %244, %243, !dbg !44
|
260 |
+
%246 = fadd float %238, %245, !dbg !45
|
261 |
+
%247 = fadd float %221, %242, !dbg !46
|
262 |
+
%248 = fmul float %243, %243, !dbg !47
|
263 |
+
%249 = fmul float %248, 2.000000e+00, !dbg !50
|
264 |
+
%250 = fmul float %244, %249, !dbg !48
|
265 |
+
%251 = fadd float %247, %250, !dbg !49
|
266 |
+
%252 = fsub float %174, %246, !dbg !39
|
267 |
+
%253 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 4.000000e+00) #6, !dbg !43
|
268 |
+
%254 = fmul float %253, %252, !dbg !44
|
269 |
+
%255 = fadd float %246, %254, !dbg !45
|
270 |
+
%256 = fadd float %222, %251, !dbg !46
|
271 |
+
%257 = fmul float %252, %252, !dbg !47
|
272 |
+
%258 = fmul float %257, 3.000000e+00, !dbg !50
|
273 |
+
%259 = fmul float %253, %258, !dbg !48
|
274 |
+
%260 = fadd float %256, %259, !dbg !49
|
275 |
+
%261 = fsub float %175, %255, !dbg !39
|
276 |
+
%262 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 5.000000e+00) #6, !dbg !43
|
277 |
+
%263 = fmul float %262, %261, !dbg !44
|
278 |
+
%264 = fadd float %255, %263, !dbg !45
|
279 |
+
%265 = fadd float %223, %260, !dbg !46
|
280 |
+
%266 = fmul float %261, %261, !dbg !47
|
281 |
+
%267 = fmul float %266, 4.000000e+00, !dbg !50
|
282 |
+
%268 = fmul float %262, %267, !dbg !48
|
283 |
+
%269 = fadd float %265, %268, !dbg !49
|
284 |
+
%270 = fsub float %176, %264, !dbg !39
|
285 |
+
%271 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 6.000000e+00) #6, !dbg !43
|
286 |
+
%272 = fmul float %271, %270, !dbg !44
|
287 |
+
%273 = fadd float %264, %272, !dbg !45
|
288 |
+
%274 = fadd float %224, %269, !dbg !46
|
289 |
+
%275 = fmul float %270, %270, !dbg !47
|
290 |
+
%276 = fmul float %275, 5.000000e+00, !dbg !50
|
291 |
+
%277 = fmul float %271, %276, !dbg !48
|
292 |
+
%278 = fadd float %274, %277, !dbg !49
|
293 |
+
%279 = fsub float %177, %273, !dbg !39
|
294 |
+
%280 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 7.000000e+00) #6, !dbg !43
|
295 |
+
%281 = fmul float %280, %279, !dbg !44
|
296 |
+
%282 = fadd float %273, %281, !dbg !45
|
297 |
+
%283 = fadd float %225, %278, !dbg !46
|
298 |
+
%284 = fmul float %279, %279, !dbg !47
|
299 |
+
%285 = fmul float %284, 6.000000e+00, !dbg !50
|
300 |
+
%286 = fmul float %280, %285, !dbg !48
|
301 |
+
%287 = fadd float %283, %286, !dbg !49
|
302 |
+
%288 = fsub float %178, %282, !dbg !39
|
303 |
+
%289 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 8.000000e+00) #6, !dbg !43
|
304 |
+
%290 = fmul float %289, %288, !dbg !44
|
305 |
+
%291 = fadd float %282, %290, !dbg !45
|
306 |
+
%292 = fadd float %226, %287, !dbg !46
|
307 |
+
%293 = fmul float %288, %288, !dbg !47
|
308 |
+
%294 = fmul float %293, 7.000000e+00, !dbg !50
|
309 |
+
%295 = fmul float %289, %294, !dbg !48
|
310 |
+
%296 = fadd float %292, %295, !dbg !49
|
311 |
+
%297 = fsub float %180, %179, !dbg !39
|
312 |
+
%298 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 2.000000e+00) #6, !dbg !43
|
313 |
+
%299 = fmul float %297, %298, !dbg !44
|
314 |
+
%300 = fadd float %179, %299, !dbg !45
|
315 |
+
%301 = fadd float %227, %228, !dbg !46
|
316 |
+
%302 = fmul float %297, %297, !dbg !47
|
317 |
+
%303 = fmul float %302, %298, !dbg !48
|
318 |
+
%304 = fadd float %301, %303, !dbg !49
|
319 |
+
%305 = fsub float %181, %300, !dbg !39
|
320 |
+
%306 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 3.000000e+00) #6, !dbg !43
|
321 |
+
%307 = fmul float %306, %305, !dbg !44
|
322 |
+
%308 = fadd float %300, %307, !dbg !45
|
323 |
+
%309 = fadd float %229, %304, !dbg !46
|
324 |
+
%310 = fmul float %305, %305, !dbg !47
|
325 |
+
%311 = fmul float %310, 2.000000e+00, !dbg !50
|
326 |
+
%312 = fmul float %306, %311, !dbg !48
|
327 |
+
%313 = fadd float %309, %312, !dbg !49
|
328 |
+
%314 = fsub float %182, %308, !dbg !39
|
329 |
+
%315 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 4.000000e+00) #6, !dbg !43
|
330 |
+
%316 = fmul float %315, %314, !dbg !44
|
331 |
+
%317 = fadd float %308, %316, !dbg !45
|
332 |
+
%318 = fadd float %230, %313, !dbg !46
|
333 |
+
%319 = fmul float %314, %314, !dbg !47
|
334 |
+
%320 = fmul float %319, 3.000000e+00, !dbg !50
|
335 |
+
%321 = fmul float %315, %320, !dbg !48
|
336 |
+
%322 = fadd float %318, %321, !dbg !49
|
337 |
+
%323 = fsub float %183, %317, !dbg !39
|
338 |
+
%324 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 5.000000e+00) #6, !dbg !43
|
339 |
+
%325 = fmul float %324, %323, !dbg !44
|
340 |
+
%326 = fadd float %317, %325, !dbg !45
|
341 |
+
%327 = fadd float %231, %322, !dbg !46
|
342 |
+
%328 = fmul float %323, %323, !dbg !47
|
343 |
+
%329 = fmul float %328, 4.000000e+00, !dbg !50
|
344 |
+
%330 = fmul float %324, %329, !dbg !48
|
345 |
+
%331 = fadd float %327, %330, !dbg !49
|
346 |
+
%332 = fsub float %184, %326, !dbg !39
|
347 |
+
%333 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 6.000000e+00) #6, !dbg !43
|
348 |
+
%334 = fmul float %333, %332, !dbg !44
|
349 |
+
%335 = fadd float %326, %334, !dbg !45
|
350 |
+
%336 = fadd float %232, %331, !dbg !46
|
351 |
+
%337 = fmul float %332, %332, !dbg !47
|
352 |
+
%338 = fmul float %337, 5.000000e+00, !dbg !50
|
353 |
+
%339 = fmul float %333, %338, !dbg !48
|
354 |
+
%340 = fadd float %336, %339, !dbg !49
|
355 |
+
%341 = fsub float %185, %335, !dbg !39
|
356 |
+
%342 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 7.000000e+00) #6, !dbg !43
|
357 |
+
%343 = fmul float %342, %341, !dbg !44
|
358 |
+
%344 = fadd float %335, %343, !dbg !45
|
359 |
+
%345 = fadd float %233, %340, !dbg !46
|
360 |
+
%346 = fmul float %341, %341, !dbg !47
|
361 |
+
%347 = fmul float %346, 6.000000e+00, !dbg !50
|
362 |
+
%348 = fmul float %342, %347, !dbg !48
|
363 |
+
%349 = fadd float %345, %348, !dbg !49
|
364 |
+
%350 = fsub float %186, %344, !dbg !39
|
365 |
+
%351 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 8.000000e+00) #6, !dbg !43
|
366 |
+
%352 = fmul float %351, %350, !dbg !44
|
367 |
+
%353 = fadd float %344, %352, !dbg !45
|
368 |
+
%354 = fadd float %234, %349, !dbg !46
|
369 |
+
%355 = fmul float %350, %350, !dbg !47
|
370 |
+
%356 = fmul float %355, 7.000000e+00, !dbg !50
|
371 |
+
%357 = fmul float %351, %356, !dbg !48
|
372 |
+
%358 = fadd float %354, %357, !dbg !49
|
373 |
+
%359 = bitcast float %291 to i32, !dbg !51
|
374 |
+
%360 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %359, i32 16, i32 31), !dbg !51
|
375 |
+
%361 = bitcast i32 %360 to float, !dbg !51
|
376 |
+
%362 = bitcast float %296 to i32, !dbg !51
|
377 |
+
%363 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %362, i32 16, i32 31), !dbg !51
|
378 |
+
%364 = bitcast i32 %363 to float, !dbg !51
|
379 |
+
%365 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 1090519040, i32 16, i32 31), !dbg !51
|
380 |
+
%366 = bitcast i32 %365 to float, !dbg !51
|
381 |
+
%367 = fsub float %361, %291, !dbg !39
|
382 |
+
%368 = fadd float %366, 8.000000e+00, !dbg !53
|
383 |
+
%369 = fcmp oeq float %368, 0.000000e+00, !dbg !54
|
384 |
+
%370 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %366, float %368) #6, !dbg !43
|
385 |
+
%371 = select i1 %369, float 0.000000e+00, float %370, !dbg !55
|
386 |
+
%372 = fmul float %371, %367, !dbg !44
|
387 |
+
%373 = fadd float %291, %372, !dbg !45
|
388 |
+
%374 = fadd float %296, %364, !dbg !46
|
389 |
+
%375 = fmul float %367, %367, !dbg !47
|
390 |
+
%376 = fmul float %375, 8.000000e+00, !dbg !50
|
391 |
+
%377 = fmul float %371, %376, !dbg !48
|
392 |
+
%378 = fadd float %374, %377, !dbg !49
|
393 |
+
%379 = bitcast float %373 to i32, !dbg !51
|
394 |
+
%380 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %379, i32 8, i32 31), !dbg !51
|
395 |
+
%381 = bitcast i32 %380 to float, !dbg !51
|
396 |
+
%382 = bitcast float %378 to i32, !dbg !51
|
397 |
+
%383 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %382, i32 8, i32 31), !dbg !51
|
398 |
+
%384 = bitcast i32 %383 to float, !dbg !51
|
399 |
+
%385 = bitcast float %368 to i32, !dbg !51
|
400 |
+
%386 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %385, i32 8, i32 31), !dbg !51
|
401 |
+
%387 = bitcast i32 %386 to float, !dbg !51
|
402 |
+
%388 = fsub float %381, %373, !dbg !39
|
403 |
+
%389 = fadd float %368, %387, !dbg !53
|
404 |
+
%390 = fcmp oeq float %389, 0.000000e+00, !dbg !54
|
405 |
+
%391 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %387, float %389) #6, !dbg !43
|
406 |
+
%392 = select i1 %390, float 0.000000e+00, float %391, !dbg !55
|
407 |
+
%393 = fmul float %392, %388, !dbg !44
|
408 |
+
%394 = fadd float %373, %393, !dbg !45
|
409 |
+
%395 = fadd float %378, %384, !dbg !46
|
410 |
+
%396 = fmul float %388, %388, !dbg !47
|
411 |
+
%397 = fmul float %368, %396, !dbg !50
|
412 |
+
%398 = fmul float %392, %397, !dbg !48
|
413 |
+
%399 = fadd float %395, %398, !dbg !49
|
414 |
+
%400 = bitcast float %394 to i32, !dbg !51
|
415 |
+
%401 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %400, i32 4, i32 31), !dbg !51
|
416 |
+
%402 = bitcast i32 %401 to float, !dbg !51
|
417 |
+
%403 = bitcast float %399 to i32, !dbg !51
|
418 |
+
%404 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %403, i32 4, i32 31), !dbg !51
|
419 |
+
%405 = bitcast i32 %404 to float, !dbg !51
|
420 |
+
%406 = bitcast float %389 to i32, !dbg !51
|
421 |
+
%407 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %406, i32 4, i32 31), !dbg !51
|
422 |
+
%408 = bitcast i32 %407 to float, !dbg !51
|
423 |
+
%409 = fsub float %402, %394, !dbg !39
|
424 |
+
%410 = fadd float %389, %408, !dbg !53
|
425 |
+
%411 = fcmp oeq float %410, 0.000000e+00, !dbg !54
|
426 |
+
%412 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %408, float %410) #6, !dbg !43
|
427 |
+
%413 = select i1 %411, float 0.000000e+00, float %412, !dbg !55
|
428 |
+
%414 = fmul float %409, %413, !dbg !44
|
429 |
+
%415 = fadd float %394, %414, !dbg !45
|
430 |
+
%416 = fadd float %399, %405, !dbg !46
|
431 |
+
%417 = fmul float %409, %409, !dbg !47
|
432 |
+
%418 = fmul float %389, %417, !dbg !50
|
433 |
+
%419 = fmul float %413, %418, !dbg !48
|
434 |
+
%420 = fadd float %416, %419, !dbg !49
|
435 |
+
%421 = bitcast float %415 to i32, !dbg !51
|
436 |
+
%422 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %421, i32 2, i32 31), !dbg !51
|
437 |
+
%423 = bitcast i32 %422 to float, !dbg !51
|
438 |
+
%424 = bitcast float %420 to i32, !dbg !51
|
439 |
+
%425 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %424, i32 2, i32 31), !dbg !51
|
440 |
+
%426 = bitcast i32 %425 to float, !dbg !51
|
441 |
+
%427 = bitcast float %410 to i32, !dbg !51
|
442 |
+
%428 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %427, i32 2, i32 31), !dbg !51
|
443 |
+
%429 = bitcast i32 %428 to float, !dbg !51
|
444 |
+
%430 = fsub float %423, %415, !dbg !39
|
445 |
+
%431 = fadd float %410, %429, !dbg !53
|
446 |
+
%432 = fcmp oeq float %431, 0.000000e+00, !dbg !54
|
447 |
+
%433 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %429, float %431) #6, !dbg !43
|
448 |
+
%434 = select i1 %432, float 0.000000e+00, float %433, !dbg !55
|
449 |
+
%435 = fmul float %430, %434, !dbg !44
|
450 |
+
%436 = fadd float %415, %435, !dbg !45
|
451 |
+
%437 = fadd float %420, %426, !dbg !46
|
452 |
+
%438 = fmul float %430, %430, !dbg !47
|
453 |
+
%439 = fmul float %410, %438, !dbg !50
|
454 |
+
%440 = fmul float %434, %439, !dbg !48
|
455 |
+
%441 = fadd float %437, %440, !dbg !49
|
456 |
+
%442 = bitcast float %436 to i32, !dbg !51
|
457 |
+
%443 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %442, i32 1, i32 31), !dbg !51
|
458 |
+
%444 = bitcast float %441 to i32, !dbg !51
|
459 |
+
%445 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %444, i32 1, i32 31), !dbg !51
|
460 |
+
%446 = bitcast float %431 to i32, !dbg !51
|
461 |
+
%447 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %446, i32 1, i32 31), !dbg !51
|
462 |
+
%448 = bitcast i32 %447 to float, !dbg !51
|
463 |
+
%449 = fadd float %431, %448, !dbg !53
|
464 |
+
%450 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %448, float %449) #6, !dbg !43
|
465 |
+
%451 = bitcast float %353 to i32, !dbg !51
|
466 |
+
%452 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %451, i32 16, i32 31), !dbg !51
|
467 |
+
%453 = bitcast i32 %452 to float, !dbg !51
|
468 |
+
%454 = bitcast float %358 to i32, !dbg !51
|
469 |
+
%455 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %454, i32 16, i32 31), !dbg !51
|
470 |
+
%456 = bitcast i32 %455 to float, !dbg !51
|
471 |
+
%457 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 1090519040, i32 16, i32 31), !dbg !51
|
472 |
+
%458 = bitcast i32 %457 to float, !dbg !51
|
473 |
+
%459 = fsub float %453, %353, !dbg !39
|
474 |
+
%460 = fadd float %458, 8.000000e+00, !dbg !53
|
475 |
+
%461 = fcmp oeq float %460, 0.000000e+00, !dbg !54
|
476 |
+
%462 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %458, float %460) #6, !dbg !43
|
477 |
+
%463 = select i1 %461, float 0.000000e+00, float %462, !dbg !55
|
478 |
+
%464 = fmul float %459, %463, !dbg !44
|
479 |
+
%465 = fadd float %353, %464, !dbg !45
|
480 |
+
%466 = fadd float %358, %456, !dbg !46
|
481 |
+
%467 = fmul float %459, %459, !dbg !47
|
482 |
+
%468 = fmul float %467, 8.000000e+00, !dbg !50
|
483 |
+
%469 = fmul float %468, %463, !dbg !48
|
484 |
+
%470 = fadd float %466, %469, !dbg !49
|
485 |
+
%471 = bitcast float %465 to i32, !dbg !51
|
486 |
+
%472 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %471, i32 8, i32 31), !dbg !51
|
487 |
+
%473 = bitcast i32 %472 to float, !dbg !51
|
488 |
+
%474 = bitcast float %470 to i32, !dbg !51
|
489 |
+
%475 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %474, i32 8, i32 31), !dbg !51
|
490 |
+
%476 = bitcast i32 %475 to float, !dbg !51
|
491 |
+
%477 = bitcast float %460 to i32, !dbg !51
|
492 |
+
%478 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %477, i32 8, i32 31), !dbg !51
|
493 |
+
%479 = bitcast i32 %478 to float, !dbg !51
|
494 |
+
%480 = fsub float %473, %465, !dbg !39
|
495 |
+
%481 = fadd float %460, %479, !dbg !53
|
496 |
+
%482 = fcmp oeq float %481, 0.000000e+00, !dbg !54
|
497 |
+
%483 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %479, float %481) #6, !dbg !43
|
498 |
+
%484 = select i1 %482, float 0.000000e+00, float %483, !dbg !55
|
499 |
+
%485 = fmul float %480, %484, !dbg !44
|
500 |
+
%486 = fadd float %465, %485, !dbg !45
|
501 |
+
%487 = fadd float %470, %476, !dbg !46
|
502 |
+
%488 = fmul float %480, %480, !dbg !47
|
503 |
+
%489 = fmul float %460, %488, !dbg !50
|
504 |
+
%490 = fmul float %484, %489, !dbg !48
|
505 |
+
%491 = fadd float %487, %490, !dbg !49
|
506 |
+
%492 = bitcast float %486 to i32, !dbg !51
|
507 |
+
%493 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %492, i32 4, i32 31), !dbg !51
|
508 |
+
%494 = bitcast i32 %493 to float, !dbg !51
|
509 |
+
%495 = bitcast float %491 to i32, !dbg !51
|
510 |
+
%496 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %495, i32 4, i32 31), !dbg !51
|
511 |
+
%497 = bitcast i32 %496 to float, !dbg !51
|
512 |
+
%498 = bitcast float %481 to i32, !dbg !51
|
513 |
+
%499 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %498, i32 4, i32 31), !dbg !51
|
514 |
+
%500 = bitcast i32 %499 to float, !dbg !51
|
515 |
+
%501 = fsub float %494, %486, !dbg !39
|
516 |
+
%502 = fadd float %481, %500, !dbg !53
|
517 |
+
%503 = fcmp oeq float %502, 0.000000e+00, !dbg !54
|
518 |
+
%504 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %500, float %502) #6, !dbg !43
|
519 |
+
%505 = select i1 %503, float 0.000000e+00, float %504, !dbg !55
|
520 |
+
%506 = fmul float %501, %505, !dbg !44
|
521 |
+
%507 = fadd float %486, %506, !dbg !45
|
522 |
+
%508 = fadd float %491, %497, !dbg !46
|
523 |
+
%509 = fmul float %501, %501, !dbg !47
|
524 |
+
%510 = fmul float %481, %509, !dbg !50
|
525 |
+
%511 = fmul float %505, %510, !dbg !48
|
526 |
+
%512 = fadd float %508, %511, !dbg !49
|
527 |
+
%513 = bitcast float %507 to i32, !dbg !51
|
528 |
+
%514 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %513, i32 2, i32 31), !dbg !51
|
529 |
+
%515 = bitcast i32 %514 to float, !dbg !51
|
530 |
+
%516 = bitcast float %512 to i32, !dbg !51
|
531 |
+
%517 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %516, i32 2, i32 31), !dbg !51
|
532 |
+
%518 = bitcast i32 %517 to float, !dbg !51
|
533 |
+
%519 = bitcast float %502 to i32, !dbg !51
|
534 |
+
%520 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %519, i32 2, i32 31), !dbg !51
|
535 |
+
%521 = bitcast i32 %520 to float, !dbg !51
|
536 |
+
%522 = fsub float %515, %507, !dbg !39
|
537 |
+
%523 = fadd float %502, %521, !dbg !53
|
538 |
+
%524 = fcmp oeq float %523, 0.000000e+00, !dbg !54
|
539 |
+
%525 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %521, float %523) #6, !dbg !43
|
540 |
+
%526 = select i1 %524, float 0.000000e+00, float %525, !dbg !55
|
541 |
+
%527 = fmul float %522, %526, !dbg !44
|
542 |
+
%528 = fadd float %507, %527, !dbg !45
|
543 |
+
%529 = fadd float %512, %518, !dbg !46
|
544 |
+
%530 = fmul float %522, %522, !dbg !47
|
545 |
+
%531 = fmul float %502, %530, !dbg !50
|
546 |
+
%532 = fmul float %526, %531, !dbg !48
|
547 |
+
%533 = fadd float %529, %532, !dbg !49
|
548 |
+
%534 = bitcast float %528 to i32, !dbg !51
|
549 |
+
%535 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %534, i32 1, i32 31), !dbg !51
|
550 |
+
%536 = bitcast float %533 to i32, !dbg !51
|
551 |
+
%537 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %536, i32 1, i32 31), !dbg !51
|
552 |
+
%538 = bitcast float %523 to i32, !dbg !51
|
553 |
+
%539 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %538, i32 1, i32 31), !dbg !51
|
554 |
+
%540 = bitcast i32 %539 to float, !dbg !51
|
555 |
+
%541 = fadd float %523, %540, !dbg !53
|
556 |
+
%542 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %540, float %541) #6, !dbg !43
|
557 |
+
%543 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %52, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
|
558 |
+
%544 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %54, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
|
559 |
+
%545 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %56, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
|
560 |
+
%546 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %58, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
|
561 |
+
%547 = zext nneg i32 %urem to i64, !dbg !57
|
562 |
+
%548 = getelementptr float, ptr addrspace(1) %3, i64 %547, !dbg !57
|
563 |
+
%549 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];\0A\09@!$4 mov.u32 $0, $3;", "=r,l,b,r,b"(ptr addrspace(1) %548, i1 true, i32 0, i1 true) #6, !dbg !58
|
564 |
+
br i1 %100, label %550, label %551, !dbg !59
|
565 |
+
|
566 |
+
550: ; preds = %102
|
567 |
+
tail call void @__assertfail(ptr nonnull @assertMessage_1, ptr nonnull @assertFile_1, i32 883, ptr nonnull @assertFunc_1, i64 1), !dbg !59
|
568 |
+
br label %551, !dbg !59
|
569 |
+
|
570 |
+
551: ; preds = %550, %102
|
571 |
+
%552 = bitcast i32 %537 to float, !dbg !51
|
572 |
+
%553 = fadd float %533, %552, !dbg !46
|
573 |
+
%554 = bitcast i32 %535 to float, !dbg !51
|
574 |
+
%555 = fsub float %554, %528, !dbg !39
|
575 |
+
%556 = fmul float %555, %555, !dbg !47
|
576 |
+
%557 = fmul float %523, %556, !dbg !50
|
577 |
+
%558 = fcmp oeq float %541, 0.000000e+00, !dbg !54
|
578 |
+
%559 = select i1 %558, float 0.000000e+00, float %542, !dbg !55
|
579 |
+
%560 = fmul float %559, %557, !dbg !48
|
580 |
+
%561 = fadd float %553, %560, !dbg !49
|
581 |
+
%562 = bitcast i32 %445 to float, !dbg !51
|
582 |
+
%563 = fadd float %441, %562, !dbg !46
|
583 |
+
%564 = bitcast i32 %443 to float, !dbg !51
|
584 |
+
%565 = fsub float %564, %436, !dbg !39
|
585 |
+
%566 = fmul float %565, %565, !dbg !47
|
586 |
+
%567 = fmul float %431, %566, !dbg !50
|
587 |
+
%568 = fcmp oeq float %449, 0.000000e+00, !dbg !54
|
588 |
+
%569 = select i1 %568, float 0.000000e+00, float %450, !dbg !55
|
589 |
+
%570 = fmul float %569, %567, !dbg !48
|
590 |
+
%571 = fadd float %563, %570, !dbg !49
|
591 |
+
%572 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %115, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
|
592 |
+
%573 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %116, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
|
593 |
+
%574 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %117, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
|
594 |
+
%575 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %118, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
|
595 |
+
%576 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
596 |
+
%577 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
597 |
+
%578 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
598 |
+
%579 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
599 |
+
%580 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
600 |
+
%581 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
601 |
+
%582 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
602 |
+
%583 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
|
603 |
+
%584 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
604 |
+
%585 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
605 |
+
%586 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
606 |
+
%587 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
607 |
+
%588 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
608 |
+
%589 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
609 |
+
%590 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
610 |
+
%591 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
|
611 |
+
%592 = fadd float %576, 0x3EE4F8B580000000, !dbg !62
|
612 |
+
%593 = fadd float %584, 0x3EE4F8B580000000, !dbg !62
|
613 |
+
%594 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
614 |
+
%.not.i = icmp eq i32 %594, 0, !dbg !63
|
615 |
+
br i1 %.not.i, label %597, label %595, !dbg !63
|
616 |
+
|
617 |
+
595: ; preds = %551
|
618 |
+
%596 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %592), !dbg !63
|
619 |
+
br label %__nv_rsqrtf.exit, !dbg !63
|
620 |
+
|
621 |
+
597: ; preds = %551
|
622 |
+
%598 = tail call float @llvm.nvvm.rsqrt.approx.f(float %592), !dbg !63
|
623 |
+
br label %__nv_rsqrtf.exit, !dbg !63
|
624 |
+
|
625 |
+
__nv_rsqrtf.exit: ; preds = %595, %597
|
626 |
+
%.0.i = phi float [ %596, %595 ], [ %598, %597 ], !dbg !63
|
627 |
+
%599 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
628 |
+
%600 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
629 |
+
%601 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
630 |
+
%602 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
631 |
+
%603 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
632 |
+
%604 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
633 |
+
%605 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
634 |
+
%606 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
635 |
+
%.not.i22 = icmp eq i32 %606, 0, !dbg !63
|
636 |
+
br i1 %.not.i22, label %609, label %607, !dbg !63
|
637 |
+
|
638 |
+
607: ; preds = %__nv_rsqrtf.exit
|
639 |
+
%608 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %593), !dbg !63
|
640 |
+
br label %__nv_rsqrtf.exit24, !dbg !63
|
641 |
+
|
642 |
+
609: ; preds = %__nv_rsqrtf.exit
|
643 |
+
%610 = tail call float @llvm.nvvm.rsqrt.approx.f(float %593), !dbg !63
|
644 |
+
br label %__nv_rsqrtf.exit24, !dbg !63
|
645 |
+
|
646 |
+
__nv_rsqrtf.exit24: ; preds = %607, %609
|
647 |
+
%.0.i23 = phi float [ %608, %607 ], [ %610, %609 ], !dbg !63
|
648 |
+
%611 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
649 |
+
%612 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
650 |
+
%613 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
651 |
+
%614 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
652 |
+
%615 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
653 |
+
%616 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
654 |
+
%617 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
|
655 |
+
%618 = extractvalue { i32, i32, i32, i32 } %575, 3, !dbg !60
|
656 |
+
%619 = bitcast i32 %618 to float, !dbg !60
|
657 |
+
%620 = extractvalue { i32, i32, i32, i32 } %546, 3, !dbg !56
|
658 |
+
%621 = bitcast i32 %620 to float, !dbg !56
|
659 |
+
%622 = fadd float %621, %619, !dbg !64
|
660 |
+
%623 = fmul float %555, %559, !dbg !44
|
661 |
+
%624 = fadd float %528, %623, !dbg !45
|
662 |
+
%625 = fsub float %622, %624, !dbg !65
|
663 |
+
%626 = extractvalue { i32, i32, i32, i32 } %575, 2, !dbg !60
|
664 |
+
%627 = bitcast i32 %626 to float, !dbg !60
|
665 |
+
%628 = extractvalue { i32, i32, i32, i32 } %546, 2, !dbg !56
|
666 |
+
%629 = bitcast i32 %628 to float, !dbg !56
|
667 |
+
%630 = fadd float %629, %627, !dbg !64
|
668 |
+
%631 = fsub float %630, %624, !dbg !65
|
669 |
+
%632 = extractvalue { i32, i32, i32, i32 } %575, 1, !dbg !60
|
670 |
+
%633 = bitcast i32 %632 to float, !dbg !60
|
671 |
+
%634 = extractvalue { i32, i32, i32, i32 } %546, 1, !dbg !56
|
672 |
+
%635 = bitcast i32 %634 to float, !dbg !56
|
673 |
+
%636 = fadd float %635, %633, !dbg !64
|
674 |
+
%637 = fsub float %636, %624, !dbg !65
|
675 |
+
%638 = extractvalue { i32, i32, i32, i32 } %575, 0, !dbg !60
|
676 |
+
%639 = bitcast i32 %638 to float, !dbg !60
|
677 |
+
%640 = extractvalue { i32, i32, i32, i32 } %546, 0, !dbg !56
|
678 |
+
%641 = bitcast i32 %640 to float, !dbg !56
|
679 |
+
%642 = fadd float %641, %639, !dbg !64
|
680 |
+
%643 = fsub float %642, %624, !dbg !65
|
681 |
+
%644 = extractvalue { i32, i32, i32, i32 } %574, 3, !dbg !60
|
682 |
+
%645 = bitcast i32 %644 to float, !dbg !60
|
683 |
+
%646 = extractvalue { i32, i32, i32, i32 } %545, 3, !dbg !56
|
684 |
+
%647 = bitcast i32 %646 to float, !dbg !56
|
685 |
+
%648 = fadd float %647, %645, !dbg !64
|
686 |
+
%649 = fsub float %648, %624, !dbg !65
|
687 |
+
%650 = extractvalue { i32, i32, i32, i32 } %574, 2, !dbg !60
|
688 |
+
%651 = bitcast i32 %650 to float, !dbg !60
|
689 |
+
%652 = extractvalue { i32, i32, i32, i32 } %545, 2, !dbg !56
|
690 |
+
%653 = bitcast i32 %652 to float, !dbg !56
|
691 |
+
%654 = fadd float %653, %651, !dbg !64
|
692 |
+
%655 = fsub float %654, %624, !dbg !65
|
693 |
+
%656 = extractvalue { i32, i32, i32, i32 } %574, 1, !dbg !60
|
694 |
+
%657 = bitcast i32 %656 to float, !dbg !60
|
695 |
+
%658 = extractvalue { i32, i32, i32, i32 } %545, 1, !dbg !56
|
696 |
+
%659 = bitcast i32 %658 to float, !dbg !56
|
697 |
+
%660 = fadd float %659, %657, !dbg !64
|
698 |
+
%661 = fsub float %660, %624, !dbg !65
|
699 |
+
%662 = extractvalue { i32, i32, i32, i32 } %574, 0, !dbg !60
|
700 |
+
%663 = bitcast i32 %662 to float, !dbg !60
|
701 |
+
%664 = extractvalue { i32, i32, i32, i32 } %545, 0, !dbg !56
|
702 |
+
%665 = bitcast i32 %664 to float, !dbg !56
|
703 |
+
%666 = fadd float %665, %663, !dbg !64
|
704 |
+
%667 = fsub float %666, %624, !dbg !65
|
705 |
+
%668 = extractvalue { i32, i32, i32, i32 } %573, 3, !dbg !60
|
706 |
+
%669 = bitcast i32 %668 to float, !dbg !60
|
707 |
+
%670 = extractvalue { i32, i32, i32, i32 } %544, 3, !dbg !56
|
708 |
+
%671 = bitcast i32 %670 to float, !dbg !56
|
709 |
+
%672 = fadd float %671, %669, !dbg !64
|
710 |
+
%673 = fmul float %565, %569, !dbg !44
|
711 |
+
%674 = fadd float %436, %673, !dbg !45
|
712 |
+
%675 = fsub float %672, %674, !dbg !65
|
713 |
+
%676 = extractvalue { i32, i32, i32, i32 } %573, 2, !dbg !60
|
714 |
+
%677 = bitcast i32 %676 to float, !dbg !60
|
715 |
+
%678 = extractvalue { i32, i32, i32, i32 } %544, 2, !dbg !56
|
716 |
+
%679 = bitcast i32 %678 to float, !dbg !56
|
717 |
+
%680 = fadd float %679, %677, !dbg !64
|
718 |
+
%681 = fsub float %680, %674, !dbg !65
|
719 |
+
%682 = extractvalue { i32, i32, i32, i32 } %573, 1, !dbg !60
|
720 |
+
%683 = bitcast i32 %682 to float, !dbg !60
|
721 |
+
%684 = extractvalue { i32, i32, i32, i32 } %544, 1, !dbg !56
|
722 |
+
%685 = bitcast i32 %684 to float, !dbg !56
|
723 |
+
%686 = fadd float %685, %683, !dbg !64
|
724 |
+
%687 = fsub float %686, %674, !dbg !65
|
725 |
+
%688 = extractvalue { i32, i32, i32, i32 } %573, 0, !dbg !60
|
726 |
+
%689 = bitcast i32 %688 to float, !dbg !60
|
727 |
+
%690 = extractvalue { i32, i32, i32, i32 } %544, 0, !dbg !56
|
728 |
+
%691 = bitcast i32 %690 to float, !dbg !56
|
729 |
+
%692 = fadd float %691, %689, !dbg !64
|
730 |
+
%693 = fsub float %692, %674, !dbg !65
|
731 |
+
%694 = extractvalue { i32, i32, i32, i32 } %572, 3, !dbg !60
|
732 |
+
%695 = bitcast i32 %694 to float, !dbg !60
|
733 |
+
%696 = extractvalue { i32, i32, i32, i32 } %543, 3, !dbg !56
|
734 |
+
%697 = bitcast i32 %696 to float, !dbg !56
|
735 |
+
%698 = fadd float %697, %695, !dbg !64
|
736 |
+
%699 = fsub float %698, %674, !dbg !65
|
737 |
+
%700 = extractvalue { i32, i32, i32, i32 } %572, 2, !dbg !60
|
738 |
+
%701 = bitcast i32 %700 to float, !dbg !60
|
739 |
+
%702 = extractvalue { i32, i32, i32, i32 } %543, 2, !dbg !56
|
740 |
+
%703 = bitcast i32 %702 to float, !dbg !56
|
741 |
+
%704 = fadd float %703, %701, !dbg !64
|
742 |
+
%705 = fsub float %704, %674, !dbg !65
|
743 |
+
%706 = extractvalue { i32, i32, i32, i32 } %572, 1, !dbg !60
|
744 |
+
%707 = bitcast i32 %706 to float, !dbg !60
|
745 |
+
%708 = extractvalue { i32, i32, i32, i32 } %543, 1, !dbg !56
|
746 |
+
%709 = bitcast i32 %708 to float, !dbg !56
|
747 |
+
%710 = fadd float %709, %707, !dbg !64
|
748 |
+
%711 = fsub float %710, %674, !dbg !65
|
749 |
+
%712 = extractvalue { i32, i32, i32, i32 } %572, 0, !dbg !60
|
750 |
+
%713 = bitcast i32 %712 to float, !dbg !60
|
751 |
+
%714 = extractvalue { i32, i32, i32, i32 } %543, 0, !dbg !56
|
752 |
+
%715 = bitcast i32 %714 to float, !dbg !56
|
753 |
+
%716 = fadd float %715, %713, !dbg !64
|
754 |
+
%717 = fsub float %716, %674, !dbg !65
|
755 |
+
%718 = fmul float %717, %.0.i, !dbg !66
|
756 |
+
%719 = fmul float %711, %.0.i, !dbg !66
|
757 |
+
%720 = fmul float %705, %.0.i, !dbg !66
|
758 |
+
%721 = fmul float %699, %.0.i, !dbg !66
|
759 |
+
%722 = fmul float %693, %.0.i, !dbg !66
|
760 |
+
%723 = fmul float %687, %.0.i, !dbg !66
|
761 |
+
%724 = fmul float %681, %.0.i, !dbg !66
|
762 |
+
%725 = fmul float %675, %.0.i, !dbg !66
|
763 |
+
%726 = fmul float %667, %.0.i23, !dbg !66
|
764 |
+
%727 = fmul float %661, %.0.i23, !dbg !66
|
765 |
+
%728 = fmul float %655, %.0.i23, !dbg !66
|
766 |
+
%729 = fmul float %649, %.0.i23, !dbg !66
|
767 |
+
%730 = fmul float %643, %.0.i23, !dbg !66
|
768 |
+
%731 = fmul float %637, %.0.i23, !dbg !66
|
769 |
+
%732 = fmul float %631, %.0.i23, !dbg !66
|
770 |
+
%733 = fmul float %625, %.0.i23, !dbg !66
|
771 |
+
%734 = getelementptr float, ptr addrspace(3) @global_smem, i64 %547, !dbg !67
|
772 |
+
store i32 %549, ptr addrspace(3) %734, align 4, !dbg !67
|
773 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !67
|
774 |
+
%735 = getelementptr float, ptr addrspace(3) @global_smem, i64 %109, !dbg !67
|
775 |
+
%736 = load float, ptr addrspace(3) %735, align 32, !dbg !67
|
776 |
+
%737 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 1, !dbg !67
|
777 |
+
%738 = load float, ptr addrspace(3) %737, align 4, !dbg !67
|
778 |
+
%739 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 2, !dbg !67
|
779 |
+
%740 = load float, ptr addrspace(3) %739, align 8, !dbg !67
|
780 |
+
%741 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 3, !dbg !67
|
781 |
+
%742 = load float, ptr addrspace(3) %741, align 4, !dbg !67
|
782 |
+
%743 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 4, !dbg !67
|
783 |
+
%744 = load float, ptr addrspace(3) %743, align 16, !dbg !67
|
784 |
+
%745 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 5, !dbg !67
|
785 |
+
%746 = load float, ptr addrspace(3) %745, align 4, !dbg !67
|
786 |
+
%747 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 6, !dbg !67
|
787 |
+
%748 = load float, ptr addrspace(3) %747, align 8, !dbg !67
|
788 |
+
%749 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 7, !dbg !67
|
789 |
+
%750 = load float, ptr addrspace(3) %749, align 4, !dbg !67
|
790 |
+
%751 = fmul float %718, %736, !dbg !67
|
791 |
+
%752 = fmul float %719, %738, !dbg !67
|
792 |
+
%753 = fmul float %720, %740, !dbg !67
|
793 |
+
%754 = fmul float %721, %742, !dbg !67
|
794 |
+
%755 = fmul float %722, %744, !dbg !67
|
795 |
+
%756 = fmul float %723, %746, !dbg !67
|
796 |
+
%757 = fmul float %724, %748, !dbg !67
|
797 |
+
%758 = fmul float %725, %750, !dbg !67
|
798 |
+
%759 = fmul float %726, %736, !dbg !67
|
799 |
+
%760 = fmul float %727, %738, !dbg !67
|
800 |
+
%761 = fmul float %728, %740, !dbg !67
|
801 |
+
%762 = fmul float %729, %742, !dbg !67
|
802 |
+
%763 = fmul float %730, %744, !dbg !67
|
803 |
+
%764 = fmul float %731, %746, !dbg !67
|
804 |
+
%765 = fmul float %732, %748, !dbg !67
|
805 |
+
%766 = fmul float %733, %750, !dbg !67
|
806 |
+
%767 = shl i32 %17, 8, !dbg !68
|
807 |
+
%768 = shl i32 %18, 8, !dbg !68
|
808 |
+
%769 = or i32 %767, %13, !dbg !69
|
809 |
+
%770 = or i32 %768, %13, !dbg !69
|
810 |
+
%771 = sext i32 %769 to i64, !dbg !70
|
811 |
+
%772 = getelementptr i16, ptr addrspace(1) %4, i64 %771, !dbg !70
|
812 |
+
%773 = sext i32 %770 to i64, !dbg !70
|
813 |
+
%774 = getelementptr i16, ptr addrspace(1) %4, i64 %773, !dbg !70
|
814 |
+
%775 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %751) #6, !dbg !71
|
815 |
+
%776 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %752) #6, !dbg !71
|
816 |
+
%777 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %753) #6, !dbg !71
|
817 |
+
%778 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %754) #6, !dbg !71
|
818 |
+
%779 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %755) #6, !dbg !71
|
819 |
+
%780 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %756) #6, !dbg !71
|
820 |
+
%781 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %757) #6, !dbg !71
|
821 |
+
%782 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %758) #6, !dbg !71
|
822 |
+
%783 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %759) #6, !dbg !71
|
823 |
+
%784 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %760) #6, !dbg !71
|
824 |
+
%785 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %761) #6, !dbg !71
|
825 |
+
%786 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %762) #6, !dbg !71
|
826 |
+
%787 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %763) #6, !dbg !71
|
827 |
+
%788 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %764) #6, !dbg !71
|
828 |
+
%789 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %765) #6, !dbg !71
|
829 |
+
%790 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %766) #6, !dbg !71
|
830 |
+
%791 = insertelement <2 x i16> undef, i16 %775, i64 0, !dbg !71
|
831 |
+
%792 = insertelement <2 x i16> %791, i16 %776, i64 1, !dbg !71
|
832 |
+
%793 = bitcast <2 x i16> %792 to i32, !dbg !71
|
833 |
+
%794 = insertelement <2 x i16> undef, i16 %777, i64 0, !dbg !71
|
834 |
+
%795 = insertelement <2 x i16> %794, i16 %778, i64 1, !dbg !71
|
835 |
+
%796 = bitcast <2 x i16> %795 to i32, !dbg !71
|
836 |
+
%797 = insertelement <2 x i16> undef, i16 %779, i64 0, !dbg !71
|
837 |
+
%798 = insertelement <2 x i16> %797, i16 %780, i64 1, !dbg !71
|
838 |
+
%799 = bitcast <2 x i16> %798 to i32, !dbg !71
|
839 |
+
%800 = insertelement <2 x i16> undef, i16 %781, i64 0, !dbg !71
|
840 |
+
%801 = insertelement <2 x i16> %800, i16 %782, i64 1, !dbg !71
|
841 |
+
%802 = bitcast <2 x i16> %801 to i32, !dbg !71
|
842 |
+
tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %793, i32 %796, i32 %799, i32 %802, ptr addrspace(1) %772, i1 true) #6, !dbg !71
|
843 |
+
%803 = insertelement <2 x i16> undef, i16 %783, i64 0, !dbg !71
|
844 |
+
%804 = insertelement <2 x i16> %803, i16 %784, i64 1, !dbg !71
|
845 |
+
%805 = bitcast <2 x i16> %804 to i32, !dbg !71
|
846 |
+
%806 = insertelement <2 x i16> undef, i16 %785, i64 0, !dbg !71
|
847 |
+
%807 = insertelement <2 x i16> %806, i16 %786, i64 1, !dbg !71
|
848 |
+
%808 = bitcast <2 x i16> %807 to i32, !dbg !71
|
849 |
+
%809 = insertelement <2 x i16> undef, i16 %787, i64 0, !dbg !71
|
850 |
+
%810 = insertelement <2 x i16> %809, i16 %788, i64 1, !dbg !71
|
851 |
+
%811 = bitcast <2 x i16> %810 to i32, !dbg !71
|
852 |
+
%812 = insertelement <2 x i16> undef, i16 %789, i64 0, !dbg !71
|
853 |
+
%813 = insertelement <2 x i16> %812, i16 %790, i64 1, !dbg !71
|
854 |
+
%814 = bitcast <2 x i16> %813 to i32, !dbg !71
|
855 |
+
tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %805, i32 %808, i32 %811, i32 %814, ptr addrspace(1) %774, i1 true) #6, !dbg !71
|
856 |
+
ret void, !dbg !72
|
857 |
+
}
|
858 |
+
|
859 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
|
860 |
+
declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
|
861 |
+
|
862 |
+
; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
|
863 |
+
declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
|
864 |
+
|
865 |
+
; Function Attrs: convergent nocallback nounwind
|
866 |
+
declare void @llvm.nvvm.barrier0() #2
|
867 |
+
|
868 |
+
; Function Attrs: alwaysinline nounwind
|
869 |
+
define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
|
870 |
+
%1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
|
871 |
+
%.not = icmp eq i32 %1, 0
|
872 |
+
br i1 %.not, label %4, label %2
|
873 |
+
|
874 |
+
2: ; preds = %0
|
875 |
+
%3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
|
876 |
+
br label %6
|
877 |
+
|
878 |
+
4: ; preds = %0
|
879 |
+
%5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
|
880 |
+
br label %6
|
881 |
+
|
882 |
+
6: ; preds = %4, %2
|
883 |
+
%.0 = phi float [ %3, %2 ], [ %5, %4 ]
|
884 |
+
ret float %.0
|
885 |
+
}
|
886 |
+
|
887 |
+
declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
|
888 |
+
|
889 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
890 |
+
declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
|
891 |
+
|
892 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
893 |
+
declare float @llvm.nvvm.rsqrt.approx.f(float) #5
|
894 |
+
|
895 |
+
attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
|
896 |
+
attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
|
897 |
+
attributes #2 = { convergent nocallback nounwind }
|
898 |
+
attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
899 |
+
attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
900 |
+
attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
|
901 |
+
attributes #6 = { nounwind }
|
902 |
+
|
903 |
+
!llvm.module.flags = !{!0, !1}
|
904 |
+
!llvm.dbg.cu = !{!2}
|
905 |
+
!nvvm.annotations = !{!4, !5, !5, !4}
|
906 |
+
!llvm.ident = !{!6}
|
907 |
+
|
908 |
+
!0 = !{i32 2, !"Debug Info Version", i32 3}
|
909 |
+
!1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
|
910 |
+
!2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
|
911 |
+
!3 = !DIFile(filename: "clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py", directory: "/tmp/torchinductor_root/lh")
|
912 |
+
!4 = !{ptr @triton__0d1d2d3d4d5de6de, !"kernel", i32 1}
|
913 |
+
!5 = !{ptr @triton__0d1d2d3d4d5de6de, !"maxntidx", i32 256}
|
914 |
+
!6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
|
915 |
+
!7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5de6de", linkageName: "triton__0d1d2d3d4d5de6de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
|
916 |
+
!8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
|
917 |
+
!9 = !{}
|
918 |
+
!10 = !DILocation(line: 22, column: 44, scope: !7)
|
919 |
+
!11 = !DILocation(line: 24, column: 33, scope: !7)
|
920 |
+
!12 = !DILocation(line: 21, column: 28, scope: !7)
|
921 |
+
!13 = !DILocation(line: 21, column: 33, scope: !7)
|
922 |
+
!14 = !DILocation(line: 22, column: 23, scope: !7)
|
923 |
+
!15 = !DILocation(line: 26, column: 30, scope: !7)
|
924 |
+
!16 = !DILocation(line: 26, column: 35, scope: !7)
|
925 |
+
!17 = !DILocation(line: 27, column: 18, scope: !7)
|
926 |
+
!18 = !DILocation(line: 35, column: 44, scope: !7)
|
927 |
+
!19 = !DILocation(line: 35, column: 40, scope: !7)
|
928 |
+
!20 = !DILocation(line: 35, column: 34, scope: !7)
|
929 |
+
!21 = !DILocation(line: 35, column: 50, scope: !7)
|
930 |
+
!22 = !DILocation(line: 36, column: 22, scope: !7)
|
931 |
+
!23 = !DILocation(line: 37, column: 22, scope: !7)
|
932 |
+
!24 = !DILocation(line: 38, column: 36, scope: !7)
|
933 |
+
!25 = !DILocation(line: 39, column: 40, scope: !7)
|
934 |
+
!26 = !DILocation(line: 39, column: 55, scope: !7)
|
935 |
+
!27 = !DILocation(line: 40, column: 44, scope: !7)
|
936 |
+
!28 = !DILocation(line: 40, column: 40, scope: !7)
|
937 |
+
!29 = !DILocation(line: 40, column: 34, scope: !7)
|
938 |
+
!30 = !DILocation(line: 40, column: 52, scope: !7)
|
939 |
+
!31 = !DILocation(line: 41, column: 22, scope: !7)
|
940 |
+
!32 = !DILocation(line: 98, column: 22, scope: !33, inlinedAt: !35)
|
941 |
+
!33 = distinct !DILexicalBlockFile(scope: !7, file: !34, discriminator: 0)
|
942 |
+
!34 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
|
943 |
+
!35 = !DILocation(line: 44, column: 38, scope: !33)
|
944 |
+
!36 = !DILocation(line: 101, column: 30, scope: !33, inlinedAt: !35)
|
945 |
+
!37 = !DILocation(line: 101, column: 22, scope: !33, inlinedAt: !35)
|
946 |
+
!38 = !DILocation(line: 101, column: 13, scope: !33, inlinedAt: !35)
|
947 |
+
!39 = !DILocation(line: 108, column: 21, scope: !40, inlinedAt: !41)
|
948 |
+
!40 = distinct !DILexicalBlockFile(scope: !33, file: !34, discriminator: 0)
|
949 |
+
!41 = !DILocation(line: 120, column: 46, scope: !40, inlinedAt: !42)
|
950 |
+
!42 = !DILocation(line: 50, column: 41, scope: !40)
|
951 |
+
!43 = !DILocation(line: 110, column: 60, scope: !40, inlinedAt: !41)
|
952 |
+
!44 = !DILocation(line: 112, column: 25, scope: !40, inlinedAt: !41)
|
953 |
+
!45 = !DILocation(line: 112, column: 17, scope: !40, inlinedAt: !41)
|
954 |
+
!46 = !DILocation(line: 113, column: 15, scope: !40, inlinedAt: !41)
|
955 |
+
!47 = !DILocation(line: 113, column: 30, scope: !40, inlinedAt: !41)
|
956 |
+
!48 = !DILocation(line: 113, column: 49, scope: !40, inlinedAt: !41)
|
957 |
+
!49 = !DILocation(line: 113, column: 22, scope: !40, inlinedAt: !41)
|
958 |
+
!50 = !DILocation(line: 113, column: 38, scope: !40, inlinedAt: !41)
|
959 |
+
!51 = !DILocation(line: 120, column: 46, scope: !33, inlinedAt: !52)
|
960 |
+
!52 = !DILocation(line: 50, column: 41, scope: !33)
|
961 |
+
!53 = !DILocation(line: 109, column: 28, scope: !40, inlinedAt: !41)
|
962 |
+
!54 = !DILocation(line: 110, column: 39, scope: !40, inlinedAt: !41)
|
963 |
+
!55 = !DILocation(line: 110, column: 49, scope: !40, inlinedAt: !41)
|
964 |
+
!56 = !DILocation(line: 59, column: 51, scope: !7)
|
965 |
+
!57 = !DILocation(line: 60, column: 35, scope: !7)
|
966 |
+
!58 = !DILocation(line: 60, column: 40, scope: !7)
|
967 |
+
!59 = !DILocation(line: 64, column: 57, scope: !7)
|
968 |
+
!60 = !DILocation(line: 65, column: 54, scope: !7)
|
969 |
+
!61 = !DILocation(line: 69, column: 23, scope: !7)
|
970 |
+
!62 = !DILocation(line: 71, column: 24, scope: !7)
|
971 |
+
!63 = !DILocation(line: 72, column: 30, scope: !7)
|
972 |
+
!64 = !DILocation(line: 66, column: 24, scope: !7)
|
973 |
+
!65 = !DILocation(line: 67, column: 24, scope: !7)
|
974 |
+
!66 = !DILocation(line: 73, column: 24, scope: !7)
|
975 |
+
!67 = !DILocation(line: 74, column: 24, scope: !7)
|
976 |
+
!68 = !DILocation(line: 76, column: 39, scope: !7)
|
977 |
+
!69 = !DILocation(line: 76, column: 35, scope: !7)
|
978 |
+
!70 = !DILocation(line: 76, column: 29, scope: !7)
|
979 |
+
!71 = !DILocation(line: 76, column: 52, scope: !7)
|
980 |
+
!72 = !DILocation(line: 55, column: 4, scope: !7)
|
.triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ptx
ADDED
@@ -0,0 +1,1654 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
//
|
2 |
+
// Generated by LLVM NVPTX Back-End
|
3 |
+
//
|
4 |
+
|
5 |
+
.version 8.2
|
6 |
+
.target sm_89
|
7 |
+
.address_size 64
|
8 |
+
|
9 |
+
// .globl triton__0d1d2d3d4d5de6de
|
10 |
+
.extern .func __assertfail
|
11 |
+
(
|
12 |
+
.param .b64 __assertfail_param_0,
|
13 |
+
.param .b64 __assertfail_param_1,
|
14 |
+
.param .b32 __assertfail_param_2,
|
15 |
+
.param .b64 __assertfail_param_3,
|
16 |
+
.param .b64 __assertfail_param_4
|
17 |
+
)
|
18 |
+
;
|
19 |
+
.global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
20 |
+
.global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
21 |
+
.global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
22 |
+
.global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
23 |
+
.global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
24 |
+
.global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
25 |
+
.extern .shared .align 1 .b8 global_smem[];
|
26 |
+
.global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
|
27 |
+
|
28 |
+
.visible .entry triton__0d1d2d3d4d5de6de(
|
29 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_0,
|
30 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_1,
|
31 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_2,
|
32 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_3,
|
33 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_4,
|
34 |
+
.param .u32 triton__0d1d2d3d4d5de6de_param_5,
|
35 |
+
.param .u32 triton__0d1d2d3d4d5de6de_param_6
|
36 |
+
)
|
37 |
+
.maxntid 256, 1, 1
|
38 |
+
{
|
39 |
+
.reg .pred %p<117>;
|
40 |
+
.reg .b16 %rs<17>;
|
41 |
+
.reg .b32 %r<375>;
|
42 |
+
.reg .f32 %f<423>;
|
43 |
+
.reg .b64 %rd<113>;
|
44 |
+
.loc 1 18 0
|
45 |
+
$L__func_begin0:
|
46 |
+
.loc 1 18 0
|
47 |
+
|
48 |
+
ld.param.u64 %rd13, [triton__0d1d2d3d4d5de6de_param_3];
|
49 |
+
ld.param.u64 %rd12, [triton__0d1d2d3d4d5de6de_param_1];
|
50 |
+
ld.param.u64 %rd53, [triton__0d1d2d3d4d5de6de_param_0];
|
51 |
+
$L__tmp0:
|
52 |
+
.loc 1 22 44
|
53 |
+
mov.u32 %r59, %tid.x;
|
54 |
+
ld.param.u64 %rd54, [triton__0d1d2d3d4d5de6de_param_2];
|
55 |
+
bfe.u32 %r60, %r59, 5, 3;
|
56 |
+
and.b32 %r61, %r59, 15;
|
57 |
+
.loc 1 24 33
|
58 |
+
shl.b32 %r62, %r59, 3;
|
59 |
+
and.b32 %r1, %r62, 248;
|
60 |
+
and.b32 %r2, %r59, 255;
|
61 |
+
.loc 1 21 28
|
62 |
+
mov.u32 %r26, %ctaid.x;
|
63 |
+
.loc 1 21 33
|
64 |
+
shl.b32 %r63, %r26, 4;
|
65 |
+
.loc 1 22 23
|
66 |
+
or.b32 %r3, %r63, %r60;
|
67 |
+
or.b32 %r4, %r3, 8;
|
68 |
+
or.b32 %r64, %r63, %r61;
|
69 |
+
.loc 1 26 30
|
70 |
+
mul.wide.s32 %rd55, %r3, 8;
|
71 |
+
add.s64 %rd16, %rd53, %rd55;
|
72 |
+
add.s64 %rd32, %rd16, 64;
|
73 |
+
mul.wide.s32 %rd56, %r64, 8;
|
74 |
+
add.s64 %rd48, %rd53, %rd56;
|
75 |
+
mov.pred %p93, -1;
|
76 |
+
.loc 1 26 35
|
77 |
+
mov.u64 %rd15, 0x0;
|
78 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd15 }, [ %rd16 + 0 ];
|
79 |
+
mov.u64 %rd17, 0x0;
|
80 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd17 }, [ %rd16 + 0 ];
|
81 |
+
mov.u64 %rd19, 0x0;
|
82 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd19 }, [ %rd16 + 0 ];
|
83 |
+
mov.u64 %rd21, 0x0;
|
84 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd21 }, [ %rd16 + 0 ];
|
85 |
+
mov.u64 %rd23, 0x0;
|
86 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd23 }, [ %rd16 + 0 ];
|
87 |
+
mov.u64 %rd25, 0x0;
|
88 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd25 }, [ %rd16 + 0 ];
|
89 |
+
mov.u64 %rd27, 0x0;
|
90 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd27 }, [ %rd16 + 0 ];
|
91 |
+
mov.u64 %rd29, 0x0;
|
92 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd29 }, [ %rd16 + 0 ];
|
93 |
+
mov.u64 %rd31, 0x0;
|
94 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd31 }, [ %rd32 + 0 ];
|
95 |
+
mov.u64 %rd33, 0x0;
|
96 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd33 }, [ %rd32 + 0 ];
|
97 |
+
mov.u64 %rd35, 0x0;
|
98 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd35 }, [ %rd32 + 0 ];
|
99 |
+
mov.u64 %rd37, 0x0;
|
100 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd37 }, [ %rd32 + 0 ];
|
101 |
+
mov.u64 %rd39, 0x0;
|
102 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd39 }, [ %rd32 + 0 ];
|
103 |
+
mov.u64 %rd41, 0x0;
|
104 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd41 }, [ %rd32 + 0 ];
|
105 |
+
mov.u64 %rd43, 0x0;
|
106 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd43 }, [ %rd32 + 0 ];
|
107 |
+
mov.u64 %rd45, 0x0;
|
108 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd45 }, [ %rd32 + 0 ];
|
109 |
+
mov.u64 %rd47, 0x0;
|
110 |
+
@%p93 ld.global.L1::evict_last.b64 { %rd47 }, [ %rd48 + 0 ];
|
111 |
+
.loc 1 27 18
|
112 |
+
bfe.s32 %r65, %r26, 27, 1;
|
113 |
+
shr.u32 %r66, %r65, 23;
|
114 |
+
add.s32 %r67, %r3, %r66;
|
115 |
+
and.b32 %r68, %r67, 16776704;
|
116 |
+
sub.s32 %r69, %r3, %r68;
|
117 |
+
add.s32 %r70, %r4, %r66;
|
118 |
+
and.b32 %r71, %r70, 16776704;
|
119 |
+
sub.s32 %r72, %r4, %r71;
|
120 |
+
.loc 1 35 44
|
121 |
+
shl.b32 %r73, %r69, 8;
|
122 |
+
shl.b32 %r74, %r72, 8;
|
123 |
+
.loc 1 35 40
|
124 |
+
or.b32 %r75, %r73, %r1;
|
125 |
+
or.b32 %r76, %r74, %r1;
|
126 |
+
.loc 1 35 34
|
127 |
+
mul.wide.s32 %rd57, %r75, 4;
|
128 |
+
add.s64 %rd80, %rd54, %rd57;
|
129 |
+
cvt.s64.s32 %rd58, %r73;
|
130 |
+
cvt.u64.u32 %rd59, %r1;
|
131 |
+
or.b64 %rd60, %rd58, %rd59;
|
132 |
+
shl.b64 %rd61, %rd60, 2;
|
133 |
+
add.s64 %rd62, %rd54, %rd61;
|
134 |
+
add.s64 %rd81, %rd62, 16;
|
135 |
+
mul.wide.s32 %rd63, %r76, 4;
|
136 |
+
add.s64 %rd82, %rd54, %rd63;
|
137 |
+
cvt.s64.s32 %rd64, %r74;
|
138 |
+
or.b64 %rd65, %rd64, %rd59;
|
139 |
+
shl.b64 %rd66, %rd65, 2;
|
140 |
+
add.s64 %rd67, %rd54, %rd66;
|
141 |
+
add.s64 %rd83, %rd67, 16;
|
142 |
+
mov.b32 %r257, 0;
|
143 |
+
.loc 1 35 50
|
144 |
+
mov.u32 %r27, 0x0;
|
145 |
+
mov.u32 %r28, 0x0;
|
146 |
+
mov.u32 %r29, 0x0;
|
147 |
+
mov.u32 %r30, 0x0;
|
148 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r27, %r28, %r29, %r30 }, [ %rd80 + 0 ];
|
149 |
+
@!%p93 mov.u32 %r27, %r257;
|
150 |
+
@!%p93 mov.u32 %r28, %r257;
|
151 |
+
@!%p93 mov.u32 %r29, %r257;
|
152 |
+
@!%p93 mov.u32 %r30, %r257;
|
153 |
+
mov.b32 %f1, %r27;
|
154 |
+
mov.b32 %f2, %r28;
|
155 |
+
mov.b32 %f3, %r29;
|
156 |
+
mov.b32 %f4, %r30;
|
157 |
+
mov.u32 %r35, 0x0;
|
158 |
+
mov.u32 %r36, 0x0;
|
159 |
+
mov.u32 %r37, 0x0;
|
160 |
+
mov.u32 %r38, 0x0;
|
161 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r35, %r36, %r37, %r38 }, [ %rd81 + 0 ];
|
162 |
+
@!%p93 mov.u32 %r35, %r257;
|
163 |
+
@!%p93 mov.u32 %r36, %r257;
|
164 |
+
@!%p93 mov.u32 %r37, %r257;
|
165 |
+
@!%p93 mov.u32 %r38, %r257;
|
166 |
+
mov.b32 %f5, %r35;
|
167 |
+
mov.b32 %f6, %r36;
|
168 |
+
mov.b32 %f7, %r37;
|
169 |
+
mov.b32 %f8, %r38;
|
170 |
+
mov.u32 %r43, 0x0;
|
171 |
+
mov.u32 %r44, 0x0;
|
172 |
+
mov.u32 %r45, 0x0;
|
173 |
+
mov.u32 %r46, 0x0;
|
174 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r43, %r44, %r45, %r46 }, [ %rd82 + 0 ];
|
175 |
+
@!%p93 mov.u32 %r43, %r257;
|
176 |
+
@!%p93 mov.u32 %r44, %r257;
|
177 |
+
@!%p93 mov.u32 %r45, %r257;
|
178 |
+
@!%p93 mov.u32 %r46, %r257;
|
179 |
+
mov.b32 %f9, %r43;
|
180 |
+
mov.b32 %f10, %r44;
|
181 |
+
mov.b32 %f11, %r45;
|
182 |
+
mov.b32 %f12, %r46;
|
183 |
+
mov.u32 %r51, 0x0;
|
184 |
+
mov.u32 %r52, 0x0;
|
185 |
+
mov.u32 %r53, 0x0;
|
186 |
+
mov.u32 %r54, 0x0;
|
187 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r51, %r52, %r53, %r54 }, [ %rd83 + 0 ];
|
188 |
+
@!%p93 mov.u32 %r51, %r257;
|
189 |
+
@!%p93 mov.u32 %r52, %r257;
|
190 |
+
@!%p93 mov.u32 %r53, %r257;
|
191 |
+
@!%p93 mov.u32 %r54, %r257;
|
192 |
+
mov.b32 %f13, %r51;
|
193 |
+
mov.b32 %f14, %r52;
|
194 |
+
mov.b32 %f15, %r53;
|
195 |
+
mov.b32 %f16, %r54;
|
196 |
+
.loc 1 36 22
|
197 |
+
add.s64 %rd68, %rd47, 50257;
|
198 |
+
.loc 1 37 22
|
199 |
+
setp.lt.s64 %p38, %rd47, 0;
|
200 |
+
.loc 1 38 36
|
201 |
+
selp.b64 %rd7, %rd68, %rd47, %p38;
|
202 |
+
.loc 1 39 40
|
203 |
+
setp.lt.u64 %p39, %rd7, 50257;
|
204 |
+
mov.b32 %r374, 883;
|
205 |
+
mov.u64 %rd112, 1;
|
206 |
+
.loc 1 39 55
|
207 |
+
@%p39 bra $L__BB0_2;
|
208 |
+
mov.u64 %rd69, assertMessage_0;
|
209 |
+
cvta.global.u64 %rd70, %rd69;
|
210 |
+
mov.u64 %rd71, assertFile_0;
|
211 |
+
cvta.global.u64 %rd72, %rd71;
|
212 |
+
mov.u64 %rd73, assertFunc_0;
|
213 |
+
cvta.global.u64 %rd74, %rd73;
|
214 |
+
{ // callseq 8, 0
|
215 |
+
.reg .b32 temp_param_reg;
|
216 |
+
.param .b64 param0;
|
217 |
+
st.param.b64 [param0+0], %rd70;
|
218 |
+
.param .b64 param1;
|
219 |
+
st.param.b64 [param1+0], %rd72;
|
220 |
+
.param .b32 param2;
|
221 |
+
st.param.b32 [param2+0], %r374;
|
222 |
+
.param .b64 param3;
|
223 |
+
st.param.b64 [param3+0], %rd74;
|
224 |
+
.param .b64 param4;
|
225 |
+
st.param.b64 [param4+0], %rd112;
|
226 |
+
call.uni
|
227 |
+
__assertfail,
|
228 |
+
(
|
229 |
+
param0,
|
230 |
+
param1,
|
231 |
+
param2,
|
232 |
+
param3,
|
233 |
+
param4
|
234 |
+
);
|
235 |
+
} // callseq 8
|
236 |
+
$L__BB0_2:
|
237 |
+
.loc 1 0 55
|
238 |
+
ld.param.u64 %rd14, [triton__0d1d2d3d4d5de6de_param_4];
|
239 |
+
.loc 1 37 22
|
240 |
+
setp.lt.s64 %p83, %rd31, 0;
|
241 |
+
setp.lt.s64 %p84, %rd15, 0;
|
242 |
+
.loc 1 40 44
|
243 |
+
shl.b64 %rd85, %rd15, 8;
|
244 |
+
add.s64 %rd86, %rd85, 12865792;
|
245 |
+
selp.b64 %rd87, %rd86, %rd85, %p84;
|
246 |
+
shl.b64 %rd88, %rd31, 8;
|
247 |
+
add.s64 %rd89, %rd88, 12865792;
|
248 |
+
selp.b64 %rd90, %rd89, %rd88, %p83;
|
249 |
+
.loc 1 40 40
|
250 |
+
or.b64 %rd92, %rd87, %rd59;
|
251 |
+
or.b64 %rd93, %rd90, %rd59;
|
252 |
+
.loc 1 40 34
|
253 |
+
shl.b64 %rd94, %rd92, 2;
|
254 |
+
add.s64 %rd104, %rd12, %rd94;
|
255 |
+
add.s64 %rd105, %rd104, 16;
|
256 |
+
shl.b64 %rd95, %rd93, 2;
|
257 |
+
add.s64 %rd106, %rd12, %rd95;
|
258 |
+
add.s64 %rd107, %rd106, 16;
|
259 |
+
.loc 1 40 52
|
260 |
+
mov.u32 %r78, 0x0;
|
261 |
+
mov.u32 %r79, 0x0;
|
262 |
+
mov.u32 %r80, 0x0;
|
263 |
+
mov.u32 %r81, 0x0;
|
264 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r78, %r79, %r80, %r81 }, [ %rd104 + 0 ];
|
265 |
+
@!%p93 mov.u32 %r78, %r257;
|
266 |
+
@!%p93 mov.u32 %r79, %r257;
|
267 |
+
@!%p93 mov.u32 %r80, %r257;
|
268 |
+
@!%p93 mov.u32 %r81, %r257;
|
269 |
+
mov.b32 %f27, %r78;
|
270 |
+
mov.b32 %f28, %r79;
|
271 |
+
mov.b32 %f29, %r80;
|
272 |
+
mov.b32 %f30, %r81;
|
273 |
+
mov.u32 %r86, 0x0;
|
274 |
+
mov.u32 %r87, 0x0;
|
275 |
+
mov.u32 %r88, 0x0;
|
276 |
+
mov.u32 %r89, 0x0;
|
277 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r86, %r87, %r88, %r89 }, [ %rd105 + 0 ];
|
278 |
+
@!%p93 mov.u32 %r86, %r257;
|
279 |
+
@!%p93 mov.u32 %r87, %r257;
|
280 |
+
@!%p93 mov.u32 %r88, %r257;
|
281 |
+
@!%p93 mov.u32 %r89, %r257;
|
282 |
+
mov.b32 %f31, %r86;
|
283 |
+
mov.b32 %f32, %r87;
|
284 |
+
mov.b32 %f33, %r88;
|
285 |
+
mov.b32 %f34, %r89;
|
286 |
+
mov.u32 %r94, 0x0;
|
287 |
+
mov.u32 %r95, 0x0;
|
288 |
+
mov.u32 %r96, 0x0;
|
289 |
+
mov.u32 %r97, 0x0;
|
290 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r94, %r95, %r96, %r97 }, [ %rd106 + 0 ];
|
291 |
+
@!%p93 mov.u32 %r94, %r257;
|
292 |
+
@!%p93 mov.u32 %r95, %r257;
|
293 |
+
@!%p93 mov.u32 %r96, %r257;
|
294 |
+
@!%p93 mov.u32 %r97, %r257;
|
295 |
+
mov.b32 %f35, %r94;
|
296 |
+
mov.b32 %f36, %r95;
|
297 |
+
mov.b32 %f37, %r96;
|
298 |
+
mov.b32 %f38, %r97;
|
299 |
+
mov.u32 %r102, 0x0;
|
300 |
+
mov.u32 %r103, 0x0;
|
301 |
+
mov.u32 %r104, 0x0;
|
302 |
+
mov.u32 %r105, 0x0;
|
303 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r102, %r103, %r104, %r105 }, [ %rd107 + 0 ];
|
304 |
+
@!%p93 mov.u32 %r102, %r257;
|
305 |
+
@!%p93 mov.u32 %r103, %r257;
|
306 |
+
@!%p93 mov.u32 %r104, %r257;
|
307 |
+
@!%p93 mov.u32 %r105, %r257;
|
308 |
+
mov.b32 %f39, %r102;
|
309 |
+
mov.b32 %f40, %r103;
|
310 |
+
mov.b32 %f41, %r104;
|
311 |
+
mov.b32 %f42, %r105;
|
312 |
+
.loc 1 41 22
|
313 |
+
add.f32 %f43, %f1, %f27;
|
314 |
+
add.f32 %f44, %f2, %f28;
|
315 |
+
add.f32 %f45, %f3, %f29;
|
316 |
+
add.f32 %f46, %f4, %f30;
|
317 |
+
add.f32 %f47, %f5, %f31;
|
318 |
+
add.f32 %f48, %f6, %f32;
|
319 |
+
add.f32 %f49, %f7, %f33;
|
320 |
+
add.f32 %f50, %f8, %f34;
|
321 |
+
add.f32 %f51, %f9, %f35;
|
322 |
+
add.f32 %f52, %f10, %f36;
|
323 |
+
add.f32 %f53, %f11, %f37;
|
324 |
+
add.f32 %f54, %f12, %f38;
|
325 |
+
add.f32 %f55, %f13, %f39;
|
326 |
+
add.f32 %f56, %f14, %f40;
|
327 |
+
add.f32 %f57, %f15, %f41;
|
328 |
+
add.f32 %f58, %f16, %f42;
|
329 |
+
$L__tmp1:
|
330 |
+
.loc 2 98 22
|
331 |
+
add.f32 %f59, %f43, 0f00000000;
|
332 |
+
add.f32 %f60, %f44, 0f00000000;
|
333 |
+
add.f32 %f61, %f45, 0f00000000;
|
334 |
+
add.f32 %f62, %f46, 0f00000000;
|
335 |
+
add.f32 %f63, %f47, 0f00000000;
|
336 |
+
add.f32 %f64, %f48, 0f00000000;
|
337 |
+
add.f32 %f65, %f49, 0f00000000;
|
338 |
+
add.f32 %f66, %f50, 0f00000000;
|
339 |
+
add.f32 %f67, %f51, 0f00000000;
|
340 |
+
add.f32 %f68, %f52, 0f00000000;
|
341 |
+
add.f32 %f69, %f53, 0f00000000;
|
342 |
+
add.f32 %f70, %f54, 0f00000000;
|
343 |
+
add.f32 %f71, %f55, 0f00000000;
|
344 |
+
add.f32 %f72, %f56, 0f00000000;
|
345 |
+
add.f32 %f73, %f57, 0f00000000;
|
346 |
+
add.f32 %f74, %f58, 0f00000000;
|
347 |
+
.loc 2 101 30
|
348 |
+
sub.f32 %f75, %f43, %f59;
|
349 |
+
sub.f32 %f76, %f44, %f60;
|
350 |
+
sub.f32 %f77, %f45, %f61;
|
351 |
+
sub.f32 %f78, %f46, %f62;
|
352 |
+
sub.f32 %f79, %f47, %f63;
|
353 |
+
sub.f32 %f80, %f48, %f64;
|
354 |
+
sub.f32 %f81, %f49, %f65;
|
355 |
+
sub.f32 %f82, %f50, %f66;
|
356 |
+
sub.f32 %f83, %f51, %f67;
|
357 |
+
sub.f32 %f84, %f52, %f68;
|
358 |
+
sub.f32 %f85, %f53, %f69;
|
359 |
+
sub.f32 %f86, %f54, %f70;
|
360 |
+
sub.f32 %f87, %f55, %f71;
|
361 |
+
sub.f32 %f88, %f56, %f72;
|
362 |
+
sub.f32 %f89, %f57, %f73;
|
363 |
+
sub.f32 %f90, %f58, %f74;
|
364 |
+
.loc 2 101 13
|
365 |
+
fma.rn.f32 %f91, %f43, %f75, 0f00000000;
|
366 |
+
fma.rn.f32 %f92, %f44, %f76, 0f00000000;
|
367 |
+
fma.rn.f32 %f93, %f45, %f77, 0f00000000;
|
368 |
+
fma.rn.f32 %f94, %f46, %f78, 0f00000000;
|
369 |
+
fma.rn.f32 %f95, %f47, %f79, 0f00000000;
|
370 |
+
fma.rn.f32 %f96, %f48, %f80, 0f00000000;
|
371 |
+
fma.rn.f32 %f97, %f49, %f81, 0f00000000;
|
372 |
+
fma.rn.f32 %f98, %f50, %f82, 0f00000000;
|
373 |
+
fma.rn.f32 %f99, %f51, %f83, 0f00000000;
|
374 |
+
fma.rn.f32 %f100, %f52, %f84, 0f00000000;
|
375 |
+
fma.rn.f32 %f101, %f53, %f85, 0f00000000;
|
376 |
+
fma.rn.f32 %f102, %f54, %f86, 0f00000000;
|
377 |
+
fma.rn.f32 %f103, %f55, %f87, 0f00000000;
|
378 |
+
fma.rn.f32 %f104, %f56, %f88, 0f00000000;
|
379 |
+
fma.rn.f32 %f105, %f57, %f89, 0f00000000;
|
380 |
+
fma.rn.f32 %f106, %f58, %f90, 0f00000000;
|
381 |
+
$L__tmp2:
|
382 |
+
.loc 2 108 21
|
383 |
+
sub.f32 %f107, %f60, %f59;
|
384 |
+
mov.b32 %r111, 1065353216;
|
385 |
+
mov.b32 %r112, 1073741824;
|
386 |
+
.loc 2 110 60
|
387 |
+
div.full.f32 %r110, %r111, %r112;
|
388 |
+
mov.b32 %f108, %r110;
|
389 |
+
.loc 2 112 17
|
390 |
+
fma.rn.f32 %f109, %f108, %f107, %f59;
|
391 |
+
.loc 2 113 15
|
392 |
+
add.f32 %f110, %f91, %f92;
|
393 |
+
.loc 2 113 30
|
394 |
+
mul.f32 %f111, %f107, %f107;
|
395 |
+
.loc 2 113 22
|
396 |
+
fma.rn.f32 %f112, %f108, %f111, %f110;
|
397 |
+
.loc 2 108 21
|
398 |
+
sub.f32 %f113, %f61, %f109;
|
399 |
+
mov.b32 %r115, 1077936128;
|
400 |
+
.loc 2 110 60
|
401 |
+
div.full.f32 %r113, %r111, %r115;
|
402 |
+
mov.b32 %f114, %r113;
|
403 |
+
.loc 2 112 17
|
404 |
+
fma.rn.f32 %f115, %f114, %f113, %f109;
|
405 |
+
.loc 2 113 15
|
406 |
+
add.f32 %f116, %f93, %f112;
|
407 |
+
.loc 2 113 30
|
408 |
+
mul.f32 %f117, %f113, %f113;
|
409 |
+
.loc 2 113 38
|
410 |
+
fma.rn.f32 %f118, %f113, %f113, %f117;
|
411 |
+
.loc 2 113 22
|
412 |
+
fma.rn.f32 %f119, %f114, %f118, %f116;
|
413 |
+
.loc 2 108 21
|
414 |
+
sub.f32 %f120, %f62, %f115;
|
415 |
+
mov.b32 %r118, 1082130432;
|
416 |
+
.loc 2 110 60
|
417 |
+
div.full.f32 %r116, %r111, %r118;
|
418 |
+
mov.b32 %f121, %r116;
|
419 |
+
.loc 2 112 17
|
420 |
+
fma.rn.f32 %f122, %f121, %f120, %f115;
|
421 |
+
.loc 2 113 15
|
422 |
+
add.f32 %f123, %f94, %f119;
|
423 |
+
.loc 2 113 30
|
424 |
+
mul.f32 %f124, %f120, %f120;
|
425 |
+
.loc 2 113 38
|
426 |
+
mul.f32 %f125, %f124, 0f40400000;
|
427 |
+
.loc 2 113 22
|
428 |
+
fma.rn.f32 %f126, %f121, %f125, %f123;
|
429 |
+
.loc 2 108 21
|
430 |
+
sub.f32 %f127, %f63, %f122;
|
431 |
+
mov.b32 %r121, 1084227584;
|
432 |
+
.loc 2 110 60
|
433 |
+
div.full.f32 %r119, %r111, %r121;
|
434 |
+
mov.b32 %f128, %r119;
|
435 |
+
.loc 2 112 17
|
436 |
+
fma.rn.f32 %f129, %f128, %f127, %f122;
|
437 |
+
.loc 2 113 15
|
438 |
+
add.f32 %f130, %f95, %f126;
|
439 |
+
.loc 2 113 30
|
440 |
+
mul.f32 %f131, %f127, %f127;
|
441 |
+
.loc 2 113 38
|
442 |
+
mul.f32 %f132, %f131, 0f40800000;
|
443 |
+
.loc 2 113 22
|
444 |
+
fma.rn.f32 %f133, %f128, %f132, %f130;
|
445 |
+
.loc 2 108 21
|
446 |
+
sub.f32 %f134, %f64, %f129;
|
447 |
+
mov.b32 %r124, 1086324736;
|
448 |
+
.loc 2 110 60
|
449 |
+
div.full.f32 %r122, %r111, %r124;
|
450 |
+
mov.b32 %f135, %r122;
|
451 |
+
.loc 2 112 17
|
452 |
+
fma.rn.f32 %f136, %f135, %f134, %f129;
|
453 |
+
.loc 2 113 15
|
454 |
+
add.f32 %f137, %f96, %f133;
|
455 |
+
.loc 2 113 30
|
456 |
+
mul.f32 %f138, %f134, %f134;
|
457 |
+
.loc 2 113 38
|
458 |
+
mul.f32 %f139, %f138, 0f40A00000;
|
459 |
+
.loc 2 113 22
|
460 |
+
fma.rn.f32 %f140, %f135, %f139, %f137;
|
461 |
+
.loc 2 108 21
|
462 |
+
sub.f32 %f141, %f65, %f136;
|
463 |
+
mov.b32 %r127, 1088421888;
|
464 |
+
.loc 2 110 60
|
465 |
+
div.full.f32 %r125, %r111, %r127;
|
466 |
+
mov.b32 %f142, %r125;
|
467 |
+
.loc 2 112 17
|
468 |
+
fma.rn.f32 %f143, %f142, %f141, %f136;
|
469 |
+
.loc 2 113 15
|
470 |
+
add.f32 %f144, %f97, %f140;
|
471 |
+
.loc 2 113 30
|
472 |
+
mul.f32 %f145, %f141, %f141;
|
473 |
+
.loc 2 113 38
|
474 |
+
mul.f32 %f146, %f145, 0f40C00000;
|
475 |
+
.loc 2 113 22
|
476 |
+
fma.rn.f32 %f147, %f142, %f146, %f144;
|
477 |
+
.loc 2 108 21
|
478 |
+
sub.f32 %f148, %f66, %f143;
|
479 |
+
mov.b32 %r130, 1090519040;
|
480 |
+
.loc 2 110 60
|
481 |
+
div.full.f32 %r128, %r111, %r130;
|
482 |
+
mov.b32 %f149, %r128;
|
483 |
+
.loc 2 112 17
|
484 |
+
fma.rn.f32 %f150, %f149, %f148, %f143;
|
485 |
+
.loc 2 113 15
|
486 |
+
add.f32 %f151, %f98, %f147;
|
487 |
+
.loc 2 113 30
|
488 |
+
mul.f32 %f152, %f148, %f148;
|
489 |
+
.loc 2 113 38
|
490 |
+
mul.f32 %f153, %f152, 0f40E00000;
|
491 |
+
.loc 2 113 22
|
492 |
+
fma.rn.f32 %f154, %f149, %f153, %f151;
|
493 |
+
.loc 2 108 21
|
494 |
+
sub.f32 %f155, %f68, %f67;
|
495 |
+
.loc 2 110 60
|
496 |
+
div.full.f32 %r131, %r111, %r112;
|
497 |
+
mov.b32 %f156, %r131;
|
498 |
+
.loc 2 112 17
|
499 |
+
fma.rn.f32 %f157, %f155, %f156, %f67;
|
500 |
+
.loc 2 113 15
|
501 |
+
add.f32 %f158, %f99, %f100;
|
502 |
+
.loc 2 113 30
|
503 |
+
mul.f32 %f159, %f155, %f155;
|
504 |
+
.loc 2 113 22
|
505 |
+
fma.rn.f32 %f160, %f159, %f156, %f158;
|
506 |
+
.loc 2 108 21
|
507 |
+
sub.f32 %f161, %f69, %f157;
|
508 |
+
.loc 2 110 60
|
509 |
+
div.full.f32 %r134, %r111, %r115;
|
510 |
+
mov.b32 %f162, %r134;
|
511 |
+
.loc 2 112 17
|
512 |
+
fma.rn.f32 %f163, %f162, %f161, %f157;
|
513 |
+
.loc 2 113 15
|
514 |
+
add.f32 %f164, %f101, %f160;
|
515 |
+
.loc 2 113 30
|
516 |
+
mul.f32 %f165, %f161, %f161;
|
517 |
+
.loc 2 113 38
|
518 |
+
fma.rn.f32 %f166, %f161, %f161, %f165;
|
519 |
+
.loc 2 113 22
|
520 |
+
fma.rn.f32 %f167, %f162, %f166, %f164;
|
521 |
+
.loc 2 108 21
|
522 |
+
sub.f32 %f168, %f70, %f163;
|
523 |
+
.loc 2 110 60
|
524 |
+
div.full.f32 %r137, %r111, %r118;
|
525 |
+
mov.b32 %f169, %r137;
|
526 |
+
.loc 2 112 17
|
527 |
+
fma.rn.f32 %f170, %f169, %f168, %f163;
|
528 |
+
.loc 2 113 15
|
529 |
+
add.f32 %f171, %f102, %f167;
|
530 |
+
.loc 2 113 30
|
531 |
+
mul.f32 %f172, %f168, %f168;
|
532 |
+
.loc 2 113 38
|
533 |
+
mul.f32 %f173, %f172, 0f40400000;
|
534 |
+
.loc 2 113 22
|
535 |
+
fma.rn.f32 %f174, %f169, %f173, %f171;
|
536 |
+
.loc 2 108 21
|
537 |
+
sub.f32 %f175, %f71, %f170;
|
538 |
+
.loc 2 110 60
|
539 |
+
div.full.f32 %r140, %r111, %r121;
|
540 |
+
mov.b32 %f176, %r140;
|
541 |
+
.loc 2 112 17
|
542 |
+
fma.rn.f32 %f177, %f176, %f175, %f170;
|
543 |
+
.loc 2 113 15
|
544 |
+
add.f32 %f178, %f103, %f174;
|
545 |
+
.loc 2 113 30
|
546 |
+
mul.f32 %f179, %f175, %f175;
|
547 |
+
.loc 2 113 38
|
548 |
+
mul.f32 %f180, %f179, 0f40800000;
|
549 |
+
.loc 2 113 22
|
550 |
+
fma.rn.f32 %f181, %f176, %f180, %f178;
|
551 |
+
.loc 2 108 21
|
552 |
+
sub.f32 %f182, %f72, %f177;
|
553 |
+
.loc 2 110 60
|
554 |
+
div.full.f32 %r143, %r111, %r124;
|
555 |
+
mov.b32 %f183, %r143;
|
556 |
+
.loc 2 112 17
|
557 |
+
fma.rn.f32 %f184, %f183, %f182, %f177;
|
558 |
+
.loc 2 113 15
|
559 |
+
add.f32 %f185, %f104, %f181;
|
560 |
+
.loc 2 113 30
|
561 |
+
mul.f32 %f186, %f182, %f182;
|
562 |
+
.loc 2 113 38
|
563 |
+
mul.f32 %f187, %f186, 0f40A00000;
|
564 |
+
.loc 2 113 22
|
565 |
+
fma.rn.f32 %f188, %f183, %f187, %f185;
|
566 |
+
.loc 2 108 21
|
567 |
+
sub.f32 %f189, %f73, %f184;
|
568 |
+
.loc 2 110 60
|
569 |
+
div.full.f32 %r146, %r111, %r127;
|
570 |
+
mov.b32 %f190, %r146;
|
571 |
+
.loc 2 112 17
|
572 |
+
fma.rn.f32 %f191, %f190, %f189, %f184;
|
573 |
+
.loc 2 113 15
|
574 |
+
add.f32 %f192, %f105, %f188;
|
575 |
+
.loc 2 113 30
|
576 |
+
mul.f32 %f193, %f189, %f189;
|
577 |
+
.loc 2 113 38
|
578 |
+
mul.f32 %f194, %f193, 0f40C00000;
|
579 |
+
.loc 2 113 22
|
580 |
+
fma.rn.f32 %f195, %f190, %f194, %f192;
|
581 |
+
.loc 2 108 21
|
582 |
+
sub.f32 %f196, %f74, %f191;
|
583 |
+
.loc 2 110 60
|
584 |
+
div.full.f32 %r149, %r111, %r130;
|
585 |
+
mov.b32 %f197, %r149;
|
586 |
+
.loc 2 112 17
|
587 |
+
fma.rn.f32 %f198, %f197, %f196, %f191;
|
588 |
+
.loc 2 113 15
|
589 |
+
add.f32 %f199, %f106, %f195;
|
590 |
+
.loc 2 113 30
|
591 |
+
mul.f32 %f200, %f196, %f196;
|
592 |
+
.loc 2 113 38
|
593 |
+
mul.f32 %f201, %f200, 0f40E00000;
|
594 |
+
.loc 2 113 22
|
595 |
+
fma.rn.f32 %f202, %f197, %f201, %f199;
|
596 |
+
$L__tmp3:
|
597 |
+
.loc 2 120 46
|
598 |
+
mov.b32 %r216, %f150;
|
599 |
+
shfl.sync.bfly.b32 %r217, %r216, 16, 31, -1;
|
600 |
+
mov.b32 %f203, %r217;
|
601 |
+
mov.b32 %r218, %f154;
|
602 |
+
shfl.sync.bfly.b32 %r219, %r218, 16, 31, -1;
|
603 |
+
mov.b32 %f204, %r219;
|
604 |
+
shfl.sync.bfly.b32 %r153, %r130, 16, 31, -1;
|
605 |
+
mov.b32 %f205, %r153;
|
606 |
+
$L__tmp4:
|
607 |
+
.loc 2 108 21
|
608 |
+
sub.f32 %f206, %f203, %f150;
|
609 |
+
.loc 2 109 28
|
610 |
+
add.f32 %f207, %f205, 0f41000000;
|
611 |
+
.loc 2 110 39
|
612 |
+
setp.eq.f32 %p85, %f207, 0f00000000;
|
613 |
+
.loc 2 110 60
|
614 |
+
mov.b32 %r154, %f207;
|
615 |
+
div.full.f32 %r152, %r153, %r154;
|
616 |
+
mov.b32 %f208, %r152;
|
617 |
+
.loc 2 110 49
|
618 |
+
selp.f32 %f209, 0f00000000, %f208, %p85;
|
619 |
+
.loc 2 112 17
|
620 |
+
fma.rn.f32 %f210, %f209, %f206, %f150;
|
621 |
+
.loc 2 113 15
|
622 |
+
add.f32 %f211, %f154, %f204;
|
623 |
+
.loc 2 113 30
|
624 |
+
mul.f32 %f212, %f206, %f206;
|
625 |
+
.loc 2 113 38
|
626 |
+
mul.f32 %f213, %f212, 0f41000000;
|
627 |
+
.loc 2 113 22
|
628 |
+
fma.rn.f32 %f214, %f209, %f213, %f211;
|
629 |
+
$L__tmp5:
|
630 |
+
.loc 2 120 46
|
631 |
+
mov.b32 %r220, %f210;
|
632 |
+
shfl.sync.bfly.b32 %r221, %r220, 8, 31, -1;
|
633 |
+
mov.b32 %f215, %r221;
|
634 |
+
mov.b32 %r222, %f214;
|
635 |
+
shfl.sync.bfly.b32 %r223, %r222, 8, 31, -1;
|
636 |
+
mov.b32 %f216, %r223;
|
637 |
+
shfl.sync.bfly.b32 %r156, %r154, 8, 31, -1;
|
638 |
+
mov.b32 %f217, %r156;
|
639 |
+
$L__tmp6:
|
640 |
+
.loc 2 108 21
|
641 |
+
sub.f32 %f218, %f215, %f210;
|
642 |
+
.loc 2 109 28
|
643 |
+
add.f32 %f219, %f207, %f217;
|
644 |
+
.loc 2 110 39
|
645 |
+
setp.eq.f32 %p86, %f219, 0f00000000;
|
646 |
+
.loc 2 110 60
|
647 |
+
mov.b32 %r157, %f219;
|
648 |
+
div.full.f32 %r155, %r156, %r157;
|
649 |
+
mov.b32 %f220, %r155;
|
650 |
+
.loc 2 110 49
|
651 |
+
selp.f32 %f221, 0f00000000, %f220, %p86;
|
652 |
+
.loc 2 112 17
|
653 |
+
fma.rn.f32 %f222, %f221, %f218, %f210;
|
654 |
+
.loc 2 113 15
|
655 |
+
add.f32 %f223, %f214, %f216;
|
656 |
+
.loc 2 113 30
|
657 |
+
mul.f32 %f224, %f218, %f218;
|
658 |
+
.loc 2 113 38
|
659 |
+
mul.f32 %f225, %f207, %f224;
|
660 |
+
.loc 2 113 22
|
661 |
+
fma.rn.f32 %f226, %f221, %f225, %f223;
|
662 |
+
$L__tmp7:
|
663 |
+
.loc 2 120 46
|
664 |
+
mov.b32 %r224, %f222;
|
665 |
+
shfl.sync.bfly.b32 %r225, %r224, 4, 31, -1;
|
666 |
+
mov.b32 %f227, %r225;
|
667 |
+
mov.b32 %r226, %f226;
|
668 |
+
shfl.sync.bfly.b32 %r227, %r226, 4, 31, -1;
|
669 |
+
mov.b32 %f228, %r227;
|
670 |
+
shfl.sync.bfly.b32 %r159, %r157, 4, 31, -1;
|
671 |
+
mov.b32 %f229, %r159;
|
672 |
+
$L__tmp8:
|
673 |
+
.loc 2 108 21
|
674 |
+
sub.f32 %f230, %f227, %f222;
|
675 |
+
.loc 2 109 28
|
676 |
+
add.f32 %f231, %f219, %f229;
|
677 |
+
.loc 2 110 39
|
678 |
+
setp.eq.f32 %p87, %f231, 0f00000000;
|
679 |
+
.loc 2 110 60
|
680 |
+
mov.b32 %r160, %f231;
|
681 |
+
div.full.f32 %r158, %r159, %r160;
|
682 |
+
mov.b32 %f232, %r158;
|
683 |
+
.loc 2 110 49
|
684 |
+
selp.f32 %f233, 0f00000000, %f232, %p87;
|
685 |
+
.loc 2 112 17
|
686 |
+
fma.rn.f32 %f234, %f230, %f233, %f222;
|
687 |
+
.loc 2 113 15
|
688 |
+
add.f32 %f235, %f226, %f228;
|
689 |
+
.loc 2 113 30
|
690 |
+
mul.f32 %f236, %f230, %f230;
|
691 |
+
.loc 2 113 38
|
692 |
+
mul.f32 %f237, %f219, %f236;
|
693 |
+
.loc 2 113 22
|
694 |
+
fma.rn.f32 %f238, %f233, %f237, %f235;
|
695 |
+
$L__tmp9:
|
696 |
+
.loc 2 120 46
|
697 |
+
mov.b32 %r228, %f234;
|
698 |
+
shfl.sync.bfly.b32 %r229, %r228, 2, 31, -1;
|
699 |
+
mov.b32 %f239, %r229;
|
700 |
+
mov.b32 %r230, %f238;
|
701 |
+
shfl.sync.bfly.b32 %r231, %r230, 2, 31, -1;
|
702 |
+
mov.b32 %f240, %r231;
|
703 |
+
shfl.sync.bfly.b32 %r162, %r160, 2, 31, -1;
|
704 |
+
mov.b32 %f241, %r162;
|
705 |
+
$L__tmp10:
|
706 |
+
.loc 2 108 21
|
707 |
+
sub.f32 %f242, %f239, %f234;
|
708 |
+
.loc 2 109 28
|
709 |
+
add.f32 %f17, %f231, %f241;
|
710 |
+
.loc 2 110 39
|
711 |
+
setp.eq.f32 %p88, %f17, 0f00000000;
|
712 |
+
.loc 2 110 60
|
713 |
+
mov.b32 %r163, %f17;
|
714 |
+
div.full.f32 %r161, %r162, %r163;
|
715 |
+
mov.b32 %f243, %r161;
|
716 |
+
.loc 2 110 49
|
717 |
+
selp.f32 %f244, 0f00000000, %f243, %p88;
|
718 |
+
.loc 2 112 17
|
719 |
+
fma.rn.f32 %f18, %f242, %f244, %f234;
|
720 |
+
.loc 2 113 15
|
721 |
+
add.f32 %f245, %f238, %f240;
|
722 |
+
.loc 2 113 30
|
723 |
+
mul.f32 %f246, %f242, %f242;
|
724 |
+
.loc 2 113 38
|
725 |
+
mul.f32 %f247, %f231, %f246;
|
726 |
+
.loc 2 113 22
|
727 |
+
fma.rn.f32 %f19, %f244, %f247, %f245;
|
728 |
+
$L__tmp11:
|
729 |
+
.loc 2 120 46
|
730 |
+
mov.b32 %r232, %f18;
|
731 |
+
shfl.sync.bfly.b32 %r5, %r232, 1, 31, -1;
|
732 |
+
mov.b32 %r233, %f19;
|
733 |
+
shfl.sync.bfly.b32 %r6, %r233, 1, 31, -1;
|
734 |
+
shfl.sync.bfly.b32 %r165, %r163, 1, 31, -1;
|
735 |
+
mov.b32 %f248, %r165;
|
736 |
+
$L__tmp12:
|
737 |
+
.loc 2 109 28
|
738 |
+
add.f32 %f20, %f17, %f248;
|
739 |
+
.loc 2 110 60
|
740 |
+
mov.b32 %r166, %f20;
|
741 |
+
div.full.f32 %r164, %r165, %r166;
|
742 |
+
mov.b32 %f21, %r164;
|
743 |
+
$L__tmp13:
|
744 |
+
.loc 2 120 46
|
745 |
+
mov.b32 %r234, %f198;
|
746 |
+
shfl.sync.bfly.b32 %r235, %r234, 16, 31, -1;
|
747 |
+
mov.b32 %f249, %r235;
|
748 |
+
mov.b32 %r236, %f202;
|
749 |
+
shfl.sync.bfly.b32 %r237, %r236, 16, 31, -1;
|
750 |
+
mov.b32 %f250, %r237;
|
751 |
+
shfl.sync.bfly.b32 %r168, %r130, 16, 31, -1;
|
752 |
+
mov.b32 %f251, %r168;
|
753 |
+
$L__tmp14:
|
754 |
+
.loc 2 108 21
|
755 |
+
sub.f32 %f252, %f249, %f198;
|
756 |
+
.loc 2 109 28
|
757 |
+
add.f32 %f253, %f251, 0f41000000;
|
758 |
+
.loc 2 110 39
|
759 |
+
setp.eq.f32 %p89, %f253, 0f00000000;
|
760 |
+
.loc 2 110 60
|
761 |
+
mov.b32 %r169, %f253;
|
762 |
+
div.full.f32 %r167, %r168, %r169;
|
763 |
+
mov.b32 %f254, %r167;
|
764 |
+
.loc 2 110 49
|
765 |
+
selp.f32 %f255, 0f00000000, %f254, %p89;
|
766 |
+
.loc 2 112 17
|
767 |
+
fma.rn.f32 %f256, %f252, %f255, %f198;
|
768 |
+
.loc 2 113 15
|
769 |
+
add.f32 %f257, %f202, %f250;
|
770 |
+
.loc 2 113 30
|
771 |
+
mul.f32 %f258, %f252, %f252;
|
772 |
+
.loc 2 113 38
|
773 |
+
mul.f32 %f259, %f258, 0f41000000;
|
774 |
+
.loc 2 113 22
|
775 |
+
fma.rn.f32 %f260, %f259, %f255, %f257;
|
776 |
+
$L__tmp15:
|
777 |
+
.loc 2 120 46
|
778 |
+
mov.b32 %r238, %f256;
|
779 |
+
shfl.sync.bfly.b32 %r239, %r238, 8, 31, -1;
|
780 |
+
mov.b32 %f261, %r239;
|
781 |
+
mov.b32 %r240, %f260;
|
782 |
+
shfl.sync.bfly.b32 %r241, %r240, 8, 31, -1;
|
783 |
+
mov.b32 %f262, %r241;
|
784 |
+
shfl.sync.bfly.b32 %r171, %r169, 8, 31, -1;
|
785 |
+
mov.b32 %f263, %r171;
|
786 |
+
$L__tmp16:
|
787 |
+
.loc 2 108 21
|
788 |
+
sub.f32 %f264, %f261, %f256;
|
789 |
+
.loc 2 109 28
|
790 |
+
add.f32 %f265, %f253, %f263;
|
791 |
+
.loc 2 110 39
|
792 |
+
setp.eq.f32 %p90, %f265, 0f00000000;
|
793 |
+
.loc 2 110 60
|
794 |
+
mov.b32 %r172, %f265;
|
795 |
+
div.full.f32 %r170, %r171, %r172;
|
796 |
+
mov.b32 %f266, %r170;
|
797 |
+
.loc 2 110 49
|
798 |
+
selp.f32 %f267, 0f00000000, %f266, %p90;
|
799 |
+
.loc 2 112 17
|
800 |
+
fma.rn.f32 %f268, %f264, %f267, %f256;
|
801 |
+
.loc 2 113 15
|
802 |
+
add.f32 %f269, %f260, %f262;
|
803 |
+
.loc 2 113 30
|
804 |
+
mul.f32 %f270, %f264, %f264;
|
805 |
+
.loc 2 113 38
|
806 |
+
mul.f32 %f271, %f253, %f270;
|
807 |
+
.loc 2 113 22
|
808 |
+
fma.rn.f32 %f272, %f267, %f271, %f269;
|
809 |
+
$L__tmp17:
|
810 |
+
.loc 2 120 46
|
811 |
+
mov.b32 %r242, %f268;
|
812 |
+
shfl.sync.bfly.b32 %r243, %r242, 4, 31, -1;
|
813 |
+
mov.b32 %f273, %r243;
|
814 |
+
mov.b32 %r244, %f272;
|
815 |
+
shfl.sync.bfly.b32 %r245, %r244, 4, 31, -1;
|
816 |
+
mov.b32 %f274, %r245;
|
817 |
+
shfl.sync.bfly.b32 %r174, %r172, 4, 31, -1;
|
818 |
+
mov.b32 %f275, %r174;
|
819 |
+
$L__tmp18:
|
820 |
+
.loc 2 108 21
|
821 |
+
sub.f32 %f276, %f273, %f268;
|
822 |
+
.loc 2 109 28
|
823 |
+
add.f32 %f277, %f265, %f275;
|
824 |
+
.loc 2 110 39
|
825 |
+
setp.eq.f32 %p91, %f277, 0f00000000;
|
826 |
+
.loc 2 110 60
|
827 |
+
mov.b32 %r175, %f277;
|
828 |
+
div.full.f32 %r173, %r174, %r175;
|
829 |
+
mov.b32 %f278, %r173;
|
830 |
+
.loc 2 110 49
|
831 |
+
selp.f32 %f279, 0f00000000, %f278, %p91;
|
832 |
+
.loc 2 112 17
|
833 |
+
fma.rn.f32 %f280, %f276, %f279, %f268;
|
834 |
+
.loc 2 113 15
|
835 |
+
add.f32 %f281, %f272, %f274;
|
836 |
+
.loc 2 113 30
|
837 |
+
mul.f32 %f282, %f276, %f276;
|
838 |
+
.loc 2 113 38
|
839 |
+
mul.f32 %f283, %f265, %f282;
|
840 |
+
.loc 2 113 22
|
841 |
+
fma.rn.f32 %f284, %f279, %f283, %f281;
|
842 |
+
$L__tmp19:
|
843 |
+
.loc 2 120 46
|
844 |
+
mov.b32 %r246, %f280;
|
845 |
+
shfl.sync.bfly.b32 %r247, %r246, 2, 31, -1;
|
846 |
+
mov.b32 %f285, %r247;
|
847 |
+
mov.b32 %r248, %f284;
|
848 |
+
shfl.sync.bfly.b32 %r249, %r248, 2, 31, -1;
|
849 |
+
mov.b32 %f286, %r249;
|
850 |
+
shfl.sync.bfly.b32 %r177, %r175, 2, 31, -1;
|
851 |
+
mov.b32 %f287, %r177;
|
852 |
+
$L__tmp20:
|
853 |
+
.loc 2 108 21
|
854 |
+
sub.f32 %f288, %f285, %f280;
|
855 |
+
.loc 2 109 28
|
856 |
+
add.f32 %f22, %f277, %f287;
|
857 |
+
.loc 2 110 39
|
858 |
+
setp.eq.f32 %p92, %f22, 0f00000000;
|
859 |
+
.loc 2 110 60
|
860 |
+
mov.b32 %r178, %f22;
|
861 |
+
div.full.f32 %r176, %r177, %r178;
|
862 |
+
mov.b32 %f289, %r176;
|
863 |
+
.loc 2 110 49
|
864 |
+
selp.f32 %f290, 0f00000000, %f289, %p92;
|
865 |
+
.loc 2 112 17
|
866 |
+
fma.rn.f32 %f23, %f288, %f290, %f280;
|
867 |
+
.loc 2 113 15
|
868 |
+
add.f32 %f291, %f284, %f286;
|
869 |
+
.loc 2 113 30
|
870 |
+
mul.f32 %f292, %f288, %f288;
|
871 |
+
.loc 2 113 38
|
872 |
+
mul.f32 %f293, %f277, %f292;
|
873 |
+
.loc 2 113 22
|
874 |
+
fma.rn.f32 %f24, %f290, %f293, %f291;
|
875 |
+
$L__tmp21:
|
876 |
+
.loc 2 120 46
|
877 |
+
mov.b32 %r250, %f23;
|
878 |
+
shfl.sync.bfly.b32 %r7, %r250, 1, 31, -1;
|
879 |
+
mov.b32 %r251, %f24;
|
880 |
+
shfl.sync.bfly.b32 %r8, %r251, 1, 31, -1;
|
881 |
+
shfl.sync.bfly.b32 %r180, %r178, 1, 31, -1;
|
882 |
+
mov.b32 %f294, %r180;
|
883 |
+
$L__tmp22:
|
884 |
+
.loc 2 109 28
|
885 |
+
add.f32 %f25, %f22, %f294;
|
886 |
+
.loc 2 110 60
|
887 |
+
mov.b32 %r181, %f25;
|
888 |
+
div.full.f32 %r179, %r180, %r181;
|
889 |
+
mov.b32 %f26, %r179;
|
890 |
+
$L__tmp23:
|
891 |
+
.loc 1 59 51
|
892 |
+
mov.u32 %r182, 0x0;
|
893 |
+
mov.u32 %r183, 0x0;
|
894 |
+
mov.u32 %r184, 0x0;
|
895 |
+
mov.u32 %r185, 0x0;
|
896 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r182, %r183, %r184, %r185 }, [ %rd80 + 0 ];
|
897 |
+
@!%p93 mov.u32 %r182, %r257;
|
898 |
+
@!%p93 mov.u32 %r183, %r257;
|
899 |
+
@!%p93 mov.u32 %r184, %r257;
|
900 |
+
@!%p93 mov.u32 %r185, %r257;
|
901 |
+
mov.u32 %r190, 0x0;
|
902 |
+
mov.u32 %r191, 0x0;
|
903 |
+
mov.u32 %r192, 0x0;
|
904 |
+
mov.u32 %r193, 0x0;
|
905 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r190, %r191, %r192, %r193 }, [ %rd81 + 0 ];
|
906 |
+
@!%p93 mov.u32 %r190, %r257;
|
907 |
+
@!%p93 mov.u32 %r191, %r257;
|
908 |
+
@!%p93 mov.u32 %r192, %r257;
|
909 |
+
@!%p93 mov.u32 %r193, %r257;
|
910 |
+
mov.u32 %r198, 0x0;
|
911 |
+
mov.u32 %r199, 0x0;
|
912 |
+
mov.u32 %r200, 0x0;
|
913 |
+
mov.u32 %r201, 0x0;
|
914 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r198, %r199, %r200, %r201 }, [ %rd82 + 0 ];
|
915 |
+
@!%p93 mov.u32 %r198, %r257;
|
916 |
+
@!%p93 mov.u32 %r199, %r257;
|
917 |
+
@!%p93 mov.u32 %r200, %r257;
|
918 |
+
@!%p93 mov.u32 %r201, %r257;
|
919 |
+
mov.u32 %r206, 0x0;
|
920 |
+
mov.u32 %r207, 0x0;
|
921 |
+
mov.u32 %r208, 0x0;
|
922 |
+
mov.u32 %r209, 0x0;
|
923 |
+
@%p93 ld.global.L1::evict_last.v4.b32 { %r206, %r207, %r208, %r209 }, [ %rd83 + 0 ];
|
924 |
+
@!%p93 mov.u32 %r206, %r257;
|
925 |
+
@!%p93 mov.u32 %r207, %r257;
|
926 |
+
@!%p93 mov.u32 %r208, %r257;
|
927 |
+
@!%p93 mov.u32 %r209, %r257;
|
928 |
+
.loc 1 60 35
|
929 |
+
mul.wide.u32 %rd96, %r2, 4;
|
930 |
+
add.s64 %rd84, %rd13, %rd96;
|
931 |
+
.loc 1 60 40
|
932 |
+
mov.u32 %r214, 0x0;
|
933 |
+
@%p93 ld.global.L1::evict_last.b32 { %r214 }, [ %rd84 + 0 ];
|
934 |
+
@!%p93 mov.u32 %r214, %r257;
|
935 |
+
.loc 1 64 57
|
936 |
+
@%p39 bra $L__BB0_4;
|
937 |
+
mov.u64 %rd97, assertMessage_1;
|
938 |
+
cvta.global.u64 %rd98, %rd97;
|
939 |
+
mov.u64 %rd99, assertFile_1;
|
940 |
+
cvta.global.u64 %rd100, %rd99;
|
941 |
+
mov.u64 %rd101, assertFunc_1;
|
942 |
+
cvta.global.u64 %rd102, %rd101;
|
943 |
+
{ // callseq 9, 0
|
944 |
+
.reg .b32 temp_param_reg;
|
945 |
+
.param .b64 param0;
|
946 |
+
st.param.b64 [param0+0], %rd98;
|
947 |
+
.param .b64 param1;
|
948 |
+
st.param.b64 [param1+0], %rd100;
|
949 |
+
.param .b32 param2;
|
950 |
+
st.param.b32 [param2+0], %r374;
|
951 |
+
.param .b64 param3;
|
952 |
+
st.param.b64 [param3+0], %rd102;
|
953 |
+
.param .b64 param4;
|
954 |
+
st.param.b64 [param4+0], %rd112;
|
955 |
+
call.uni
|
956 |
+
__assertfail,
|
957 |
+
(
|
958 |
+
param0,
|
959 |
+
param1,
|
960 |
+
param2,
|
961 |
+
param3,
|
962 |
+
param4
|
963 |
+
);
|
964 |
+
} // callseq 9
|
965 |
+
$L__BB0_4:
|
966 |
+
$L__tmp24:
|
967 |
+
.loc 2 120 46
|
968 |
+
mov.b32 %f295, %r8;
|
969 |
+
$L__tmp25:
|
970 |
+
.loc 2 113 15
|
971 |
+
add.f32 %f296, %f24, %f295;
|
972 |
+
$L__tmp26:
|
973 |
+
.loc 2 120 46
|
974 |
+
mov.b32 %f297, %r7;
|
975 |
+
$L__tmp27:
|
976 |
+
.loc 2 108 21
|
977 |
+
sub.f32 %f298, %f297, %f23;
|
978 |
+
.loc 2 113 30
|
979 |
+
mul.f32 %f299, %f298, %f298;
|
980 |
+
.loc 2 113 38
|
981 |
+
mul.f32 %f300, %f22, %f299;
|
982 |
+
.loc 2 110 39
|
983 |
+
setp.eq.f32 %p115, %f25, 0f00000000;
|
984 |
+
.loc 2 110 49
|
985 |
+
selp.f32 %f301, 0f00000000, %f26, %p115;
|
986 |
+
.loc 2 113 22
|
987 |
+
fma.rn.f32 %f302, %f301, %f300, %f296;
|
988 |
+
$L__tmp28:
|
989 |
+
.loc 2 120 46
|
990 |
+
mov.b32 %f303, %r6;
|
991 |
+
$L__tmp29:
|
992 |
+
.loc 2 113 15
|
993 |
+
add.f32 %f304, %f19, %f303;
|
994 |
+
$L__tmp30:
|
995 |
+
.loc 2 120 46
|
996 |
+
mov.b32 %f305, %r5;
|
997 |
+
$L__tmp31:
|
998 |
+
.loc 2 108 21
|
999 |
+
sub.f32 %f306, %f305, %f18;
|
1000 |
+
.loc 2 113 30
|
1001 |
+
mul.f32 %f307, %f306, %f306;
|
1002 |
+
.loc 2 113 38
|
1003 |
+
mul.f32 %f308, %f17, %f307;
|
1004 |
+
.loc 2 110 39
|
1005 |
+
setp.eq.f32 %p116, %f20, 0f00000000;
|
1006 |
+
.loc 2 110 49
|
1007 |
+
selp.f32 %f309, 0f00000000, %f21, %p116;
|
1008 |
+
.loc 2 113 22
|
1009 |
+
fma.rn.f32 %f310, %f309, %f308, %f304;
|
1010 |
+
$L__tmp32:
|
1011 |
+
.loc 1 65 54
|
1012 |
+
mov.u32 %r253, 0x0;
|
1013 |
+
mov.u32 %r254, 0x0;
|
1014 |
+
mov.u32 %r255, 0x0;
|
1015 |
+
mov.u32 %r256, 0x0;
|
1016 |
+
@%p93 ld.global.L1::evict_first.v4.b32 { %r253, %r254, %r255, %r256 }, [ %rd104 + 0 ];
|
1017 |
+
@!%p93 mov.u32 %r253, %r257;
|
1018 |
+
@!%p93 mov.u32 %r254, %r257;
|
1019 |
+
@!%p93 mov.u32 %r255, %r257;
|
1020 |
+
@!%p93 mov.u32 %r256, %r257;
|
1021 |
+
mov.u32 %r261, 0x0;
|
1022 |
+
mov.u32 %r262, 0x0;
|
1023 |
+
mov.u32 %r263, 0x0;
|
1024 |
+
mov.u32 %r264, 0x0;
|
1025 |
+
@%p93 ld.global.L1::evict_first.v4.b32 { %r261, %r262, %r263, %r264 }, [ %rd105 + 0 ];
|
1026 |
+
@!%p93 mov.u32 %r261, %r257;
|
1027 |
+
@!%p93 mov.u32 %r262, %r257;
|
1028 |
+
@!%p93 mov.u32 %r263, %r257;
|
1029 |
+
@!%p93 mov.u32 %r264, %r257;
|
1030 |
+
mov.u32 %r269, 0x0;
|
1031 |
+
mov.u32 %r270, 0x0;
|
1032 |
+
mov.u32 %r271, 0x0;
|
1033 |
+
mov.u32 %r272, 0x0;
|
1034 |
+
@%p93 ld.global.L1::evict_first.v4.b32 { %r269, %r270, %r271, %r272 }, [ %rd106 + 0 ];
|
1035 |
+
@!%p93 mov.u32 %r269, %r257;
|
1036 |
+
@!%p93 mov.u32 %r270, %r257;
|
1037 |
+
@!%p93 mov.u32 %r271, %r257;
|
1038 |
+
@!%p93 mov.u32 %r272, %r257;
|
1039 |
+
mov.u32 %r277, 0x0;
|
1040 |
+
mov.u32 %r278, 0x0;
|
1041 |
+
mov.u32 %r279, 0x0;
|
1042 |
+
mov.u32 %r280, 0x0;
|
1043 |
+
@%p93 ld.global.L1::evict_first.v4.b32 { %r277, %r278, %r279, %r280 }, [ %rd107 + 0 ];
|
1044 |
+
@!%p93 mov.u32 %r277, %r257;
|
1045 |
+
@!%p93 mov.u32 %r278, %r257;
|
1046 |
+
@!%p93 mov.u32 %r279, %r257;
|
1047 |
+
@!%p93 mov.u32 %r280, %r257;
|
1048 |
+
.loc 1 69 23
|
1049 |
+
mov.b32 %r286, %f310;
|
1050 |
+
mov.b32 %r287, 1132462080;
|
1051 |
+
div.full.f32 %r285, %r286, %r287;
|
1052 |
+
mov.b32 %f311, %r285;
|
1053 |
+
mov.b32 %r310, %f302;
|
1054 |
+
div.full.f32 %r309, %r310, %r287;
|
1055 |
+
mov.b32 %f312, %r309;
|
1056 |
+
.loc 1 71 24
|
1057 |
+
add.f32 %f313, %f311, 0f3727C5AC;
|
1058 |
+
add.f32 %f314, %f312, 0f3727C5AC;
|
1059 |
+
.loc 1 72 30
|
1060 |
+
rsqrt.approx.ftz.f32 %f315, %f313;
|
1061 |
+
rsqrt.approx.ftz.f32 %f316, %f314;
|
1062 |
+
.loc 1 65 54
|
1063 |
+
mov.b32 %f317, %r280;
|
1064 |
+
.loc 1 59 51
|
1065 |
+
mov.b32 %f318, %r209;
|
1066 |
+
.loc 1 66 24
|
1067 |
+
add.f32 %f319, %f318, %f317;
|
1068 |
+
$L__tmp33:
|
1069 |
+
.loc 2 112 17
|
1070 |
+
fma.rn.f32 %f320, %f298, %f301, %f23;
|
1071 |
+
$L__tmp34:
|
1072 |
+
.loc 1 67 24
|
1073 |
+
sub.f32 %f321, %f319, %f320;
|
1074 |
+
.loc 1 65 54
|
1075 |
+
mov.b32 %f322, %r279;
|
1076 |
+
.loc 1 59 51
|
1077 |
+
mov.b32 %f323, %r208;
|
1078 |
+
.loc 1 66 24
|
1079 |
+
add.f32 %f324, %f323, %f322;
|
1080 |
+
.loc 1 67 24
|
1081 |
+
sub.f32 %f325, %f324, %f320;
|
1082 |
+
.loc 1 65 54
|
1083 |
+
mov.b32 %f326, %r278;
|
1084 |
+
.loc 1 59 51
|
1085 |
+
mov.b32 %f327, %r207;
|
1086 |
+
.loc 1 66 24
|
1087 |
+
add.f32 %f328, %f327, %f326;
|
1088 |
+
.loc 1 67 24
|
1089 |
+
sub.f32 %f329, %f328, %f320;
|
1090 |
+
.loc 1 65 54
|
1091 |
+
mov.b32 %f330, %r277;
|
1092 |
+
.loc 1 59 51
|
1093 |
+
mov.b32 %f331, %r206;
|
1094 |
+
.loc 1 66 24
|
1095 |
+
add.f32 %f332, %f331, %f330;
|
1096 |
+
.loc 1 67 24
|
1097 |
+
sub.f32 %f333, %f332, %f320;
|
1098 |
+
.loc 1 65 54
|
1099 |
+
mov.b32 %f334, %r272;
|
1100 |
+
.loc 1 59 51
|
1101 |
+
mov.b32 %f335, %r201;
|
1102 |
+
.loc 1 66 24
|
1103 |
+
add.f32 %f336, %f335, %f334;
|
1104 |
+
.loc 1 67 24
|
1105 |
+
sub.f32 %f337, %f336, %f320;
|
1106 |
+
.loc 1 65 54
|
1107 |
+
mov.b32 %f338, %r271;
|
1108 |
+
.loc 1 59 51
|
1109 |
+
mov.b32 %f339, %r200;
|
1110 |
+
.loc 1 66 24
|
1111 |
+
add.f32 %f340, %f339, %f338;
|
1112 |
+
.loc 1 67 24
|
1113 |
+
sub.f32 %f341, %f340, %f320;
|
1114 |
+
.loc 1 65 54
|
1115 |
+
mov.b32 %f342, %r270;
|
1116 |
+
.loc 1 59 51
|
1117 |
+
mov.b32 %f343, %r199;
|
1118 |
+
.loc 1 66 24
|
1119 |
+
add.f32 %f344, %f343, %f342;
|
1120 |
+
.loc 1 67 24
|
1121 |
+
sub.f32 %f345, %f344, %f320;
|
1122 |
+
.loc 1 65 54
|
1123 |
+
mov.b32 %f346, %r269;
|
1124 |
+
.loc 1 59 51
|
1125 |
+
mov.b32 %f347, %r198;
|
1126 |
+
.loc 1 66 24
|
1127 |
+
add.f32 %f348, %f347, %f346;
|
1128 |
+
.loc 1 67 24
|
1129 |
+
sub.f32 %f349, %f348, %f320;
|
1130 |
+
.loc 1 65 54
|
1131 |
+
mov.b32 %f350, %r264;
|
1132 |
+
.loc 1 59 51
|
1133 |
+
mov.b32 %f351, %r193;
|
1134 |
+
.loc 1 66 24
|
1135 |
+
add.f32 %f352, %f351, %f350;
|
1136 |
+
$L__tmp35:
|
1137 |
+
.loc 2 112 17
|
1138 |
+
fma.rn.f32 %f353, %f306, %f309, %f18;
|
1139 |
+
$L__tmp36:
|
1140 |
+
.loc 1 67 24
|
1141 |
+
sub.f32 %f354, %f352, %f353;
|
1142 |
+
.loc 1 65 54
|
1143 |
+
mov.b32 %f355, %r263;
|
1144 |
+
.loc 1 59 51
|
1145 |
+
mov.b32 %f356, %r192;
|
1146 |
+
.loc 1 66 24
|
1147 |
+
add.f32 %f357, %f356, %f355;
|
1148 |
+
.loc 1 67 24
|
1149 |
+
sub.f32 %f358, %f357, %f353;
|
1150 |
+
.loc 1 65 54
|
1151 |
+
mov.b32 %f359, %r262;
|
1152 |
+
.loc 1 59 51
|
1153 |
+
mov.b32 %f360, %r191;
|
1154 |
+
.loc 1 66 24
|
1155 |
+
add.f32 %f361, %f360, %f359;
|
1156 |
+
.loc 1 67 24
|
1157 |
+
sub.f32 %f362, %f361, %f353;
|
1158 |
+
.loc 1 65 54
|
1159 |
+
mov.b32 %f363, %r261;
|
1160 |
+
.loc 1 59 51
|
1161 |
+
mov.b32 %f364, %r190;
|
1162 |
+
.loc 1 66 24
|
1163 |
+
add.f32 %f365, %f364, %f363;
|
1164 |
+
.loc 1 67 24
|
1165 |
+
sub.f32 %f366, %f365, %f353;
|
1166 |
+
.loc 1 65 54
|
1167 |
+
mov.b32 %f367, %r256;
|
1168 |
+
.loc 1 59 51
|
1169 |
+
mov.b32 %f368, %r185;
|
1170 |
+
.loc 1 66 24
|
1171 |
+
add.f32 %f369, %f368, %f367;
|
1172 |
+
.loc 1 67 24
|
1173 |
+
sub.f32 %f370, %f369, %f353;
|
1174 |
+
.loc 1 65 54
|
1175 |
+
mov.b32 %f371, %r255;
|
1176 |
+
.loc 1 59 51
|
1177 |
+
mov.b32 %f372, %r184;
|
1178 |
+
.loc 1 66 24
|
1179 |
+
add.f32 %f373, %f372, %f371;
|
1180 |
+
.loc 1 67 24
|
1181 |
+
sub.f32 %f374, %f373, %f353;
|
1182 |
+
.loc 1 65 54
|
1183 |
+
mov.b32 %f375, %r254;
|
1184 |
+
.loc 1 59 51
|
1185 |
+
mov.b32 %f376, %r183;
|
1186 |
+
.loc 1 66 24
|
1187 |
+
add.f32 %f377, %f376, %f375;
|
1188 |
+
.loc 1 67 24
|
1189 |
+
sub.f32 %f378, %f377, %f353;
|
1190 |
+
.loc 1 65 54
|
1191 |
+
mov.b32 %f379, %r253;
|
1192 |
+
.loc 1 59 51
|
1193 |
+
mov.b32 %f380, %r182;
|
1194 |
+
.loc 1 66 24
|
1195 |
+
add.f32 %f381, %f380, %f379;
|
1196 |
+
.loc 1 67 24
|
1197 |
+
sub.f32 %f382, %f381, %f353;
|
1198 |
+
.loc 1 73 24
|
1199 |
+
mul.f32 %f383, %f382, %f315;
|
1200 |
+
mul.f32 %f384, %f378, %f315;
|
1201 |
+
mul.f32 %f385, %f374, %f315;
|
1202 |
+
mul.f32 %f386, %f370, %f315;
|
1203 |
+
mul.f32 %f387, %f366, %f315;
|
1204 |
+
mul.f32 %f388, %f362, %f315;
|
1205 |
+
mul.f32 %f389, %f358, %f315;
|
1206 |
+
mul.f32 %f390, %f354, %f315;
|
1207 |
+
mul.f32 %f391, %f349, %f316;
|
1208 |
+
mul.f32 %f392, %f345, %f316;
|
1209 |
+
mul.f32 %f393, %f341, %f316;
|
1210 |
+
mul.f32 %f394, %f337, %f316;
|
1211 |
+
mul.f32 %f395, %f333, %f316;
|
1212 |
+
mul.f32 %f396, %f329, %f316;
|
1213 |
+
mul.f32 %f397, %f325, %f316;
|
1214 |
+
mul.f32 %f398, %f321, %f316;
|
1215 |
+
.loc 1 74 24
|
1216 |
+
shl.b32 %r357, %r2, 2;
|
1217 |
+
mov.u32 %r358, global_smem;
|
1218 |
+
add.s32 %r359, %r358, %r357;
|
1219 |
+
st.shared.u32 [%r359], %r214;
|
1220 |
+
bar.sync 0;
|
1221 |
+
shl.b32 %r360, %r1, 2;
|
1222 |
+
add.s32 %r361, %r358, %r360;
|
1223 |
+
ld.shared.v4.f32 {%f399, %f400, %f401, %f402}, [%r361];
|
1224 |
+
ld.shared.v4.f32 {%f403, %f404, %f405, %f406}, [%r361+16];
|
1225 |
+
mul.f32 %f407, %f383, %f399;
|
1226 |
+
mul.f32 %f408, %f384, %f400;
|
1227 |
+
mul.f32 %f409, %f385, %f401;
|
1228 |
+
mul.f32 %f410, %f386, %f402;
|
1229 |
+
mul.f32 %f411, %f387, %f403;
|
1230 |
+
mul.f32 %f412, %f388, %f404;
|
1231 |
+
mul.f32 %f413, %f389, %f405;
|
1232 |
+
mul.f32 %f414, %f390, %f406;
|
1233 |
+
mul.f32 %f415, %f391, %f399;
|
1234 |
+
mul.f32 %f416, %f392, %f400;
|
1235 |
+
mul.f32 %f417, %f393, %f401;
|
1236 |
+
mul.f32 %f418, %f394, %f402;
|
1237 |
+
mul.f32 %f419, %f395, %f403;
|
1238 |
+
mul.f32 %f420, %f396, %f404;
|
1239 |
+
mul.f32 %f421, %f397, %f405;
|
1240 |
+
mul.f32 %f422, %f398, %f406;
|
1241 |
+
.loc 1 76 39
|
1242 |
+
shl.b32 %r362, %r3, 8;
|
1243 |
+
shl.b32 %r363, %r4, 8;
|
1244 |
+
.loc 1 76 35
|
1245 |
+
or.b32 %r364, %r362, %r1;
|
1246 |
+
or.b32 %r365, %r363, %r1;
|
1247 |
+
.loc 1 76 29
|
1248 |
+
mul.wide.s32 %rd110, %r364, 2;
|
1249 |
+
add.s64 %rd108, %rd14, %rd110;
|
1250 |
+
mul.wide.s32 %rd111, %r365, 2;
|
1251 |
+
add.s64 %rd109, %rd14, %rd111;
|
1252 |
+
.loc 1 76 52
|
1253 |
+
mov.b32 %r333, %f407;
|
1254 |
+
cvt.rn.bf16.f32 %rs1, %r333;
|
1255 |
+
mov.b32 %r334, %f408;
|
1256 |
+
cvt.rn.bf16.f32 %rs2, %r334;
|
1257 |
+
mov.b32 %r335, %f409;
|
1258 |
+
cvt.rn.bf16.f32 %rs3, %r335;
|
1259 |
+
mov.b32 %r336, %f410;
|
1260 |
+
cvt.rn.bf16.f32 %rs4, %r336;
|
1261 |
+
mov.b32 %r337, %f411;
|
1262 |
+
cvt.rn.bf16.f32 %rs5, %r337;
|
1263 |
+
mov.b32 %r338, %f412;
|
1264 |
+
cvt.rn.bf16.f32 %rs6, %r338;
|
1265 |
+
mov.b32 %r339, %f413;
|
1266 |
+
cvt.rn.bf16.f32 %rs7, %r339;
|
1267 |
+
mov.b32 %r340, %f414;
|
1268 |
+
cvt.rn.bf16.f32 %rs8, %r340;
|
1269 |
+
mov.b32 %r341, %f415;
|
1270 |
+
cvt.rn.bf16.f32 %rs9, %r341;
|
1271 |
+
mov.b32 %r342, %f416;
|
1272 |
+
cvt.rn.bf16.f32 %rs10, %r342;
|
1273 |
+
mov.b32 %r343, %f417;
|
1274 |
+
cvt.rn.bf16.f32 %rs11, %r343;
|
1275 |
+
mov.b32 %r344, %f418;
|
1276 |
+
cvt.rn.bf16.f32 %rs12, %r344;
|
1277 |
+
mov.b32 %r345, %f419;
|
1278 |
+
cvt.rn.bf16.f32 %rs13, %r345;
|
1279 |
+
mov.b32 %r346, %f420;
|
1280 |
+
cvt.rn.bf16.f32 %rs14, %r346;
|
1281 |
+
mov.b32 %r347, %f421;
|
1282 |
+
cvt.rn.bf16.f32 %rs15, %r347;
|
1283 |
+
mov.b32 %r348, %f422;
|
1284 |
+
cvt.rn.bf16.f32 %rs16, %r348;
|
1285 |
+
mov.b32 %r366, {%rs1, %rs2};
|
1286 |
+
mov.b32 %r367, {%rs3, %rs4};
|
1287 |
+
mov.b32 %r368, {%rs5, %rs6};
|
1288 |
+
mov.b32 %r369, {%rs7, %rs8};
|
1289 |
+
@%p93 st.global.v4.b32 [ %rd108 + 0 ], { %r366, %r367, %r368, %r369 };
|
1290 |
+
mov.b32 %r370, {%rs9, %rs10};
|
1291 |
+
mov.b32 %r371, {%rs11, %rs12};
|
1292 |
+
mov.b32 %r372, {%rs13, %rs14};
|
1293 |
+
mov.b32 %r373, {%rs15, %rs16};
|
1294 |
+
@%p93 st.global.v4.b32 [ %rd109 + 0 ], { %r370, %r371, %r372, %r373 };
|
1295 |
+
.loc 1 55 4
|
1296 |
+
ret;
|
1297 |
+
$L__tmp37:
|
1298 |
+
$L__func_end0:
|
1299 |
+
|
1300 |
+
}
|
1301 |
+
// .globl __nv_rsqrtf
|
1302 |
+
.visible .func (.param .b32 func_retval0) __nv_rsqrtf(
|
1303 |
+
.param .b32 __nv_rsqrtf_param_0
|
1304 |
+
)
|
1305 |
+
{
|
1306 |
+
.reg .f32 %f<3>;
|
1307 |
+
$L__func_begin1:
|
1308 |
+
|
1309 |
+
ld.param.f32 %f1, [__nv_rsqrtf_param_0];
|
1310 |
+
rsqrt.approx.ftz.f32 %f2, %f1;
|
1311 |
+
st.param.f32 [func_retval0+0], %f2;
|
1312 |
+
ret;
|
1313 |
+
$L__func_end1:
|
1314 |
+
|
1315 |
+
}
|
1316 |
+
.file 1 "/tmp/torchinductor_root/lh/clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py"
|
1317 |
+
.file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
|
1318 |
+
.section .debug_abbrev
|
1319 |
+
{
|
1320 |
+
.b8 1
|
1321 |
+
.b8 17
|
1322 |
+
.b8 1
|
1323 |
+
.b8 37
|
1324 |
+
.b8 8
|
1325 |
+
.b8 19
|
1326 |
+
.b8 5
|
1327 |
+
.b8 3
|
1328 |
+
.b8 8
|
1329 |
+
.b8 16
|
1330 |
+
.b8 6
|
1331 |
+
.b8 27
|
1332 |
+
.b8 8
|
1333 |
+
.b8 180
|
1334 |
+
.b8 66
|
1335 |
+
.b8 12
|
1336 |
+
.b8 17
|
1337 |
+
.b8 1
|
1338 |
+
.b8 18
|
1339 |
+
.b8 1
|
1340 |
+
.b8 0
|
1341 |
+
.b8 0
|
1342 |
+
.b8 2
|
1343 |
+
.b8 46
|
1344 |
+
.b8 0
|
1345 |
+
.b8 135
|
1346 |
+
.b8 64
|
1347 |
+
.b8 8
|
1348 |
+
.b8 3
|
1349 |
+
.b8 8
|
1350 |
+
.b8 58
|
1351 |
+
.b8 11
|
1352 |
+
.b8 59
|
1353 |
+
.b8 11
|
1354 |
+
.b8 63
|
1355 |
+
.b8 12
|
1356 |
+
.b8 32
|
1357 |
+
.b8 11
|
1358 |
+
.b8 0
|
1359 |
+
.b8 0
|
1360 |
+
.b8 3
|
1361 |
+
.b8 46
|
1362 |
+
.b8 1
|
1363 |
+
.b8 17
|
1364 |
+
.b8 1
|
1365 |
+
.b8 18
|
1366 |
+
.b8 1
|
1367 |
+
.b8 64
|
1368 |
+
.b8 10
|
1369 |
+
.b8 49
|
1370 |
+
.b8 19
|
1371 |
+
.b8 0
|
1372 |
+
.b8 0
|
1373 |
+
.b8 4
|
1374 |
+
.b8 29
|
1375 |
+
.b8 0
|
1376 |
+
.b8 49
|
1377 |
+
.b8 19
|
1378 |
+
.b8 17
|
1379 |
+
.b8 1
|
1380 |
+
.b8 18
|
1381 |
+
.b8 1
|
1382 |
+
.b8 88
|
1383 |
+
.b8 11
|
1384 |
+
.b8 89
|
1385 |
+
.b8 11
|
1386 |
+
.b8 87
|
1387 |
+
.b8 11
|
1388 |
+
.b8 0
|
1389 |
+
.b8 0
|
1390 |
+
.b8 5
|
1391 |
+
.b8 29
|
1392 |
+
.b8 1
|
1393 |
+
.b8 49
|
1394 |
+
.b8 19
|
1395 |
+
.b8 17
|
1396 |
+
.b8 1
|
1397 |
+
.b8 18
|
1398 |
+
.b8 1
|
1399 |
+
.b8 88
|
1400 |
+
.b8 11
|
1401 |
+
.b8 89
|
1402 |
+
.b8 11
|
1403 |
+
.b8 87
|
1404 |
+
.b8 11
|
1405 |
+
.b8 0
|
1406 |
+
.b8 0
|
1407 |
+
.b8 0
|
1408 |
+
}
|
1409 |
+
.section .debug_info
|
1410 |
+
{
|
1411 |
+
.b32 298
|
1412 |
+
.b8 2
|
1413 |
+
.b8 0
|
1414 |
+
.b32 .debug_abbrev
|
1415 |
+
.b8 8
|
1416 |
+
.b8 1
|
1417 |
+
.b8 116
|
1418 |
+
.b8 114
|
1419 |
+
.b8 105
|
1420 |
+
.b8 116
|
1421 |
+
.b8 111
|
1422 |
+
.b8 110
|
1423 |
+
.b8 0
|
1424 |
+
.b8 2
|
1425 |
+
.b8 0
|
1426 |
+
.b8 99
|
1427 |
+
.b8 108
|
1428 |
+
.b8 104
|
1429 |
+
.b8 101
|
1430 |
+
.b8 52
|
1431 |
+
.b8 97
|
1432 |
+
.b8 51
|
1433 |
+
.b8 115
|
1434 |
+
.b8 116
|
1435 |
+
.b8 118
|
1436 |
+
.b8 117
|
1437 |
+
.b8 102
|
1438 |
+
.b8 120
|
1439 |
+
.b8 97
|
1440 |
+
.b8 102
|
1441 |
+
.b8 109
|
1442 |
+
.b8 113
|
1443 |
+
.b8 51
|
1444 |
+
.b8 107
|
1445 |
+
.b8 107
|
1446 |
+
.b8 53
|
1447 |
+
.b8 104
|
1448 |
+
.b8 111
|
1449 |
+
.b8 100
|
1450 |
+
.b8 97
|
1451 |
+
.b8 122
|
1452 |
+
.b8 122
|
1453 |
+
.b8 50
|
1454 |
+
.b8 101
|
1455 |
+
.b8 102
|
1456 |
+
.b8 99
|
1457 |
+
.b8 116
|
1458 |
+
.b8 102
|
1459 |
+
.b8 102
|
1460 |
+
.b8 116
|
1461 |
+
.b8 101
|
1462 |
+
.b8 54
|
1463 |
+
.b8 52
|
1464 |
+
.b8 54
|
1465 |
+
.b8 122
|
1466 |
+
.b8 110
|
1467 |
+
.b8 106
|
1468 |
+
.b8 100
|
1469 |
+
.b8 110
|
1470 |
+
.b8 118
|
1471 |
+
.b8 51
|
1472 |
+
.b8 108
|
1473 |
+
.b8 113
|
1474 |
+
.b8 105
|
1475 |
+
.b8 53
|
1476 |
+
.b8 111
|
1477 |
+
.b8 97
|
1478 |
+
.b8 46
|
1479 |
+
.b8 112
|
1480 |
+
.b8 121
|
1481 |
+
.b8 0
|
1482 |
+
.b32 .debug_line
|
1483 |
+
.b8 47
|
1484 |
+
.b8 116
|
1485 |
+
.b8 109
|
1486 |
+
.b8 112
|
1487 |
+
.b8 47
|
1488 |
+
.b8 116
|
1489 |
+
.b8 111
|
1490 |
+
.b8 114
|
1491 |
+
.b8 99
|
1492 |
+
.b8 104
|
1493 |
+
.b8 105
|
1494 |
+
.b8 110
|
1495 |
+
.b8 100
|
1496 |
+
.b8 117
|
1497 |
+
.b8 99
|
1498 |
+
.b8 116
|
1499 |
+
.b8 111
|
1500 |
+
.b8 114
|
1501 |
+
.b8 95
|
1502 |
+
.b8 114
|
1503 |
+
.b8 111
|
1504 |
+
.b8 111
|
1505 |
+
.b8 116
|
1506 |
+
.b8 47
|
1507 |
+
.b8 108
|
1508 |
+
.b8 104
|
1509 |
+
.b8 0
|
1510 |
+
.b8 1
|
1511 |
+
.b64 $L__func_begin0
|
1512 |
+
.b64 $L__func_end0
|
1513 |
+
.b8 2
|
1514 |
+
.b8 116
|
1515 |
+
.b8 114
|
1516 |
+
.b8 105
|
1517 |
+
.b8 116
|
1518 |
+
.b8 111
|
1519 |
+
.b8 110
|
1520 |
+
.b8 95
|
1521 |
+
.b8 95
|
1522 |
+
.b8 48
|
1523 |
+
.b8 100
|
1524 |
+
.b8 49
|
1525 |
+
.b8 100
|
1526 |
+
.b8 50
|
1527 |
+
.b8 100
|
1528 |
+
.b8 51
|
1529 |
+
.b8 100
|
1530 |
+
.b8 52
|
1531 |
+
.b8 100
|
1532 |
+
.b8 53
|
1533 |
+
.b8 100
|
1534 |
+
.b8 101
|
1535 |
+
.b8 54
|
1536 |
+
.b8 100
|
1537 |
+
.b8 101
|
1538 |
+
.b8 0
|
1539 |
+
.b8 116
|
1540 |
+
.b8 114
|
1541 |
+
.b8 105
|
1542 |
+
.b8 116
|
1543 |
+
.b8 111
|
1544 |
+
.b8 110
|
1545 |
+
.b8 95
|
1546 |
+
.b8 95
|
1547 |
+
.b8 48
|
1548 |
+
.b8 100
|
1549 |
+
.b8 49
|
1550 |
+
.b8 100
|
1551 |
+
.b8 50
|
1552 |
+
.b8 100
|
1553 |
+
.b8 51
|
1554 |
+
.b8 100
|
1555 |
+
.b8 52
|
1556 |
+
.b8 100
|
1557 |
+
.b8 53
|
1558 |
+
.b8 100
|
1559 |
+
.b8 101
|
1560 |
+
.b8 54
|
1561 |
+
.b8 100
|
1562 |
+
.b8 101
|
1563 |
+
.b8 0
|
1564 |
+
.b8 1
|
1565 |
+
.b8 18
|
1566 |
+
.b8 1
|
1567 |
+
.b8 1
|
1568 |
+
.b8 3
|
1569 |
+
.b64 $L__func_begin0
|
1570 |
+
.b64 $L__func_end0
|
1571 |
+
.b8 1
|
1572 |
+
.b8 156
|
1573 |
+
.b32 125
|
1574 |
+
.b8 4
|
1575 |
+
.b32 125
|
1576 |
+
.b64 $L__tmp1
|
1577 |
+
.b64 $L__tmp2
|
1578 |
+
.b8 2
|
1579 |
+
.b8 44
|
1580 |
+
.b8 38
|
1581 |
+
.b8 5
|
1582 |
+
.b32 125
|
1583 |
+
.b64 $L__tmp2
|
1584 |
+
.b64 $L__tmp36
|
1585 |
+
.b8 2
|
1586 |
+
.b8 50
|
1587 |
+
.b8 41
|
1588 |
+
.b8 4
|
1589 |
+
.b32 125
|
1590 |
+
.b64 $L__tmp2
|
1591 |
+
.b64 $L__tmp36
|
1592 |
+
.b8 2
|
1593 |
+
.b8 120
|
1594 |
+
.b8 46
|
1595 |
+
.b8 0
|
1596 |
+
.b8 4
|
1597 |
+
.b32 125
|
1598 |
+
.b64 $L__tmp3
|
1599 |
+
.b64 $L__tmp31
|
1600 |
+
.b8 2
|
1601 |
+
.b8 50
|
1602 |
+
.b8 41
|
1603 |
+
.b8 0
|
1604 |
+
.b8 0
|
1605 |
+
}
|
1606 |
+
.section .debug_pubnames
|
1607 |
+
{
|
1608 |
+
.b32 $L__pubNames_end0-$L__pubNames_start0
|
1609 |
+
$L__pubNames_start0:
|
1610 |
+
.b8 2
|
1611 |
+
.b8 0
|
1612 |
+
.b32 .debug_info
|
1613 |
+
.b32 302
|
1614 |
+
.b32 125
|
1615 |
+
.b8 116
|
1616 |
+
.b8 114
|
1617 |
+
.b8 105
|
1618 |
+
.b8 116
|
1619 |
+
.b8 111
|
1620 |
+
.b8 110
|
1621 |
+
.b8 95
|
1622 |
+
.b8 95
|
1623 |
+
.b8 48
|
1624 |
+
.b8 100
|
1625 |
+
.b8 49
|
1626 |
+
.b8 100
|
1627 |
+
.b8 50
|
1628 |
+
.b8 100
|
1629 |
+
.b8 51
|
1630 |
+
.b8 100
|
1631 |
+
.b8 52
|
1632 |
+
.b8 100
|
1633 |
+
.b8 53
|
1634 |
+
.b8 100
|
1635 |
+
.b8 101
|
1636 |
+
.b8 54
|
1637 |
+
.b8 100
|
1638 |
+
.b8 101
|
1639 |
+
.b8 0
|
1640 |
+
.b32 0
|
1641 |
+
$L__pubNames_end0:
|
1642 |
+
}
|
1643 |
+
.section .debug_pubtypes
|
1644 |
+
{
|
1645 |
+
.b32 $L__pubTypes_end0-$L__pubTypes_start0
|
1646 |
+
$L__pubTypes_start0:
|
1647 |
+
.b8 2
|
1648 |
+
.b8 0
|
1649 |
+
.b32 .debug_info
|
1650 |
+
.b32 302
|
1651 |
+
.b32 0
|
1652 |
+
$L__pubTypes_end0:
|
1653 |
+
}
|
1654 |
+
.section .debug_loc { }
|
.triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ttir
ADDED
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
module {
|
2 |
+
tt.func public @triton__0d1d2d3d4d5de6de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
3 |
+
%cst = arith.constant dense<1.000000e+00> : tensor<1x256xf32>
|
4 |
+
%cst_0 = arith.constant dense<0.000000e+00> : tensor<1x256xf32>
|
5 |
+
%cst_1 = arith.constant 0.000000e+00 : f32
|
6 |
+
%cst_2 = arith.constant dense<256> : tensor<16x1xi64>
|
7 |
+
%cst_3 = arith.constant dense<50257> : tensor<16x1xi64>
|
8 |
+
%cst_4 = arith.constant dense<0> : tensor<16x1xi64>
|
9 |
+
%cst_5 = arith.constant dense<9.99999974E-6> : tensor<16x1xf32>
|
10 |
+
%cst_6 = arith.constant dense<2.560000e+02> : tensor<16x1xf32>
|
11 |
+
%cst_7 = arith.constant dense<0.000000e+00> : tensor<16x256xf32>
|
12 |
+
%cst_8 = arith.constant dense<256> : tensor<16x1xi32>
|
13 |
+
%cst_9 = arith.constant dense<256> : tensor<1x256xi32>
|
14 |
+
%cst_10 = arith.constant dense<512> : tensor<16x1xi32>
|
15 |
+
%c16_i32 = arith.constant 16 : i32
|
16 |
+
%0 = tt.get_program_id x : i32
|
17 |
+
%1 = arith.muli %0, %c16_i32 : i32
|
18 |
+
%2 = tt.make_range {end = 16 : i32, start = 0 : i32} : tensor<16xi32>
|
19 |
+
%3 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<16xi32>) -> tensor<16x1xi32>
|
20 |
+
%4 = tt.splat %1 : (i32) -> tensor<16x1xi32>
|
21 |
+
%5 = arith.addi %4, %3 : tensor<16x1xi32>
|
22 |
+
%6 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32>
|
23 |
+
%7 = tt.expand_dims %6 {axis = 0 : i32} : (tensor<256xi32>) -> tensor<1x256xi32>
|
24 |
+
%8 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<16x1x!tt.ptr<i64, 1>>
|
25 |
+
%9 = tt.addptr %8, %5 : tensor<16x1x!tt.ptr<i64, 1>>, tensor<16x1xi32>
|
26 |
+
%10 = tt.load %9 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x1xi64>
|
27 |
+
%11 = arith.remsi %5, %cst_10 : tensor<16x1xi32>
|
28 |
+
%12 = arith.cmpi slt, %7, %cst_9 : tensor<1x256xi32>
|
29 |
+
%13 = arith.muli %11, %cst_8 : tensor<16x1xi32>
|
30 |
+
%14 = tt.broadcast %7 : (tensor<1x256xi32>) -> tensor<16x256xi32>
|
31 |
+
%15 = tt.broadcast %13 : (tensor<16x1xi32>) -> tensor<16x256xi32>
|
32 |
+
%16 = arith.addi %14, %15 : tensor<16x256xi32>
|
33 |
+
%17 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<16x256x!tt.ptr<f32, 1>>
|
34 |
+
%18 = tt.addptr %17, %16 : tensor<16x256x!tt.ptr<f32, 1>>, tensor<16x256xi32>
|
35 |
+
%19 = tt.broadcast %12 : (tensor<1x256xi1>) -> tensor<16x256xi1>
|
36 |
+
%20 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x256xf32>
|
37 |
+
%21 = arith.addi %10, %cst_3 : tensor<16x1xi64>
|
38 |
+
%22 = arith.cmpi slt, %10, %cst_4 : tensor<16x1xi64>
|
39 |
+
%23 = arith.select %22, %21, %10 : tensor<16x1xi1>, tensor<16x1xi64>
|
40 |
+
%24 = arith.cmpi sge, %23, %cst_4 : tensor<16x1xi64>
|
41 |
+
%25 = arith.cmpi slt, %23, %cst_3 : tensor<16x1xi64>
|
42 |
+
%26 = arith.andi %24, %25 : tensor<16x1xi1>
|
43 |
+
tt.assert %26, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<16x1xi1>
|
44 |
+
%27 = arith.muli %23, %cst_2 : tensor<16x1xi64>
|
45 |
+
%28 = tt.broadcast %27 : (tensor<16x1xi64>) -> tensor<16x256xi64>
|
46 |
+
%29 = arith.extsi %7 : tensor<1x256xi32> to tensor<1x256xi64>
|
47 |
+
%30 = tt.broadcast %29 : (tensor<1x256xi64>) -> tensor<16x256xi64>
|
48 |
+
%31 = arith.addi %30, %28 : tensor<16x256xi64>
|
49 |
+
%32 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<16x256x!tt.ptr<f32, 1>>
|
50 |
+
%33 = tt.addptr %32, %31 : tensor<16x256x!tt.ptr<f32, 1>>, tensor<16x256xi64>
|
51 |
+
%34 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x256xf32>
|
52 |
+
%35 = arith.addf %34, %20 : tensor<16x256xf32>
|
53 |
+
%36 = arith.addf %35, %cst_7 : tensor<16x256xf32>
|
54 |
+
%37 = arith.subf %35, %36 : tensor<16x256xf32>
|
55 |
+
%38 = arith.mulf %35, %37 : tensor<16x256xf32>
|
56 |
+
%39 = arith.addf %38, %cst_7 : tensor<16x256xf32>
|
57 |
+
%40 = arith.select %19, %36, %cst_7 : tensor<16x256xi1>, tensor<16x256xf32>
|
58 |
+
%41 = arith.select %19, %39, %cst_7 : tensor<16x256xi1>, tensor<16x256xf32>
|
59 |
+
%42 = arith.select %12, %cst, %cst_0 : tensor<1x256xi1>, tensor<1x256xf32>
|
60 |
+
%43 = tt.broadcast %42 : (tensor<1x256xf32>) -> tensor<16x256xf32>
|
61 |
+
%44:3 = "tt.reduce"(%40, %41, %43) <{axis = 1 : i32}> ({
|
62 |
+
^bb0(%arg7: f32, %arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32):
|
63 |
+
%68 = arith.subf %arg10, %arg7 : f32
|
64 |
+
%69 = arith.addf %arg9, %arg12 : f32
|
65 |
+
%70 = arith.cmpf oeq, %69, %cst_1 : f32
|
66 |
+
%71 = arith.divf %arg12, %69 : f32
|
67 |
+
%72 = arith.select %70, %cst_1, %71 : f32
|
68 |
+
%73 = arith.mulf %68, %72 : f32
|
69 |
+
%74 = arith.addf %arg7, %73 : f32
|
70 |
+
%75 = arith.addf %arg8, %arg11 : f32
|
71 |
+
%76 = arith.mulf %68, %68 : f32
|
72 |
+
%77 = arith.mulf %76, %arg9 : f32
|
73 |
+
%78 = arith.mulf %77, %72 : f32
|
74 |
+
%79 = arith.addf %75, %78 : f32
|
75 |
+
tt.reduce.return %74, %79, %69 : f32, f32, f32
|
76 |
+
}) : (tensor<16x256xf32>, tensor<16x256xf32>, tensor<16x256xf32>) -> (tensor<16xf32>, tensor<16xf32>, tensor<16xf32>)
|
77 |
+
%45 = tt.expand_dims %44#0 {axis = 1 : i32} : (tensor<16xf32>) -> tensor<16x1xf32>
|
78 |
+
%46 = tt.expand_dims %44#1 {axis = 1 : i32} : (tensor<16xf32>) -> tensor<16x1xf32>
|
79 |
+
%47 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x256xf32>
|
80 |
+
%48 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x256x!tt.ptr<f32, 1>>
|
81 |
+
%49 = tt.addptr %48, %7 : tensor<1x256x!tt.ptr<f32, 1>>, tensor<1x256xi32>
|
82 |
+
%50 = tt.load %49, %12, %cst_0 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x256xf32>
|
83 |
+
tt.assert %26, "index out of bounds: 0 <= tmp13 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<16x1xi1>
|
84 |
+
%51 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<16x256xf32>
|
85 |
+
%52 = arith.addf %51, %47 : tensor<16x256xf32>
|
86 |
+
%53 = tt.broadcast %45 : (tensor<16x1xf32>) -> tensor<16x256xf32>
|
87 |
+
%54 = arith.subf %52, %53 : tensor<16x256xf32>
|
88 |
+
%55 = arith.divf %46, %cst_6 : tensor<16x1xf32>
|
89 |
+
%56 = arith.addf %55, %cst_5 : tensor<16x1xf32>
|
90 |
+
%57 = tt.extern_elementwise %56 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<16x1xf32>) -> tensor<16x1xf32>
|
91 |
+
%58 = tt.broadcast %57 : (tensor<16x1xf32>) -> tensor<16x256xf32>
|
92 |
+
%59 = arith.mulf %54, %58 : tensor<16x256xf32>
|
93 |
+
%60 = tt.broadcast %50 : (tensor<1x256xf32>) -> tensor<16x256xf32>
|
94 |
+
%61 = arith.mulf %59, %60 : tensor<16x256xf32>
|
95 |
+
%62 = arith.muli %5, %cst_8 : tensor<16x1xi32>
|
96 |
+
%63 = tt.broadcast %62 : (tensor<16x1xi32>) -> tensor<16x256xi32>
|
97 |
+
%64 = arith.addi %14, %63 : tensor<16x256xi32>
|
98 |
+
%65 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<16x256x!tt.ptr<bf16, 1>>
|
99 |
+
%66 = tt.addptr %65, %64 : tensor<16x256x!tt.ptr<bf16, 1>>, tensor<16x256xi32>
|
100 |
+
%67 = arith.truncf %61 : tensor<16x256xf32> to tensor<16x256xbf16>
|
101 |
+
tt.store %66, %67, %19 {cache = 1 : i32, evict = 1 : i32} : tensor<16x256xbf16>
|
102 |
+
tt.return
|
103 |
+
}
|
104 |
+
}
|
.triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ptx
ADDED
@@ -0,0 +1,465 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
//
|
2 |
+
// Generated by LLVM NVPTX Back-End
|
3 |
+
//
|
4 |
+
|
5 |
+
.version 8.2
|
6 |
+
.target sm_89
|
7 |
+
.address_size 64
|
8 |
+
|
9 |
+
// .globl triton__0d1d2d34e
|
10 |
+
|
11 |
+
.visible .entry triton__0d1d2d34e(
|
12 |
+
.param .u64 triton__0d1d2d34e_param_0,
|
13 |
+
.param .u64 triton__0d1d2d34e_param_1,
|
14 |
+
.param .u64 triton__0d1d2d34e_param_2,
|
15 |
+
.param .u32 triton__0d1d2d34e_param_3,
|
16 |
+
.param .u32 triton__0d1d2d34e_param_4
|
17 |
+
)
|
18 |
+
.maxntid 64, 1, 1
|
19 |
+
{
|
20 |
+
.reg .pred %p<6>;
|
21 |
+
.reg .b32 %r<27>;
|
22 |
+
.reg .f32 %f<9>;
|
23 |
+
.reg .b64 %rd<24>;
|
24 |
+
.loc 1 18 0
|
25 |
+
$L__func_begin0:
|
26 |
+
.loc 1 18 0
|
27 |
+
|
28 |
+
ld.param.u64 %rd4, [triton__0d1d2d34e_param_0];
|
29 |
+
ld.param.u64 %rd5, [triton__0d1d2d34e_param_1];
|
30 |
+
$L__tmp0:
|
31 |
+
.loc 1 25 34
|
32 |
+
mov.u32 %r7, %tid.x;
|
33 |
+
and.b32 %r8, %r7, 7;
|
34 |
+
ld.param.u64 %rd6, [triton__0d1d2d34e_param_2];
|
35 |
+
.loc 1 28 30
|
36 |
+
mul.wide.u32 %rd7, %r8, 4;
|
37 |
+
add.s64 %rd1, %rd5, %rd7;
|
38 |
+
mov.b32 %r2, 0;
|
39 |
+
mov.pred %p1, -1;
|
40 |
+
.loc 1 28 35
|
41 |
+
mov.u32 %r1, 0x0;
|
42 |
+
@%p1 ld.global.b32 { %r1 }, [ %rd1 + 0 ];
|
43 |
+
@!%p1 mov.u32 %r1, %r2;
|
44 |
+
mov.b32 %f1, %r1;
|
45 |
+
.loc 1 29 30
|
46 |
+
mul.wide.u32 %rd8, %r8, 8;
|
47 |
+
add.s64 %rd3, %rd6, %rd8;
|
48 |
+
.loc 1 29 35
|
49 |
+
mov.u64 %rd2, 0x0;
|
50 |
+
@%p1 ld.global.b64 { %rd2 }, [ %rd3 + 0 ];
|
51 |
+
@!%p1 mov.u64 %rd2, 0x0;
|
52 |
+
$L__tmp1:
|
53 |
+
.loc 2 243 36
|
54 |
+
shfl.sync.bfly.b32 %r9, %r1, 4, 31, -1;
|
55 |
+
mov.b32 %f2, %r9;
|
56 |
+
$L__tmp2:
|
57 |
+
.loc 2 233 15
|
58 |
+
add.f32 %f3, %f1, %f2;
|
59 |
+
$L__tmp3:
|
60 |
+
.loc 2 243 36
|
61 |
+
mov.b32 %r10, %f3;
|
62 |
+
shfl.sync.bfly.b32 %r11, %r10, 2, 31, -1;
|
63 |
+
mov.b32 %f4, %r11;
|
64 |
+
$L__tmp4:
|
65 |
+
.loc 2 233 15
|
66 |
+
add.f32 %f5, %f3, %f4;
|
67 |
+
$L__tmp5:
|
68 |
+
.loc 2 243 36
|
69 |
+
mov.b32 %r12, %f5;
|
70 |
+
shfl.sync.bfly.b32 %r13, %r12, 1, 31, -1;
|
71 |
+
mov.b32 %f6, %r13;
|
72 |
+
$L__tmp6:
|
73 |
+
.loc 2 233 15
|
74 |
+
add.f32 %f7, %f5, %f6;
|
75 |
+
$L__tmp7:
|
76 |
+
.loc 2 243 36
|
77 |
+
cvt.u32.u64 %r14, %rd2;
|
78 |
+
shfl.sync.bfly.b32 %r15, %r14, 4, 31, -1;
|
79 |
+
{ .reg .b32 tmp; mov.b64 {tmp, %r16}, %rd2; }
|
80 |
+
shfl.sync.bfly.b32 %r17, %r16, 4, 31, -1;
|
81 |
+
cvt.u64.u32 %rd9, %r15;
|
82 |
+
cvt.u64.u32 %rd10, %r17;
|
83 |
+
shl.b64 %rd11, %rd10, 32;
|
84 |
+
or.b64 %rd12, %rd9, %rd11;
|
85 |
+
$L__tmp8:
|
86 |
+
.loc 2 233 15
|
87 |
+
add.s64 %rd13, %rd2, %rd12;
|
88 |
+
$L__tmp9:
|
89 |
+
.loc 2 243 36
|
90 |
+
cvt.u32.u64 %r18, %rd13;
|
91 |
+
shfl.sync.bfly.b32 %r19, %r18, 2, 31, -1;
|
92 |
+
{ .reg .b32 tmp; mov.b64 {tmp, %r20}, %rd13; }
|
93 |
+
shfl.sync.bfly.b32 %r21, %r20, 2, 31, -1;
|
94 |
+
cvt.u64.u32 %rd14, %r19;
|
95 |
+
cvt.u64.u32 %rd15, %r21;
|
96 |
+
shl.b64 %rd16, %rd15, 32;
|
97 |
+
or.b64 %rd17, %rd14, %rd16;
|
98 |
+
$L__tmp10:
|
99 |
+
.loc 2 233 15
|
100 |
+
add.s64 %rd18, %rd13, %rd17;
|
101 |
+
$L__tmp11:
|
102 |
+
.loc 2 243 36
|
103 |
+
cvt.u32.u64 %r22, %rd18;
|
104 |
+
shfl.sync.bfly.b32 %r23, %r22, 1, 31, -1;
|
105 |
+
{ .reg .b32 tmp; mov.b64 {tmp, %r24}, %rd18; }
|
106 |
+
shfl.sync.bfly.b32 %r25, %r24, 1, 31, -1;
|
107 |
+
cvt.u64.u32 %rd19, %r23;
|
108 |
+
cvt.u64.u32 %rd20, %r25;
|
109 |
+
shl.b64 %rd21, %rd20, 32;
|
110 |
+
or.b64 %rd22, %rd19, %rd21;
|
111 |
+
$L__tmp12:
|
112 |
+
.loc 2 233 15
|
113 |
+
add.s64 %rd23, %rd18, %rd22;
|
114 |
+
$L__tmp13:
|
115 |
+
.loc 1 36 20
|
116 |
+
cvt.rn.f32.s64 %f8, %rd23;
|
117 |
+
.loc 1 37 19
|
118 |
+
mov.b32 %r4, %f7;
|
119 |
+
mov.b32 %r5, %f8;
|
120 |
+
div.full.f32 %r6, %r4, %r5;
|
121 |
+
.loc 1 38 4
|
122 |
+
bar.sync 0;
|
123 |
+
.loc 1 39 71
|
124 |
+
and.b32 %r26, %r7, 63;
|
125 |
+
setp.eq.s32 %p5, %r26, 0;
|
126 |
+
@%p5 st.global.b32 [ %rd4 + 0 ], { %r6 };
|
127 |
+
.loc 1 39 4
|
128 |
+
ret;
|
129 |
+
$L__tmp14:
|
130 |
+
$L__func_end0:
|
131 |
+
|
132 |
+
}
|
133 |
+
.file 1 "/tmp/torchinductor_root/2q/c2qomesxoic3sfzpdzftrhej7z6hhd6pritis2f4ye2ckqoetmyt.py"
|
134 |
+
.file 2 "/usr/local/lib/python3.10/dist-packages/triton/language/standard.py"
|
135 |
+
.section .debug_abbrev
|
136 |
+
{
|
137 |
+
.b8 1
|
138 |
+
.b8 17
|
139 |
+
.b8 1
|
140 |
+
.b8 37
|
141 |
+
.b8 8
|
142 |
+
.b8 19
|
143 |
+
.b8 5
|
144 |
+
.b8 3
|
145 |
+
.b8 8
|
146 |
+
.b8 16
|
147 |
+
.b8 6
|
148 |
+
.b8 27
|
149 |
+
.b8 8
|
150 |
+
.b8 180
|
151 |
+
.b8 66
|
152 |
+
.b8 12
|
153 |
+
.b8 17
|
154 |
+
.b8 1
|
155 |
+
.b8 18
|
156 |
+
.b8 1
|
157 |
+
.b8 0
|
158 |
+
.b8 0
|
159 |
+
.b8 2
|
160 |
+
.b8 46
|
161 |
+
.b8 0
|
162 |
+
.b8 135
|
163 |
+
.b8 64
|
164 |
+
.b8 8
|
165 |
+
.b8 3
|
166 |
+
.b8 8
|
167 |
+
.b8 58
|
168 |
+
.b8 11
|
169 |
+
.b8 59
|
170 |
+
.b8 11
|
171 |
+
.b8 63
|
172 |
+
.b8 12
|
173 |
+
.b8 32
|
174 |
+
.b8 11
|
175 |
+
.b8 0
|
176 |
+
.b8 0
|
177 |
+
.b8 3
|
178 |
+
.b8 46
|
179 |
+
.b8 1
|
180 |
+
.b8 17
|
181 |
+
.b8 1
|
182 |
+
.b8 18
|
183 |
+
.b8 1
|
184 |
+
.b8 64
|
185 |
+
.b8 10
|
186 |
+
.b8 49
|
187 |
+
.b8 19
|
188 |
+
.b8 0
|
189 |
+
.b8 0
|
190 |
+
.b8 4
|
191 |
+
.b8 29
|
192 |
+
.b8 0
|
193 |
+
.b8 49
|
194 |
+
.b8 19
|
195 |
+
.b8 17
|
196 |
+
.b8 1
|
197 |
+
.b8 18
|
198 |
+
.b8 1
|
199 |
+
.b8 88
|
200 |
+
.b8 11
|
201 |
+
.b8 89
|
202 |
+
.b8 11
|
203 |
+
.b8 87
|
204 |
+
.b8 11
|
205 |
+
.b8 0
|
206 |
+
.b8 0
|
207 |
+
.b8 5
|
208 |
+
.b8 29
|
209 |
+
.b8 1
|
210 |
+
.b8 49
|
211 |
+
.b8 19
|
212 |
+
.b8 17
|
213 |
+
.b8 1
|
214 |
+
.b8 18
|
215 |
+
.b8 1
|
216 |
+
.b8 88
|
217 |
+
.b8 11
|
218 |
+
.b8 89
|
219 |
+
.b8 11
|
220 |
+
.b8 87
|
221 |
+
.b8 11
|
222 |
+
.b8 0
|
223 |
+
.b8 0
|
224 |
+
.b8 0
|
225 |
+
}
|
226 |
+
.section .debug_info
|
227 |
+
{
|
228 |
+
.b32 333
|
229 |
+
.b8 2
|
230 |
+
.b8 0
|
231 |
+
.b32 .debug_abbrev
|
232 |
+
.b8 8
|
233 |
+
.b8 1
|
234 |
+
.b8 116
|
235 |
+
.b8 114
|
236 |
+
.b8 105
|
237 |
+
.b8 116
|
238 |
+
.b8 111
|
239 |
+
.b8 110
|
240 |
+
.b8 0
|
241 |
+
.b8 2
|
242 |
+
.b8 0
|
243 |
+
.b8 99
|
244 |
+
.b8 50
|
245 |
+
.b8 113
|
246 |
+
.b8 111
|
247 |
+
.b8 109
|
248 |
+
.b8 101
|
249 |
+
.b8 115
|
250 |
+
.b8 120
|
251 |
+
.b8 111
|
252 |
+
.b8 105
|
253 |
+
.b8 99
|
254 |
+
.b8 51
|
255 |
+
.b8 115
|
256 |
+
.b8 102
|
257 |
+
.b8 122
|
258 |
+
.b8 112
|
259 |
+
.b8 100
|
260 |
+
.b8 122
|
261 |
+
.b8 102
|
262 |
+
.b8 116
|
263 |
+
.b8 114
|
264 |
+
.b8 104
|
265 |
+
.b8 101
|
266 |
+
.b8 106
|
267 |
+
.b8 55
|
268 |
+
.b8 122
|
269 |
+
.b8 54
|
270 |
+
.b8 104
|
271 |
+
.b8 104
|
272 |
+
.b8 100
|
273 |
+
.b8 54
|
274 |
+
.b8 112
|
275 |
+
.b8 114
|
276 |
+
.b8 105
|
277 |
+
.b8 116
|
278 |
+
.b8 105
|
279 |
+
.b8 115
|
280 |
+
.b8 50
|
281 |
+
.b8 102
|
282 |
+
.b8 52
|
283 |
+
.b8 121
|
284 |
+
.b8 101
|
285 |
+
.b8 50
|
286 |
+
.b8 99
|
287 |
+
.b8 107
|
288 |
+
.b8 113
|
289 |
+
.b8 111
|
290 |
+
.b8 101
|
291 |
+
.b8 116
|
292 |
+
.b8 109
|
293 |
+
.b8 121
|
294 |
+
.b8 116
|
295 |
+
.b8 46
|
296 |
+
.b8 112
|
297 |
+
.b8 121
|
298 |
+
.b8 0
|
299 |
+
.b32 .debug_line
|
300 |
+
.b8 47
|
301 |
+
.b8 116
|
302 |
+
.b8 109
|
303 |
+
.b8 112
|
304 |
+
.b8 47
|
305 |
+
.b8 116
|
306 |
+
.b8 111
|
307 |
+
.b8 114
|
308 |
+
.b8 99
|
309 |
+
.b8 104
|
310 |
+
.b8 105
|
311 |
+
.b8 110
|
312 |
+
.b8 100
|
313 |
+
.b8 117
|
314 |
+
.b8 99
|
315 |
+
.b8 116
|
316 |
+
.b8 111
|
317 |
+
.b8 114
|
318 |
+
.b8 95
|
319 |
+
.b8 114
|
320 |
+
.b8 111
|
321 |
+
.b8 111
|
322 |
+
.b8 116
|
323 |
+
.b8 47
|
324 |
+
.b8 50
|
325 |
+
.b8 113
|
326 |
+
.b8 0
|
327 |
+
.b8 1
|
328 |
+
.b64 $L__func_begin0
|
329 |
+
.b64 $L__func_end0
|
330 |
+
.b8 2
|
331 |
+
.b8 116
|
332 |
+
.b8 114
|
333 |
+
.b8 105
|
334 |
+
.b8 116
|
335 |
+
.b8 111
|
336 |
+
.b8 110
|
337 |
+
.b8 95
|
338 |
+
.b8 95
|
339 |
+
.b8 48
|
340 |
+
.b8 100
|
341 |
+
.b8 49
|
342 |
+
.b8 100
|
343 |
+
.b8 50
|
344 |
+
.b8 100
|
345 |
+
.b8 51
|
346 |
+
.b8 52
|
347 |
+
.b8 101
|
348 |
+
.b8 0
|
349 |
+
.b8 116
|
350 |
+
.b8 114
|
351 |
+
.b8 105
|
352 |
+
.b8 116
|
353 |
+
.b8 111
|
354 |
+
.b8 110
|
355 |
+
.b8 95
|
356 |
+
.b8 95
|
357 |
+
.b8 48
|
358 |
+
.b8 100
|
359 |
+
.b8 49
|
360 |
+
.b8 100
|
361 |
+
.b8 50
|
362 |
+
.b8 100
|
363 |
+
.b8 51
|
364 |
+
.b8 52
|
365 |
+
.b8 101
|
366 |
+
.b8 0
|
367 |
+
.b8 1
|
368 |
+
.b8 18
|
369 |
+
.b8 1
|
370 |
+
.b8 1
|
371 |
+
.b8 3
|
372 |
+
.b64 $L__func_begin0
|
373 |
+
.b64 $L__func_end0
|
374 |
+
.b8 1
|
375 |
+
.b8 156
|
376 |
+
.b32 125
|
377 |
+
.b8 4
|
378 |
+
.b32 125
|
379 |
+
.b64 $L__tmp1
|
380 |
+
.b64 $L__tmp6
|
381 |
+
.b8 2
|
382 |
+
.b8 32
|
383 |
+
.b8 24
|
384 |
+
.b8 5
|
385 |
+
.b32 125
|
386 |
+
.b64 $L__tmp2
|
387 |
+
.b64 $L__tmp7
|
388 |
+
.b8 2
|
389 |
+
.b8 32
|
390 |
+
.b8 24
|
391 |
+
.b8 4
|
392 |
+
.b32 125
|
393 |
+
.b64 $L__tmp2
|
394 |
+
.b64 $L__tmp7
|
395 |
+
.b8 2
|
396 |
+
.b8 243
|
397 |
+
.b8 36
|
398 |
+
.b8 0
|
399 |
+
.b8 4
|
400 |
+
.b32 125
|
401 |
+
.b64 $L__tmp7
|
402 |
+
.b64 $L__tmp12
|
403 |
+
.b8 2
|
404 |
+
.b8 35
|
405 |
+
.b8 24
|
406 |
+
.b8 5
|
407 |
+
.b32 125
|
408 |
+
.b64 $L__tmp8
|
409 |
+
.b64 $L__tmp13
|
410 |
+
.b8 2
|
411 |
+
.b8 35
|
412 |
+
.b8 24
|
413 |
+
.b8 4
|
414 |
+
.b32 125
|
415 |
+
.b64 $L__tmp8
|
416 |
+
.b64 $L__tmp13
|
417 |
+
.b8 2
|
418 |
+
.b8 243
|
419 |
+
.b8 36
|
420 |
+
.b8 0
|
421 |
+
.b8 0
|
422 |
+
.b8 0
|
423 |
+
}
|
424 |
+
.section .debug_pubnames
|
425 |
+
{
|
426 |
+
.b32 $L__pubNames_end0-$L__pubNames_start0
|
427 |
+
$L__pubNames_start0:
|
428 |
+
.b8 2
|
429 |
+
.b8 0
|
430 |
+
.b32 .debug_info
|
431 |
+
.b32 337
|
432 |
+
.b32 125
|
433 |
+
.b8 116
|
434 |
+
.b8 114
|
435 |
+
.b8 105
|
436 |
+
.b8 116
|
437 |
+
.b8 111
|
438 |
+
.b8 110
|
439 |
+
.b8 95
|
440 |
+
.b8 95
|
441 |
+
.b8 48
|
442 |
+
.b8 100
|
443 |
+
.b8 49
|
444 |
+
.b8 100
|
445 |
+
.b8 50
|
446 |
+
.b8 100
|
447 |
+
.b8 51
|
448 |
+
.b8 52
|
449 |
+
.b8 101
|
450 |
+
.b8 0
|
451 |
+
.b32 0
|
452 |
+
$L__pubNames_end0:
|
453 |
+
}
|
454 |
+
.section .debug_pubtypes
|
455 |
+
{
|
456 |
+
.b32 $L__pubTypes_end0-$L__pubTypes_start0
|
457 |
+
$L__pubTypes_start0:
|
458 |
+
.b8 2
|
459 |
+
.b8 0
|
460 |
+
.b32 .debug_info
|
461 |
+
.b32 337
|
462 |
+
.b32 0
|
463 |
+
$L__pubTypes_end0:
|
464 |
+
}
|
465 |
+
.section .debug_loc { }
|
.triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttgir
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 2], order = [0, 1], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
2 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
3 |
+
tt.func public @triton__0d1d2d34e(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg3: i32, %arg4: i32 {tt.max_divisibility = 8 : i32}) attributes {noinline = false} {
|
4 |
+
%cst = arith.constant dense<0> : tensor<1x8xi64, #blocked>
|
5 |
+
%cst_0 = arith.constant dense<0.000000e+00> : tensor<1x8xf32, #blocked>
|
6 |
+
%cst_1 = arith.constant dense<8> : tensor<1x8xi32, #blocked>
|
7 |
+
%c0_i32 = arith.constant 0 : i32
|
8 |
+
%0 = tt.make_range {end = 8 : i32, start = 0 : i32} : tensor<8xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
|
9 |
+
%1 = tt.expand_dims %0 {axis = 0 : i32} : (tensor<8xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x8xi32, #blocked>
|
10 |
+
%2 = arith.cmpi slt, %1, %cst_1 : tensor<1x8xi32, #blocked>
|
11 |
+
%3 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<1x8x!tt.ptr<f32, 1>, #blocked>
|
12 |
+
%4 = tt.addptr %3, %1 : tensor<1x8x!tt.ptr<f32, 1>, #blocked>, tensor<1x8xi32, #blocked>
|
13 |
+
%5 = tt.load %4, %2, %cst_0 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xf32, #blocked>
|
14 |
+
%6 = tt.splat %arg2 : (!tt.ptr<i64, 1>) -> tensor<1x8x!tt.ptr<i64, 1>, #blocked>
|
15 |
+
%7 = tt.addptr %6, %1 : tensor<1x8x!tt.ptr<i64, 1>, #blocked>, tensor<1x8xi32, #blocked>
|
16 |
+
%8 = tt.load %7, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xi64, #blocked>
|
17 |
+
%9 = arith.select %2, %5, %cst_0 : tensor<1x8xi1, #blocked>, tensor<1x8xf32, #blocked>
|
18 |
+
%10 = "tt.reduce"(%9) <{axis = 1 : i32}> ({
|
19 |
+
^bb0(%arg5: f32, %arg6: f32):
|
20 |
+
%19 = arith.addf %arg5, %arg6 : f32
|
21 |
+
tt.reduce.return %19 : f32
|
22 |
+
}) : (tensor<1x8xf32, #blocked>) -> tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
|
23 |
+
%11 = tt.expand_dims %10 {axis = 1 : i32} : (tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<1x1xf32, #blocked>
|
24 |
+
%12 = arith.select %2, %8, %cst : tensor<1x8xi1, #blocked>, tensor<1x8xi64, #blocked>
|
25 |
+
%13 = "tt.reduce"(%12) <{axis = 1 : i32}> ({
|
26 |
+
^bb0(%arg5: i64, %arg6: i64):
|
27 |
+
%19 = arith.addi %arg5, %arg6 : i64
|
28 |
+
tt.reduce.return %19 : i64
|
29 |
+
}) : (tensor<1x8xi64, #blocked>) -> tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
|
30 |
+
%14 = tt.expand_dims %13 {axis = 1 : i32} : (tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<1x1xi64, #blocked>
|
31 |
+
%15 = arith.sitofp %14 : tensor<1x1xi64, #blocked> to tensor<1x1xf32, #blocked>
|
32 |
+
%16 = arith.divf %11, %15 : tensor<1x1xf32, #blocked>
|
33 |
+
gpu.barrier
|
34 |
+
%17 = tt.addptr %arg0, %c0_i32 : !tt.ptr<f32, 1>, i32
|
35 |
+
%18 = tt.splat %17 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>, #blocked>
|
36 |
+
tt.store %18, %16 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32, #blocked>
|
37 |
+
tt.return
|
38 |
+
}
|
39 |
+
}
|
.triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttir
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
module {
|
2 |
+
tt.func public @triton__0d1d2d34e(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg3: i32, %arg4: i32 {tt.max_divisibility = 8 : i32}) attributes {noinline = false} {
|
3 |
+
%c0_i32 = arith.constant 0 : i32
|
4 |
+
%cst = arith.constant dense<0> : tensor<1x8xi64>
|
5 |
+
%cst_0 = arith.constant dense<0.000000e+00> : tensor<1x8xf32>
|
6 |
+
%cst_1 = arith.constant dense<8> : tensor<1x8xi32>
|
7 |
+
%0 = tt.make_range {end = 8 : i32, start = 0 : i32} : tensor<8xi32>
|
8 |
+
%1 = tt.expand_dims %0 {axis = 0 : i32} : (tensor<8xi32>) -> tensor<1x8xi32>
|
9 |
+
%2 = arith.cmpi slt, %1, %cst_1 : tensor<1x8xi32>
|
10 |
+
%3 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<1x8x!tt.ptr<f32, 1>>
|
11 |
+
%4 = tt.addptr %3, %1 : tensor<1x8x!tt.ptr<f32, 1>>, tensor<1x8xi32>
|
12 |
+
%5 = tt.load %4, %2, %cst_0 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xf32>
|
13 |
+
%6 = tt.splat %arg2 : (!tt.ptr<i64, 1>) -> tensor<1x8x!tt.ptr<i64, 1>>
|
14 |
+
%7 = tt.addptr %6, %1 : tensor<1x8x!tt.ptr<i64, 1>>, tensor<1x8xi32>
|
15 |
+
%8 = tt.load %7, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xi64>
|
16 |
+
%9 = arith.select %2, %5, %cst_0 : tensor<1x8xi1>, tensor<1x8xf32>
|
17 |
+
%10 = "tt.reduce"(%9) <{axis = 1 : i32}> ({
|
18 |
+
^bb0(%arg5: f32, %arg6: f32):
|
19 |
+
%19 = arith.addf %arg5, %arg6 : f32
|
20 |
+
tt.reduce.return %19 : f32
|
21 |
+
}) : (tensor<1x8xf32>) -> tensor<1xf32>
|
22 |
+
%11 = tt.expand_dims %10 {axis = 1 : i32} : (tensor<1xf32>) -> tensor<1x1xf32>
|
23 |
+
%12 = arith.select %2, %8, %cst : tensor<1x8xi1>, tensor<1x8xi64>
|
24 |
+
%13 = "tt.reduce"(%12) <{axis = 1 : i32}> ({
|
25 |
+
^bb0(%arg5: i64, %arg6: i64):
|
26 |
+
%19 = arith.addi %arg5, %arg6 : i64
|
27 |
+
tt.reduce.return %19 : i64
|
28 |
+
}) : (tensor<1x8xi64>) -> tensor<1xi64>
|
29 |
+
%14 = tt.expand_dims %13 {axis = 1 : i32} : (tensor<1xi64>) -> tensor<1x1xi64>
|
30 |
+
%15 = arith.sitofp %14 : tensor<1x1xi64> to tensor<1x1xf32>
|
31 |
+
%16 = arith.divf %11, %15 : tensor<1x1xf32>
|
32 |
+
gpu.barrier
|
33 |
+
%17 = tt.addptr %arg0, %c0_i32 : !tt.ptr<f32, 1>, i32
|
34 |
+
%18 = tt.splat %17 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>>
|
35 |
+
tt.store %18, %16 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32>
|
36 |
+
tt.return
|
37 |
+
}
|
38 |
+
}
|
.triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.cubin
ADDED
Binary file (31.3 kB). View file
|
|
.triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.llir
ADDED
@@ -0,0 +1,550 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
; ModuleID = 'LLVMDialectModule'
|
2 |
+
source_filename = "LLVMDialectModule"
|
3 |
+
|
4 |
+
@assertFunc_1 = internal constant [25 x i8] c"_call_with_frames_removed"
|
5 |
+
@assertFile_1 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
|
6 |
+
@assertMessage_1 = internal constant [39 x i8] c"index out of bounds: 0 <= tmp16 < 50257"
|
7 |
+
@assertFunc_0 = internal constant [25 x i8] c"_call_with_frames_removed"
|
8 |
+
@assertFile_0 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
|
9 |
+
@assertMessage_0 = internal constant [38 x i8] c"index out of bounds: 0 <= tmp3 < 50257"
|
10 |
+
@global_smem = external addrspace(3) global [0 x i8]
|
11 |
+
@.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
|
12 |
+
|
13 |
+
declare void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr
|
14 |
+
|
15 |
+
define void @triton__0d1d2d3d4d5d6de7de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, i32 %6, i32 %7) local_unnamed_addr !dbg !7 {
|
16 |
+
%9 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
|
17 |
+
%10 = and i32 %9, 31, !dbg !10
|
18 |
+
%11 = lshr i32 %9, 5, !dbg !10
|
19 |
+
%12 = lshr i32 %9, 6, !dbg !10
|
20 |
+
%13 = and i32 %12, 1, !dbg !10
|
21 |
+
%14 = and i32 %9, 1, !dbg !10
|
22 |
+
%15 = and i32 %11, 1, !dbg !11
|
23 |
+
%urem = shl i32 %9, 2, !dbg !11
|
24 |
+
%16 = and i32 %urem, 252, !dbg !11
|
25 |
+
%17 = shl i32 %9, 1, !dbg !11
|
26 |
+
%18 = and i32 %17, 254, !dbg !11
|
27 |
+
%19 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !12
|
28 |
+
%20 = shl i32 %19, 1, !dbg !13
|
29 |
+
%21 = or i32 %20, %13, !dbg !14
|
30 |
+
%22 = or i32 %20, %14, !dbg !14
|
31 |
+
%23 = sext i32 %21 to i64, !dbg !15
|
32 |
+
%24 = getelementptr i64, ptr addrspace(1) %0, i64 %23, !dbg !15
|
33 |
+
%25 = sext i32 %22 to i64, !dbg !15
|
34 |
+
%26 = getelementptr i64, ptr addrspace(1) %0, i64 %25, !dbg !15
|
35 |
+
%27 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
|
36 |
+
%28 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
|
37 |
+
%29 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
|
38 |
+
%30 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
|
39 |
+
%31 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %26, i1 true) #6, !dbg !16
|
40 |
+
%32 = srem i32 %21, 512, !dbg !17
|
41 |
+
%33 = shl nsw i32 %32, 8, !dbg !18
|
42 |
+
%34 = or i32 %33, %16, !dbg !19
|
43 |
+
%35 = sext i32 %34 to i64, !dbg !20
|
44 |
+
%36 = getelementptr float, ptr addrspace(1) %2, i64 %35, !dbg !20
|
45 |
+
%37 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %36, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
|
46 |
+
%38 = extractvalue { i32, i32, i32, i32 } %37, 0, !dbg !21
|
47 |
+
%39 = extractvalue { i32, i32, i32, i32 } %37, 1, !dbg !21
|
48 |
+
%40 = extractvalue { i32, i32, i32, i32 } %37, 2, !dbg !21
|
49 |
+
%41 = extractvalue { i32, i32, i32, i32 } %37, 3, !dbg !21
|
50 |
+
%42 = insertelement <2 x i32> poison, i32 %39, i64 0, !dbg !21
|
51 |
+
%43 = insertelement <2 x i32> %42, i32 %38, i64 1, !dbg !21
|
52 |
+
%44 = bitcast <2 x i32> %43 to <2 x float>, !dbg !21
|
53 |
+
%45 = bitcast i32 %40 to float, !dbg !21
|
54 |
+
%46 = bitcast i32 %41 to float, !dbg !21
|
55 |
+
%47 = shl i32 %21, 8, !dbg !22
|
56 |
+
%48 = or i32 %47, %16, !dbg !23
|
57 |
+
%49 = sext i32 %48 to i64, !dbg !24
|
58 |
+
%50 = getelementptr i16, ptr addrspace(1) %3, i64 %49, !dbg !24
|
59 |
+
%51 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.L1::evict_last.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %50, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !25
|
60 |
+
%52 = extractvalue { i32, i32 } %51, 0, !dbg !25
|
61 |
+
%53 = extractvalue { i32, i32 } %51, 1, !dbg !25
|
62 |
+
%54 = trunc i32 %52 to i16, !dbg !25
|
63 |
+
%extelt.offset = lshr i32 %52, 16, !dbg !25
|
64 |
+
%55 = trunc i32 %extelt.offset to i16, !dbg !25
|
65 |
+
%56 = trunc i32 %53 to i16, !dbg !25
|
66 |
+
%extelt.offset1 = lshr i32 %53, 16, !dbg !25
|
67 |
+
%57 = trunc i32 %extelt.offset1 to i16, !dbg !25
|
68 |
+
%58 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %54) #6, !dbg !26
|
69 |
+
%59 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %55) #6, !dbg !26
|
70 |
+
%60 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %56) #6, !dbg !26
|
71 |
+
%61 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %57) #6, !dbg !26
|
72 |
+
%62 = add i64 %31, 50257, !dbg !27
|
73 |
+
%63 = icmp slt i64 %27, 0, !dbg !28
|
74 |
+
%64 = icmp slt i64 %31, 0, !dbg !28
|
75 |
+
%65 = select i1 %64, i64 %62, i64 %31, !dbg !29
|
76 |
+
%66 = icmp ugt i64 %65, 50256, !dbg !30
|
77 |
+
br i1 %66, label %67, label %68, !dbg !31
|
78 |
+
|
79 |
+
67: ; preds = %8
|
80 |
+
tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 883, ptr nonnull @assertFunc_0, i64 1), !dbg !31
|
81 |
+
br label %68, !dbg !31
|
82 |
+
|
83 |
+
68: ; preds = %67, %8
|
84 |
+
%69 = shl i64 %27, 8, !dbg !32
|
85 |
+
%70 = add i64 %69, 12865792, !dbg !32
|
86 |
+
%71 = select i1 %63, i64 %70, i64 %69, !dbg !32
|
87 |
+
%72 = zext nneg i32 %16 to i64
|
88 |
+
%73 = or i64 %71, %72, !dbg !33
|
89 |
+
%74 = getelementptr float, ptr addrspace(1) %1, i64 %73, !dbg !34
|
90 |
+
%75 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %74, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !35
|
91 |
+
%76 = extractvalue { i32, i32, i32, i32 } %75, 0, !dbg !35
|
92 |
+
%77 = extractvalue { i32, i32, i32, i32 } %75, 1, !dbg !35
|
93 |
+
%78 = extractvalue { i32, i32, i32, i32 } %75, 2, !dbg !35
|
94 |
+
%79 = extractvalue { i32, i32, i32, i32 } %75, 3, !dbg !35
|
95 |
+
%80 = bitcast i32 %78 to float, !dbg !35
|
96 |
+
%81 = bitcast i32 %79 to float, !dbg !35
|
97 |
+
%82 = fadd float %45, %80, !dbg !36
|
98 |
+
%83 = fadd float %46, %81, !dbg !36
|
99 |
+
%84 = fadd float %60, %82, !dbg !37
|
100 |
+
%85 = fadd float %61, %83, !dbg !37
|
101 |
+
%86 = insertelement <2 x i32> poison, i32 %77, i64 0, !dbg !35
|
102 |
+
%87 = insertelement <2 x i32> %86, i32 %76, i64 1, !dbg !35
|
103 |
+
%88 = bitcast <2 x i32> %87 to <2 x float>, !dbg !35
|
104 |
+
%89 = fadd <2 x float> %44, %88, !dbg !36
|
105 |
+
%90 = insertelement <2 x float> poison, float %59, i64 0, !dbg !37
|
106 |
+
%91 = insertelement <2 x float> %90, float %58, i64 1, !dbg !37
|
107 |
+
%92 = fadd <2 x float> %91, %89, !dbg !37
|
108 |
+
%93 = fadd <2 x float> %92, zeroinitializer, !dbg !38
|
109 |
+
%94 = fadd float %84, 0.000000e+00, !dbg !38
|
110 |
+
%95 = fadd float %85, 0.000000e+00, !dbg !38
|
111 |
+
%96 = extractelement <2 x float> %93, i64 1, !dbg !42
|
112 |
+
%97 = extractelement <2 x float> %92, i64 1, !dbg !46
|
113 |
+
%98 = fsub float %97, %96, !dbg !47
|
114 |
+
%99 = extractelement <2 x float> %93, i64 0, !dbg !42
|
115 |
+
%100 = extractelement <2 x float> %92, i64 0, !dbg !46
|
116 |
+
%101 = fsub float %100, %99, !dbg !47
|
117 |
+
%102 = fsub float %84, %94, !dbg !47
|
118 |
+
%103 = fsub float %85, %95, !dbg !47
|
119 |
+
%104 = fmul float %97, %98, !dbg !46
|
120 |
+
%105 = fmul float %100, %101, !dbg !46
|
121 |
+
%106 = fmul float %84, %102, !dbg !46
|
122 |
+
%107 = fmul float %85, %103, !dbg !46
|
123 |
+
%108 = fadd float %104, 0.000000e+00, !dbg !48
|
124 |
+
%109 = fadd float %105, 0.000000e+00, !dbg !48
|
125 |
+
%110 = fadd float %106, 0.000000e+00, !dbg !48
|
126 |
+
%111 = fadd float %107, 0.000000e+00, !dbg !48
|
127 |
+
%112 = fsub float %99, %96, !dbg !42
|
128 |
+
%113 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 2.000000e+00) #6, !dbg !49
|
129 |
+
%114 = fmul float %113, %112, !dbg !50
|
130 |
+
%115 = fadd float %96, %114, !dbg !51
|
131 |
+
%116 = fadd float %108, %109, !dbg !52
|
132 |
+
%117 = fmul float %112, %112, !dbg !53
|
133 |
+
%118 = fmul float %113, %117, !dbg !54
|
134 |
+
%119 = fadd float %118, %116, !dbg !55
|
135 |
+
%120 = fsub float %94, %115, !dbg !42
|
136 |
+
%121 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 3.000000e+00) #6, !dbg !49
|
137 |
+
%122 = fmul float %121, %120, !dbg !50
|
138 |
+
%123 = fadd float %115, %122, !dbg !51
|
139 |
+
%124 = fadd float %110, %119, !dbg !52
|
140 |
+
%125 = fmul float %120, %120, !dbg !53
|
141 |
+
%126 = fmul float %125, 2.000000e+00, !dbg !56
|
142 |
+
%127 = fmul float %121, %126, !dbg !54
|
143 |
+
%128 = fadd float %124, %127, !dbg !55
|
144 |
+
%129 = fsub float %95, %123, !dbg !42
|
145 |
+
%130 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 4.000000e+00) #6, !dbg !49
|
146 |
+
%131 = fmul float %130, %129, !dbg !50
|
147 |
+
%132 = fadd float %123, %131, !dbg !51
|
148 |
+
%133 = fadd float %111, %128, !dbg !52
|
149 |
+
%134 = fmul float %129, %129, !dbg !53
|
150 |
+
%135 = fmul float %134, 3.000000e+00, !dbg !56
|
151 |
+
%136 = fmul float %130, %135, !dbg !54
|
152 |
+
%137 = fadd float %133, %136, !dbg !55
|
153 |
+
%138 = bitcast float %132 to i32, !dbg !57
|
154 |
+
%139 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %138, i32 16, i32 31), !dbg !57
|
155 |
+
%140 = bitcast i32 %139 to float, !dbg !57
|
156 |
+
%141 = bitcast float %137 to i32, !dbg !57
|
157 |
+
%142 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %141, i32 16, i32 31), !dbg !57
|
158 |
+
%143 = bitcast i32 %142 to float, !dbg !57
|
159 |
+
%144 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 1082130432, i32 16, i32 31), !dbg !57
|
160 |
+
%145 = bitcast i32 %144 to float, !dbg !57
|
161 |
+
%146 = fsub float %140, %132, !dbg !42
|
162 |
+
%147 = fadd float %145, 4.000000e+00, !dbg !59
|
163 |
+
%148 = fcmp oeq float %147, 0.000000e+00, !dbg !60
|
164 |
+
%149 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %145, float %147) #6, !dbg !49
|
165 |
+
%150 = select i1 %148, float 0.000000e+00, float %149, !dbg !61
|
166 |
+
%151 = fmul float %150, %146, !dbg !50
|
167 |
+
%152 = fadd float %132, %151, !dbg !51
|
168 |
+
%153 = fadd float %137, %143, !dbg !52
|
169 |
+
%154 = fmul float %146, %146, !dbg !53
|
170 |
+
%155 = fmul float %154, 4.000000e+00, !dbg !56
|
171 |
+
%156 = fmul float %150, %155, !dbg !54
|
172 |
+
%157 = fadd float %153, %156, !dbg !55
|
173 |
+
%158 = bitcast float %152 to i32, !dbg !57
|
174 |
+
%159 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %158, i32 8, i32 31), !dbg !57
|
175 |
+
%160 = bitcast i32 %159 to float, !dbg !57
|
176 |
+
%161 = bitcast float %157 to i32, !dbg !57
|
177 |
+
%162 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %161, i32 8, i32 31), !dbg !57
|
178 |
+
%163 = bitcast i32 %162 to float, !dbg !57
|
179 |
+
%164 = bitcast float %147 to i32, !dbg !57
|
180 |
+
%165 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %164, i32 8, i32 31), !dbg !57
|
181 |
+
%166 = bitcast i32 %165 to float, !dbg !57
|
182 |
+
%167 = fsub float %160, %152, !dbg !42
|
183 |
+
%168 = fadd float %147, %166, !dbg !59
|
184 |
+
%169 = fcmp oeq float %168, 0.000000e+00, !dbg !60
|
185 |
+
%170 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %166, float %168) #6, !dbg !49
|
186 |
+
%171 = select i1 %169, float 0.000000e+00, float %170, !dbg !61
|
187 |
+
%172 = fmul float %171, %167, !dbg !50
|
188 |
+
%173 = fadd float %152, %172, !dbg !51
|
189 |
+
%174 = fadd float %157, %163, !dbg !52
|
190 |
+
%175 = fmul float %167, %167, !dbg !53
|
191 |
+
%176 = fmul float %147, %175, !dbg !56
|
192 |
+
%177 = fmul float %171, %176, !dbg !54
|
193 |
+
%178 = fadd float %174, %177, !dbg !55
|
194 |
+
%179 = bitcast float %173 to i32, !dbg !57
|
195 |
+
%180 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %179, i32 4, i32 31), !dbg !57
|
196 |
+
%181 = bitcast i32 %180 to float, !dbg !57
|
197 |
+
%182 = bitcast float %178 to i32, !dbg !57
|
198 |
+
%183 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %182, i32 4, i32 31), !dbg !57
|
199 |
+
%184 = bitcast i32 %183 to float, !dbg !57
|
200 |
+
%185 = bitcast float %168 to i32, !dbg !57
|
201 |
+
%186 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %185, i32 4, i32 31), !dbg !57
|
202 |
+
%187 = bitcast i32 %186 to float, !dbg !57
|
203 |
+
%188 = fsub float %181, %173, !dbg !42
|
204 |
+
%189 = fadd float %168, %187, !dbg !59
|
205 |
+
%190 = fcmp oeq float %189, 0.000000e+00, !dbg !60
|
206 |
+
%191 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %187, float %189) #6, !dbg !49
|
207 |
+
%192 = select i1 %190, float 0.000000e+00, float %191, !dbg !61
|
208 |
+
%193 = fmul float %192, %188, !dbg !50
|
209 |
+
%194 = fadd float %173, %193, !dbg !51
|
210 |
+
%195 = fadd float %178, %184, !dbg !52
|
211 |
+
%196 = fmul float %188, %188, !dbg !53
|
212 |
+
%197 = fmul float %168, %196, !dbg !56
|
213 |
+
%198 = fmul float %192, %197, !dbg !54
|
214 |
+
%199 = fadd float %195, %198, !dbg !55
|
215 |
+
%200 = bitcast float %194 to i32, !dbg !57
|
216 |
+
%201 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %200, i32 2, i32 31), !dbg !57
|
217 |
+
%202 = bitcast i32 %201 to float, !dbg !57
|
218 |
+
%203 = bitcast float %199 to i32, !dbg !57
|
219 |
+
%204 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %203, i32 2, i32 31), !dbg !57
|
220 |
+
%205 = bitcast i32 %204 to float, !dbg !57
|
221 |
+
%206 = bitcast float %189 to i32, !dbg !57
|
222 |
+
%207 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %206, i32 2, i32 31), !dbg !57
|
223 |
+
%208 = bitcast i32 %207 to float, !dbg !57
|
224 |
+
%209 = fsub float %202, %194, !dbg !42
|
225 |
+
%210 = fadd float %189, %208, !dbg !59
|
226 |
+
%211 = fcmp oeq float %210, 0.000000e+00, !dbg !60
|
227 |
+
%212 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %208, float %210) #6, !dbg !49
|
228 |
+
%213 = select i1 %211, float 0.000000e+00, float %212, !dbg !61
|
229 |
+
%214 = fmul float %213, %209, !dbg !50
|
230 |
+
%215 = fadd float %194, %214, !dbg !51
|
231 |
+
%216 = fadd float %199, %205, !dbg !52
|
232 |
+
%217 = fmul float %209, %209, !dbg !53
|
233 |
+
%218 = fmul float %189, %217, !dbg !56
|
234 |
+
%219 = fmul float %213, %218, !dbg !54
|
235 |
+
%220 = fadd float %216, %219, !dbg !55
|
236 |
+
%221 = bitcast float %215 to i32, !dbg !57
|
237 |
+
%222 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %221, i32 1, i32 31), !dbg !57
|
238 |
+
%223 = bitcast i32 %222 to float, !dbg !57
|
239 |
+
%224 = bitcast float %220 to i32, !dbg !57
|
240 |
+
%225 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %224, i32 1, i32 31), !dbg !57
|
241 |
+
%226 = bitcast i32 %225 to float, !dbg !57
|
242 |
+
%227 = bitcast float %210 to i32, !dbg !57
|
243 |
+
%228 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %227, i32 1, i32 31), !dbg !57
|
244 |
+
%229 = bitcast i32 %228 to float, !dbg !57
|
245 |
+
%230 = fsub float %223, %215, !dbg !42
|
246 |
+
%231 = fadd float %210, %229, !dbg !59
|
247 |
+
%232 = fcmp oeq float %231, 0.000000e+00, !dbg !60
|
248 |
+
%233 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %229, float %231) #6, !dbg !49
|
249 |
+
%234 = select i1 %232, float 0.000000e+00, float %233, !dbg !61
|
250 |
+
%235 = fmul float %234, %230, !dbg !50
|
251 |
+
%236 = fadd float %215, %235, !dbg !51
|
252 |
+
%237 = fadd float %220, %226, !dbg !52
|
253 |
+
%238 = fmul float %230, %230, !dbg !53
|
254 |
+
%239 = fmul float %210, %238, !dbg !56
|
255 |
+
%240 = fmul float %234, %239, !dbg !54
|
256 |
+
%241 = fadd float %237, %240, !dbg !55
|
257 |
+
%242 = icmp eq i32 %10, 0, !dbg !57
|
258 |
+
%243 = shl nuw nsw i32 %13, 1, !dbg !57
|
259 |
+
%244 = or i32 %243, %15, !dbg !57
|
260 |
+
%245 = zext nneg i32 %244 to i64, !dbg !57
|
261 |
+
%246 = getelementptr float, ptr addrspace(3) @global_smem, i64 %245, !dbg !57
|
262 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %246, float %236, i1 %242) #6, !dbg !57
|
263 |
+
%247 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 16), i64 %245, !dbg !57
|
264 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %247, float %241, i1 %242) #6, !dbg !57
|
265 |
+
%248 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 32), i64 %245, !dbg !57
|
266 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %248, float %231, i1 %242) #6, !dbg !57
|
267 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !57
|
268 |
+
%249 = icmp slt i32 %9, 4, !dbg !57
|
269 |
+
%250 = sext i32 %9 to i64, !dbg !57
|
270 |
+
%251 = getelementptr float, ptr addrspace(3) @global_smem, i64 %250, !dbg !57
|
271 |
+
%252 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %251, i1 %249) #6, !dbg !57
|
272 |
+
%253 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 16), i64 %250, !dbg !57
|
273 |
+
%254 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %253, i1 %249) #6, !dbg !57
|
274 |
+
%255 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 32), i64 %250, !dbg !57
|
275 |
+
%256 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %255, i1 %249) #6, !dbg !57
|
276 |
+
%257 = bitcast float %252 to i32, !dbg !57
|
277 |
+
%258 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %257, i32 1, i32 31), !dbg !57
|
278 |
+
%259 = bitcast i32 %258 to float, !dbg !57
|
279 |
+
%260 = bitcast float %254 to i32, !dbg !57
|
280 |
+
%261 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %260, i32 1, i32 31), !dbg !57
|
281 |
+
%262 = bitcast i32 %261 to float, !dbg !57
|
282 |
+
%263 = bitcast float %256 to i32, !dbg !57
|
283 |
+
%264 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %263, i32 1, i32 31), !dbg !57
|
284 |
+
%265 = bitcast i32 %264 to float, !dbg !57
|
285 |
+
%266 = fsub float %259, %252, !dbg !42
|
286 |
+
%267 = fadd float %256, %265, !dbg !59
|
287 |
+
%268 = fcmp oeq float %267, 0.000000e+00, !dbg !60
|
288 |
+
%269 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %265, float %267) #6, !dbg !49
|
289 |
+
%270 = select i1 %268, float 0.000000e+00, float %269, !dbg !61
|
290 |
+
%271 = fmul float %266, %270, !dbg !50
|
291 |
+
%272 = fadd float %252, %271, !dbg !51
|
292 |
+
%273 = fadd float %254, %262, !dbg !52
|
293 |
+
%274 = fmul float %266, %266, !dbg !53
|
294 |
+
%275 = fmul float %256, %274, !dbg !56
|
295 |
+
%276 = fmul float %275, %270, !dbg !54
|
296 |
+
%277 = fadd float %273, %276, !dbg !55
|
297 |
+
%278 = icmp eq i32 %14, 0, !dbg !57
|
298 |
+
%279 = and i1 %249, %278, !dbg !57
|
299 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %251, float %272, i1 %279) #6, !dbg !57
|
300 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %253, float %277, i1 %279) #6, !dbg !57
|
301 |
+
tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %255, float %267, i1 %279) #6, !dbg !57
|
302 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !57
|
303 |
+
%280 = zext nneg i32 %243 to i64, !dbg !57
|
304 |
+
%281 = getelementptr float, ptr addrspace(3) @global_smem, i64 %280, !dbg !57
|
305 |
+
%282 = load float, ptr addrspace(3) %281, align 4, !dbg !57
|
306 |
+
%283 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 16), i64 %280, !dbg !57
|
307 |
+
%284 = load float, ptr addrspace(3) %283, align 4, !dbg !57
|
308 |
+
%285 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %36, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !62
|
309 |
+
%286 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.L1::evict_first.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %50, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !63
|
310 |
+
%287 = extractvalue { i32, i32 } %286, 0, !dbg !63
|
311 |
+
%288 = extractvalue { i32, i32 } %286, 1, !dbg !63
|
312 |
+
%289 = trunc i32 %287 to i16, !dbg !63
|
313 |
+
%extelt.offset2 = lshr i32 %287, 16, !dbg !63
|
314 |
+
%290 = trunc i32 %extelt.offset2 to i16, !dbg !63
|
315 |
+
%291 = trunc i32 %288 to i16, !dbg !63
|
316 |
+
%extelt.offset3 = lshr i32 %288, 16, !dbg !63
|
317 |
+
%292 = trunc i32 %extelt.offset3 to i16, !dbg !63
|
318 |
+
%293 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %289) #6, !dbg !64
|
319 |
+
%294 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %290) #6, !dbg !64
|
320 |
+
%295 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %291) #6, !dbg !64
|
321 |
+
%296 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %292) #6, !dbg !64
|
322 |
+
%297 = zext nneg i32 %18 to i64, !dbg !65
|
323 |
+
%298 = getelementptr float, ptr addrspace(1) %4, i64 %297, !dbg !65
|
324 |
+
%299 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.L1::evict_last.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %298, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !66
|
325 |
+
br i1 %66, label %300, label %301, !dbg !67
|
326 |
+
|
327 |
+
300: ; preds = %68
|
328 |
+
tail call void @__assertfail(ptr nonnull @assertMessage_1, ptr nonnull @assertFile_1, i32 883, ptr nonnull @assertFunc_1, i64 1), !dbg !67
|
329 |
+
br label %301, !dbg !67
|
330 |
+
|
331 |
+
301: ; preds = %300, %68
|
332 |
+
%302 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %74, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !68
|
333 |
+
%303 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
|
334 |
+
%304 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
|
335 |
+
%305 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
|
336 |
+
%306 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
|
337 |
+
%307 = fadd float %303, 0x3EE4F8B580000000, !dbg !70
|
338 |
+
%308 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
|
339 |
+
%.not.i = icmp eq i32 %308, 0, !dbg !71
|
340 |
+
br i1 %.not.i, label %311, label %309, !dbg !71
|
341 |
+
|
342 |
+
309: ; preds = %301
|
343 |
+
%310 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %307), !dbg !71
|
344 |
+
br label %__nv_rsqrtf.exit, !dbg !71
|
345 |
+
|
346 |
+
311: ; preds = %301
|
347 |
+
%312 = tail call float @llvm.nvvm.rsqrt.approx.f(float %307), !dbg !71
|
348 |
+
br label %__nv_rsqrtf.exit, !dbg !71
|
349 |
+
|
350 |
+
__nv_rsqrtf.exit: ; preds = %309, %311
|
351 |
+
%.0.i = phi float [ %310, %309 ], [ %312, %311 ], !dbg !71
|
352 |
+
%313 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
|
353 |
+
%314 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
|
354 |
+
%315 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
|
355 |
+
%316 = extractvalue { i32, i32, i32, i32 } %302, 3, !dbg !68
|
356 |
+
%317 = bitcast i32 %316 to float, !dbg !68
|
357 |
+
%318 = extractvalue { i32, i32, i32, i32 } %285, 3, !dbg !62
|
358 |
+
%319 = bitcast i32 %318 to float, !dbg !62
|
359 |
+
%320 = fadd float %319, %317, !dbg !72
|
360 |
+
%321 = fadd float %296, %320, !dbg !73
|
361 |
+
%322 = fsub float %321, %282, !dbg !74
|
362 |
+
%323 = extractvalue { i32, i32, i32, i32 } %302, 2, !dbg !68
|
363 |
+
%324 = bitcast i32 %323 to float, !dbg !68
|
364 |
+
%325 = extractvalue { i32, i32, i32, i32 } %285, 2, !dbg !62
|
365 |
+
%326 = bitcast i32 %325 to float, !dbg !62
|
366 |
+
%327 = fadd float %326, %324, !dbg !72
|
367 |
+
%328 = fadd float %295, %327, !dbg !73
|
368 |
+
%329 = fsub float %328, %282, !dbg !74
|
369 |
+
%330 = extractvalue { i32, i32, i32, i32 } %302, 1, !dbg !68
|
370 |
+
%331 = bitcast i32 %330 to float, !dbg !68
|
371 |
+
%332 = extractvalue { i32, i32, i32, i32 } %285, 1, !dbg !62
|
372 |
+
%333 = bitcast i32 %332 to float, !dbg !62
|
373 |
+
%334 = fadd float %333, %331, !dbg !72
|
374 |
+
%335 = fadd float %294, %334, !dbg !73
|
375 |
+
%336 = fsub float %335, %282, !dbg !74
|
376 |
+
%337 = extractvalue { i32, i32, i32, i32 } %302, 0, !dbg !68
|
377 |
+
%338 = bitcast i32 %337 to float, !dbg !68
|
378 |
+
%339 = extractvalue { i32, i32, i32, i32 } %285, 0, !dbg !62
|
379 |
+
%340 = bitcast i32 %339 to float, !dbg !62
|
380 |
+
%341 = fadd float %340, %338, !dbg !72
|
381 |
+
%342 = fadd float %293, %341, !dbg !73
|
382 |
+
%343 = fsub float %342, %282, !dbg !74
|
383 |
+
%344 = extractvalue { i32, i32 } %299, 0, !dbg !66
|
384 |
+
%345 = extractvalue { i32, i32 } %299, 1, !dbg !66
|
385 |
+
%346 = fmul float %343, %.0.i, !dbg !75
|
386 |
+
%347 = fmul float %336, %.0.i, !dbg !75
|
387 |
+
%348 = fmul float %329, %.0.i, !dbg !75
|
388 |
+
%349 = fmul float %322, %.0.i, !dbg !75
|
389 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !76
|
390 |
+
%350 = getelementptr float, ptr addrspace(3) @global_smem, i64 %297, !dbg !76
|
391 |
+
%351 = insertelement <2 x i32> undef, i32 %344, i64 0, !dbg !76
|
392 |
+
%352 = insertelement <2 x i32> %351, i32 %345, i64 1, !dbg !76
|
393 |
+
store <2 x i32> %352, ptr addrspace(3) %350, align 8, !dbg !76
|
394 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !76
|
395 |
+
%353 = getelementptr float, ptr addrspace(3) @global_smem, i64 %72, !dbg !76
|
396 |
+
%354 = load float, ptr addrspace(3) %353, align 16, !dbg !76
|
397 |
+
%355 = getelementptr inbounds <4 x float>, ptr addrspace(3) %353, i64 0, i64 1, !dbg !76
|
398 |
+
%356 = load float, ptr addrspace(3) %355, align 4, !dbg !76
|
399 |
+
%357 = getelementptr inbounds <4 x float>, ptr addrspace(3) %353, i64 0, i64 2, !dbg !76
|
400 |
+
%358 = load float, ptr addrspace(3) %357, align 8, !dbg !76
|
401 |
+
%359 = getelementptr inbounds <4 x float>, ptr addrspace(3) %353, i64 0, i64 3, !dbg !76
|
402 |
+
%360 = load float, ptr addrspace(3) %359, align 4, !dbg !76
|
403 |
+
%361 = fmul float %346, %354, !dbg !76
|
404 |
+
%362 = fmul float %347, %356, !dbg !76
|
405 |
+
%363 = fmul float %348, %358, !dbg !76
|
406 |
+
%364 = fmul float %349, %360, !dbg !76
|
407 |
+
%365 = getelementptr i16, ptr addrspace(1) %5, i64 %49, !dbg !77
|
408 |
+
%366 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %361) #6, !dbg !78
|
409 |
+
%367 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %362) #6, !dbg !78
|
410 |
+
%368 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %363) #6, !dbg !78
|
411 |
+
%369 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %364) #6, !dbg !78
|
412 |
+
%370 = insertelement <2 x i16> undef, i16 %366, i64 0, !dbg !78
|
413 |
+
%371 = insertelement <2 x i16> %370, i16 %367, i64 1, !dbg !78
|
414 |
+
%372 = bitcast <2 x i16> %371 to i32, !dbg !78
|
415 |
+
%373 = insertelement <2 x i16> undef, i16 %368, i64 0, !dbg !78
|
416 |
+
%374 = insertelement <2 x i16> %373, i16 %369, i64 1, !dbg !78
|
417 |
+
%375 = bitcast <2 x i16> %374 to i32, !dbg !78
|
418 |
+
tail call void asm sideeffect "@$3 st.global.v2.b32 [ $2 + 0 ], { $0, $1 };", "r,r,l,b"(i32 %372, i32 %375, ptr addrspace(1) %365, i1 true) #6, !dbg !78
|
419 |
+
ret void, !dbg !79
|
420 |
+
}
|
421 |
+
|
422 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
|
423 |
+
declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
|
424 |
+
|
425 |
+
; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
|
426 |
+
declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
|
427 |
+
|
428 |
+
; Function Attrs: convergent nocallback nounwind
|
429 |
+
declare void @llvm.nvvm.barrier0() #2
|
430 |
+
|
431 |
+
; Function Attrs: alwaysinline nounwind
|
432 |
+
define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
|
433 |
+
%1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
|
434 |
+
%.not = icmp eq i32 %1, 0
|
435 |
+
br i1 %.not, label %4, label %2
|
436 |
+
|
437 |
+
2: ; preds = %0
|
438 |
+
%3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
|
439 |
+
br label %6
|
440 |
+
|
441 |
+
4: ; preds = %0
|
442 |
+
%5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
|
443 |
+
br label %6
|
444 |
+
|
445 |
+
6: ; preds = %4, %2
|
446 |
+
%.0 = phi float [ %3, %2 ], [ %5, %4 ]
|
447 |
+
ret float %.0
|
448 |
+
}
|
449 |
+
|
450 |
+
declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
|
451 |
+
|
452 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
453 |
+
declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
|
454 |
+
|
455 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
456 |
+
declare float @llvm.nvvm.rsqrt.approx.f(float) #5
|
457 |
+
|
458 |
+
attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
|
459 |
+
attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
|
460 |
+
attributes #2 = { convergent nocallback nounwind }
|
461 |
+
attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
462 |
+
attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
463 |
+
attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
|
464 |
+
attributes #6 = { nounwind }
|
465 |
+
|
466 |
+
!llvm.module.flags = !{!0, !1}
|
467 |
+
!llvm.dbg.cu = !{!2}
|
468 |
+
!nvvm.annotations = !{!4, !5, !5, !4}
|
469 |
+
!llvm.ident = !{!6}
|
470 |
+
|
471 |
+
!0 = !{i32 2, !"Debug Info Version", i32 3}
|
472 |
+
!1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
|
473 |
+
!2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
|
474 |
+
!3 = !DIFile(filename: "ccig6fki6p4lxrdmgg6eudahiexcvueeol2p4qp532pvve2y463y.py", directory: "/tmp/torchinductor_root/ci")
|
475 |
+
!4 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"kernel", i32 1}
|
476 |
+
!5 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"maxntidx", i32 128}
|
477 |
+
!6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
|
478 |
+
!7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5d6de7de", linkageName: "triton__0d1d2d3d4d5d6de7de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
|
479 |
+
!8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
|
480 |
+
!9 = !{}
|
481 |
+
!10 = !DILocation(line: 22, column: 44, scope: !7)
|
482 |
+
!11 = !DILocation(line: 24, column: 33, scope: !7)
|
483 |
+
!12 = !DILocation(line: 21, column: 28, scope: !7)
|
484 |
+
!13 = !DILocation(line: 21, column: 33, scope: !7)
|
485 |
+
!14 = !DILocation(line: 22, column: 23, scope: !7)
|
486 |
+
!15 = !DILocation(line: 26, column: 30, scope: !7)
|
487 |
+
!16 = !DILocation(line: 26, column: 35, scope: !7)
|
488 |
+
!17 = !DILocation(line: 27, column: 18, scope: !7)
|
489 |
+
!18 = !DILocation(line: 35, column: 44, scope: !7)
|
490 |
+
!19 = !DILocation(line: 35, column: 40, scope: !7)
|
491 |
+
!20 = !DILocation(line: 35, column: 34, scope: !7)
|
492 |
+
!21 = !DILocation(line: 35, column: 50, scope: !7)
|
493 |
+
!22 = !DILocation(line: 36, column: 44, scope: !7)
|
494 |
+
!23 = !DILocation(line: 36, column: 40, scope: !7)
|
495 |
+
!24 = !DILocation(line: 36, column: 34, scope: !7)
|
496 |
+
!25 = !DILocation(line: 36, column: 50, scope: !7)
|
497 |
+
!26 = !DILocation(line: 36, column: 101, scope: !7)
|
498 |
+
!27 = !DILocation(line: 37, column: 22, scope: !7)
|
499 |
+
!28 = !DILocation(line: 38, column: 22, scope: !7)
|
500 |
+
!29 = !DILocation(line: 39, column: 36, scope: !7)
|
501 |
+
!30 = !DILocation(line: 40, column: 40, scope: !7)
|
502 |
+
!31 = !DILocation(line: 40, column: 55, scope: !7)
|
503 |
+
!32 = !DILocation(line: 41, column: 44, scope: !7)
|
504 |
+
!33 = !DILocation(line: 41, column: 40, scope: !7)
|
505 |
+
!34 = !DILocation(line: 41, column: 34, scope: !7)
|
506 |
+
!35 = !DILocation(line: 41, column: 52, scope: !7)
|
507 |
+
!36 = !DILocation(line: 42, column: 22, scope: !7)
|
508 |
+
!37 = !DILocation(line: 44, column: 22, scope: !7)
|
509 |
+
!38 = !DILocation(line: 98, column: 22, scope: !39, inlinedAt: !41)
|
510 |
+
!39 = distinct !DILexicalBlockFile(scope: !7, file: !40, discriminator: 0)
|
511 |
+
!40 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
|
512 |
+
!41 = !DILocation(line: 47, column: 41, scope: !39)
|
513 |
+
!42 = !DILocation(line: 108, column: 21, scope: !43, inlinedAt: !44)
|
514 |
+
!43 = distinct !DILexicalBlockFile(scope: !39, file: !40, discriminator: 0)
|
515 |
+
!44 = !DILocation(line: 120, column: 46, scope: !43, inlinedAt: !45)
|
516 |
+
!45 = !DILocation(line: 53, column: 44, scope: !43)
|
517 |
+
!46 = !DILocation(line: 101, column: 22, scope: !39, inlinedAt: !41)
|
518 |
+
!47 = !DILocation(line: 101, column: 30, scope: !39, inlinedAt: !41)
|
519 |
+
!48 = !DILocation(line: 101, column: 13, scope: !39, inlinedAt: !41)
|
520 |
+
!49 = !DILocation(line: 110, column: 60, scope: !43, inlinedAt: !44)
|
521 |
+
!50 = !DILocation(line: 112, column: 25, scope: !43, inlinedAt: !44)
|
522 |
+
!51 = !DILocation(line: 112, column: 17, scope: !43, inlinedAt: !44)
|
523 |
+
!52 = !DILocation(line: 113, column: 15, scope: !43, inlinedAt: !44)
|
524 |
+
!53 = !DILocation(line: 113, column: 30, scope: !43, inlinedAt: !44)
|
525 |
+
!54 = !DILocation(line: 113, column: 49, scope: !43, inlinedAt: !44)
|
526 |
+
!55 = !DILocation(line: 113, column: 22, scope: !43, inlinedAt: !44)
|
527 |
+
!56 = !DILocation(line: 113, column: 38, scope: !43, inlinedAt: !44)
|
528 |
+
!57 = !DILocation(line: 120, column: 46, scope: !39, inlinedAt: !58)
|
529 |
+
!58 = !DILocation(line: 53, column: 44, scope: !39)
|
530 |
+
!59 = !DILocation(line: 109, column: 28, scope: !43, inlinedAt: !44)
|
531 |
+
!60 = !DILocation(line: 110, column: 39, scope: !43, inlinedAt: !44)
|
532 |
+
!61 = !DILocation(line: 110, column: 49, scope: !43, inlinedAt: !44)
|
533 |
+
!62 = !DILocation(line: 62, column: 51, scope: !7)
|
534 |
+
!63 = !DILocation(line: 63, column: 51, scope: !7)
|
535 |
+
!64 = !DILocation(line: 63, column: 103, scope: !7)
|
536 |
+
!65 = !DILocation(line: 64, column: 35, scope: !7)
|
537 |
+
!66 = !DILocation(line: 64, column: 40, scope: !7)
|
538 |
+
!67 = !DILocation(line: 68, column: 57, scope: !7)
|
539 |
+
!68 = !DILocation(line: 69, column: 54, scope: !7)
|
540 |
+
!69 = !DILocation(line: 75, column: 24, scope: !7)
|
541 |
+
!70 = !DILocation(line: 77, column: 24, scope: !7)
|
542 |
+
!71 = !DILocation(line: 78, column: 30, scope: !7)
|
543 |
+
!72 = !DILocation(line: 70, column: 24, scope: !7)
|
544 |
+
!73 = !DILocation(line: 72, column: 24, scope: !7)
|
545 |
+
!74 = !DILocation(line: 73, column: 24, scope: !7)
|
546 |
+
!75 = !DILocation(line: 79, column: 24, scope: !7)
|
547 |
+
!76 = !DILocation(line: 80, column: 24, scope: !7)
|
548 |
+
!77 = !DILocation(line: 82, column: 29, scope: !7)
|
549 |
+
!78 = !DILocation(line: 82, column: 52, scope: !7)
|
550 |
+
!79 = !DILocation(line: 58, column: 4, scope: !7)
|
.triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.cubin
ADDED
Binary file (5.54 kB). View file
|
|
.triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.llir
ADDED
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
; ModuleID = 'LLVMDialectModule'
|
2 |
+
source_filename = "LLVMDialectModule"
|
3 |
+
|
4 |
+
define void @triton__0d1d2de(ptr addrspace(1) %0, ptr addrspace(1) %1, i32 %2) local_unnamed_addr !dbg !5 {
|
5 |
+
%4 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !8
|
6 |
+
%5 = shl i32 %4, 3, !dbg !8
|
7 |
+
%6 = and i32 %5, 1016, !dbg !8
|
8 |
+
%7 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #1, !dbg !9
|
9 |
+
%8 = shl i32 %7, 10, !dbg !10
|
10 |
+
%9 = or i32 %8, %6, !dbg !11
|
11 |
+
%10 = or i32 %9, 4, !dbg !11
|
12 |
+
%11 = sext i32 %9 to i64, !dbg !12
|
13 |
+
%12 = getelementptr float, ptr addrspace(1) %0, i64 %11, !dbg !12
|
14 |
+
%13 = sext i32 %10 to i64, !dbg !12
|
15 |
+
%14 = getelementptr float, ptr addrspace(1) %0, i64 %13, !dbg !12
|
16 |
+
%15 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];", "=r,=r,=r,=r,l,b"(ptr addrspace(1) %12, i1 true) #1, !dbg !13
|
17 |
+
%16 = extractvalue { i32, i32, i32, i32 } %15, 0, !dbg !13
|
18 |
+
%17 = extractvalue { i32, i32, i32, i32 } %15, 1, !dbg !13
|
19 |
+
%18 = extractvalue { i32, i32, i32, i32 } %15, 2, !dbg !13
|
20 |
+
%19 = extractvalue { i32, i32, i32, i32 } %15, 3, !dbg !13
|
21 |
+
%20 = bitcast i32 %16 to float, !dbg !13
|
22 |
+
%21 = bitcast i32 %17 to float, !dbg !13
|
23 |
+
%22 = bitcast i32 %18 to float, !dbg !13
|
24 |
+
%23 = bitcast i32 %19 to float, !dbg !13
|
25 |
+
%24 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];", "=r,=r,=r,=r,l,b"(ptr addrspace(1) %14, i1 true) #1, !dbg !13
|
26 |
+
%25 = extractvalue { i32, i32, i32, i32 } %24, 0, !dbg !13
|
27 |
+
%26 = extractvalue { i32, i32, i32, i32 } %24, 1, !dbg !13
|
28 |
+
%27 = extractvalue { i32, i32, i32, i32 } %24, 2, !dbg !13
|
29 |
+
%28 = extractvalue { i32, i32, i32, i32 } %24, 3, !dbg !13
|
30 |
+
%29 = bitcast i32 %25 to float, !dbg !13
|
31 |
+
%30 = bitcast i32 %26 to float, !dbg !13
|
32 |
+
%31 = bitcast i32 %27 to float, !dbg !13
|
33 |
+
%32 = bitcast i32 %28 to float, !dbg !13
|
34 |
+
%33 = getelementptr i16, ptr addrspace(1) %1, i64 %11, !dbg !14
|
35 |
+
%34 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %20) #1, !dbg !15
|
36 |
+
%35 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %21) #1, !dbg !15
|
37 |
+
%36 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %22) #1, !dbg !15
|
38 |
+
%37 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %23) #1, !dbg !15
|
39 |
+
%38 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %29) #1, !dbg !15
|
40 |
+
%39 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %30) #1, !dbg !15
|
41 |
+
%40 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %31) #1, !dbg !15
|
42 |
+
%41 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %32) #1, !dbg !15
|
43 |
+
%42 = insertelement <2 x i16> undef, i16 %34, i64 0, !dbg !15
|
44 |
+
%43 = insertelement <2 x i16> %42, i16 %35, i64 1, !dbg !15
|
45 |
+
%44 = bitcast <2 x i16> %43 to i32, !dbg !15
|
46 |
+
%45 = insertelement <2 x i16> undef, i16 %36, i64 0, !dbg !15
|
47 |
+
%46 = insertelement <2 x i16> %45, i16 %37, i64 1, !dbg !15
|
48 |
+
%47 = bitcast <2 x i16> %46 to i32, !dbg !15
|
49 |
+
%48 = insertelement <2 x i16> undef, i16 %38, i64 0, !dbg !15
|
50 |
+
%49 = insertelement <2 x i16> %48, i16 %39, i64 1, !dbg !15
|
51 |
+
%50 = bitcast <2 x i16> %49 to i32, !dbg !15
|
52 |
+
%51 = insertelement <2 x i16> undef, i16 %40, i64 0, !dbg !15
|
53 |
+
%52 = insertelement <2 x i16> %51, i16 %41, i64 1, !dbg !15
|
54 |
+
%53 = bitcast <2 x i16> %52 to i32, !dbg !15
|
55 |
+
tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %44, i32 %47, i32 %50, i32 %53, ptr addrspace(1) %33, i1 true) #1, !dbg !15
|
56 |
+
ret void, !dbg !16
|
57 |
+
}
|
58 |
+
|
59 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
|
60 |
+
declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
|
61 |
+
|
62 |
+
attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
|
63 |
+
attributes #1 = { nounwind }
|
64 |
+
|
65 |
+
!llvm.module.flags = !{!0}
|
66 |
+
!llvm.dbg.cu = !{!1}
|
67 |
+
!nvvm.annotations = !{!3, !4, !4, !3}
|
68 |
+
|
69 |
+
!0 = !{i32 2, !"Debug Info Version", i32 3}
|
70 |
+
!1 = distinct !DICompileUnit(language: DW_LANG_C, file: !2, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
|
71 |
+
!2 = !DIFile(filename: "czjxjqxojsyyr4zmce6q6twysnucw6p4l5ujgp6ts2ecrm3ue3ex.py", directory: "/tmp/torchinductor_root/zj")
|
72 |
+
!3 = !{ptr @triton__0d1d2de, !"kernel", i32 1}
|
73 |
+
!4 = !{ptr @triton__0d1d2de, !"maxntidx", i32 128}
|
74 |
+
!5 = distinct !DISubprogram(name: "triton__0d1d2de", linkageName: "triton__0d1d2de", scope: !2, file: !2, line: 18, type: !6, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !1)
|
75 |
+
!6 = !DISubroutineType(cc: DW_CC_normal, types: !7)
|
76 |
+
!7 = !{}
|
77 |
+
!8 = !DILocation(line: 21, column: 36, scope: !5)
|
78 |
+
!9 = !DILocation(line: 20, column: 28, scope: !5)
|
79 |
+
!10 = !DILocation(line: 20, column: 33, scope: !5)
|
80 |
+
!11 = !DILocation(line: 21, column: 23, scope: !5)
|
81 |
+
!12 = !DILocation(line: 24, column: 30, scope: !5)
|
82 |
+
!13 = !DILocation(line: 24, column: 35, scope: !5)
|
83 |
+
!14 = !DILocation(line: 26, column: 25, scope: !5)
|
84 |
+
!15 = !DILocation(line: 26, column: 36, scope: !5)
|
85 |
+
!16 = !DILocation(line: 26, column: 4, scope: !5)
|
.triton/dump/510522bb05917b836ed253751364fcad/triton_.cubin
ADDED
Binary file (66.2 kB). View file
|
|
.triton/dump/510522bb05917b836ed253751364fcad/triton_.ptx
ADDED
@@ -0,0 +1,1810 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
//
|
2 |
+
// Generated by LLVM NVPTX Back-End
|
3 |
+
//
|
4 |
+
|
5 |
+
.version 8.2
|
6 |
+
.target sm_89
|
7 |
+
.address_size 64
|
8 |
+
|
9 |
+
// .globl triton__0d1d2d3d4d5de6de
|
10 |
+
.extern .func __assertfail
|
11 |
+
(
|
12 |
+
.param .b64 __assertfail_param_0,
|
13 |
+
.param .b64 __assertfail_param_1,
|
14 |
+
.param .b32 __assertfail_param_2,
|
15 |
+
.param .b64 __assertfail_param_3,
|
16 |
+
.param .b64 __assertfail_param_4
|
17 |
+
)
|
18 |
+
;
|
19 |
+
.global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
20 |
+
.global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
21 |
+
.global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
22 |
+
.global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
23 |
+
.global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
24 |
+
.global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
25 |
+
.extern .shared .align 1 .b8 global_smem[];
|
26 |
+
.global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
|
27 |
+
|
28 |
+
.visible .entry triton__0d1d2d3d4d5de6de(
|
29 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_0,
|
30 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_1,
|
31 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_2,
|
32 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_3,
|
33 |
+
.param .u64 triton__0d1d2d3d4d5de6de_param_4,
|
34 |
+
.param .u32 triton__0d1d2d3d4d5de6de_param_5,
|
35 |
+
.param .u32 triton__0d1d2d3d4d5de6de_param_6
|
36 |
+
)
|
37 |
+
.maxntid 256, 1, 1
|
38 |
+
{
|
39 |
+
.reg .pred %p<137>;
|
40 |
+
.reg .b16 %rs<17>;
|
41 |
+
.reg .b32 %r<408>;
|
42 |
+
.reg .f32 %f<614>;
|
43 |
+
.reg .b64 %rd<107>;
|
44 |
+
.loc 1 18 0
|
45 |
+
$L__func_begin0:
|
46 |
+
.loc 1 18 0
|
47 |
+
|
48 |
+
ld.param.u64 %rd13, [triton__0d1d2d3d4d5de6de_param_4];
|
49 |
+
ld.param.u64 %rd12, [triton__0d1d2d3d4d5de6de_param_3];
|
50 |
+
ld.param.u64 %rd49, [triton__0d1d2d3d4d5de6de_param_0];
|
51 |
+
ld.param.u64 %rd50, [triton__0d1d2d3d4d5de6de_param_1];
|
52 |
+
$L__tmp0:
|
53 |
+
.loc 1 22 44
|
54 |
+
mov.u32 %r13, %tid.x;
|
55 |
+
ld.param.u64 %rd51, [triton__0d1d2d3d4d5de6de_param_2];
|
56 |
+
bfe.u32 %r1, %r13, 3, 5;
|
57 |
+
and.b32 %r2, %r13, 63;
|
58 |
+
.loc 1 24 33
|
59 |
+
shl.b32 %r14, %r13, 3;
|
60 |
+
and.b32 %r3, %r14, 56;
|
61 |
+
.loc 1 31 36
|
62 |
+
shr.u32 %r4, %r13, 6;
|
63 |
+
.loc 1 21 28
|
64 |
+
mov.u32 %r11, %ctaid.x;
|
65 |
+
.loc 1 21 33
|
66 |
+
shl.b32 %r15, %r11, 6;
|
67 |
+
.loc 1 22 23
|
68 |
+
or.b32 %r16, %r15, %r1;
|
69 |
+
or.b32 %r17, %r16, 32;
|
70 |
+
or.b32 %r18, %r15, %r2;
|
71 |
+
.loc 1 26 30
|
72 |
+
mul.wide.s32 %rd52, %r16, 8;
|
73 |
+
add.s64 %rd15, %rd49, %rd52;
|
74 |
+
add.s64 %rd31, %rd15, 256;
|
75 |
+
mul.wide.s32 %rd53, %r18, 8;
|
76 |
+
add.s64 %rd47, %rd49, %rd53;
|
77 |
+
mov.pred %p1, -1;
|
78 |
+
.loc 1 26 35
|
79 |
+
mov.u64 %rd14, 0x0;
|
80 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd14 }, [ %rd15 + 0 ];
|
81 |
+
mov.u64 %rd16, 0x0;
|
82 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd16 }, [ %rd15 + 0 ];
|
83 |
+
mov.u64 %rd18, 0x0;
|
84 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd18 }, [ %rd15 + 0 ];
|
85 |
+
mov.u64 %rd20, 0x0;
|
86 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd20 }, [ %rd15 + 0 ];
|
87 |
+
mov.u64 %rd22, 0x0;
|
88 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd22 }, [ %rd15 + 0 ];
|
89 |
+
mov.u64 %rd24, 0x0;
|
90 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd24 }, [ %rd15 + 0 ];
|
91 |
+
mov.u64 %rd26, 0x0;
|
92 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd26 }, [ %rd15 + 0 ];
|
93 |
+
mov.u64 %rd28, 0x0;
|
94 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd28 }, [ %rd15 + 0 ];
|
95 |
+
mov.u64 %rd30, 0x0;
|
96 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd30 }, [ %rd31 + 0 ];
|
97 |
+
mov.u64 %rd32, 0x0;
|
98 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd32 }, [ %rd31 + 0 ];
|
99 |
+
mov.u64 %rd34, 0x0;
|
100 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd34 }, [ %rd31 + 0 ];
|
101 |
+
mov.u64 %rd36, 0x0;
|
102 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd36 }, [ %rd31 + 0 ];
|
103 |
+
mov.u64 %rd38, 0x0;
|
104 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd38 }, [ %rd31 + 0 ];
|
105 |
+
mov.u64 %rd40, 0x0;
|
106 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd40 }, [ %rd31 + 0 ];
|
107 |
+
mov.u64 %rd42, 0x0;
|
108 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd42 }, [ %rd31 + 0 ];
|
109 |
+
mov.u64 %rd44, 0x0;
|
110 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd44 }, [ %rd31 + 0 ];
|
111 |
+
mov.u64 %rd46, 0x0;
|
112 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd46 }, [ %rd47 + 0 ];
|
113 |
+
.loc 1 27 18
|
114 |
+
bfe.s32 %r19, %r11, 25, 1;
|
115 |
+
shr.u32 %r20, %r19, 23;
|
116 |
+
add.s32 %r21, %r16, %r20;
|
117 |
+
and.b32 %r22, %r21, 16776704;
|
118 |
+
sub.s32 %r23, %r16, %r22;
|
119 |
+
add.s32 %r24, %r17, %r20;
|
120 |
+
and.b32 %r25, %r24, 16776704;
|
121 |
+
sub.s32 %r26, %r17, %r25;
|
122 |
+
.loc 1 35 44
|
123 |
+
shl.b32 %r27, %r23, 8;
|
124 |
+
shl.b32 %r28, %r26, 8;
|
125 |
+
.loc 1 36 22
|
126 |
+
add.s64 %rd54, %rd46, 50257;
|
127 |
+
.loc 1 37 22
|
128 |
+
setp.lt.s64 %p18, %rd14, 0;
|
129 |
+
setp.lt.s64 %p19, %rd30, 0;
|
130 |
+
setp.lt.s64 %p20, %rd46, 0;
|
131 |
+
.loc 1 38 36
|
132 |
+
selp.b64 %rd1, %rd54, %rd46, %p20;
|
133 |
+
.loc 1 40 44
|
134 |
+
shl.b64 %rd55, %rd14, 8;
|
135 |
+
add.s64 %rd56, %rd55, 12865792;
|
136 |
+
selp.b64 %rd57, %rd56, %rd55, %p18;
|
137 |
+
shl.b64 %rd58, %rd30, 8;
|
138 |
+
add.s64 %rd59, %rd58, 12865792;
|
139 |
+
selp.b64 %rd60, %rd59, %rd58, %p19;
|
140 |
+
.loc 1 31 36
|
141 |
+
and.b32 %r29, %r13, 7;
|
142 |
+
mul.wide.u32 %rd2, %r29, 32;
|
143 |
+
shl.b64 %rd61, %rd60, 2;
|
144 |
+
or.b64 %rd62, %rd2, %rd61;
|
145 |
+
add.s64 %rd3, %rd50, %rd62;
|
146 |
+
shl.b64 %rd63, %rd57, 2;
|
147 |
+
or.b64 %rd64, %rd2, %rd63;
|
148 |
+
add.s64 %rd4, %rd50, %rd64;
|
149 |
+
or.b32 %r30, %r28, %r3;
|
150 |
+
mul.wide.s32 %rd65, %r30, 4;
|
151 |
+
add.s64 %rd5, %rd51, %rd65;
|
152 |
+
or.b32 %r31, %r27, %r3;
|
153 |
+
mul.wide.s32 %rd66, %r31, 4;
|
154 |
+
add.s64 %rd6, %rd51, %rd66;
|
155 |
+
mov.f32 %f550, 0f00000000;
|
156 |
+
mov.u64 %rd105, 0;
|
157 |
+
mov.b32 %r406, -64;
|
158 |
+
mov.f32 %f551, %f550;
|
159 |
+
mov.f32 %f552, %f550;
|
160 |
+
mov.f32 %f553, %f550;
|
161 |
+
mov.f32 %f554, %f550;
|
162 |
+
mov.f32 %f555, %f550;
|
163 |
+
mov.f32 %f556, %f550;
|
164 |
+
mov.f32 %f557, %f550;
|
165 |
+
mov.f32 %f558, %f550;
|
166 |
+
mov.f32 %f559, %f550;
|
167 |
+
mov.f32 %f560, %f550;
|
168 |
+
mov.f32 %f561, %f550;
|
169 |
+
mov.f32 %f562, %f550;
|
170 |
+
mov.f32 %f563, %f550;
|
171 |
+
mov.f32 %f564, %f550;
|
172 |
+
mov.f32 %f565, %f550;
|
173 |
+
mov.f32 %f566, %f550;
|
174 |
+
mov.f32 %f567, %f550;
|
175 |
+
mov.f32 %f568, %f550;
|
176 |
+
mov.f32 %f569, %f550;
|
177 |
+
mov.f32 %f570, %f550;
|
178 |
+
mov.f32 %f571, %f550;
|
179 |
+
mov.f32 %f572, %f550;
|
180 |
+
mov.f32 %f573, %f550;
|
181 |
+
mov.f32 %f574, %f550;
|
182 |
+
mov.f32 %f575, %f550;
|
183 |
+
mov.f32 %f576, %f550;
|
184 |
+
mov.f32 %f577, %f550;
|
185 |
+
mov.f32 %f578, %f550;
|
186 |
+
mov.f32 %f579, %f550;
|
187 |
+
mov.f32 %f580, %f550;
|
188 |
+
mov.f32 %f581, %f550;
|
189 |
+
mov.f32 %f582, %f550;
|
190 |
+
mov.f32 %f583, %f550;
|
191 |
+
mov.f32 %f584, %f550;
|
192 |
+
mov.f32 %f585, %f550;
|
193 |
+
mov.f32 %f586, %f550;
|
194 |
+
mov.f32 %f587, %f550;
|
195 |
+
mov.f32 %f588, %f550;
|
196 |
+
mov.f32 %f589, %f550;
|
197 |
+
mov.f32 %f590, %f550;
|
198 |
+
mov.f32 %f591, %f550;
|
199 |
+
mov.f32 %f592, %f550;
|
200 |
+
mov.f32 %f593, %f550;
|
201 |
+
mov.f32 %f594, %f550;
|
202 |
+
mov.f32 %f595, %f550;
|
203 |
+
mov.f32 %f596, %f550;
|
204 |
+
mov.f32 %f597, %f550;
|
205 |
+
mov.f32 %f598, %f550;
|
206 |
+
mov.f32 %f599, %f550;
|
207 |
+
mov.f32 %f600, %f550;
|
208 |
+
mov.f32 %f601, %f550;
|
209 |
+
mov.f32 %f602, %f550;
|
210 |
+
mov.f32 %f603, %f550;
|
211 |
+
mov.f32 %f604, %f550;
|
212 |
+
mov.f32 %f605, %f550;
|
213 |
+
mov.f32 %f606, %f550;
|
214 |
+
mov.f32 %f607, %f550;
|
215 |
+
mov.f32 %f608, %f550;
|
216 |
+
mov.f32 %f609, %f550;
|
217 |
+
mov.f32 %f610, %f550;
|
218 |
+
mov.f32 %f611, %f550;
|
219 |
+
mov.f32 %f612, %f550;
|
220 |
+
mov.f32 %f613, %f550;
|
221 |
+
bra.uni $L__BB0_1;
|
222 |
+
$L__BB0_3:
|
223 |
+
.loc 1 40 40
|
224 |
+
add.s64 %rd78, %rd4, %rd105;
|
225 |
+
.loc 1 40 34
|
226 |
+
add.s64 %rd79, %rd78, 16;
|
227 |
+
add.s64 %rd80, %rd3, %rd105;
|
228 |
+
.loc 1 40 52
|
229 |
+
add.s64 %rd81, %rd80, 16;
|
230 |
+
mov.u32 %r65, 0x0;
|
231 |
+
mov.u32 %r66, 0x0;
|
232 |
+
mov.u32 %r67, 0x0;
|
233 |
+
mov.u32 %r68, 0x0;
|
234 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r65, %r66, %r67, %r68 }, [ %rd78 + 0 ];
|
235 |
+
@!%p1 mov.u32 %r65, %r342;
|
236 |
+
@!%p1 mov.u32 %r66, %r342;
|
237 |
+
@!%p1 mov.u32 %r67, %r342;
|
238 |
+
@!%p1 mov.u32 %r68, %r342;
|
239 |
+
mov.b32 %f174, %r65;
|
240 |
+
mov.b32 %f175, %r66;
|
241 |
+
mov.b32 %f176, %r67;
|
242 |
+
mov.b32 %f177, %r68;
|
243 |
+
mov.u32 %r73, 0x0;
|
244 |
+
mov.u32 %r74, 0x0;
|
245 |
+
mov.u32 %r75, 0x0;
|
246 |
+
mov.u32 %r76, 0x0;
|
247 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r73, %r74, %r75, %r76 }, [ %rd79 + 0 ];
|
248 |
+
@!%p1 mov.u32 %r73, %r342;
|
249 |
+
@!%p1 mov.u32 %r74, %r342;
|
250 |
+
@!%p1 mov.u32 %r75, %r342;
|
251 |
+
@!%p1 mov.u32 %r76, %r342;
|
252 |
+
mov.b32 %f178, %r73;
|
253 |
+
mov.b32 %f179, %r74;
|
254 |
+
mov.b32 %f180, %r75;
|
255 |
+
mov.b32 %f181, %r76;
|
256 |
+
mov.u32 %r81, 0x0;
|
257 |
+
mov.u32 %r82, 0x0;
|
258 |
+
mov.u32 %r83, 0x0;
|
259 |
+
mov.u32 %r84, 0x0;
|
260 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r81, %r82, %r83, %r84 }, [ %rd80 + 0 ];
|
261 |
+
@!%p1 mov.u32 %r81, %r342;
|
262 |
+
@!%p1 mov.u32 %r82, %r342;
|
263 |
+
@!%p1 mov.u32 %r83, %r342;
|
264 |
+
@!%p1 mov.u32 %r84, %r342;
|
265 |
+
mov.b32 %f182, %r81;
|
266 |
+
mov.b32 %f183, %r82;
|
267 |
+
mov.b32 %f184, %r83;
|
268 |
+
mov.b32 %f185, %r84;
|
269 |
+
mov.u32 %r89, 0x0;
|
270 |
+
mov.u32 %r90, 0x0;
|
271 |
+
mov.u32 %r91, 0x0;
|
272 |
+
mov.u32 %r92, 0x0;
|
273 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r89, %r90, %r91, %r92 }, [ %rd81 + 0 ];
|
274 |
+
@!%p1 mov.u32 %r89, %r342;
|
275 |
+
@!%p1 mov.u32 %r90, %r342;
|
276 |
+
@!%p1 mov.u32 %r91, %r342;
|
277 |
+
@!%p1 mov.u32 %r92, %r342;
|
278 |
+
mov.b32 %f186, %r89;
|
279 |
+
mov.b32 %f187, %r90;
|
280 |
+
mov.b32 %f188, %r91;
|
281 |
+
mov.b32 %f189, %r92;
|
282 |
+
.loc 1 41 22
|
283 |
+
add.f32 %f190, %f65, %f174;
|
284 |
+
add.f32 %f191, %f66, %f175;
|
285 |
+
add.f32 %f192, %f67, %f176;
|
286 |
+
add.f32 %f193, %f68, %f177;
|
287 |
+
add.f32 %f194, %f69, %f178;
|
288 |
+
add.f32 %f195, %f70, %f179;
|
289 |
+
add.f32 %f196, %f71, %f180;
|
290 |
+
add.f32 %f197, %f72, %f181;
|
291 |
+
add.f32 %f198, %f73, %f182;
|
292 |
+
add.f32 %f199, %f74, %f183;
|
293 |
+
add.f32 %f200, %f75, %f184;
|
294 |
+
add.f32 %f201, %f76, %f185;
|
295 |
+
add.f32 %f202, %f77, %f186;
|
296 |
+
add.f32 %f203, %f78, %f187;
|
297 |
+
add.f32 %f204, %f79, %f188;
|
298 |
+
add.f32 %f205, %f80, %f189;
|
299 |
+
$L__tmp1:
|
300 |
+
.loc 2 96 20
|
301 |
+
sub.f32 %f206, %f190, %f598;
|
302 |
+
sub.f32 %f207, %f191, %f599;
|
303 |
+
sub.f32 %f208, %f192, %f600;
|
304 |
+
sub.f32 %f209, %f193, %f601;
|
305 |
+
sub.f32 %f210, %f194, %f602;
|
306 |
+
sub.f32 %f211, %f195, %f603;
|
307 |
+
sub.f32 %f212, %f196, %f604;
|
308 |
+
sub.f32 %f213, %f197, %f605;
|
309 |
+
sub.f32 %f214, %f198, %f606;
|
310 |
+
sub.f32 %f215, %f199, %f607;
|
311 |
+
sub.f32 %f216, %f200, %f608;
|
312 |
+
sub.f32 %f217, %f201, %f609;
|
313 |
+
sub.f32 %f218, %f202, %f610;
|
314 |
+
sub.f32 %f219, %f203, %f611;
|
315 |
+
sub.f32 %f220, %f204, %f612;
|
316 |
+
sub.f32 %f221, %f205, %f613;
|
317 |
+
.loc 2 97 26
|
318 |
+
add.f32 %f550, %f550, 0f3F800000;
|
319 |
+
add.f32 %f551, %f551, 0f3F800000;
|
320 |
+
add.f32 %f552, %f552, 0f3F800000;
|
321 |
+
add.f32 %f553, %f553, 0f3F800000;
|
322 |
+
add.f32 %f554, %f554, 0f3F800000;
|
323 |
+
add.f32 %f555, %f555, 0f3F800000;
|
324 |
+
add.f32 %f556, %f556, 0f3F800000;
|
325 |
+
add.f32 %f557, %f557, 0f3F800000;
|
326 |
+
add.f32 %f558, %f558, 0f3F800000;
|
327 |
+
add.f32 %f559, %f559, 0f3F800000;
|
328 |
+
add.f32 %f560, %f560, 0f3F800000;
|
329 |
+
add.f32 %f561, %f561, 0f3F800000;
|
330 |
+
add.f32 %f562, %f562, 0f3F800000;
|
331 |
+
add.f32 %f563, %f563, 0f3F800000;
|
332 |
+
add.f32 %f564, %f564, 0f3F800000;
|
333 |
+
add.f32 %f565, %f565, 0f3F800000;
|
334 |
+
add.f32 %f566, %f566, 0f3F800000;
|
335 |
+
add.f32 %f567, %f567, 0f3F800000;
|
336 |
+
add.f32 %f568, %f568, 0f3F800000;
|
337 |
+
add.f32 %f569, %f569, 0f3F800000;
|
338 |
+
add.f32 %f570, %f570, 0f3F800000;
|
339 |
+
add.f32 %f571, %f571, 0f3F800000;
|
340 |
+
add.f32 %f572, %f572, 0f3F800000;
|
341 |
+
add.f32 %f573, %f573, 0f3F800000;
|
342 |
+
add.f32 %f574, %f574, 0f3F800000;
|
343 |
+
add.f32 %f575, %f575, 0f3F800000;
|
344 |
+
add.f32 %f576, %f576, 0f3F800000;
|
345 |
+
add.f32 %f577, %f577, 0f3F800000;
|
346 |
+
add.f32 %f578, %f578, 0f3F800000;
|
347 |
+
add.f32 %f579, %f579, 0f3F800000;
|
348 |
+
add.f32 %f580, %f580, 0f3F800000;
|
349 |
+
add.f32 %f581, %f581, 0f3F800000;
|
350 |
+
.loc 2 98 30
|
351 |
+
mov.b32 %r98, %f206;
|
352 |
+
mov.b32 %r99, %f550;
|
353 |
+
div.full.f32 %r97, %r98, %r99;
|
354 |
+
mov.b32 %f222, %r97;
|
355 |
+
mov.b32 %r101, %f207;
|
356 |
+
mov.b32 %r102, %f551;
|
357 |
+
div.full.f32 %r100, %r101, %r102;
|
358 |
+
mov.b32 %f223, %r100;
|
359 |
+
mov.b32 %r104, %f208;
|
360 |
+
mov.b32 %r105, %f552;
|
361 |
+
div.full.f32 %r103, %r104, %r105;
|
362 |
+
mov.b32 %f224, %r103;
|
363 |
+
mov.b32 %r107, %f209;
|
364 |
+
mov.b32 %r108, %f553;
|
365 |
+
div.full.f32 %r106, %r107, %r108;
|
366 |
+
mov.b32 %f225, %r106;
|
367 |
+
mov.b32 %r110, %f210;
|
368 |
+
mov.b32 %r111, %f554;
|
369 |
+
div.full.f32 %r109, %r110, %r111;
|
370 |
+
mov.b32 %f226, %r109;
|
371 |
+
mov.b32 %r113, %f211;
|
372 |
+
mov.b32 %r114, %f555;
|
373 |
+
div.full.f32 %r112, %r113, %r114;
|
374 |
+
mov.b32 %f227, %r112;
|
375 |
+
mov.b32 %r116, %f212;
|
376 |
+
mov.b32 %r117, %f556;
|
377 |
+
div.full.f32 %r115, %r116, %r117;
|
378 |
+
mov.b32 %f228, %r115;
|
379 |
+
mov.b32 %r119, %f213;
|
380 |
+
mov.b32 %r120, %f557;
|
381 |
+
div.full.f32 %r118, %r119, %r120;
|
382 |
+
mov.b32 %f229, %r118;
|
383 |
+
mov.b32 %r122, %f214;
|
384 |
+
mov.b32 %r123, %f558;
|
385 |
+
div.full.f32 %r121, %r122, %r123;
|
386 |
+
mov.b32 %f230, %r121;
|
387 |
+
mov.b32 %r125, %f215;
|
388 |
+
mov.b32 %r126, %f559;
|
389 |
+
div.full.f32 %r124, %r125, %r126;
|
390 |
+
mov.b32 %f231, %r124;
|
391 |
+
mov.b32 %r128, %f216;
|
392 |
+
mov.b32 %r129, %f560;
|
393 |
+
div.full.f32 %r127, %r128, %r129;
|
394 |
+
mov.b32 %f232, %r127;
|
395 |
+
mov.b32 %r131, %f217;
|
396 |
+
mov.b32 %r132, %f561;
|
397 |
+
div.full.f32 %r130, %r131, %r132;
|
398 |
+
mov.b32 %f233, %r130;
|
399 |
+
mov.b32 %r134, %f218;
|
400 |
+
mov.b32 %r135, %f562;
|
401 |
+
div.full.f32 %r133, %r134, %r135;
|
402 |
+
mov.b32 %f234, %r133;
|
403 |
+
mov.b32 %r137, %f219;
|
404 |
+
mov.b32 %r138, %f563;
|
405 |
+
div.full.f32 %r136, %r137, %r138;
|
406 |
+
mov.b32 %f235, %r136;
|
407 |
+
mov.b32 %r140, %f220;
|
408 |
+
mov.b32 %r141, %f564;
|
409 |
+
div.full.f32 %r139, %r140, %r141;
|
410 |
+
mov.b32 %f236, %r139;
|
411 |
+
mov.b32 %r143, %f221;
|
412 |
+
mov.b32 %r144, %f565;
|
413 |
+
div.full.f32 %r142, %r143, %r144;
|
414 |
+
mov.b32 %f237, %r142;
|
415 |
+
.loc 2 98 22
|
416 |
+
add.f32 %f598, %f598, %f222;
|
417 |
+
add.f32 %f599, %f599, %f223;
|
418 |
+
add.f32 %f600, %f600, %f224;
|
419 |
+
add.f32 %f601, %f601, %f225;
|
420 |
+
add.f32 %f602, %f602, %f226;
|
421 |
+
add.f32 %f603, %f603, %f227;
|
422 |
+
add.f32 %f604, %f604, %f228;
|
423 |
+
add.f32 %f605, %f605, %f229;
|
424 |
+
add.f32 %f606, %f606, %f230;
|
425 |
+
add.f32 %f607, %f607, %f231;
|
426 |
+
add.f32 %f608, %f608, %f232;
|
427 |
+
add.f32 %f609, %f609, %f233;
|
428 |
+
add.f32 %f610, %f610, %f234;
|
429 |
+
add.f32 %f611, %f611, %f235;
|
430 |
+
add.f32 %f612, %f612, %f236;
|
431 |
+
add.f32 %f613, %f613, %f237;
|
432 |
+
.loc 2 101 30
|
433 |
+
sub.f32 %f238, %f190, %f598;
|
434 |
+
sub.f32 %f239, %f191, %f599;
|
435 |
+
sub.f32 %f240, %f192, %f600;
|
436 |
+
sub.f32 %f241, %f193, %f601;
|
437 |
+
sub.f32 %f242, %f194, %f602;
|
438 |
+
sub.f32 %f243, %f195, %f603;
|
439 |
+
sub.f32 %f244, %f196, %f604;
|
440 |
+
sub.f32 %f245, %f197, %f605;
|
441 |
+
sub.f32 %f246, %f198, %f606;
|
442 |
+
sub.f32 %f247, %f199, %f607;
|
443 |
+
sub.f32 %f248, %f200, %f608;
|
444 |
+
sub.f32 %f249, %f201, %f609;
|
445 |
+
sub.f32 %f250, %f202, %f610;
|
446 |
+
sub.f32 %f251, %f203, %f611;
|
447 |
+
sub.f32 %f252, %f204, %f612;
|
448 |
+
sub.f32 %f253, %f205, %f613;
|
449 |
+
$L__tmp2:
|
450 |
+
.loc 1 47 48
|
451 |
+
fma.rn.f32 %f582, %f206, %f238, %f582;
|
452 |
+
fma.rn.f32 %f583, %f207, %f239, %f583;
|
453 |
+
fma.rn.f32 %f584, %f208, %f240, %f584;
|
454 |
+
fma.rn.f32 %f585, %f209, %f241, %f585;
|
455 |
+
fma.rn.f32 %f586, %f210, %f242, %f586;
|
456 |
+
fma.rn.f32 %f587, %f211, %f243, %f587;
|
457 |
+
fma.rn.f32 %f588, %f212, %f244, %f588;
|
458 |
+
fma.rn.f32 %f589, %f213, %f245, %f589;
|
459 |
+
fma.rn.f32 %f590, %f214, %f246, %f590;
|
460 |
+
fma.rn.f32 %f591, %f215, %f247, %f591;
|
461 |
+
fma.rn.f32 %f592, %f216, %f248, %f592;
|
462 |
+
fma.rn.f32 %f593, %f217, %f249, %f593;
|
463 |
+
fma.rn.f32 %f594, %f218, %f250, %f594;
|
464 |
+
fma.rn.f32 %f595, %f219, %f251, %f595;
|
465 |
+
fma.rn.f32 %f596, %f220, %f252, %f596;
|
466 |
+
fma.rn.f32 %f597, %f221, %f253, %f597;
|
467 |
+
.loc 1 31 36
|
468 |
+
add.s64 %rd105, %rd105, 256;
|
469 |
+
add.s32 %r406, %r406, 64;
|
470 |
+
setp.lt.u32 %p62, %r406, 192;
|
471 |
+
@%p62 bra $L__BB0_1;
|
472 |
+
bra.uni $L__BB0_4;
|
473 |
+
$L__BB0_1:
|
474 |
+
.loc 1 39 40
|
475 |
+
setp.lt.u64 %p41, %rd1, 50257;
|
476 |
+
.loc 1 35 34
|
477 |
+
add.s64 %rd67, %rd6, %rd105;
|
478 |
+
add.s64 %rd68, %rd67, 16;
|
479 |
+
add.s64 %rd69, %rd5, %rd105;
|
480 |
+
.loc 1 35 50
|
481 |
+
add.s64 %rd70, %rd69, 16;
|
482 |
+
mov.b32 %r342, 0;
|
483 |
+
mov.u32 %r32, 0x0;
|
484 |
+
mov.u32 %r33, 0x0;
|
485 |
+
mov.u32 %r34, 0x0;
|
486 |
+
mov.u32 %r35, 0x0;
|
487 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r32, %r33, %r34, %r35 }, [ %rd67 + 0 ];
|
488 |
+
@!%p1 mov.u32 %r32, %r342;
|
489 |
+
@!%p1 mov.u32 %r33, %r342;
|
490 |
+
@!%p1 mov.u32 %r34, %r342;
|
491 |
+
@!%p1 mov.u32 %r35, %r342;
|
492 |
+
mov.b32 %f65, %r32;
|
493 |
+
mov.b32 %f66, %r33;
|
494 |
+
mov.b32 %f67, %r34;
|
495 |
+
mov.b32 %f68, %r35;
|
496 |
+
mov.u32 %r40, 0x0;
|
497 |
+
mov.u32 %r41, 0x0;
|
498 |
+
mov.u32 %r42, 0x0;
|
499 |
+
mov.u32 %r43, 0x0;
|
500 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r40, %r41, %r42, %r43 }, [ %rd68 + 0 ];
|
501 |
+
@!%p1 mov.u32 %r40, %r342;
|
502 |
+
@!%p1 mov.u32 %r41, %r342;
|
503 |
+
@!%p1 mov.u32 %r42, %r342;
|
504 |
+
@!%p1 mov.u32 %r43, %r342;
|
505 |
+
mov.b32 %f69, %r40;
|
506 |
+
mov.b32 %f70, %r41;
|
507 |
+
mov.b32 %f71, %r42;
|
508 |
+
mov.b32 %f72, %r43;
|
509 |
+
mov.u32 %r48, 0x0;
|
510 |
+
mov.u32 %r49, 0x0;
|
511 |
+
mov.u32 %r50, 0x0;
|
512 |
+
mov.u32 %r51, 0x0;
|
513 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r48, %r49, %r50, %r51 }, [ %rd69 + 0 ];
|
514 |
+
@!%p1 mov.u32 %r48, %r342;
|
515 |
+
@!%p1 mov.u32 %r49, %r342;
|
516 |
+
@!%p1 mov.u32 %r50, %r342;
|
517 |
+
@!%p1 mov.u32 %r51, %r342;
|
518 |
+
mov.b32 %f73, %r48;
|
519 |
+
mov.b32 %f74, %r49;
|
520 |
+
mov.b32 %f75, %r50;
|
521 |
+
mov.b32 %f76, %r51;
|
522 |
+
mov.u32 %r56, 0x0;
|
523 |
+
mov.u32 %r57, 0x0;
|
524 |
+
mov.u32 %r58, 0x0;
|
525 |
+
mov.u32 %r59, 0x0;
|
526 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r56, %r57, %r58, %r59 }, [ %rd70 + 0 ];
|
527 |
+
@!%p1 mov.u32 %r56, %r342;
|
528 |
+
@!%p1 mov.u32 %r57, %r342;
|
529 |
+
@!%p1 mov.u32 %r58, %r342;
|
530 |
+
@!%p1 mov.u32 %r59, %r342;
|
531 |
+
mov.b32 %f77, %r56;
|
532 |
+
mov.b32 %f78, %r57;
|
533 |
+
mov.b32 %f79, %r58;
|
534 |
+
mov.b32 %f80, %r59;
|
535 |
+
mov.b32 %r405, 883;
|
536 |
+
mov.u64 %rd104, 1;
|
537 |
+
.loc 1 39 55
|
538 |
+
@%p41 bra $L__BB0_3;
|
539 |
+
mov.u64 %rd71, assertMessage_0;
|
540 |
+
cvta.global.u64 %rd72, %rd71;
|
541 |
+
mov.u64 %rd73, assertFile_0;
|
542 |
+
cvta.global.u64 %rd74, %rd73;
|
543 |
+
mov.u64 %rd75, assertFunc_0;
|
544 |
+
cvta.global.u64 %rd76, %rd75;
|
545 |
+
{ // callseq 6, 0
|
546 |
+
.reg .b32 temp_param_reg;
|
547 |
+
.param .b64 param0;
|
548 |
+
st.param.b64 [param0+0], %rd72;
|
549 |
+
.param .b64 param1;
|
550 |
+
st.param.b64 [param1+0], %rd74;
|
551 |
+
.param .b32 param2;
|
552 |
+
st.param.b32 [param2+0], %r405;
|
553 |
+
.param .b64 param3;
|
554 |
+
st.param.b64 [param3+0], %rd76;
|
555 |
+
.param .b64 param4;
|
556 |
+
st.param.b64 [param4+0], %rd104;
|
557 |
+
call.uni
|
558 |
+
__assertfail,
|
559 |
+
(
|
560 |
+
param0,
|
561 |
+
param1,
|
562 |
+
param2,
|
563 |
+
param3,
|
564 |
+
param4
|
565 |
+
);
|
566 |
+
} // callseq 6
|
567 |
+
bra.uni $L__BB0_3;
|
568 |
+
$L__BB0_4:
|
569 |
+
.loc 1 31 36
|
570 |
+
and.b32 %r254, %r4, 3;
|
571 |
+
mad.lo.s32 %r255, %r254, 72, %r2;
|
572 |
+
shl.b32 %r256, %r255, 2;
|
573 |
+
mov.u32 %r257, global_smem;
|
574 |
+
add.s32 %r258, %r257, %r256;
|
575 |
+
st.shared.f32 [%r258], %f566;
|
576 |
+
st.shared.f32 [%r258+1152], %f567;
|
577 |
+
st.shared.f32 [%r258+2304], %f568;
|
578 |
+
st.shared.f32 [%r258+3456], %f569;
|
579 |
+
st.shared.f32 [%r258+4608], %f570;
|
580 |
+
st.shared.f32 [%r258+5760], %f571;
|
581 |
+
st.shared.f32 [%r258+6912], %f572;
|
582 |
+
st.shared.f32 [%r258+8064], %f573;
|
583 |
+
bar.sync 0;
|
584 |
+
mad.lo.s32 %r259, %r1, 72, %r3;
|
585 |
+
shl.b32 %r260, %r259, 2;
|
586 |
+
add.s32 %r261, %r257, %r260;
|
587 |
+
ld.shared.v4.f32 {%f254, %f255, %f256, %f257}, [%r261];
|
588 |
+
ld.shared.v4.f32 {%f258, %f259, %f260, %f261}, [%r261+16];
|
589 |
+
bar.sync 0;
|
590 |
+
st.shared.f32 [%r258], %f574;
|
591 |
+
st.shared.f32 [%r258+1152], %f575;
|
592 |
+
st.shared.f32 [%r258+2304], %f576;
|
593 |
+
st.shared.f32 [%r258+3456], %f577;
|
594 |
+
st.shared.f32 [%r258+4608], %f578;
|
595 |
+
st.shared.f32 [%r258+5760], %f579;
|
596 |
+
st.shared.f32 [%r258+6912], %f580;
|
597 |
+
st.shared.f32 [%r258+8064], %f581;
|
598 |
+
bar.sync 0;
|
599 |
+
ld.shared.v4.f32 {%f262, %f263, %f264, %f265}, [%r261];
|
600 |
+
ld.shared.v4.f32 {%f266, %f267, %f268, %f269}, [%r261+16];
|
601 |
+
$L__tmp3:
|
602 |
+
.loc 2 108 21
|
603 |
+
sub.f32 %f270, %f599, %f598;
|
604 |
+
.loc 2 109 28
|
605 |
+
add.f32 %f271, %f254, %f255;
|
606 |
+
.loc 2 110 39
|
607 |
+
setp.eq.f32 %p63, %f271, 0f00000000;
|
608 |
+
.loc 2 110 60
|
609 |
+
mov.b32 %r146, %f255;
|
610 |
+
mov.b32 %r147, %f271;
|
611 |
+
div.full.f32 %r145, %r146, %r147;
|
612 |
+
mov.b32 %f272, %r145;
|
613 |
+
.loc 2 110 49
|
614 |
+
selp.f32 %f273, 0f00000000, %f272, %p63;
|
615 |
+
.loc 2 112 17
|
616 |
+
fma.rn.f32 %f274, %f270, %f273, %f598;
|
617 |
+
.loc 2 113 15
|
618 |
+
add.f32 %f275, %f582, %f583;
|
619 |
+
.loc 2 113 30
|
620 |
+
mul.f32 %f276, %f270, %f270;
|
621 |
+
.loc 2 113 38
|
622 |
+
mul.f32 %f277, %f276, %f254;
|
623 |
+
.loc 2 113 22
|
624 |
+
fma.rn.f32 %f278, %f277, %f273, %f275;
|
625 |
+
.loc 2 108 21
|
626 |
+
sub.f32 %f279, %f600, %f274;
|
627 |
+
.loc 2 109 28
|
628 |
+
add.f32 %f280, %f256, %f271;
|
629 |
+
.loc 2 110 39
|
630 |
+
setp.eq.f32 %p64, %f280, 0f00000000;
|
631 |
+
.loc 2 110 60
|
632 |
+
mov.b32 %r150, %f280;
|
633 |
+
mov.b32 %r149, %f256;
|
634 |
+
div.full.f32 %r148, %r149, %r150;
|
635 |
+
mov.b32 %f281, %r148;
|
636 |
+
.loc 2 110 49
|
637 |
+
selp.f32 %f282, 0f00000000, %f281, %p64;
|
638 |
+
.loc 2 112 17
|
639 |
+
fma.rn.f32 %f283, %f282, %f279, %f274;
|
640 |
+
.loc 2 113 15
|
641 |
+
add.f32 %f284, %f584, %f278;
|
642 |
+
.loc 2 113 30
|
643 |
+
mul.f32 %f285, %f279, %f279;
|
644 |
+
.loc 2 113 38
|
645 |
+
mul.f32 %f286, %f271, %f285;
|
646 |
+
.loc 2 113 22
|
647 |
+
fma.rn.f32 %f287, %f282, %f286, %f284;
|
648 |
+
.loc 2 108 21
|
649 |
+
sub.f32 %f288, %f601, %f283;
|
650 |
+
.loc 2 109 28
|
651 |
+
add.f32 %f289, %f257, %f280;
|
652 |
+
.loc 2 110 39
|
653 |
+
setp.eq.f32 %p65, %f289, 0f00000000;
|
654 |
+
.loc 2 110 60
|
655 |
+
mov.b32 %r153, %f289;
|
656 |
+
mov.b32 %r152, %f257;
|
657 |
+
div.full.f32 %r151, %r152, %r153;
|
658 |
+
mov.b32 %f290, %r151;
|
659 |
+
.loc 2 110 49
|
660 |
+
selp.f32 %f291, 0f00000000, %f290, %p65;
|
661 |
+
.loc 2 112 17
|
662 |
+
fma.rn.f32 %f292, %f291, %f288, %f283;
|
663 |
+
.loc 2 113 15
|
664 |
+
add.f32 %f293, %f585, %f287;
|
665 |
+
.loc 2 113 30
|
666 |
+
mul.f32 %f294, %f288, %f288;
|
667 |
+
.loc 2 113 38
|
668 |
+
mul.f32 %f295, %f280, %f294;
|
669 |
+
.loc 2 113 22
|
670 |
+
fma.rn.f32 %f296, %f291, %f295, %f293;
|
671 |
+
.loc 2 108 21
|
672 |
+
sub.f32 %f297, %f602, %f292;
|
673 |
+
.loc 2 109 28
|
674 |
+
add.f32 %f298, %f258, %f289;
|
675 |
+
.loc 2 110 39
|
676 |
+
setp.eq.f32 %p66, %f298, 0f00000000;
|
677 |
+
.loc 2 110 60
|
678 |
+
mov.b32 %r156, %f298;
|
679 |
+
mov.b32 %r155, %f258;
|
680 |
+
div.full.f32 %r154, %r155, %r156;
|
681 |
+
mov.b32 %f299, %r154;
|
682 |
+
.loc 2 110 49
|
683 |
+
selp.f32 %f300, 0f00000000, %f299, %p66;
|
684 |
+
.loc 2 112 17
|
685 |
+
fma.rn.f32 %f301, %f300, %f297, %f292;
|
686 |
+
.loc 2 113 15
|
687 |
+
add.f32 %f302, %f586, %f296;
|
688 |
+
.loc 2 113 30
|
689 |
+
mul.f32 %f303, %f297, %f297;
|
690 |
+
.loc 2 113 38
|
691 |
+
mul.f32 %f304, %f289, %f303;
|
692 |
+
.loc 2 113 22
|
693 |
+
fma.rn.f32 %f305, %f300, %f304, %f302;
|
694 |
+
.loc 2 108 21
|
695 |
+
sub.f32 %f306, %f603, %f301;
|
696 |
+
.loc 2 109 28
|
697 |
+
add.f32 %f307, %f259, %f298;
|
698 |
+
.loc 2 110 39
|
699 |
+
setp.eq.f32 %p67, %f307, 0f00000000;
|
700 |
+
.loc 2 110 60
|
701 |
+
mov.b32 %r159, %f307;
|
702 |
+
mov.b32 %r158, %f259;
|
703 |
+
div.full.f32 %r157, %r158, %r159;
|
704 |
+
mov.b32 %f308, %r157;
|
705 |
+
.loc 2 110 49
|
706 |
+
selp.f32 %f309, 0f00000000, %f308, %p67;
|
707 |
+
.loc 2 112 17
|
708 |
+
fma.rn.f32 %f310, %f309, %f306, %f301;
|
709 |
+
.loc 2 113 15
|
710 |
+
add.f32 %f311, %f587, %f305;
|
711 |
+
.loc 2 113 30
|
712 |
+
mul.f32 %f312, %f306, %f306;
|
713 |
+
.loc 2 113 38
|
714 |
+
mul.f32 %f313, %f298, %f312;
|
715 |
+
.loc 2 113 22
|
716 |
+
fma.rn.f32 %f314, %f309, %f313, %f311;
|
717 |
+
.loc 2 108 21
|
718 |
+
sub.f32 %f315, %f604, %f310;
|
719 |
+
.loc 2 109 28
|
720 |
+
add.f32 %f316, %f260, %f307;
|
721 |
+
.loc 2 110 39
|
722 |
+
setp.eq.f32 %p68, %f316, 0f00000000;
|
723 |
+
.loc 2 110 60
|
724 |
+
mov.b32 %r162, %f316;
|
725 |
+
mov.b32 %r161, %f260;
|
726 |
+
div.full.f32 %r160, %r161, %r162;
|
727 |
+
mov.b32 %f317, %r160;
|
728 |
+
.loc 2 110 49
|
729 |
+
selp.f32 %f318, 0f00000000, %f317, %p68;
|
730 |
+
.loc 2 112 17
|
731 |
+
fma.rn.f32 %f319, %f318, %f315, %f310;
|
732 |
+
.loc 2 113 15
|
733 |
+
add.f32 %f320, %f588, %f314;
|
734 |
+
.loc 2 113 30
|
735 |
+
mul.f32 %f321, %f315, %f315;
|
736 |
+
.loc 2 113 38
|
737 |
+
mul.f32 %f322, %f307, %f321;
|
738 |
+
.loc 2 113 22
|
739 |
+
fma.rn.f32 %f323, %f318, %f322, %f320;
|
740 |
+
.loc 2 108 21
|
741 |
+
sub.f32 %f324, %f605, %f319;
|
742 |
+
.loc 2 109 28
|
743 |
+
add.f32 %f325, %f261, %f316;
|
744 |
+
.loc 2 110 39
|
745 |
+
setp.eq.f32 %p69, %f325, 0f00000000;
|
746 |
+
.loc 2 110 60
|
747 |
+
mov.b32 %r165, %f325;
|
748 |
+
mov.b32 %r164, %f261;
|
749 |
+
div.full.f32 %r163, %r164, %r165;
|
750 |
+
mov.b32 %f326, %r163;
|
751 |
+
.loc 2 110 49
|
752 |
+
selp.f32 %f327, 0f00000000, %f326, %p69;
|
753 |
+
.loc 2 112 17
|
754 |
+
fma.rn.f32 %f328, %f327, %f324, %f319;
|
755 |
+
.loc 2 113 15
|
756 |
+
add.f32 %f329, %f589, %f323;
|
757 |
+
.loc 2 113 30
|
758 |
+
mul.f32 %f330, %f324, %f324;
|
759 |
+
.loc 2 113 38
|
760 |
+
mul.f32 %f331, %f316, %f330;
|
761 |
+
.loc 2 113 22
|
762 |
+
fma.rn.f32 %f332, %f327, %f331, %f329;
|
763 |
+
.loc 2 108 21
|
764 |
+
sub.f32 %f333, %f607, %f606;
|
765 |
+
.loc 2 109 28
|
766 |
+
add.f32 %f334, %f262, %f263;
|
767 |
+
.loc 2 110 39
|
768 |
+
setp.eq.f32 %p70, %f334, 0f00000000;
|
769 |
+
.loc 2 110 60
|
770 |
+
mov.b32 %r167, %f263;
|
771 |
+
mov.b32 %r168, %f334;
|
772 |
+
div.full.f32 %r166, %r167, %r168;
|
773 |
+
mov.b32 %f335, %r166;
|
774 |
+
.loc 2 110 49
|
775 |
+
selp.f32 %f336, 0f00000000, %f335, %p70;
|
776 |
+
.loc 2 112 17
|
777 |
+
fma.rn.f32 %f337, %f333, %f336, %f606;
|
778 |
+
.loc 2 113 15
|
779 |
+
add.f32 %f338, %f590, %f591;
|
780 |
+
.loc 2 113 30
|
781 |
+
mul.f32 %f339, %f333, %f333;
|
782 |
+
.loc 2 113 38
|
783 |
+
mul.f32 %f340, %f339, %f262;
|
784 |
+
.loc 2 113 22
|
785 |
+
fma.rn.f32 %f341, %f340, %f336, %f338;
|
786 |
+
.loc 2 108 21
|
787 |
+
sub.f32 %f342, %f608, %f337;
|
788 |
+
.loc 2 109 28
|
789 |
+
add.f32 %f343, %f264, %f334;
|
790 |
+
.loc 2 110 39
|
791 |
+
setp.eq.f32 %p71, %f343, 0f00000000;
|
792 |
+
.loc 2 110 60
|
793 |
+
mov.b32 %r171, %f343;
|
794 |
+
mov.b32 %r170, %f264;
|
795 |
+
div.full.f32 %r169, %r170, %r171;
|
796 |
+
mov.b32 %f344, %r169;
|
797 |
+
.loc 2 110 49
|
798 |
+
selp.f32 %f345, 0f00000000, %f344, %p71;
|
799 |
+
.loc 2 112 17
|
800 |
+
fma.rn.f32 %f346, %f345, %f342, %f337;
|
801 |
+
.loc 2 113 15
|
802 |
+
add.f32 %f347, %f592, %f341;
|
803 |
+
.loc 2 113 30
|
804 |
+
mul.f32 %f348, %f342, %f342;
|
805 |
+
.loc 2 113 38
|
806 |
+
mul.f32 %f349, %f334, %f348;
|
807 |
+
.loc 2 113 22
|
808 |
+
fma.rn.f32 %f350, %f345, %f349, %f347;
|
809 |
+
.loc 2 108 21
|
810 |
+
sub.f32 %f351, %f609, %f346;
|
811 |
+
.loc 2 109 28
|
812 |
+
add.f32 %f352, %f265, %f343;
|
813 |
+
.loc 2 110 39
|
814 |
+
setp.eq.f32 %p72, %f352, 0f00000000;
|
815 |
+
.loc 2 110 60
|
816 |
+
mov.b32 %r174, %f352;
|
817 |
+
mov.b32 %r173, %f265;
|
818 |
+
div.full.f32 %r172, %r173, %r174;
|
819 |
+
mov.b32 %f353, %r172;
|
820 |
+
.loc 2 110 49
|
821 |
+
selp.f32 %f354, 0f00000000, %f353, %p72;
|
822 |
+
.loc 2 112 17
|
823 |
+
fma.rn.f32 %f355, %f354, %f351, %f346;
|
824 |
+
.loc 2 113 15
|
825 |
+
add.f32 %f356, %f593, %f350;
|
826 |
+
.loc 2 113 30
|
827 |
+
mul.f32 %f357, %f351, %f351;
|
828 |
+
.loc 2 113 38
|
829 |
+
mul.f32 %f358, %f343, %f357;
|
830 |
+
.loc 2 113 22
|
831 |
+
fma.rn.f32 %f359, %f354, %f358, %f356;
|
832 |
+
.loc 2 108 21
|
833 |
+
sub.f32 %f360, %f610, %f355;
|
834 |
+
.loc 2 109 28
|
835 |
+
add.f32 %f361, %f266, %f352;
|
836 |
+
.loc 2 110 39
|
837 |
+
setp.eq.f32 %p73, %f361, 0f00000000;
|
838 |
+
.loc 2 110 60
|
839 |
+
mov.b32 %r177, %f361;
|
840 |
+
mov.b32 %r176, %f266;
|
841 |
+
div.full.f32 %r175, %r176, %r177;
|
842 |
+
mov.b32 %f362, %r175;
|
843 |
+
.loc 2 110 49
|
844 |
+
selp.f32 %f363, 0f00000000, %f362, %p73;
|
845 |
+
.loc 2 112 17
|
846 |
+
fma.rn.f32 %f364, %f363, %f360, %f355;
|
847 |
+
.loc 2 113 15
|
848 |
+
add.f32 %f365, %f594, %f359;
|
849 |
+
.loc 2 113 30
|
850 |
+
mul.f32 %f366, %f360, %f360;
|
851 |
+
.loc 2 113 38
|
852 |
+
mul.f32 %f367, %f352, %f366;
|
853 |
+
.loc 2 113 22
|
854 |
+
fma.rn.f32 %f368, %f363, %f367, %f365;
|
855 |
+
.loc 2 108 21
|
856 |
+
sub.f32 %f369, %f611, %f364;
|
857 |
+
.loc 2 109 28
|
858 |
+
add.f32 %f370, %f267, %f361;
|
859 |
+
.loc 2 110 39
|
860 |
+
setp.eq.f32 %p74, %f370, 0f00000000;
|
861 |
+
.loc 2 110 60
|
862 |
+
mov.b32 %r180, %f370;
|
863 |
+
mov.b32 %r179, %f267;
|
864 |
+
div.full.f32 %r178, %r179, %r180;
|
865 |
+
mov.b32 %f371, %r178;
|
866 |
+
.loc 2 110 49
|
867 |
+
selp.f32 %f372, 0f00000000, %f371, %p74;
|
868 |
+
.loc 2 112 17
|
869 |
+
fma.rn.f32 %f373, %f372, %f369, %f364;
|
870 |
+
.loc 2 113 15
|
871 |
+
add.f32 %f374, %f595, %f368;
|
872 |
+
.loc 2 113 30
|
873 |
+
mul.f32 %f375, %f369, %f369;
|
874 |
+
.loc 2 113 38
|
875 |
+
mul.f32 %f376, %f361, %f375;
|
876 |
+
.loc 2 113 22
|
877 |
+
fma.rn.f32 %f377, %f372, %f376, %f374;
|
878 |
+
.loc 2 108 21
|
879 |
+
sub.f32 %f378, %f612, %f373;
|
880 |
+
.loc 2 109 28
|
881 |
+
add.f32 %f379, %f268, %f370;
|
882 |
+
.loc 2 110 39
|
883 |
+
setp.eq.f32 %p75, %f379, 0f00000000;
|
884 |
+
.loc 2 110 60
|
885 |
+
mov.b32 %r183, %f379;
|
886 |
+
mov.b32 %r182, %f268;
|
887 |
+
div.full.f32 %r181, %r182, %r183;
|
888 |
+
mov.b32 %f380, %r181;
|
889 |
+
.loc 2 110 49
|
890 |
+
selp.f32 %f381, 0f00000000, %f380, %p75;
|
891 |
+
.loc 2 112 17
|
892 |
+
fma.rn.f32 %f382, %f381, %f378, %f373;
|
893 |
+
.loc 2 113 15
|
894 |
+
add.f32 %f383, %f596, %f377;
|
895 |
+
.loc 2 113 30
|
896 |
+
mul.f32 %f384, %f378, %f378;
|
897 |
+
.loc 2 113 38
|
898 |
+
mul.f32 %f385, %f370, %f384;
|
899 |
+
.loc 2 113 22
|
900 |
+
fma.rn.f32 %f386, %f381, %f385, %f383;
|
901 |
+
.loc 2 108 21
|
902 |
+
sub.f32 %f387, %f613, %f382;
|
903 |
+
.loc 2 109 28
|
904 |
+
add.f32 %f388, %f269, %f379;
|
905 |
+
.loc 2 110 39
|
906 |
+
setp.eq.f32 %p76, %f388, 0f00000000;
|
907 |
+
.loc 2 110 60
|
908 |
+
mov.b32 %r186, %f388;
|
909 |
+
mov.b32 %r185, %f269;
|
910 |
+
div.full.f32 %r184, %r185, %r186;
|
911 |
+
mov.b32 %f389, %r184;
|
912 |
+
.loc 2 110 49
|
913 |
+
selp.f32 %f390, 0f00000000, %f389, %p76;
|
914 |
+
.loc 2 112 17
|
915 |
+
fma.rn.f32 %f391, %f390, %f387, %f382;
|
916 |
+
.loc 2 113 15
|
917 |
+
add.f32 %f392, %f597, %f386;
|
918 |
+
.loc 2 113 30
|
919 |
+
mul.f32 %f393, %f387, %f387;
|
920 |
+
.loc 2 113 38
|
921 |
+
mul.f32 %f394, %f379, %f393;
|
922 |
+
.loc 2 113 22
|
923 |
+
fma.rn.f32 %f395, %f390, %f394, %f392;
|
924 |
+
$L__tmp4:
|
925 |
+
.loc 2 120 46
|
926 |
+
mov.b32 %r262, %f328;
|
927 |
+
shfl.sync.bfly.b32 %r263, %r262, 4, 31, -1;
|
928 |
+
mov.b32 %f396, %r263;
|
929 |
+
mov.b32 %r264, %f332;
|
930 |
+
shfl.sync.bfly.b32 %r265, %r264, 4, 31, -1;
|
931 |
+
mov.b32 %f397, %r265;
|
932 |
+
shfl.sync.bfly.b32 %r188, %r165, 4, 31, -1;
|
933 |
+
mov.b32 %f398, %r188;
|
934 |
+
$L__tmp5:
|
935 |
+
.loc 2 108 21
|
936 |
+
sub.f32 %f399, %f396, %f328;
|
937 |
+
.loc 2 109 28
|
938 |
+
add.f32 %f400, %f325, %f398;
|
939 |
+
.loc 2 110 39
|
940 |
+
setp.eq.f32 %p77, %f400, 0f00000000;
|
941 |
+
.loc 2 110 60
|
942 |
+
mov.b32 %r189, %f400;
|
943 |
+
div.full.f32 %r187, %r188, %r189;
|
944 |
+
mov.b32 %f401, %r187;
|
945 |
+
.loc 2 110 49
|
946 |
+
selp.f32 %f402, 0f00000000, %f401, %p77;
|
947 |
+
.loc 2 112 17
|
948 |
+
fma.rn.f32 %f403, %f402, %f399, %f328;
|
949 |
+
.loc 2 113 15
|
950 |
+
add.f32 %f404, %f332, %f397;
|
951 |
+
.loc 2 113 30
|
952 |
+
mul.f32 %f405, %f399, %f399;
|
953 |
+
.loc 2 113 38
|
954 |
+
mul.f32 %f406, %f325, %f405;
|
955 |
+
.loc 2 113 22
|
956 |
+
fma.rn.f32 %f407, %f402, %f406, %f404;
|
957 |
+
$L__tmp6:
|
958 |
+
.loc 2 120 46
|
959 |
+
mov.b32 %r266, %f403;
|
960 |
+
shfl.sync.bfly.b32 %r267, %r266, 2, 31, -1;
|
961 |
+
mov.b32 %f408, %r267;
|
962 |
+
mov.b32 %r268, %f407;
|
963 |
+
shfl.sync.bfly.b32 %r269, %r268, 2, 31, -1;
|
964 |
+
mov.b32 %f409, %r269;
|
965 |
+
shfl.sync.bfly.b32 %r191, %r189, 2, 31, -1;
|
966 |
+
mov.b32 %f410, %r191;
|
967 |
+
$L__tmp7:
|
968 |
+
.loc 2 108 21
|
969 |
+
sub.f32 %f411, %f408, %f403;
|
970 |
+
.loc 2 109 28
|
971 |
+
add.f32 %f412, %f400, %f410;
|
972 |
+
.loc 2 110 39
|
973 |
+
setp.eq.f32 %p78, %f412, 0f00000000;
|
974 |
+
.loc 2 110 60
|
975 |
+
mov.b32 %r192, %f412;
|
976 |
+
div.full.f32 %r190, %r191, %r192;
|
977 |
+
mov.b32 %f413, %r190;
|
978 |
+
.loc 2 110 49
|
979 |
+
selp.f32 %f414, 0f00000000, %f413, %p78;
|
980 |
+
.loc 2 112 17
|
981 |
+
fma.rn.f32 %f415, %f414, %f411, %f403;
|
982 |
+
.loc 2 113 15
|
983 |
+
add.f32 %f416, %f407, %f409;
|
984 |
+
.loc 2 113 30
|
985 |
+
mul.f32 %f417, %f411, %f411;
|
986 |
+
.loc 2 113 38
|
987 |
+
mul.f32 %f418, %f400, %f417;
|
988 |
+
.loc 2 113 22
|
989 |
+
fma.rn.f32 %f419, %f414, %f418, %f416;
|
990 |
+
$L__tmp8:
|
991 |
+
.loc 2 120 46
|
992 |
+
mov.b32 %r270, %f415;
|
993 |
+
shfl.sync.bfly.b32 %r271, %r270, 1, 31, -1;
|
994 |
+
mov.b32 %f420, %r271;
|
995 |
+
mov.b32 %r272, %f419;
|
996 |
+
shfl.sync.bfly.b32 %r273, %r272, 1, 31, -1;
|
997 |
+
mov.b32 %f421, %r273;
|
998 |
+
shfl.sync.bfly.b32 %r194, %r192, 1, 31, -1;
|
999 |
+
mov.b32 %f422, %r194;
|
1000 |
+
$L__tmp9:
|
1001 |
+
.loc 2 108 21
|
1002 |
+
sub.f32 %f423, %f420, %f415;
|
1003 |
+
.loc 2 109 28
|
1004 |
+
add.f32 %f424, %f412, %f422;
|
1005 |
+
.loc 2 110 39
|
1006 |
+
setp.eq.f32 %p79, %f424, 0f00000000;
|
1007 |
+
.loc 2 110 60
|
1008 |
+
mov.b32 %r195, %f424;
|
1009 |
+
div.full.f32 %r193, %r194, %r195;
|
1010 |
+
mov.b32 %f425, %r193;
|
1011 |
+
.loc 2 110 49
|
1012 |
+
selp.f32 %f426, 0f00000000, %f425, %p79;
|
1013 |
+
.loc 2 112 17
|
1014 |
+
fma.rn.f32 %f145, %f423, %f426, %f415;
|
1015 |
+
.loc 2 113 15
|
1016 |
+
add.f32 %f427, %f419, %f421;
|
1017 |
+
.loc 2 113 30
|
1018 |
+
mul.f32 %f428, %f423, %f423;
|
1019 |
+
.loc 2 113 38
|
1020 |
+
mul.f32 %f429, %f412, %f428;
|
1021 |
+
.loc 2 113 22
|
1022 |
+
fma.rn.f32 %f430, %f426, %f429, %f427;
|
1023 |
+
$L__tmp10:
|
1024 |
+
.loc 2 120 46
|
1025 |
+
mov.b32 %r274, %f391;
|
1026 |
+
shfl.sync.bfly.b32 %r275, %r274, 4, 31, -1;
|
1027 |
+
mov.b32 %f431, %r275;
|
1028 |
+
mov.b32 %r276, %f395;
|
1029 |
+
shfl.sync.bfly.b32 %r277, %r276, 4, 31, -1;
|
1030 |
+
mov.b32 %f432, %r277;
|
1031 |
+
shfl.sync.bfly.b32 %r197, %r186, 4, 31, -1;
|
1032 |
+
mov.b32 %f433, %r197;
|
1033 |
+
$L__tmp11:
|
1034 |
+
.loc 2 108 21
|
1035 |
+
sub.f32 %f434, %f431, %f391;
|
1036 |
+
.loc 2 109 28
|
1037 |
+
add.f32 %f435, %f388, %f433;
|
1038 |
+
.loc 2 110 39
|
1039 |
+
setp.eq.f32 %p80, %f435, 0f00000000;
|
1040 |
+
.loc 2 110 60
|
1041 |
+
mov.b32 %r198, %f435;
|
1042 |
+
div.full.f32 %r196, %r197, %r198;
|
1043 |
+
mov.b32 %f436, %r196;
|
1044 |
+
.loc 2 110 49
|
1045 |
+
selp.f32 %f437, 0f00000000, %f436, %p80;
|
1046 |
+
.loc 2 112 17
|
1047 |
+
fma.rn.f32 %f438, %f434, %f437, %f391;
|
1048 |
+
.loc 2 113 15
|
1049 |
+
add.f32 %f439, %f395, %f432;
|
1050 |
+
.loc 2 113 30
|
1051 |
+
mul.f32 %f440, %f434, %f434;
|
1052 |
+
.loc 2 113 38
|
1053 |
+
mul.f32 %f441, %f388, %f440;
|
1054 |
+
.loc 2 113 22
|
1055 |
+
fma.rn.f32 %f442, %f441, %f437, %f439;
|
1056 |
+
$L__tmp12:
|
1057 |
+
.loc 2 120 46
|
1058 |
+
mov.b32 %r278, %f438;
|
1059 |
+
shfl.sync.bfly.b32 %r279, %r278, 2, 31, -1;
|
1060 |
+
mov.b32 %f443, %r279;
|
1061 |
+
mov.b32 %r280, %f442;
|
1062 |
+
shfl.sync.bfly.b32 %r281, %r280, 2, 31, -1;
|
1063 |
+
mov.b32 %f444, %r281;
|
1064 |
+
shfl.sync.bfly.b32 %r200, %r198, 2, 31, -1;
|
1065 |
+
mov.b32 %f445, %r200;
|
1066 |
+
$L__tmp13:
|
1067 |
+
.loc 2 108 21
|
1068 |
+
sub.f32 %f446, %f443, %f438;
|
1069 |
+
.loc 2 109 28
|
1070 |
+
add.f32 %f447, %f435, %f445;
|
1071 |
+
.loc 2 110 39
|
1072 |
+
setp.eq.f32 %p81, %f447, 0f00000000;
|
1073 |
+
.loc 2 110 60
|
1074 |
+
mov.b32 %r201, %f447;
|
1075 |
+
div.full.f32 %r199, %r200, %r201;
|
1076 |
+
mov.b32 %f448, %r199;
|
1077 |
+
.loc 2 110 49
|
1078 |
+
selp.f32 %f449, 0f00000000, %f448, %p81;
|
1079 |
+
.loc 2 112 17
|
1080 |
+
fma.rn.f32 %f450, %f446, %f449, %f438;
|
1081 |
+
.loc 2 113 15
|
1082 |
+
add.f32 %f451, %f442, %f444;
|
1083 |
+
.loc 2 113 30
|
1084 |
+
mul.f32 %f452, %f446, %f446;
|
1085 |
+
.loc 2 113 38
|
1086 |
+
mul.f32 %f453, %f435, %f452;
|
1087 |
+
.loc 2 113 22
|
1088 |
+
fma.rn.f32 %f454, %f449, %f453, %f451;
|
1089 |
+
$L__tmp14:
|
1090 |
+
.loc 2 120 46
|
1091 |
+
mov.b32 %r282, %f450;
|
1092 |
+
shfl.sync.bfly.b32 %r283, %r282, 1, 31, -1;
|
1093 |
+
mov.b32 %f455, %r283;
|
1094 |
+
mov.b32 %r284, %f454;
|
1095 |
+
shfl.sync.bfly.b32 %r285, %r284, 1, 31, -1;
|
1096 |
+
mov.b32 %f456, %r285;
|
1097 |
+
shfl.sync.bfly.b32 %r203, %r201, 1, 31, -1;
|
1098 |
+
mov.b32 %f457, %r203;
|
1099 |
+
$L__tmp15:
|
1100 |
+
.loc 2 108 21
|
1101 |
+
sub.f32 %f458, %f455, %f450;
|
1102 |
+
.loc 2 109 28
|
1103 |
+
add.f32 %f459, %f447, %f457;
|
1104 |
+
.loc 2 110 39
|
1105 |
+
setp.eq.f32 %p82, %f459, 0f00000000;
|
1106 |
+
.loc 2 110 60
|
1107 |
+
mov.b32 %r204, %f459;
|
1108 |
+
div.full.f32 %r202, %r203, %r204;
|
1109 |
+
mov.b32 %f460, %r202;
|
1110 |
+
.loc 2 110 49
|
1111 |
+
selp.f32 %f461, 0f00000000, %f460, %p82;
|
1112 |
+
.loc 2 112 17
|
1113 |
+
fma.rn.f32 %f146, %f458, %f461, %f450;
|
1114 |
+
.loc 2 113 15
|
1115 |
+
add.f32 %f462, %f454, %f456;
|
1116 |
+
.loc 2 113 30
|
1117 |
+
mul.f32 %f463, %f458, %f458;
|
1118 |
+
.loc 2 113 38
|
1119 |
+
mul.f32 %f464, %f447, %f463;
|
1120 |
+
.loc 2 113 22
|
1121 |
+
fma.rn.f32 %f465, %f461, %f464, %f462;
|
1122 |
+
$L__tmp16:
|
1123 |
+
.loc 1 69 23
|
1124 |
+
mov.b32 %r206, %f430;
|
1125 |
+
mov.b32 %r207, 1132462080;
|
1126 |
+
div.full.f32 %r205, %r206, %r207;
|
1127 |
+
mov.b32 %f466, %r205;
|
1128 |
+
mov.b32 %r230, %f465;
|
1129 |
+
div.full.f32 %r229, %r230, %r207;
|
1130 |
+
mov.b32 %f467, %r229;
|
1131 |
+
.loc 1 71 24
|
1132 |
+
add.f32 %f147, %f466, 0f3727C5AC;
|
1133 |
+
add.f32 %f148, %f467, 0f3727C5AC;
|
1134 |
+
.loc 1 55 36
|
1135 |
+
add.s64 %rd9, %rd12, %rd2;
|
1136 |
+
shl.b32 %r286, %r11, 14;
|
1137 |
+
shl.b32 %r287, %r1, 8;
|
1138 |
+
or.b32 %r288, %r286, %r287;
|
1139 |
+
or.b32 %r8, %r288, %r3;
|
1140 |
+
mov.u64 %rd106, 0;
|
1141 |
+
mov.b32 %r407, -64;
|
1142 |
+
rsqrt.approx.ftz.f32 %f516, %f147;
|
1143 |
+
rsqrt.approx.ftz.f32 %f517, %f148;
|
1144 |
+
bra.uni $L__BB0_5;
|
1145 |
+
$L__BB0_7:
|
1146 |
+
.loc 1 65 35
|
1147 |
+
add.s64 %rd96, %rd4, %rd106;
|
1148 |
+
add.s64 %rd97, %rd96, 16;
|
1149 |
+
add.s64 %rd98, %rd3, %rd106;
|
1150 |
+
.loc 1 65 54
|
1151 |
+
add.s64 %rd99, %rd98, 16;
|
1152 |
+
mov.u32 %r338, 0x0;
|
1153 |
+
mov.u32 %r339, 0x0;
|
1154 |
+
mov.u32 %r340, 0x0;
|
1155 |
+
mov.u32 %r341, 0x0;
|
1156 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r338, %r339, %r340, %r341 }, [ %rd96 + 0 ];
|
1157 |
+
@!%p1 mov.u32 %r338, %r342;
|
1158 |
+
@!%p1 mov.u32 %r339, %r342;
|
1159 |
+
@!%p1 mov.u32 %r340, %r342;
|
1160 |
+
@!%p1 mov.u32 %r341, %r342;
|
1161 |
+
mov.b32 %f468, %r338;
|
1162 |
+
mov.b32 %f469, %r339;
|
1163 |
+
mov.b32 %f470, %r340;
|
1164 |
+
mov.b32 %f471, %r341;
|
1165 |
+
mov.u32 %r346, 0x0;
|
1166 |
+
mov.u32 %r347, 0x0;
|
1167 |
+
mov.u32 %r348, 0x0;
|
1168 |
+
mov.u32 %r349, 0x0;
|
1169 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r346, %r347, %r348, %r349 }, [ %rd97 + 0 ];
|
1170 |
+
@!%p1 mov.u32 %r346, %r342;
|
1171 |
+
@!%p1 mov.u32 %r347, %r342;
|
1172 |
+
@!%p1 mov.u32 %r348, %r342;
|
1173 |
+
@!%p1 mov.u32 %r349, %r342;
|
1174 |
+
mov.b32 %f472, %r346;
|
1175 |
+
mov.b32 %f473, %r347;
|
1176 |
+
mov.b32 %f474, %r348;
|
1177 |
+
mov.b32 %f475, %r349;
|
1178 |
+
mov.u32 %r354, 0x0;
|
1179 |
+
mov.u32 %r355, 0x0;
|
1180 |
+
mov.u32 %r356, 0x0;
|
1181 |
+
mov.u32 %r357, 0x0;
|
1182 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r354, %r355, %r356, %r357 }, [ %rd98 + 0 ];
|
1183 |
+
@!%p1 mov.u32 %r354, %r342;
|
1184 |
+
@!%p1 mov.u32 %r355, %r342;
|
1185 |
+
@!%p1 mov.u32 %r356, %r342;
|
1186 |
+
@!%p1 mov.u32 %r357, %r342;
|
1187 |
+
mov.b32 %f476, %r354;
|
1188 |
+
mov.b32 %f477, %r355;
|
1189 |
+
mov.b32 %f478, %r356;
|
1190 |
+
mov.b32 %f479, %r357;
|
1191 |
+
mov.u32 %r362, 0x0;
|
1192 |
+
mov.u32 %r363, 0x0;
|
1193 |
+
mov.u32 %r364, 0x0;
|
1194 |
+
mov.u32 %r365, 0x0;
|
1195 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r362, %r363, %r364, %r365 }, [ %rd99 + 0 ];
|
1196 |
+
@!%p1 mov.u32 %r362, %r342;
|
1197 |
+
@!%p1 mov.u32 %r363, %r342;
|
1198 |
+
@!%p1 mov.u32 %r364, %r342;
|
1199 |
+
@!%p1 mov.u32 %r365, %r342;
|
1200 |
+
mov.b32 %f480, %r362;
|
1201 |
+
mov.b32 %f481, %r363;
|
1202 |
+
mov.b32 %f482, %r364;
|
1203 |
+
mov.b32 %f483, %r365;
|
1204 |
+
.loc 1 66 24
|
1205 |
+
add.f32 %f484, %f149, %f468;
|
1206 |
+
add.f32 %f485, %f150, %f469;
|
1207 |
+
add.f32 %f486, %f151, %f470;
|
1208 |
+
add.f32 %f487, %f152, %f471;
|
1209 |
+
add.f32 %f488, %f153, %f472;
|
1210 |
+
add.f32 %f489, %f154, %f473;
|
1211 |
+
add.f32 %f490, %f155, %f474;
|
1212 |
+
add.f32 %f491, %f156, %f475;
|
1213 |
+
add.f32 %f492, %f157, %f476;
|
1214 |
+
add.f32 %f493, %f158, %f477;
|
1215 |
+
add.f32 %f494, %f159, %f478;
|
1216 |
+
add.f32 %f495, %f160, %f479;
|
1217 |
+
add.f32 %f496, %f161, %f480;
|
1218 |
+
add.f32 %f497, %f162, %f481;
|
1219 |
+
add.f32 %f498, %f163, %f482;
|
1220 |
+
add.f32 %f499, %f164, %f483;
|
1221 |
+
.loc 1 67 24
|
1222 |
+
sub.f32 %f500, %f484, %f145;
|
1223 |
+
sub.f32 %f501, %f485, %f145;
|
1224 |
+
sub.f32 %f502, %f486, %f145;
|
1225 |
+
sub.f32 %f503, %f487, %f145;
|
1226 |
+
sub.f32 %f504, %f488, %f145;
|
1227 |
+
sub.f32 %f505, %f489, %f145;
|
1228 |
+
sub.f32 %f506, %f490, %f145;
|
1229 |
+
sub.f32 %f507, %f491, %f145;
|
1230 |
+
sub.f32 %f508, %f492, %f146;
|
1231 |
+
sub.f32 %f509, %f493, %f146;
|
1232 |
+
sub.f32 %f510, %f494, %f146;
|
1233 |
+
sub.f32 %f511, %f495, %f146;
|
1234 |
+
sub.f32 %f512, %f496, %f146;
|
1235 |
+
sub.f32 %f513, %f497, %f146;
|
1236 |
+
sub.f32 %f514, %f498, %f146;
|
1237 |
+
sub.f32 %f515, %f499, %f146;
|
1238 |
+
.loc 1 73 24
|
1239 |
+
mul.f32 %f518, %f500, %f516;
|
1240 |
+
mul.f32 %f519, %f501, %f516;
|
1241 |
+
mul.f32 %f520, %f502, %f516;
|
1242 |
+
mul.f32 %f521, %f503, %f516;
|
1243 |
+
mul.f32 %f522, %f504, %f516;
|
1244 |
+
mul.f32 %f523, %f505, %f516;
|
1245 |
+
mul.f32 %f524, %f506, %f516;
|
1246 |
+
mul.f32 %f525, %f507, %f516;
|
1247 |
+
mul.f32 %f526, %f508, %f517;
|
1248 |
+
mul.f32 %f527, %f509, %f517;
|
1249 |
+
mul.f32 %f528, %f510, %f517;
|
1250 |
+
mul.f32 %f529, %f511, %f517;
|
1251 |
+
mul.f32 %f530, %f512, %f517;
|
1252 |
+
mul.f32 %f531, %f513, %f517;
|
1253 |
+
mul.f32 %f532, %f514, %f517;
|
1254 |
+
mul.f32 %f533, %f515, %f517;
|
1255 |
+
.loc 1 74 24
|
1256 |
+
mul.f32 %f534, %f518, %f165;
|
1257 |
+
mul.f32 %f535, %f519, %f166;
|
1258 |
+
mul.f32 %f536, %f520, %f167;
|
1259 |
+
mul.f32 %f537, %f521, %f168;
|
1260 |
+
mul.f32 %f538, %f522, %f169;
|
1261 |
+
mul.f32 %f539, %f523, %f170;
|
1262 |
+
mul.f32 %f540, %f524, %f171;
|
1263 |
+
mul.f32 %f541, %f525, %f172;
|
1264 |
+
mul.f32 %f542, %f526, %f165;
|
1265 |
+
mul.f32 %f543, %f527, %f166;
|
1266 |
+
mul.f32 %f544, %f528, %f167;
|
1267 |
+
mul.f32 %f545, %f529, %f168;
|
1268 |
+
mul.f32 %f546, %f530, %f169;
|
1269 |
+
mul.f32 %f547, %f531, %f170;
|
1270 |
+
mul.f32 %f548, %f532, %f171;
|
1271 |
+
mul.f32 %f549, %f533, %f172;
|
1272 |
+
.loc 1 76 35
|
1273 |
+
add.s32 %r394, %r8, %r407;
|
1274 |
+
add.s32 %r395, %r394, 64;
|
1275 |
+
.loc 1 76 29
|
1276 |
+
add.s32 %r396, %r394, 8256;
|
1277 |
+
mul.wide.s32 %rd102, %r395, 2;
|
1278 |
+
add.s64 %rd100, %rd13, %rd102;
|
1279 |
+
mul.wide.s32 %rd103, %r396, 2;
|
1280 |
+
add.s64 %rd101, %rd13, %rd103;
|
1281 |
+
.loc 1 76 52
|
1282 |
+
mov.b32 %r370, %f534;
|
1283 |
+
cvt.rn.bf16.f32 %rs1, %r370;
|
1284 |
+
mov.b32 %r371, %f535;
|
1285 |
+
cvt.rn.bf16.f32 %rs2, %r371;
|
1286 |
+
mov.b32 %r372, %f536;
|
1287 |
+
cvt.rn.bf16.f32 %rs3, %r372;
|
1288 |
+
mov.b32 %r373, %f537;
|
1289 |
+
cvt.rn.bf16.f32 %rs4, %r373;
|
1290 |
+
mov.b32 %r374, %f538;
|
1291 |
+
cvt.rn.bf16.f32 %rs5, %r374;
|
1292 |
+
mov.b32 %r375, %f539;
|
1293 |
+
cvt.rn.bf16.f32 %rs6, %r375;
|
1294 |
+
mov.b32 %r376, %f540;
|
1295 |
+
cvt.rn.bf16.f32 %rs7, %r376;
|
1296 |
+
mov.b32 %r377, %f541;
|
1297 |
+
cvt.rn.bf16.f32 %rs8, %r377;
|
1298 |
+
mov.b32 %r378, %f542;
|
1299 |
+
cvt.rn.bf16.f32 %rs9, %r378;
|
1300 |
+
mov.b32 %r379, %f543;
|
1301 |
+
cvt.rn.bf16.f32 %rs10, %r379;
|
1302 |
+
mov.b32 %r380, %f544;
|
1303 |
+
cvt.rn.bf16.f32 %rs11, %r380;
|
1304 |
+
mov.b32 %r381, %f545;
|
1305 |
+
cvt.rn.bf16.f32 %rs12, %r381;
|
1306 |
+
mov.b32 %r382, %f546;
|
1307 |
+
cvt.rn.bf16.f32 %rs13, %r382;
|
1308 |
+
mov.b32 %r383, %f547;
|
1309 |
+
cvt.rn.bf16.f32 %rs14, %r383;
|
1310 |
+
mov.b32 %r384, %f548;
|
1311 |
+
cvt.rn.bf16.f32 %rs15, %r384;
|
1312 |
+
mov.b32 %r385, %f549;
|
1313 |
+
cvt.rn.bf16.f32 %rs16, %r385;
|
1314 |
+
mov.b32 %r397, {%rs1, %rs2};
|
1315 |
+
mov.b32 %r398, {%rs3, %rs4};
|
1316 |
+
mov.b32 %r399, {%rs5, %rs6};
|
1317 |
+
mov.b32 %r400, {%rs7, %rs8};
|
1318 |
+
@%p1 st.global.v4.b32 [ %rd100 + 0 ], { %r397, %r398, %r399, %r400 };
|
1319 |
+
mov.b32 %r401, {%rs9, %rs10};
|
1320 |
+
mov.b32 %r402, {%rs11, %rs12};
|
1321 |
+
mov.b32 %r403, {%rs13, %rs14};
|
1322 |
+
mov.b32 %r404, {%rs15, %rs16};
|
1323 |
+
@%p1 st.global.v4.b32 [ %rd101 + 0 ], { %r401, %r402, %r403, %r404 };
|
1324 |
+
.loc 1 55 36
|
1325 |
+
add.s64 %rd106, %rd106, 256;
|
1326 |
+
add.s32 %r407, %r407, 64;
|
1327 |
+
setp.lt.u32 %p136, %r407, 192;
|
1328 |
+
@%p136 bra $L__BB0_5;
|
1329 |
+
bra.uni $L__BB0_8;
|
1330 |
+
$L__BB0_5:
|
1331 |
+
.loc 1 59 35
|
1332 |
+
add.s64 %rd83, %rd6, %rd106;
|
1333 |
+
add.s64 %rd84, %rd83, 16;
|
1334 |
+
add.s64 %rd85, %rd5, %rd106;
|
1335 |
+
.loc 1 59 51
|
1336 |
+
add.s64 %rd86, %rd85, 16;
|
1337 |
+
mov.u32 %r289, 0x0;
|
1338 |
+
mov.u32 %r290, 0x0;
|
1339 |
+
mov.u32 %r291, 0x0;
|
1340 |
+
mov.u32 %r292, 0x0;
|
1341 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r289, %r290, %r291, %r292 }, [ %rd83 + 0 ];
|
1342 |
+
@!%p1 mov.u32 %r289, %r342;
|
1343 |
+
@!%p1 mov.u32 %r290, %r342;
|
1344 |
+
@!%p1 mov.u32 %r291, %r342;
|
1345 |
+
@!%p1 mov.u32 %r292, %r342;
|
1346 |
+
mov.b32 %f149, %r289;
|
1347 |
+
mov.b32 %f150, %r290;
|
1348 |
+
mov.b32 %f151, %r291;
|
1349 |
+
mov.b32 %f152, %r292;
|
1350 |
+
mov.u32 %r297, 0x0;
|
1351 |
+
mov.u32 %r298, 0x0;
|
1352 |
+
mov.u32 %r299, 0x0;
|
1353 |
+
mov.u32 %r300, 0x0;
|
1354 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r297, %r298, %r299, %r300 }, [ %rd84 + 0 ];
|
1355 |
+
@!%p1 mov.u32 %r297, %r342;
|
1356 |
+
@!%p1 mov.u32 %r298, %r342;
|
1357 |
+
@!%p1 mov.u32 %r299, %r342;
|
1358 |
+
@!%p1 mov.u32 %r300, %r342;
|
1359 |
+
mov.b32 %f153, %r297;
|
1360 |
+
mov.b32 %f154, %r298;
|
1361 |
+
mov.b32 %f155, %r299;
|
1362 |
+
mov.b32 %f156, %r300;
|
1363 |
+
mov.u32 %r305, 0x0;
|
1364 |
+
mov.u32 %r306, 0x0;
|
1365 |
+
mov.u32 %r307, 0x0;
|
1366 |
+
mov.u32 %r308, 0x0;
|
1367 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r305, %r306, %r307, %r308 }, [ %rd85 + 0 ];
|
1368 |
+
@!%p1 mov.u32 %r305, %r342;
|
1369 |
+
@!%p1 mov.u32 %r306, %r342;
|
1370 |
+
@!%p1 mov.u32 %r307, %r342;
|
1371 |
+
@!%p1 mov.u32 %r308, %r342;
|
1372 |
+
mov.b32 %f157, %r305;
|
1373 |
+
mov.b32 %f158, %r306;
|
1374 |
+
mov.b32 %f159, %r307;
|
1375 |
+
mov.b32 %f160, %r308;
|
1376 |
+
mov.u32 %r313, 0x0;
|
1377 |
+
mov.u32 %r314, 0x0;
|
1378 |
+
mov.u32 %r315, 0x0;
|
1379 |
+
mov.u32 %r316, 0x0;
|
1380 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r313, %r314, %r315, %r316 }, [ %rd86 + 0 ];
|
1381 |
+
@!%p1 mov.u32 %r313, %r342;
|
1382 |
+
@!%p1 mov.u32 %r314, %r342;
|
1383 |
+
@!%p1 mov.u32 %r315, %r342;
|
1384 |
+
@!%p1 mov.u32 %r316, %r342;
|
1385 |
+
mov.b32 %f161, %r313;
|
1386 |
+
mov.b32 %f162, %r314;
|
1387 |
+
mov.b32 %f163, %r315;
|
1388 |
+
mov.b32 %f164, %r316;
|
1389 |
+
.loc 1 60 35
|
1390 |
+
add.s64 %rd87, %rd9, %rd106;
|
1391 |
+
.loc 1 60 40
|
1392 |
+
add.s64 %rd88, %rd87, 16;
|
1393 |
+
mov.u32 %r321, 0x0;
|
1394 |
+
mov.u32 %r322, 0x0;
|
1395 |
+
mov.u32 %r323, 0x0;
|
1396 |
+
mov.u32 %r324, 0x0;
|
1397 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r321, %r322, %r323, %r324 }, [ %rd87 + 0 ];
|
1398 |
+
@!%p1 mov.u32 %r321, %r342;
|
1399 |
+
@!%p1 mov.u32 %r322, %r342;
|
1400 |
+
@!%p1 mov.u32 %r323, %r342;
|
1401 |
+
@!%p1 mov.u32 %r324, %r342;
|
1402 |
+
mov.b32 %f165, %r321;
|
1403 |
+
mov.b32 %f166, %r322;
|
1404 |
+
mov.b32 %f167, %r323;
|
1405 |
+
mov.b32 %f168, %r324;
|
1406 |
+
mov.u32 %r329, 0x0;
|
1407 |
+
mov.u32 %r330, 0x0;
|
1408 |
+
mov.u32 %r331, 0x0;
|
1409 |
+
mov.u32 %r332, 0x0;
|
1410 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r329, %r330, %r331, %r332 }, [ %rd88 + 0 ];
|
1411 |
+
@!%p1 mov.u32 %r329, %r342;
|
1412 |
+
@!%p1 mov.u32 %r330, %r342;
|
1413 |
+
@!%p1 mov.u32 %r331, %r342;
|
1414 |
+
@!%p1 mov.u32 %r332, %r342;
|
1415 |
+
mov.b32 %f169, %r329;
|
1416 |
+
mov.b32 %f170, %r330;
|
1417 |
+
mov.b32 %f171, %r331;
|
1418 |
+
mov.b32 %f172, %r332;
|
1419 |
+
.loc 1 64 57
|
1420 |
+
@%p41 bra $L__BB0_7;
|
1421 |
+
mov.u64 %rd89, assertMessage_1;
|
1422 |
+
cvta.global.u64 %rd90, %rd89;
|
1423 |
+
mov.u64 %rd91, assertFile_1;
|
1424 |
+
cvta.global.u64 %rd92, %rd91;
|
1425 |
+
mov.u64 %rd93, assertFunc_1;
|
1426 |
+
cvta.global.u64 %rd94, %rd93;
|
1427 |
+
{ // callseq 7, 0
|
1428 |
+
.reg .b32 temp_param_reg;
|
1429 |
+
.param .b64 param0;
|
1430 |
+
st.param.b64 [param0+0], %rd90;
|
1431 |
+
.param .b64 param1;
|
1432 |
+
st.param.b64 [param1+0], %rd92;
|
1433 |
+
.param .b32 param2;
|
1434 |
+
st.param.b32 [param2+0], %r405;
|
1435 |
+
.param .b64 param3;
|
1436 |
+
st.param.b64 [param3+0], %rd94;
|
1437 |
+
.param .b64 param4;
|
1438 |
+
st.param.b64 [param4+0], %rd104;
|
1439 |
+
call.uni
|
1440 |
+
__assertfail,
|
1441 |
+
(
|
1442 |
+
param0,
|
1443 |
+
param1,
|
1444 |
+
param2,
|
1445 |
+
param3,
|
1446 |
+
param4
|
1447 |
+
);
|
1448 |
+
} // callseq 7
|
1449 |
+
bra.uni $L__BB0_7;
|
1450 |
+
$L__BB0_8:
|
1451 |
+
.loc 1 55 4
|
1452 |
+
ret;
|
1453 |
+
$L__tmp17:
|
1454 |
+
$L__func_end0:
|
1455 |
+
|
1456 |
+
}
|
1457 |
+
// .globl __nv_rsqrtf
|
1458 |
+
.visible .func (.param .b32 func_retval0) __nv_rsqrtf(
|
1459 |
+
.param .b32 __nv_rsqrtf_param_0
|
1460 |
+
)
|
1461 |
+
{
|
1462 |
+
.reg .f32 %f<3>;
|
1463 |
+
$L__func_begin1:
|
1464 |
+
|
1465 |
+
ld.param.f32 %f1, [__nv_rsqrtf_param_0];
|
1466 |
+
rsqrt.approx.ftz.f32 %f2, %f1;
|
1467 |
+
st.param.f32 [func_retval0+0], %f2;
|
1468 |
+
ret;
|
1469 |
+
$L__func_end1:
|
1470 |
+
|
1471 |
+
}
|
1472 |
+
.file 1 "/tmp/torchinductor_root/lh/clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py"
|
1473 |
+
.file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
|
1474 |
+
.section .debug_abbrev
|
1475 |
+
{
|
1476 |
+
.b8 1
|
1477 |
+
.b8 17
|
1478 |
+
.b8 1
|
1479 |
+
.b8 37
|
1480 |
+
.b8 8
|
1481 |
+
.b8 19
|
1482 |
+
.b8 5
|
1483 |
+
.b8 3
|
1484 |
+
.b8 8
|
1485 |
+
.b8 16
|
1486 |
+
.b8 6
|
1487 |
+
.b8 27
|
1488 |
+
.b8 8
|
1489 |
+
.b8 180
|
1490 |
+
.b8 66
|
1491 |
+
.b8 12
|
1492 |
+
.b8 17
|
1493 |
+
.b8 1
|
1494 |
+
.b8 18
|
1495 |
+
.b8 1
|
1496 |
+
.b8 0
|
1497 |
+
.b8 0
|
1498 |
+
.b8 2
|
1499 |
+
.b8 46
|
1500 |
+
.b8 0
|
1501 |
+
.b8 135
|
1502 |
+
.b8 64
|
1503 |
+
.b8 8
|
1504 |
+
.b8 3
|
1505 |
+
.b8 8
|
1506 |
+
.b8 58
|
1507 |
+
.b8 11
|
1508 |
+
.b8 59
|
1509 |
+
.b8 11
|
1510 |
+
.b8 63
|
1511 |
+
.b8 12
|
1512 |
+
.b8 32
|
1513 |
+
.b8 11
|
1514 |
+
.b8 0
|
1515 |
+
.b8 0
|
1516 |
+
.b8 3
|
1517 |
+
.b8 46
|
1518 |
+
.b8 1
|
1519 |
+
.b8 17
|
1520 |
+
.b8 1
|
1521 |
+
.b8 18
|
1522 |
+
.b8 1
|
1523 |
+
.b8 64
|
1524 |
+
.b8 10
|
1525 |
+
.b8 49
|
1526 |
+
.b8 19
|
1527 |
+
.b8 0
|
1528 |
+
.b8 0
|
1529 |
+
.b8 4
|
1530 |
+
.b8 29
|
1531 |
+
.b8 0
|
1532 |
+
.b8 49
|
1533 |
+
.b8 19
|
1534 |
+
.b8 17
|
1535 |
+
.b8 1
|
1536 |
+
.b8 18
|
1537 |
+
.b8 1
|
1538 |
+
.b8 88
|
1539 |
+
.b8 11
|
1540 |
+
.b8 89
|
1541 |
+
.b8 11
|
1542 |
+
.b8 87
|
1543 |
+
.b8 11
|
1544 |
+
.b8 0
|
1545 |
+
.b8 0
|
1546 |
+
.b8 5
|
1547 |
+
.b8 29
|
1548 |
+
.b8 1
|
1549 |
+
.b8 49
|
1550 |
+
.b8 19
|
1551 |
+
.b8 17
|
1552 |
+
.b8 1
|
1553 |
+
.b8 18
|
1554 |
+
.b8 1
|
1555 |
+
.b8 88
|
1556 |
+
.b8 11
|
1557 |
+
.b8 89
|
1558 |
+
.b8 11
|
1559 |
+
.b8 87
|
1560 |
+
.b8 11
|
1561 |
+
.b8 0
|
1562 |
+
.b8 0
|
1563 |
+
.b8 0
|
1564 |
+
}
|
1565 |
+
.section .debug_info
|
1566 |
+
{
|
1567 |
+
.b32 298
|
1568 |
+
.b8 2
|
1569 |
+
.b8 0
|
1570 |
+
.b32 .debug_abbrev
|
1571 |
+
.b8 8
|
1572 |
+
.b8 1
|
1573 |
+
.b8 116
|
1574 |
+
.b8 114
|
1575 |
+
.b8 105
|
1576 |
+
.b8 116
|
1577 |
+
.b8 111
|
1578 |
+
.b8 110
|
1579 |
+
.b8 0
|
1580 |
+
.b8 2
|
1581 |
+
.b8 0
|
1582 |
+
.b8 99
|
1583 |
+
.b8 108
|
1584 |
+
.b8 104
|
1585 |
+
.b8 101
|
1586 |
+
.b8 52
|
1587 |
+
.b8 97
|
1588 |
+
.b8 51
|
1589 |
+
.b8 115
|
1590 |
+
.b8 116
|
1591 |
+
.b8 118
|
1592 |
+
.b8 117
|
1593 |
+
.b8 102
|
1594 |
+
.b8 120
|
1595 |
+
.b8 97
|
1596 |
+
.b8 102
|
1597 |
+
.b8 109
|
1598 |
+
.b8 113
|
1599 |
+
.b8 51
|
1600 |
+
.b8 107
|
1601 |
+
.b8 107
|
1602 |
+
.b8 53
|
1603 |
+
.b8 104
|
1604 |
+
.b8 111
|
1605 |
+
.b8 100
|
1606 |
+
.b8 97
|
1607 |
+
.b8 122
|
1608 |
+
.b8 122
|
1609 |
+
.b8 50
|
1610 |
+
.b8 101
|
1611 |
+
.b8 102
|
1612 |
+
.b8 99
|
1613 |
+
.b8 116
|
1614 |
+
.b8 102
|
1615 |
+
.b8 102
|
1616 |
+
.b8 116
|
1617 |
+
.b8 101
|
1618 |
+
.b8 54
|
1619 |
+
.b8 52
|
1620 |
+
.b8 54
|
1621 |
+
.b8 122
|
1622 |
+
.b8 110
|
1623 |
+
.b8 106
|
1624 |
+
.b8 100
|
1625 |
+
.b8 110
|
1626 |
+
.b8 118
|
1627 |
+
.b8 51
|
1628 |
+
.b8 108
|
1629 |
+
.b8 113
|
1630 |
+
.b8 105
|
1631 |
+
.b8 53
|
1632 |
+
.b8 111
|
1633 |
+
.b8 97
|
1634 |
+
.b8 46
|
1635 |
+
.b8 112
|
1636 |
+
.b8 121
|
1637 |
+
.b8 0
|
1638 |
+
.b32 .debug_line
|
1639 |
+
.b8 47
|
1640 |
+
.b8 116
|
1641 |
+
.b8 109
|
1642 |
+
.b8 112
|
1643 |
+
.b8 47
|
1644 |
+
.b8 116
|
1645 |
+
.b8 111
|
1646 |
+
.b8 114
|
1647 |
+
.b8 99
|
1648 |
+
.b8 104
|
1649 |
+
.b8 105
|
1650 |
+
.b8 110
|
1651 |
+
.b8 100
|
1652 |
+
.b8 117
|
1653 |
+
.b8 99
|
1654 |
+
.b8 116
|
1655 |
+
.b8 111
|
1656 |
+
.b8 114
|
1657 |
+
.b8 95
|
1658 |
+
.b8 114
|
1659 |
+
.b8 111
|
1660 |
+
.b8 111
|
1661 |
+
.b8 116
|
1662 |
+
.b8 47
|
1663 |
+
.b8 108
|
1664 |
+
.b8 104
|
1665 |
+
.b8 0
|
1666 |
+
.b8 1
|
1667 |
+
.b64 $L__func_begin0
|
1668 |
+
.b64 $L__func_end0
|
1669 |
+
.b8 2
|
1670 |
+
.b8 116
|
1671 |
+
.b8 114
|
1672 |
+
.b8 105
|
1673 |
+
.b8 116
|
1674 |
+
.b8 111
|
1675 |
+
.b8 110
|
1676 |
+
.b8 95
|
1677 |
+
.b8 95
|
1678 |
+
.b8 48
|
1679 |
+
.b8 100
|
1680 |
+
.b8 49
|
1681 |
+
.b8 100
|
1682 |
+
.b8 50
|
1683 |
+
.b8 100
|
1684 |
+
.b8 51
|
1685 |
+
.b8 100
|
1686 |
+
.b8 52
|
1687 |
+
.b8 100
|
1688 |
+
.b8 53
|
1689 |
+
.b8 100
|
1690 |
+
.b8 101
|
1691 |
+
.b8 54
|
1692 |
+
.b8 100
|
1693 |
+
.b8 101
|
1694 |
+
.b8 0
|
1695 |
+
.b8 116
|
1696 |
+
.b8 114
|
1697 |
+
.b8 105
|
1698 |
+
.b8 116
|
1699 |
+
.b8 111
|
1700 |
+
.b8 110
|
1701 |
+
.b8 95
|
1702 |
+
.b8 95
|
1703 |
+
.b8 48
|
1704 |
+
.b8 100
|
1705 |
+
.b8 49
|
1706 |
+
.b8 100
|
1707 |
+
.b8 50
|
1708 |
+
.b8 100
|
1709 |
+
.b8 51
|
1710 |
+
.b8 100
|
1711 |
+
.b8 52
|
1712 |
+
.b8 100
|
1713 |
+
.b8 53
|
1714 |
+
.b8 100
|
1715 |
+
.b8 101
|
1716 |
+
.b8 54
|
1717 |
+
.b8 100
|
1718 |
+
.b8 101
|
1719 |
+
.b8 0
|
1720 |
+
.b8 1
|
1721 |
+
.b8 18
|
1722 |
+
.b8 1
|
1723 |
+
.b8 1
|
1724 |
+
.b8 3
|
1725 |
+
.b64 $L__func_begin0
|
1726 |
+
.b64 $L__func_end0
|
1727 |
+
.b8 1
|
1728 |
+
.b8 156
|
1729 |
+
.b32 125
|
1730 |
+
.b8 4
|
1731 |
+
.b32 125
|
1732 |
+
.b64 $L__tmp1
|
1733 |
+
.b64 $L__tmp2
|
1734 |
+
.b8 2
|
1735 |
+
.b8 44
|
1736 |
+
.b8 38
|
1737 |
+
.b8 5
|
1738 |
+
.b32 125
|
1739 |
+
.b64 $L__tmp3
|
1740 |
+
.b64 $L__tmp16
|
1741 |
+
.b8 2
|
1742 |
+
.b8 50
|
1743 |
+
.b8 41
|
1744 |
+
.b8 4
|
1745 |
+
.b32 125
|
1746 |
+
.b64 $L__tmp3
|
1747 |
+
.b64 $L__tmp16
|
1748 |
+
.b8 2
|
1749 |
+
.b8 120
|
1750 |
+
.b8 46
|
1751 |
+
.b8 0
|
1752 |
+
.b8 4
|
1753 |
+
.b32 125
|
1754 |
+
.b64 $L__tmp4
|
1755 |
+
.b64 $L__tmp15
|
1756 |
+
.b8 2
|
1757 |
+
.b8 50
|
1758 |
+
.b8 41
|
1759 |
+
.b8 0
|
1760 |
+
.b8 0
|
1761 |
+
}
|
1762 |
+
.section .debug_pubnames
|
1763 |
+
{
|
1764 |
+
.b32 $L__pubNames_end0-$L__pubNames_start0
|
1765 |
+
$L__pubNames_start0:
|
1766 |
+
.b8 2
|
1767 |
+
.b8 0
|
1768 |
+
.b32 .debug_info
|
1769 |
+
.b32 302
|
1770 |
+
.b32 125
|
1771 |
+
.b8 116
|
1772 |
+
.b8 114
|
1773 |
+
.b8 105
|
1774 |
+
.b8 116
|
1775 |
+
.b8 111
|
1776 |
+
.b8 110
|
1777 |
+
.b8 95
|
1778 |
+
.b8 95
|
1779 |
+
.b8 48
|
1780 |
+
.b8 100
|
1781 |
+
.b8 49
|
1782 |
+
.b8 100
|
1783 |
+
.b8 50
|
1784 |
+
.b8 100
|
1785 |
+
.b8 51
|
1786 |
+
.b8 100
|
1787 |
+
.b8 52
|
1788 |
+
.b8 100
|
1789 |
+
.b8 53
|
1790 |
+
.b8 100
|
1791 |
+
.b8 101
|
1792 |
+
.b8 54
|
1793 |
+
.b8 100
|
1794 |
+
.b8 101
|
1795 |
+
.b8 0
|
1796 |
+
.b32 0
|
1797 |
+
$L__pubNames_end0:
|
1798 |
+
}
|
1799 |
+
.section .debug_pubtypes
|
1800 |
+
{
|
1801 |
+
.b32 $L__pubTypes_end0-$L__pubTypes_start0
|
1802 |
+
$L__pubTypes_start0:
|
1803 |
+
.b8 2
|
1804 |
+
.b8 0
|
1805 |
+
.b32 .debug_info
|
1806 |
+
.b32 302
|
1807 |
+
.b32 0
|
1808 |
+
$L__pubTypes_end0:
|
1809 |
+
}
|
1810 |
+
.section .debug_loc { }
|
.triton/dump/510522bb05917b836ed253751364fcad/triton_.ttgir
ADDED
@@ -0,0 +1,153 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [1, 8], threadsPerWarp = [4, 8], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
2 |
+
#blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
3 |
+
#blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [4, 2], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
4 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 8 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
5 |
+
tt.func public @triton__0d1d2d3d4d5de6de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
6 |
+
%cst = arith.constant dense<512> : tensor<64x1xi32, #blocked>
|
7 |
+
%cst_0 = arith.constant dense<256> : tensor<1x64xi32, #blocked>
|
8 |
+
%cst_1 = arith.constant dense<256> : tensor<64x1xi32, #blocked>
|
9 |
+
%cst_2 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked>
|
10 |
+
%cst_3 = arith.constant dense<0.000000e+00> : tensor<1x64xf32, #blocked>
|
11 |
+
%cst_4 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked>
|
12 |
+
%cst_5 = arith.constant dense<256> : tensor<64x1xi64, #blocked>
|
13 |
+
%cst_6 = arith.constant dense<0> : tensor<64x1xi64, #blocked>
|
14 |
+
%cst_7 = arith.constant dense<50257> : tensor<64x1xi64, #blocked>
|
15 |
+
%cst_8 = arith.constant dense<50257> : tensor<64x1xi64, #blocked1>
|
16 |
+
%cst_9 = arith.constant dense<0> : tensor<64x1xi64, #blocked1>
|
17 |
+
%c0_i32 = arith.constant 0 : i32
|
18 |
+
%c64_i32 = arith.constant 64 : i32
|
19 |
+
%c256_i32 = arith.constant 256 : i32
|
20 |
+
%cst_10 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked2>
|
21 |
+
%cst_11 = arith.constant 0.000000e+00 : f32
|
22 |
+
%cst_12 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked2>
|
23 |
+
%cst_13 = arith.constant dense<256> : tensor<1x64xi32, #blocked2>
|
24 |
+
%cst_14 = arith.constant dense<9.99999974E-6> : tensor<64x1xf32, #blocked>
|
25 |
+
%cst_15 = arith.constant dense<2.560000e+02> : tensor<64x1xf32, #blocked>
|
26 |
+
%0 = tt.get_program_id x : i32
|
27 |
+
%1 = arith.muli %0, %c64_i32 : i32
|
28 |
+
%2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
|
29 |
+
%3 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
|
30 |
+
%4 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xi32, #blocked>
|
31 |
+
%5 = tt.expand_dims %3 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<64x1xi32, #blocked1>
|
32 |
+
%6 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked>
|
33 |
+
%7 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked1>
|
34 |
+
%8 = arith.addi %6, %4 : tensor<64x1xi32, #blocked>
|
35 |
+
%9 = arith.addi %7, %5 : tensor<64x1xi32, #blocked1>
|
36 |
+
%10 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
|
37 |
+
%11 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>
|
38 |
+
%12 = tt.expand_dims %10 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x64xi32, #blocked>
|
39 |
+
%13 = tt.expand_dims %11 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>) -> tensor<1x64xi32, #blocked2>
|
40 |
+
%14 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked>
|
41 |
+
%15 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked1>
|
42 |
+
%16 = tt.addptr %14, %8 : tensor<64x1x!tt.ptr<i64, 1>, #blocked>, tensor<64x1xi32, #blocked>
|
43 |
+
%17 = tt.addptr %15, %9 : tensor<64x1x!tt.ptr<i64, 1>, #blocked1>, tensor<64x1xi32, #blocked1>
|
44 |
+
%18 = tt.load %16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked>
|
45 |
+
%19 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked1>
|
46 |
+
%20 = arith.remsi %8, %cst : tensor<64x1xi32, #blocked>
|
47 |
+
%21 = arith.muli %20, %cst_1 : tensor<64x1xi32, #blocked>
|
48 |
+
%22 = tt.broadcast %21 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
49 |
+
%23 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
|
50 |
+
%24 = arith.addi %18, %cst_7 : tensor<64x1xi64, #blocked>
|
51 |
+
%25 = arith.addi %19, %cst_8 : tensor<64x1xi64, #blocked1>
|
52 |
+
%26 = arith.cmpi slt, %18, %cst_6 : tensor<64x1xi64, #blocked>
|
53 |
+
%27 = arith.cmpi slt, %19, %cst_9 : tensor<64x1xi64, #blocked1>
|
54 |
+
%28 = arith.select %26, %24, %18 : tensor<64x1xi1, #blocked>, tensor<64x1xi64, #blocked>
|
55 |
+
%29 = arith.select %27, %25, %19 : tensor<64x1xi1, #blocked1>, tensor<64x1xi64, #blocked1>
|
56 |
+
%30 = arith.cmpi sge, %29, %cst_9 : tensor<64x1xi64, #blocked1>
|
57 |
+
%31 = arith.cmpi slt, %29, %cst_8 : tensor<64x1xi64, #blocked1>
|
58 |
+
%32 = arith.andi %30, %31 : tensor<64x1xi1, #blocked1>
|
59 |
+
%33 = arith.muli %28, %cst_5 : tensor<64x1xi64, #blocked>
|
60 |
+
%34 = tt.broadcast %33 : (tensor<64x1xi64, #blocked>) -> tensor<64x64xi64, #blocked>
|
61 |
+
%35 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
|
62 |
+
%36:4 = scf.for %arg7 = %c0_i32 to %c256_i32 step %c64_i32 iter_args(%arg8 = %cst_2, %arg9 = %cst_2, %arg10 = %cst_12, %arg11 = %cst_2) -> (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>) : i32 {
|
63 |
+
%48 = tt.splat %arg7 : (i32) -> tensor<1x64xi32, #blocked>
|
64 |
+
%49 = tt.splat %arg7 : (i32) -> tensor<1x64xi32, #blocked2>
|
65 |
+
%50 = arith.addi %48, %12 : tensor<1x64xi32, #blocked>
|
66 |
+
%51 = arith.addi %49, %13 : tensor<1x64xi32, #blocked2>
|
67 |
+
%52 = arith.cmpi slt, %50, %cst_0 : tensor<1x64xi32, #blocked>
|
68 |
+
%53 = arith.cmpi slt, %51, %cst_13 : tensor<1x64xi32, #blocked2>
|
69 |
+
%54 = tt.broadcast %50 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
70 |
+
%55 = arith.addi %54, %22 : tensor<64x64xi32, #blocked>
|
71 |
+
%56 = tt.addptr %23, %55 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
72 |
+
%57 = tt.broadcast %52 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
|
73 |
+
%58 = tt.broadcast %53 : (tensor<1x64xi1, #blocked2>) -> tensor<64x64xi1, #blocked2>
|
74 |
+
%59 = tt.load %56, %57, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
75 |
+
tt.assert %32, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
|
76 |
+
%60 = arith.extsi %50 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
|
77 |
+
%61 = tt.broadcast %60 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
|
78 |
+
%62 = arith.addi %61, %34 : tensor<64x64xi64, #blocked>
|
79 |
+
%63 = tt.addptr %35, %62 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
|
80 |
+
%64 = tt.load %63, %57, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
81 |
+
%65 = arith.addf %64, %59 : tensor<64x64xf32, #blocked>
|
82 |
+
%66 = arith.subf %65, %arg8 : tensor<64x64xf32, #blocked>
|
83 |
+
%67 = arith.addf %arg11, %cst_4 : tensor<64x64xf32, #blocked>
|
84 |
+
%68 = arith.addf %arg10, %cst_10 : tensor<64x64xf32, #blocked2>
|
85 |
+
%69 = arith.divf %66, %67 : tensor<64x64xf32, #blocked>
|
86 |
+
%70 = arith.addf %arg8, %69 : tensor<64x64xf32, #blocked>
|
87 |
+
%71 = arith.subf %65, %70 : tensor<64x64xf32, #blocked>
|
88 |
+
%72 = arith.mulf %66, %71 : tensor<64x64xf32, #blocked>
|
89 |
+
%73 = arith.addf %arg9, %72 : tensor<64x64xf32, #blocked>
|
90 |
+
%74 = arith.select %57, %70, %arg8 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
|
91 |
+
%75 = arith.select %57, %73, %arg9 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
|
92 |
+
%76 = arith.select %57, %67, %arg11 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
|
93 |
+
%77 = arith.select %58, %68, %arg10 : tensor<64x64xi1, #blocked2>, tensor<64x64xf32, #blocked2>
|
94 |
+
scf.yield %74, %75, %77, %76 : tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>
|
95 |
+
}
|
96 |
+
%37 = triton_gpu.convert_layout %36#2 : (tensor<64x64xf32, #blocked2>) -> tensor<64x64xf32, #blocked>
|
97 |
+
%38:3 = "tt.reduce"(%36#0, %36#1, %37) <{axis = 1 : i32}> ({
|
98 |
+
^bb0(%arg7: f32, %arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32):
|
99 |
+
%48 = arith.subf %arg10, %arg7 : f32
|
100 |
+
%49 = arith.addf %arg9, %arg12 : f32
|
101 |
+
%50 = arith.cmpf oeq, %49, %cst_11 : f32
|
102 |
+
%51 = arith.divf %arg12, %49 : f32
|
103 |
+
%52 = arith.select %50, %cst_11, %51 : f32
|
104 |
+
%53 = arith.mulf %48, %52 : f32
|
105 |
+
%54 = arith.addf %arg7, %53 : f32
|
106 |
+
%55 = arith.addf %arg8, %arg11 : f32
|
107 |
+
%56 = arith.mulf %48, %48 : f32
|
108 |
+
%57 = arith.mulf %56, %arg9 : f32
|
109 |
+
%58 = arith.mulf %57, %52 : f32
|
110 |
+
%59 = arith.addf %55, %58 : f32
|
111 |
+
tt.reduce.return %54, %59, %49 : f32, f32, f32
|
112 |
+
}) : (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>) -> (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>)
|
113 |
+
%39 = tt.expand_dims %38#0 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
|
114 |
+
%40 = tt.expand_dims %38#1 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
|
115 |
+
%41 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x64x!tt.ptr<f32, 1>, #blocked>
|
116 |
+
%42 = tt.broadcast %39 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
|
117 |
+
%43 = arith.divf %40, %cst_15 : tensor<64x1xf32, #blocked>
|
118 |
+
%44 = arith.addf %43, %cst_14 : tensor<64x1xf32, #blocked>
|
119 |
+
%45 = arith.muli %8, %cst_1 : tensor<64x1xi32, #blocked>
|
120 |
+
%46 = tt.broadcast %45 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
121 |
+
%47 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>, #blocked>
|
122 |
+
scf.for %arg7 = %c0_i32 to %c256_i32 step %c64_i32 : i32 {
|
123 |
+
%48 = tt.splat %arg7 : (i32) -> tensor<1x64xi32, #blocked>
|
124 |
+
%49 = arith.addi %48, %12 : tensor<1x64xi32, #blocked>
|
125 |
+
%50 = arith.cmpi slt, %49, %cst_0 : tensor<1x64xi32, #blocked>
|
126 |
+
%51 = tt.broadcast %49 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
127 |
+
%52 = arith.addi %51, %22 : tensor<64x64xi32, #blocked>
|
128 |
+
%53 = tt.addptr %23, %52 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
129 |
+
%54 = tt.broadcast %50 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
|
130 |
+
%55 = tt.load %53, %54, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
131 |
+
%56 = tt.addptr %41, %49 : tensor<1x64x!tt.ptr<f32, 1>, #blocked>, tensor<1x64xi32, #blocked>
|
132 |
+
%57 = tt.load %56, %50, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x64xf32, #blocked>
|
133 |
+
tt.assert %32, "index out of bounds: 0 <= tmp13 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
|
134 |
+
%58 = arith.extsi %49 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
|
135 |
+
%59 = tt.broadcast %58 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
|
136 |
+
%60 = arith.addi %59, %34 : tensor<64x64xi64, #blocked>
|
137 |
+
%61 = tt.addptr %35, %60 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
|
138 |
+
%62 = tt.load %61, %54, %cst_2 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
139 |
+
%63 = arith.addf %62, %55 : tensor<64x64xf32, #blocked>
|
140 |
+
%64 = arith.subf %63, %42 : tensor<64x64xf32, #blocked>
|
141 |
+
%65 = tt.extern_elementwise %44 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<64x1xf32, #blocked>) -> tensor<64x1xf32, #blocked>
|
142 |
+
%66 = tt.broadcast %65 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
|
143 |
+
%67 = arith.mulf %64, %66 : tensor<64x64xf32, #blocked>
|
144 |
+
%68 = tt.broadcast %57 : (tensor<1x64xf32, #blocked>) -> tensor<64x64xf32, #blocked>
|
145 |
+
%69 = arith.mulf %67, %68 : tensor<64x64xf32, #blocked>
|
146 |
+
%70 = arith.addi %51, %46 : tensor<64x64xi32, #blocked>
|
147 |
+
%71 = tt.addptr %47, %70 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
148 |
+
%72 = arith.truncf %69 : tensor<64x64xf32, #blocked> to tensor<64x64xbf16, #blocked>
|
149 |
+
tt.store %71, %72, %54 {cache = 1 : i32, evict = 1 : i32} : tensor<64x64xbf16, #blocked>
|
150 |
+
}
|
151 |
+
tt.return
|
152 |
+
}
|
153 |
+
}
|
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.cubin
ADDED
Binary file (73.7 kB). View file
|
|
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.llir
ADDED
@@ -0,0 +1,1360 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
; ModuleID = 'LLVMDialectModule'
|
2 |
+
source_filename = "LLVMDialectModule"
|
3 |
+
|
4 |
+
@assertFunc_1 = internal constant [25 x i8] c"_call_with_frames_removed"
|
5 |
+
@assertFile_1 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
|
6 |
+
@assertMessage_1 = internal constant [39 x i8] c"index out of bounds: 0 <= tmp16 < 50257"
|
7 |
+
@assertFunc_0 = internal constant [25 x i8] c"_call_with_frames_removed"
|
8 |
+
@assertFile_0 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
|
9 |
+
@assertMessage_0 = internal constant [38 x i8] c"index out of bounds: 0 <= tmp3 < 50257"
|
10 |
+
@global_smem = external local_unnamed_addr addrspace(3) global [0 x i8]
|
11 |
+
@.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
|
12 |
+
|
13 |
+
declare void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr
|
14 |
+
|
15 |
+
define void @triton__0d1d2d3d4d5d6de7de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, i32 %6, i32 %7) local_unnamed_addr !dbg !7 {
|
16 |
+
%9 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
|
17 |
+
%10 = lshr i32 %9, 3, !dbg !10
|
18 |
+
%11 = and i32 %10, 31, !dbg !10
|
19 |
+
%12 = and i32 %9, 63, !dbg !10
|
20 |
+
%13 = shl i32 %9, 3, !dbg !11
|
21 |
+
%14 = and i32 %13, 56, !dbg !11
|
22 |
+
%15 = or i32 %14, 4, !dbg !11
|
23 |
+
%16 = lshr i32 %9, 6, !dbg !12
|
24 |
+
%17 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !13
|
25 |
+
%18 = shl i32 %17, 6, !dbg !14
|
26 |
+
%19 = or i32 %18, %11, !dbg !15
|
27 |
+
%20 = or i32 %19, 32, !dbg !15
|
28 |
+
%21 = or i32 %18, %12, !dbg !15
|
29 |
+
%22 = sext i32 %19 to i64, !dbg !16
|
30 |
+
%23 = getelementptr i64, ptr addrspace(1) %0, i64 %22, !dbg !16
|
31 |
+
%24 = sext i32 %20 to i64, !dbg !16
|
32 |
+
%25 = getelementptr i64, ptr addrspace(1) %0, i64 %24, !dbg !16
|
33 |
+
%26 = sext i32 %21 to i64, !dbg !16
|
34 |
+
%27 = getelementptr i64, ptr addrspace(1) %0, i64 %26, !dbg !16
|
35 |
+
%28 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
36 |
+
%29 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
37 |
+
%30 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
38 |
+
%31 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
39 |
+
%32 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
40 |
+
%33 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
41 |
+
%34 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
42 |
+
%35 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
|
43 |
+
%36 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
44 |
+
%37 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
45 |
+
%38 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
46 |
+
%39 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
47 |
+
%40 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
48 |
+
%41 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
49 |
+
%42 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
50 |
+
%43 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
|
51 |
+
%44 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %27, i1 true) #6, !dbg !17
|
52 |
+
%45 = srem i32 %19, 512, !dbg !18
|
53 |
+
%46 = srem i32 %20, 512, !dbg !18
|
54 |
+
%47 = shl nsw i32 %45, 8, !dbg !19
|
55 |
+
%48 = shl nsw i32 %46, 8, !dbg !19
|
56 |
+
%49 = shl i32 %19, 8, !dbg !20
|
57 |
+
%50 = shl i32 %20, 8, !dbg !20
|
58 |
+
%51 = add i64 %44, 50257, !dbg !21
|
59 |
+
%52 = icmp slt i64 %28, 0, !dbg !22
|
60 |
+
%53 = icmp slt i64 %36, 0, !dbg !22
|
61 |
+
%54 = icmp slt i64 %44, 0, !dbg !22
|
62 |
+
%55 = select i1 %54, i64 %51, i64 %44, !dbg !23
|
63 |
+
%56 = icmp ugt i64 %55, 50256, !dbg !24
|
64 |
+
%57 = shl i64 %28, 8, !dbg !25
|
65 |
+
%58 = add i64 %57, 12865792, !dbg !25
|
66 |
+
%59 = select i1 %52, i64 %58, i64 %57, !dbg !25
|
67 |
+
%60 = shl i64 %36, 8, !dbg !25
|
68 |
+
%61 = add i64 %60, 12865792, !dbg !25
|
69 |
+
%62 = select i1 %53, i64 %61, i64 %60, !dbg !25
|
70 |
+
%63 = getelementptr float, ptr addrspace(1) %1, i64 %59
|
71 |
+
%64 = getelementptr float, ptr addrspace(1) %1, i64 %62
|
72 |
+
br label %65, !dbg !12
|
73 |
+
|
74 |
+
65: ; preds = %8, %230
|
75 |
+
%66 = phi float [ 0.000000e+00, %8 ], [ %321, %230 ]
|
76 |
+
%67 = phi float [ 0.000000e+00, %8 ], [ %322, %230 ]
|
77 |
+
%68 = phi float [ 0.000000e+00, %8 ], [ %323, %230 ]
|
78 |
+
%69 = phi float [ 0.000000e+00, %8 ], [ %324, %230 ]
|
79 |
+
%70 = phi float [ 0.000000e+00, %8 ], [ %325, %230 ]
|
80 |
+
%71 = phi float [ 0.000000e+00, %8 ], [ %326, %230 ]
|
81 |
+
%72 = phi float [ 0.000000e+00, %8 ], [ %327, %230 ]
|
82 |
+
%73 = phi float [ 0.000000e+00, %8 ], [ %328, %230 ]
|
83 |
+
%74 = phi float [ 0.000000e+00, %8 ], [ %329, %230 ]
|
84 |
+
%75 = phi float [ 0.000000e+00, %8 ], [ %330, %230 ]
|
85 |
+
%76 = phi float [ 0.000000e+00, %8 ], [ %331, %230 ]
|
86 |
+
%77 = phi float [ 0.000000e+00, %8 ], [ %332, %230 ]
|
87 |
+
%78 = phi float [ 0.000000e+00, %8 ], [ %333, %230 ]
|
88 |
+
%79 = phi float [ 0.000000e+00, %8 ], [ %334, %230 ]
|
89 |
+
%80 = phi float [ 0.000000e+00, %8 ], [ %335, %230 ]
|
90 |
+
%81 = phi float [ 0.000000e+00, %8 ], [ %336, %230 ]
|
91 |
+
%82 = phi float [ 0.000000e+00, %8 ], [ %337, %230 ]
|
92 |
+
%83 = phi float [ 0.000000e+00, %8 ], [ %338, %230 ]
|
93 |
+
%84 = phi float [ 0.000000e+00, %8 ], [ %339, %230 ]
|
94 |
+
%85 = phi float [ 0.000000e+00, %8 ], [ %340, %230 ]
|
95 |
+
%86 = phi float [ 0.000000e+00, %8 ], [ %341, %230 ]
|
96 |
+
%87 = phi float [ 0.000000e+00, %8 ], [ %342, %230 ]
|
97 |
+
%88 = phi float [ 0.000000e+00, %8 ], [ %343, %230 ]
|
98 |
+
%89 = phi float [ 0.000000e+00, %8 ], [ %344, %230 ]
|
99 |
+
%90 = phi float [ 0.000000e+00, %8 ], [ %345, %230 ]
|
100 |
+
%91 = phi float [ 0.000000e+00, %8 ], [ %346, %230 ]
|
101 |
+
%92 = phi float [ 0.000000e+00, %8 ], [ %347, %230 ]
|
102 |
+
%93 = phi float [ 0.000000e+00, %8 ], [ %348, %230 ]
|
103 |
+
%94 = phi float [ 0.000000e+00, %8 ], [ %349, %230 ]
|
104 |
+
%95 = phi float [ 0.000000e+00, %8 ], [ %350, %230 ]
|
105 |
+
%96 = phi float [ 0.000000e+00, %8 ], [ %351, %230 ]
|
106 |
+
%97 = phi float [ 0.000000e+00, %8 ], [ %352, %230 ]
|
107 |
+
%98 = phi float [ 0.000000e+00, %8 ], [ %417, %230 ]
|
108 |
+
%99 = phi float [ 0.000000e+00, %8 ], [ %418, %230 ]
|
109 |
+
%100 = phi float [ 0.000000e+00, %8 ], [ %419, %230 ]
|
110 |
+
%101 = phi float [ 0.000000e+00, %8 ], [ %420, %230 ]
|
111 |
+
%102 = phi float [ 0.000000e+00, %8 ], [ %421, %230 ]
|
112 |
+
%103 = phi float [ 0.000000e+00, %8 ], [ %422, %230 ]
|
113 |
+
%104 = phi float [ 0.000000e+00, %8 ], [ %423, %230 ]
|
114 |
+
%105 = phi float [ 0.000000e+00, %8 ], [ %424, %230 ]
|
115 |
+
%106 = phi float [ 0.000000e+00, %8 ], [ %425, %230 ]
|
116 |
+
%107 = phi float [ 0.000000e+00, %8 ], [ %426, %230 ]
|
117 |
+
%108 = phi float [ 0.000000e+00, %8 ], [ %427, %230 ]
|
118 |
+
%109 = phi float [ 0.000000e+00, %8 ], [ %428, %230 ]
|
119 |
+
%110 = phi float [ 0.000000e+00, %8 ], [ %429, %230 ]
|
120 |
+
%111 = phi float [ 0.000000e+00, %8 ], [ %430, %230 ]
|
121 |
+
%112 = phi float [ 0.000000e+00, %8 ], [ %431, %230 ]
|
122 |
+
%113 = phi float [ 0.000000e+00, %8 ], [ %432, %230 ]
|
123 |
+
%114 = phi float [ 0.000000e+00, %8 ], [ %369, %230 ]
|
124 |
+
%115 = phi float [ 0.000000e+00, %8 ], [ %370, %230 ]
|
125 |
+
%116 = phi float [ 0.000000e+00, %8 ], [ %371, %230 ]
|
126 |
+
%117 = phi float [ 0.000000e+00, %8 ], [ %372, %230 ]
|
127 |
+
%118 = phi float [ 0.000000e+00, %8 ], [ %373, %230 ]
|
128 |
+
%119 = phi float [ 0.000000e+00, %8 ], [ %374, %230 ]
|
129 |
+
%120 = phi float [ 0.000000e+00, %8 ], [ %375, %230 ]
|
130 |
+
%121 = phi float [ 0.000000e+00, %8 ], [ %376, %230 ]
|
131 |
+
%122 = phi float [ 0.000000e+00, %8 ], [ %377, %230 ]
|
132 |
+
%123 = phi float [ 0.000000e+00, %8 ], [ %378, %230 ]
|
133 |
+
%124 = phi float [ 0.000000e+00, %8 ], [ %379, %230 ]
|
134 |
+
%125 = phi float [ 0.000000e+00, %8 ], [ %380, %230 ]
|
135 |
+
%126 = phi float [ 0.000000e+00, %8 ], [ %381, %230 ]
|
136 |
+
%127 = phi float [ 0.000000e+00, %8 ], [ %382, %230 ]
|
137 |
+
%128 = phi float [ 0.000000e+00, %8 ], [ %383, %230 ]
|
138 |
+
%129 = phi float [ 0.000000e+00, %8 ], [ %384, %230 ]
|
139 |
+
%130 = phi i32 [ 0, %8 ], [ %433, %230 ]
|
140 |
+
%131 = or i32 %130, %14, !dbg !26
|
141 |
+
%132 = or i32 %130, %15, !dbg !26
|
142 |
+
%133 = add i32 %131, %47, !dbg !27
|
143 |
+
%134 = add i32 %132, %47, !dbg !27
|
144 |
+
%135 = add i32 %131, %48, !dbg !27
|
145 |
+
%136 = add i32 %132, %48, !dbg !27
|
146 |
+
%137 = sext i32 %133 to i64, !dbg !28
|
147 |
+
%138 = getelementptr float, ptr addrspace(1) %2, i64 %137, !dbg !28
|
148 |
+
%139 = sext i32 %134 to i64, !dbg !28
|
149 |
+
%140 = getelementptr float, ptr addrspace(1) %2, i64 %139, !dbg !28
|
150 |
+
%141 = sext i32 %135 to i64, !dbg !28
|
151 |
+
%142 = getelementptr float, ptr addrspace(1) %2, i64 %141, !dbg !28
|
152 |
+
%143 = sext i32 %136 to i64, !dbg !28
|
153 |
+
%144 = getelementptr float, ptr addrspace(1) %2, i64 %143, !dbg !28
|
154 |
+
%145 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %138, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
|
155 |
+
%146 = extractvalue { i32, i32, i32, i32 } %145, 0, !dbg !29
|
156 |
+
%147 = extractvalue { i32, i32, i32, i32 } %145, 1, !dbg !29
|
157 |
+
%148 = extractvalue { i32, i32, i32, i32 } %145, 2, !dbg !29
|
158 |
+
%149 = extractvalue { i32, i32, i32, i32 } %145, 3, !dbg !29
|
159 |
+
%150 = bitcast i32 %146 to float, !dbg !29
|
160 |
+
%151 = bitcast i32 %147 to float, !dbg !29
|
161 |
+
%152 = bitcast i32 %148 to float, !dbg !29
|
162 |
+
%153 = bitcast i32 %149 to float, !dbg !29
|
163 |
+
%154 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %140, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
|
164 |
+
%155 = extractvalue { i32, i32, i32, i32 } %154, 0, !dbg !29
|
165 |
+
%156 = extractvalue { i32, i32, i32, i32 } %154, 1, !dbg !29
|
166 |
+
%157 = extractvalue { i32, i32, i32, i32 } %154, 2, !dbg !29
|
167 |
+
%158 = extractvalue { i32, i32, i32, i32 } %154, 3, !dbg !29
|
168 |
+
%159 = bitcast i32 %155 to float, !dbg !29
|
169 |
+
%160 = bitcast i32 %156 to float, !dbg !29
|
170 |
+
%161 = bitcast i32 %157 to float, !dbg !29
|
171 |
+
%162 = bitcast i32 %158 to float, !dbg !29
|
172 |
+
%163 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %142, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
|
173 |
+
%164 = extractvalue { i32, i32, i32, i32 } %163, 0, !dbg !29
|
174 |
+
%165 = extractvalue { i32, i32, i32, i32 } %163, 1, !dbg !29
|
175 |
+
%166 = extractvalue { i32, i32, i32, i32 } %163, 2, !dbg !29
|
176 |
+
%167 = extractvalue { i32, i32, i32, i32 } %163, 3, !dbg !29
|
177 |
+
%168 = bitcast i32 %164 to float, !dbg !29
|
178 |
+
%169 = bitcast i32 %165 to float, !dbg !29
|
179 |
+
%170 = bitcast i32 %166 to float, !dbg !29
|
180 |
+
%171 = bitcast i32 %167 to float, !dbg !29
|
181 |
+
%172 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %144, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
|
182 |
+
%173 = extractvalue { i32, i32, i32, i32 } %172, 0, !dbg !29
|
183 |
+
%174 = extractvalue { i32, i32, i32, i32 } %172, 1, !dbg !29
|
184 |
+
%175 = extractvalue { i32, i32, i32, i32 } %172, 2, !dbg !29
|
185 |
+
%176 = extractvalue { i32, i32, i32, i32 } %172, 3, !dbg !29
|
186 |
+
%177 = bitcast i32 %173 to float, !dbg !29
|
187 |
+
%178 = bitcast i32 %174 to float, !dbg !29
|
188 |
+
%179 = bitcast i32 %175 to float, !dbg !29
|
189 |
+
%180 = bitcast i32 %176 to float, !dbg !29
|
190 |
+
%181 = add i32 %131, %49, !dbg !30
|
191 |
+
%182 = add i32 %131, %50, !dbg !30
|
192 |
+
%183 = sext i32 %181 to i64, !dbg !31
|
193 |
+
%184 = getelementptr i16, ptr addrspace(1) %3, i64 %183, !dbg !31
|
194 |
+
%185 = sext i32 %182 to i64, !dbg !31
|
195 |
+
%186 = getelementptr i16, ptr addrspace(1) %3, i64 %185, !dbg !31
|
196 |
+
%187 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %184, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !32
|
197 |
+
%188 = extractvalue { i32, i32, i32, i32 } %187, 0, !dbg !32
|
198 |
+
%189 = extractvalue { i32, i32, i32, i32 } %187, 1, !dbg !32
|
199 |
+
%190 = extractvalue { i32, i32, i32, i32 } %187, 2, !dbg !32
|
200 |
+
%191 = extractvalue { i32, i32, i32, i32 } %187, 3, !dbg !32
|
201 |
+
%192 = trunc i32 %188 to i16, !dbg !32
|
202 |
+
%extelt.offset9 = lshr i32 %188, 16, !dbg !32
|
203 |
+
%193 = trunc i32 %extelt.offset9 to i16, !dbg !32
|
204 |
+
%194 = trunc i32 %189 to i16, !dbg !32
|
205 |
+
%extelt.offset10 = lshr i32 %189, 16, !dbg !32
|
206 |
+
%195 = trunc i32 %extelt.offset10 to i16, !dbg !32
|
207 |
+
%196 = trunc i32 %190 to i16, !dbg !32
|
208 |
+
%extelt.offset11 = lshr i32 %190, 16, !dbg !32
|
209 |
+
%197 = trunc i32 %extelt.offset11 to i16, !dbg !32
|
210 |
+
%198 = trunc i32 %191 to i16, !dbg !32
|
211 |
+
%extelt.offset12 = lshr i32 %191, 16, !dbg !32
|
212 |
+
%199 = trunc i32 %extelt.offset12 to i16, !dbg !32
|
213 |
+
%200 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %186, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !32
|
214 |
+
%201 = extractvalue { i32, i32, i32, i32 } %200, 0, !dbg !32
|
215 |
+
%202 = extractvalue { i32, i32, i32, i32 } %200, 1, !dbg !32
|
216 |
+
%203 = extractvalue { i32, i32, i32, i32 } %200, 2, !dbg !32
|
217 |
+
%204 = extractvalue { i32, i32, i32, i32 } %200, 3, !dbg !32
|
218 |
+
%205 = trunc i32 %201 to i16, !dbg !32
|
219 |
+
%extelt.offset13 = lshr i32 %201, 16, !dbg !32
|
220 |
+
%206 = trunc i32 %extelt.offset13 to i16, !dbg !32
|
221 |
+
%207 = trunc i32 %202 to i16, !dbg !32
|
222 |
+
%extelt.offset14 = lshr i32 %202, 16, !dbg !32
|
223 |
+
%208 = trunc i32 %extelt.offset14 to i16, !dbg !32
|
224 |
+
%209 = trunc i32 %203 to i16, !dbg !32
|
225 |
+
%extelt.offset15 = lshr i32 %203, 16, !dbg !32
|
226 |
+
%210 = trunc i32 %extelt.offset15 to i16, !dbg !32
|
227 |
+
%211 = trunc i32 %204 to i16, !dbg !32
|
228 |
+
%extelt.offset16 = lshr i32 %204, 16, !dbg !32
|
229 |
+
%212 = trunc i32 %extelt.offset16 to i16, !dbg !32
|
230 |
+
%213 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %192) #6, !dbg !33
|
231 |
+
%214 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %193) #6, !dbg !33
|
232 |
+
%215 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %194) #6, !dbg !33
|
233 |
+
%216 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %195) #6, !dbg !33
|
234 |
+
%217 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %196) #6, !dbg !33
|
235 |
+
%218 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %197) #6, !dbg !33
|
236 |
+
%219 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %198) #6, !dbg !33
|
237 |
+
%220 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %199) #6, !dbg !33
|
238 |
+
%221 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %205) #6, !dbg !33
|
239 |
+
%222 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %206) #6, !dbg !33
|
240 |
+
%223 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %207) #6, !dbg !33
|
241 |
+
%224 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %208) #6, !dbg !33
|
242 |
+
%225 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %209) #6, !dbg !33
|
243 |
+
%226 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %210) #6, !dbg !33
|
244 |
+
%227 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %211) #6, !dbg !33
|
245 |
+
%228 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %212) #6, !dbg !33
|
246 |
+
br i1 %56, label %229, label %230, !dbg !34
|
247 |
+
|
248 |
+
229: ; preds = %65
|
249 |
+
tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 883, ptr nonnull @assertFunc_0, i64 1), !dbg !34
|
250 |
+
br label %230, !dbg !34
|
251 |
+
|
252 |
+
230: ; preds = %229, %65
|
253 |
+
%231 = zext nneg i32 %131 to i64, !dbg !35
|
254 |
+
%232 = zext nneg i32 %132 to i64, !dbg !35
|
255 |
+
%233 = getelementptr float, ptr addrspace(1) %63, i64 %231, !dbg !36
|
256 |
+
%234 = getelementptr float, ptr addrspace(1) %63, i64 %232, !dbg !36
|
257 |
+
%235 = getelementptr float, ptr addrspace(1) %64, i64 %231, !dbg !36
|
258 |
+
%236 = getelementptr float, ptr addrspace(1) %64, i64 %232, !dbg !36
|
259 |
+
%237 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %233, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
|
260 |
+
%238 = extractvalue { i32, i32, i32, i32 } %237, 0, !dbg !37
|
261 |
+
%239 = extractvalue { i32, i32, i32, i32 } %237, 1, !dbg !37
|
262 |
+
%240 = extractvalue { i32, i32, i32, i32 } %237, 2, !dbg !37
|
263 |
+
%241 = extractvalue { i32, i32, i32, i32 } %237, 3, !dbg !37
|
264 |
+
%242 = bitcast i32 %238 to float, !dbg !37
|
265 |
+
%243 = bitcast i32 %239 to float, !dbg !37
|
266 |
+
%244 = bitcast i32 %240 to float, !dbg !37
|
267 |
+
%245 = bitcast i32 %241 to float, !dbg !37
|
268 |
+
%246 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %234, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
|
269 |
+
%247 = extractvalue { i32, i32, i32, i32 } %246, 0, !dbg !37
|
270 |
+
%248 = extractvalue { i32, i32, i32, i32 } %246, 1, !dbg !37
|
271 |
+
%249 = extractvalue { i32, i32, i32, i32 } %246, 2, !dbg !37
|
272 |
+
%250 = extractvalue { i32, i32, i32, i32 } %246, 3, !dbg !37
|
273 |
+
%251 = bitcast i32 %247 to float, !dbg !37
|
274 |
+
%252 = bitcast i32 %248 to float, !dbg !37
|
275 |
+
%253 = bitcast i32 %249 to float, !dbg !37
|
276 |
+
%254 = bitcast i32 %250 to float, !dbg !37
|
277 |
+
%255 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %235, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
|
278 |
+
%256 = extractvalue { i32, i32, i32, i32 } %255, 0, !dbg !37
|
279 |
+
%257 = extractvalue { i32, i32, i32, i32 } %255, 1, !dbg !37
|
280 |
+
%258 = extractvalue { i32, i32, i32, i32 } %255, 2, !dbg !37
|
281 |
+
%259 = extractvalue { i32, i32, i32, i32 } %255, 3, !dbg !37
|
282 |
+
%260 = bitcast i32 %256 to float, !dbg !37
|
283 |
+
%261 = bitcast i32 %257 to float, !dbg !37
|
284 |
+
%262 = bitcast i32 %258 to float, !dbg !37
|
285 |
+
%263 = bitcast i32 %259 to float, !dbg !37
|
286 |
+
%264 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %236, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
|
287 |
+
%265 = extractvalue { i32, i32, i32, i32 } %264, 0, !dbg !37
|
288 |
+
%266 = extractvalue { i32, i32, i32, i32 } %264, 1, !dbg !37
|
289 |
+
%267 = extractvalue { i32, i32, i32, i32 } %264, 2, !dbg !37
|
290 |
+
%268 = extractvalue { i32, i32, i32, i32 } %264, 3, !dbg !37
|
291 |
+
%269 = bitcast i32 %265 to float, !dbg !37
|
292 |
+
%270 = bitcast i32 %266 to float, !dbg !37
|
293 |
+
%271 = bitcast i32 %267 to float, !dbg !37
|
294 |
+
%272 = bitcast i32 %268 to float, !dbg !37
|
295 |
+
%273 = fadd float %150, %242, !dbg !38
|
296 |
+
%274 = fadd float %151, %243, !dbg !38
|
297 |
+
%275 = fadd float %152, %244, !dbg !38
|
298 |
+
%276 = fadd float %153, %245, !dbg !38
|
299 |
+
%277 = fadd float %159, %251, !dbg !38
|
300 |
+
%278 = fadd float %160, %252, !dbg !38
|
301 |
+
%279 = fadd float %161, %253, !dbg !38
|
302 |
+
%280 = fadd float %162, %254, !dbg !38
|
303 |
+
%281 = fadd float %168, %260, !dbg !38
|
304 |
+
%282 = fadd float %169, %261, !dbg !38
|
305 |
+
%283 = fadd float %170, %262, !dbg !38
|
306 |
+
%284 = fadd float %171, %263, !dbg !38
|
307 |
+
%285 = fadd float %177, %269, !dbg !38
|
308 |
+
%286 = fadd float %178, %270, !dbg !38
|
309 |
+
%287 = fadd float %179, %271, !dbg !38
|
310 |
+
%288 = fadd float %180, %272, !dbg !38
|
311 |
+
%289 = fadd float %213, %273, !dbg !39
|
312 |
+
%290 = fadd float %214, %274, !dbg !39
|
313 |
+
%291 = fadd float %215, %275, !dbg !39
|
314 |
+
%292 = fadd float %216, %276, !dbg !39
|
315 |
+
%293 = fadd float %217, %277, !dbg !39
|
316 |
+
%294 = fadd float %218, %278, !dbg !39
|
317 |
+
%295 = fadd float %219, %279, !dbg !39
|
318 |
+
%296 = fadd float %220, %280, !dbg !39
|
319 |
+
%297 = fadd float %221, %281, !dbg !39
|
320 |
+
%298 = fadd float %222, %282, !dbg !39
|
321 |
+
%299 = fadd float %223, %283, !dbg !39
|
322 |
+
%300 = fadd float %224, %284, !dbg !39
|
323 |
+
%301 = fadd float %225, %285, !dbg !39
|
324 |
+
%302 = fadd float %226, %286, !dbg !39
|
325 |
+
%303 = fadd float %227, %287, !dbg !39
|
326 |
+
%304 = fadd float %228, %288, !dbg !39
|
327 |
+
%305 = fsub float %289, %114, !dbg !40
|
328 |
+
%306 = fsub float %290, %115, !dbg !40
|
329 |
+
%307 = fsub float %291, %116, !dbg !40
|
330 |
+
%308 = fsub float %292, %117, !dbg !40
|
331 |
+
%309 = fsub float %293, %118, !dbg !40
|
332 |
+
%310 = fsub float %294, %119, !dbg !40
|
333 |
+
%311 = fsub float %295, %120, !dbg !40
|
334 |
+
%312 = fsub float %296, %121, !dbg !40
|
335 |
+
%313 = fsub float %297, %122, !dbg !40
|
336 |
+
%314 = fsub float %298, %123, !dbg !40
|
337 |
+
%315 = fsub float %299, %124, !dbg !40
|
338 |
+
%316 = fsub float %300, %125, !dbg !40
|
339 |
+
%317 = fsub float %301, %126, !dbg !40
|
340 |
+
%318 = fsub float %302, %127, !dbg !40
|
341 |
+
%319 = fsub float %303, %128, !dbg !40
|
342 |
+
%320 = fsub float %304, %129, !dbg !40
|
343 |
+
%321 = fadd float %66, 1.000000e+00, !dbg !44
|
344 |
+
%322 = fadd float %67, 1.000000e+00, !dbg !44
|
345 |
+
%323 = fadd float %68, 1.000000e+00, !dbg !44
|
346 |
+
%324 = fadd float %69, 1.000000e+00, !dbg !44
|
347 |
+
%325 = fadd float %70, 1.000000e+00, !dbg !44
|
348 |
+
%326 = fadd float %71, 1.000000e+00, !dbg !44
|
349 |
+
%327 = fadd float %72, 1.000000e+00, !dbg !44
|
350 |
+
%328 = fadd float %73, 1.000000e+00, !dbg !44
|
351 |
+
%329 = fadd float %74, 1.000000e+00, !dbg !44
|
352 |
+
%330 = fadd float %75, 1.000000e+00, !dbg !44
|
353 |
+
%331 = fadd float %76, 1.000000e+00, !dbg !44
|
354 |
+
%332 = fadd float %77, 1.000000e+00, !dbg !44
|
355 |
+
%333 = fadd float %78, 1.000000e+00, !dbg !44
|
356 |
+
%334 = fadd float %79, 1.000000e+00, !dbg !44
|
357 |
+
%335 = fadd float %80, 1.000000e+00, !dbg !44
|
358 |
+
%336 = fadd float %81, 1.000000e+00, !dbg !44
|
359 |
+
%337 = fadd float %82, 1.000000e+00, !dbg !44
|
360 |
+
%338 = fadd float %83, 1.000000e+00, !dbg !44
|
361 |
+
%339 = fadd float %84, 1.000000e+00, !dbg !44
|
362 |
+
%340 = fadd float %85, 1.000000e+00, !dbg !44
|
363 |
+
%341 = fadd float %86, 1.000000e+00, !dbg !44
|
364 |
+
%342 = fadd float %87, 1.000000e+00, !dbg !44
|
365 |
+
%343 = fadd float %88, 1.000000e+00, !dbg !44
|
366 |
+
%344 = fadd float %89, 1.000000e+00, !dbg !44
|
367 |
+
%345 = fadd float %90, 1.000000e+00, !dbg !44
|
368 |
+
%346 = fadd float %91, 1.000000e+00, !dbg !44
|
369 |
+
%347 = fadd float %92, 1.000000e+00, !dbg !44
|
370 |
+
%348 = fadd float %93, 1.000000e+00, !dbg !44
|
371 |
+
%349 = fadd float %94, 1.000000e+00, !dbg !44
|
372 |
+
%350 = fadd float %95, 1.000000e+00, !dbg !44
|
373 |
+
%351 = fadd float %96, 1.000000e+00, !dbg !44
|
374 |
+
%352 = fadd float %97, 1.000000e+00, !dbg !44
|
375 |
+
%353 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %305, float %321) #6, !dbg !45
|
376 |
+
%354 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %306, float %322) #6, !dbg !45
|
377 |
+
%355 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %307, float %323) #6, !dbg !45
|
378 |
+
%356 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %308, float %324) #6, !dbg !45
|
379 |
+
%357 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %309, float %325) #6, !dbg !45
|
380 |
+
%358 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %310, float %326) #6, !dbg !45
|
381 |
+
%359 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %311, float %327) #6, !dbg !45
|
382 |
+
%360 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %312, float %328) #6, !dbg !45
|
383 |
+
%361 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %313, float %329) #6, !dbg !45
|
384 |
+
%362 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %314, float %330) #6, !dbg !45
|
385 |
+
%363 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %315, float %331) #6, !dbg !45
|
386 |
+
%364 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %316, float %332) #6, !dbg !45
|
387 |
+
%365 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %317, float %333) #6, !dbg !45
|
388 |
+
%366 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %318, float %334) #6, !dbg !45
|
389 |
+
%367 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %319, float %335) #6, !dbg !45
|
390 |
+
%368 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %320, float %336) #6, !dbg !45
|
391 |
+
%369 = fadd float %114, %353, !dbg !46
|
392 |
+
%370 = fadd float %115, %354, !dbg !46
|
393 |
+
%371 = fadd float %116, %355, !dbg !46
|
394 |
+
%372 = fadd float %117, %356, !dbg !46
|
395 |
+
%373 = fadd float %118, %357, !dbg !46
|
396 |
+
%374 = fadd float %119, %358, !dbg !46
|
397 |
+
%375 = fadd float %120, %359, !dbg !46
|
398 |
+
%376 = fadd float %121, %360, !dbg !46
|
399 |
+
%377 = fadd float %122, %361, !dbg !46
|
400 |
+
%378 = fadd float %123, %362, !dbg !46
|
401 |
+
%379 = fadd float %124, %363, !dbg !46
|
402 |
+
%380 = fadd float %125, %364, !dbg !46
|
403 |
+
%381 = fadd float %126, %365, !dbg !46
|
404 |
+
%382 = fadd float %127, %366, !dbg !46
|
405 |
+
%383 = fadd float %128, %367, !dbg !46
|
406 |
+
%384 = fadd float %129, %368, !dbg !46
|
407 |
+
%385 = fsub float %289, %369, !dbg !47
|
408 |
+
%386 = fsub float %290, %370, !dbg !47
|
409 |
+
%387 = fsub float %291, %371, !dbg !47
|
410 |
+
%388 = fsub float %292, %372, !dbg !47
|
411 |
+
%389 = fsub float %293, %373, !dbg !47
|
412 |
+
%390 = fsub float %294, %374, !dbg !47
|
413 |
+
%391 = fsub float %295, %375, !dbg !47
|
414 |
+
%392 = fsub float %296, %376, !dbg !47
|
415 |
+
%393 = fsub float %297, %377, !dbg !47
|
416 |
+
%394 = fsub float %298, %378, !dbg !47
|
417 |
+
%395 = fsub float %299, %379, !dbg !47
|
418 |
+
%396 = fsub float %300, %380, !dbg !47
|
419 |
+
%397 = fsub float %301, %381, !dbg !47
|
420 |
+
%398 = fsub float %302, %382, !dbg !47
|
421 |
+
%399 = fsub float %303, %383, !dbg !47
|
422 |
+
%400 = fsub float %304, %384, !dbg !47
|
423 |
+
%401 = fmul float %305, %385, !dbg !48
|
424 |
+
%402 = fmul float %306, %386, !dbg !48
|
425 |
+
%403 = fmul float %307, %387, !dbg !48
|
426 |
+
%404 = fmul float %308, %388, !dbg !48
|
427 |
+
%405 = fmul float %309, %389, !dbg !48
|
428 |
+
%406 = fmul float %310, %390, !dbg !48
|
429 |
+
%407 = fmul float %311, %391, !dbg !48
|
430 |
+
%408 = fmul float %312, %392, !dbg !48
|
431 |
+
%409 = fmul float %313, %393, !dbg !48
|
432 |
+
%410 = fmul float %314, %394, !dbg !48
|
433 |
+
%411 = fmul float %315, %395, !dbg !48
|
434 |
+
%412 = fmul float %316, %396, !dbg !48
|
435 |
+
%413 = fmul float %317, %397, !dbg !48
|
436 |
+
%414 = fmul float %318, %398, !dbg !48
|
437 |
+
%415 = fmul float %319, %399, !dbg !48
|
438 |
+
%416 = fmul float %320, %400, !dbg !48
|
439 |
+
%417 = fadd float %98, %401, !dbg !49
|
440 |
+
%418 = fadd float %99, %402, !dbg !49
|
441 |
+
%419 = fadd float %100, %403, !dbg !49
|
442 |
+
%420 = fadd float %101, %404, !dbg !49
|
443 |
+
%421 = fadd float %102, %405, !dbg !49
|
444 |
+
%422 = fadd float %103, %406, !dbg !49
|
445 |
+
%423 = fadd float %104, %407, !dbg !49
|
446 |
+
%424 = fadd float %105, %408, !dbg !49
|
447 |
+
%425 = fadd float %106, %409, !dbg !49
|
448 |
+
%426 = fadd float %107, %410, !dbg !49
|
449 |
+
%427 = fadd float %108, %411, !dbg !49
|
450 |
+
%428 = fadd float %109, %412, !dbg !49
|
451 |
+
%429 = fadd float %110, %413, !dbg !49
|
452 |
+
%430 = fadd float %111, %414, !dbg !49
|
453 |
+
%431 = fadd float %112, %415, !dbg !49
|
454 |
+
%432 = fadd float %113, %416, !dbg !49
|
455 |
+
%433 = add nuw nsw i32 %130, 64, !dbg !12
|
456 |
+
%434 = icmp ult i32 %130, 192, !dbg !12
|
457 |
+
br i1 %434, label %65, label %435, !dbg !12
|
458 |
+
|
459 |
+
435: ; preds = %230
|
460 |
+
%436 = and i32 %16, 3, !dbg !12
|
461 |
+
%437 = mul nuw nsw i32 %436, 72, !dbg !12
|
462 |
+
%438 = add nuw nsw i32 %437, %12, !dbg !12
|
463 |
+
%439 = zext nneg i32 %438 to i64, !dbg !12
|
464 |
+
%440 = getelementptr float, ptr addrspace(3) @global_smem, i64 %439, !dbg !12
|
465 |
+
%441 = insertelement <1 x float> undef, float %337, i64 0, !dbg !12
|
466 |
+
store <1 x float> %441, ptr addrspace(3) %440, align 4, !dbg !12
|
467 |
+
%442 = add nuw nsw i32 %12, 288, !dbg !12
|
468 |
+
%443 = add nuw nsw i32 %442, %437, !dbg !12
|
469 |
+
%444 = zext nneg i32 %443 to i64, !dbg !12
|
470 |
+
%445 = getelementptr float, ptr addrspace(3) @global_smem, i64 %444, !dbg !12
|
471 |
+
%446 = insertelement <1 x float> undef, float %338, i64 0, !dbg !12
|
472 |
+
store <1 x float> %446, ptr addrspace(3) %445, align 4, !dbg !12
|
473 |
+
%447 = or i32 %12, 576, !dbg !12
|
474 |
+
%448 = add nuw nsw i32 %447, %437, !dbg !12
|
475 |
+
%449 = zext nneg i32 %448 to i64, !dbg !12
|
476 |
+
%450 = getelementptr float, ptr addrspace(3) @global_smem, i64 %449, !dbg !12
|
477 |
+
%451 = insertelement <1 x float> undef, float %339, i64 0, !dbg !12
|
478 |
+
store <1 x float> %451, ptr addrspace(3) %450, align 4, !dbg !12
|
479 |
+
%452 = add nuw nsw i32 %12, 864, !dbg !12
|
480 |
+
%453 = add nuw nsw i32 %452, %437, !dbg !12
|
481 |
+
%454 = zext nneg i32 %453 to i64, !dbg !12
|
482 |
+
%455 = getelementptr float, ptr addrspace(3) @global_smem, i64 %454, !dbg !12
|
483 |
+
%456 = insertelement <1 x float> undef, float %340, i64 0, !dbg !12
|
484 |
+
store <1 x float> %456, ptr addrspace(3) %455, align 4, !dbg !12
|
485 |
+
%457 = or i32 %12, 1152, !dbg !12
|
486 |
+
%458 = add nuw nsw i32 %457, %437, !dbg !12
|
487 |
+
%459 = zext nneg i32 %458 to i64, !dbg !12
|
488 |
+
%460 = getelementptr float, ptr addrspace(3) @global_smem, i64 %459, !dbg !12
|
489 |
+
%461 = insertelement <1 x float> undef, float %341, i64 0, !dbg !12
|
490 |
+
store <1 x float> %461, ptr addrspace(3) %460, align 4, !dbg !12
|
491 |
+
%462 = add nuw nsw i32 %12, 1440, !dbg !12
|
492 |
+
%463 = add nuw nsw i32 %462, %437, !dbg !12
|
493 |
+
%464 = zext nneg i32 %463 to i64, !dbg !12
|
494 |
+
%465 = getelementptr float, ptr addrspace(3) @global_smem, i64 %464, !dbg !12
|
495 |
+
%466 = insertelement <1 x float> undef, float %342, i64 0, !dbg !12
|
496 |
+
store <1 x float> %466, ptr addrspace(3) %465, align 4, !dbg !12
|
497 |
+
%467 = or i32 %12, 1728, !dbg !12
|
498 |
+
%468 = add nuw nsw i32 %467, %437, !dbg !12
|
499 |
+
%469 = zext nneg i32 %468 to i64, !dbg !12
|
500 |
+
%470 = getelementptr float, ptr addrspace(3) @global_smem, i64 %469, !dbg !12
|
501 |
+
%471 = insertelement <1 x float> undef, float %343, i64 0, !dbg !12
|
502 |
+
store <1 x float> %471, ptr addrspace(3) %470, align 4, !dbg !12
|
503 |
+
%472 = add nuw nsw i32 %12, 2016, !dbg !12
|
504 |
+
%473 = add nuw nsw i32 %472, %437, !dbg !12
|
505 |
+
%474 = zext nneg i32 %473 to i64, !dbg !12
|
506 |
+
%475 = getelementptr float, ptr addrspace(3) @global_smem, i64 %474, !dbg !12
|
507 |
+
%476 = insertelement <1 x float> undef, float %344, i64 0, !dbg !12
|
508 |
+
store <1 x float> %476, ptr addrspace(3) %475, align 4, !dbg !12
|
509 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !12
|
510 |
+
%477 = mul nuw nsw i32 %11, 72, !dbg !12
|
511 |
+
%478 = add nuw nsw i32 %477, %14, !dbg !12
|
512 |
+
%479 = zext nneg i32 %478 to i64, !dbg !12
|
513 |
+
%480 = getelementptr float, ptr addrspace(3) @global_smem, i64 %479, !dbg !12
|
514 |
+
%481 = load float, ptr addrspace(3) %480, align 32, !dbg !12
|
515 |
+
%482 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 1, !dbg !12
|
516 |
+
%483 = load float, ptr addrspace(3) %482, align 4, !dbg !12
|
517 |
+
%484 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 2, !dbg !12
|
518 |
+
%485 = load float, ptr addrspace(3) %484, align 8, !dbg !12
|
519 |
+
%486 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 3, !dbg !12
|
520 |
+
%487 = load float, ptr addrspace(3) %486, align 4, !dbg !12
|
521 |
+
%488 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 4, !dbg !12
|
522 |
+
%489 = load float, ptr addrspace(3) %488, align 16, !dbg !12
|
523 |
+
%490 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 5, !dbg !12
|
524 |
+
%491 = load float, ptr addrspace(3) %490, align 4, !dbg !12
|
525 |
+
%492 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 6, !dbg !12
|
526 |
+
%493 = load float, ptr addrspace(3) %492, align 8, !dbg !12
|
527 |
+
%494 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 7, !dbg !12
|
528 |
+
%495 = load float, ptr addrspace(3) %494, align 4, !dbg !12
|
529 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !12
|
530 |
+
%496 = insertelement <1 x float> undef, float %345, i64 0, !dbg !12
|
531 |
+
store <1 x float> %496, ptr addrspace(3) %440, align 4, !dbg !12
|
532 |
+
%497 = insertelement <1 x float> undef, float %346, i64 0, !dbg !12
|
533 |
+
store <1 x float> %497, ptr addrspace(3) %445, align 4, !dbg !12
|
534 |
+
%498 = insertelement <1 x float> undef, float %347, i64 0, !dbg !12
|
535 |
+
store <1 x float> %498, ptr addrspace(3) %450, align 4, !dbg !12
|
536 |
+
%499 = insertelement <1 x float> undef, float %348, i64 0, !dbg !12
|
537 |
+
store <1 x float> %499, ptr addrspace(3) %455, align 4, !dbg !12
|
538 |
+
%500 = insertelement <1 x float> undef, float %349, i64 0, !dbg !12
|
539 |
+
store <1 x float> %500, ptr addrspace(3) %460, align 4, !dbg !12
|
540 |
+
%501 = insertelement <1 x float> undef, float %350, i64 0, !dbg !12
|
541 |
+
store <1 x float> %501, ptr addrspace(3) %465, align 4, !dbg !12
|
542 |
+
%502 = insertelement <1 x float> undef, float %351, i64 0, !dbg !12
|
543 |
+
store <1 x float> %502, ptr addrspace(3) %470, align 4, !dbg !12
|
544 |
+
%503 = insertelement <1 x float> undef, float %352, i64 0, !dbg !12
|
545 |
+
store <1 x float> %503, ptr addrspace(3) %475, align 4, !dbg !12
|
546 |
+
tail call void @llvm.nvvm.barrier0(), !dbg !12
|
547 |
+
%504 = load float, ptr addrspace(3) %480, align 32, !dbg !12
|
548 |
+
%505 = load float, ptr addrspace(3) %482, align 4, !dbg !12
|
549 |
+
%506 = load float, ptr addrspace(3) %484, align 8, !dbg !12
|
550 |
+
%507 = load float, ptr addrspace(3) %486, align 4, !dbg !12
|
551 |
+
%508 = load float, ptr addrspace(3) %488, align 16, !dbg !12
|
552 |
+
%509 = load float, ptr addrspace(3) %490, align 4, !dbg !12
|
553 |
+
%510 = load float, ptr addrspace(3) %492, align 8, !dbg !12
|
554 |
+
%511 = load float, ptr addrspace(3) %494, align 4, !dbg !12
|
555 |
+
%512 = fsub float %370, %369, !dbg !50
|
556 |
+
%513 = fadd float %481, %483, !dbg !54
|
557 |
+
%514 = fcmp oeq float %513, 0.000000e+00, !dbg !55
|
558 |
+
%515 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %483, float %513) #6, !dbg !56
|
559 |
+
%516 = select i1 %514, float 0.000000e+00, float %515, !dbg !57
|
560 |
+
%517 = fmul float %512, %516, !dbg !58
|
561 |
+
%518 = fadd float %369, %517, !dbg !59
|
562 |
+
%519 = fadd float %417, %418, !dbg !60
|
563 |
+
%520 = fmul float %512, %512, !dbg !61
|
564 |
+
%521 = fmul float %520, %481, !dbg !62
|
565 |
+
%522 = fmul float %521, %516, !dbg !63
|
566 |
+
%523 = fadd float %519, %522, !dbg !64
|
567 |
+
%524 = fsub float %371, %518, !dbg !50
|
568 |
+
%525 = fadd float %485, %513, !dbg !54
|
569 |
+
%526 = fcmp oeq float %525, 0.000000e+00, !dbg !55
|
570 |
+
%527 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %485, float %525) #6, !dbg !56
|
571 |
+
%528 = select i1 %526, float 0.000000e+00, float %527, !dbg !57
|
572 |
+
%529 = fmul float %528, %524, !dbg !58
|
573 |
+
%530 = fadd float %518, %529, !dbg !59
|
574 |
+
%531 = fadd float %419, %523, !dbg !60
|
575 |
+
%532 = fmul float %524, %524, !dbg !61
|
576 |
+
%533 = fmul float %513, %532, !dbg !62
|
577 |
+
%534 = fmul float %528, %533, !dbg !63
|
578 |
+
%535 = fadd float %531, %534, !dbg !64
|
579 |
+
%536 = fsub float %372, %530, !dbg !50
|
580 |
+
%537 = fadd float %487, %525, !dbg !54
|
581 |
+
%538 = fcmp oeq float %537, 0.000000e+00, !dbg !55
|
582 |
+
%539 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %487, float %537) #6, !dbg !56
|
583 |
+
%540 = select i1 %538, float 0.000000e+00, float %539, !dbg !57
|
584 |
+
%541 = fmul float %540, %536, !dbg !58
|
585 |
+
%542 = fadd float %530, %541, !dbg !59
|
586 |
+
%543 = fadd float %420, %535, !dbg !60
|
587 |
+
%544 = fmul float %536, %536, !dbg !61
|
588 |
+
%545 = fmul float %525, %544, !dbg !62
|
589 |
+
%546 = fmul float %540, %545, !dbg !63
|
590 |
+
%547 = fadd float %543, %546, !dbg !64
|
591 |
+
%548 = fsub float %373, %542, !dbg !50
|
592 |
+
%549 = fadd float %489, %537, !dbg !54
|
593 |
+
%550 = fcmp oeq float %549, 0.000000e+00, !dbg !55
|
594 |
+
%551 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %489, float %549) #6, !dbg !56
|
595 |
+
%552 = select i1 %550, float 0.000000e+00, float %551, !dbg !57
|
596 |
+
%553 = fmul float %552, %548, !dbg !58
|
597 |
+
%554 = fadd float %542, %553, !dbg !59
|
598 |
+
%555 = fadd float %421, %547, !dbg !60
|
599 |
+
%556 = fmul float %548, %548, !dbg !61
|
600 |
+
%557 = fmul float %537, %556, !dbg !62
|
601 |
+
%558 = fmul float %552, %557, !dbg !63
|
602 |
+
%559 = fadd float %555, %558, !dbg !64
|
603 |
+
%560 = fsub float %374, %554, !dbg !50
|
604 |
+
%561 = fadd float %491, %549, !dbg !54
|
605 |
+
%562 = fcmp oeq float %561, 0.000000e+00, !dbg !55
|
606 |
+
%563 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %491, float %561) #6, !dbg !56
|
607 |
+
%564 = select i1 %562, float 0.000000e+00, float %563, !dbg !57
|
608 |
+
%565 = fmul float %564, %560, !dbg !58
|
609 |
+
%566 = fadd float %554, %565, !dbg !59
|
610 |
+
%567 = fadd float %422, %559, !dbg !60
|
611 |
+
%568 = fmul float %560, %560, !dbg !61
|
612 |
+
%569 = fmul float %549, %568, !dbg !62
|
613 |
+
%570 = fmul float %564, %569, !dbg !63
|
614 |
+
%571 = fadd float %567, %570, !dbg !64
|
615 |
+
%572 = fsub float %375, %566, !dbg !50
|
616 |
+
%573 = fadd float %493, %561, !dbg !54
|
617 |
+
%574 = fcmp oeq float %573, 0.000000e+00, !dbg !55
|
618 |
+
%575 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %493, float %573) #6, !dbg !56
|
619 |
+
%576 = select i1 %574, float 0.000000e+00, float %575, !dbg !57
|
620 |
+
%577 = fmul float %576, %572, !dbg !58
|
621 |
+
%578 = fadd float %566, %577, !dbg !59
|
622 |
+
%579 = fadd float %423, %571, !dbg !60
|
623 |
+
%580 = fmul float %572, %572, !dbg !61
|
624 |
+
%581 = fmul float %561, %580, !dbg !62
|
625 |
+
%582 = fmul float %576, %581, !dbg !63
|
626 |
+
%583 = fadd float %579, %582, !dbg !64
|
627 |
+
%584 = fsub float %376, %578, !dbg !50
|
628 |
+
%585 = fadd float %495, %573, !dbg !54
|
629 |
+
%586 = fcmp oeq float %585, 0.000000e+00, !dbg !55
|
630 |
+
%587 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %495, float %585) #6, !dbg !56
|
631 |
+
%588 = select i1 %586, float 0.000000e+00, float %587, !dbg !57
|
632 |
+
%589 = fmul float %588, %584, !dbg !58
|
633 |
+
%590 = fadd float %578, %589, !dbg !59
|
634 |
+
%591 = fadd float %424, %583, !dbg !60
|
635 |
+
%592 = fmul float %584, %584, !dbg !61
|
636 |
+
%593 = fmul float %573, %592, !dbg !62
|
637 |
+
%594 = fmul float %588, %593, !dbg !63
|
638 |
+
%595 = fadd float %591, %594, !dbg !64
|
639 |
+
%596 = fsub float %378, %377, !dbg !50
|
640 |
+
%597 = fadd float %504, %505, !dbg !54
|
641 |
+
%598 = fcmp oeq float %597, 0.000000e+00, !dbg !55
|
642 |
+
%599 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %505, float %597) #6, !dbg !56
|
643 |
+
%600 = select i1 %598, float 0.000000e+00, float %599, !dbg !57
|
644 |
+
%601 = fmul float %596, %600, !dbg !58
|
645 |
+
%602 = fadd float %377, %601, !dbg !59
|
646 |
+
%603 = fadd float %425, %426, !dbg !60
|
647 |
+
%604 = fmul float %596, %596, !dbg !61
|
648 |
+
%605 = fmul float %604, %504, !dbg !62
|
649 |
+
%606 = fmul float %605, %600, !dbg !63
|
650 |
+
%607 = fadd float %603, %606, !dbg !64
|
651 |
+
%608 = fsub float %379, %602, !dbg !50
|
652 |
+
%609 = fadd float %506, %597, !dbg !54
|
653 |
+
%610 = fcmp oeq float %609, 0.000000e+00, !dbg !55
|
654 |
+
%611 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %506, float %609) #6, !dbg !56
|
655 |
+
%612 = select i1 %610, float 0.000000e+00, float %611, !dbg !57
|
656 |
+
%613 = fmul float %612, %608, !dbg !58
|
657 |
+
%614 = fadd float %602, %613, !dbg !59
|
658 |
+
%615 = fadd float %427, %607, !dbg !60
|
659 |
+
%616 = fmul float %608, %608, !dbg !61
|
660 |
+
%617 = fmul float %597, %616, !dbg !62
|
661 |
+
%618 = fmul float %612, %617, !dbg !63
|
662 |
+
%619 = fadd float %615, %618, !dbg !64
|
663 |
+
%620 = fsub float %380, %614, !dbg !50
|
664 |
+
%621 = fadd float %507, %609, !dbg !54
|
665 |
+
%622 = fcmp oeq float %621, 0.000000e+00, !dbg !55
|
666 |
+
%623 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %507, float %621) #6, !dbg !56
|
667 |
+
%624 = select i1 %622, float 0.000000e+00, float %623, !dbg !57
|
668 |
+
%625 = fmul float %624, %620, !dbg !58
|
669 |
+
%626 = fadd float %614, %625, !dbg !59
|
670 |
+
%627 = fadd float %428, %619, !dbg !60
|
671 |
+
%628 = fmul float %620, %620, !dbg !61
|
672 |
+
%629 = fmul float %609, %628, !dbg !62
|
673 |
+
%630 = fmul float %624, %629, !dbg !63
|
674 |
+
%631 = fadd float %627, %630, !dbg !64
|
675 |
+
%632 = fsub float %381, %626, !dbg !50
|
676 |
+
%633 = fadd float %508, %621, !dbg !54
|
677 |
+
%634 = fcmp oeq float %633, 0.000000e+00, !dbg !55
|
678 |
+
%635 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %508, float %633) #6, !dbg !56
|
679 |
+
%636 = select i1 %634, float 0.000000e+00, float %635, !dbg !57
|
680 |
+
%637 = fmul float %636, %632, !dbg !58
|
681 |
+
%638 = fadd float %626, %637, !dbg !59
|
682 |
+
%639 = fadd float %429, %631, !dbg !60
|
683 |
+
%640 = fmul float %632, %632, !dbg !61
|
684 |
+
%641 = fmul float %621, %640, !dbg !62
|
685 |
+
%642 = fmul float %636, %641, !dbg !63
|
686 |
+
%643 = fadd float %639, %642, !dbg !64
|
687 |
+
%644 = fsub float %382, %638, !dbg !50
|
688 |
+
%645 = fadd float %509, %633, !dbg !54
|
689 |
+
%646 = fcmp oeq float %645, 0.000000e+00, !dbg !55
|
690 |
+
%647 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %509, float %645) #6, !dbg !56
|
691 |
+
%648 = select i1 %646, float 0.000000e+00, float %647, !dbg !57
|
692 |
+
%649 = fmul float %648, %644, !dbg !58
|
693 |
+
%650 = fadd float %638, %649, !dbg !59
|
694 |
+
%651 = fadd float %430, %643, !dbg !60
|
695 |
+
%652 = fmul float %644, %644, !dbg !61
|
696 |
+
%653 = fmul float %633, %652, !dbg !62
|
697 |
+
%654 = fmul float %648, %653, !dbg !63
|
698 |
+
%655 = fadd float %651, %654, !dbg !64
|
699 |
+
%656 = fsub float %383, %650, !dbg !50
|
700 |
+
%657 = fadd float %510, %645, !dbg !54
|
701 |
+
%658 = fcmp oeq float %657, 0.000000e+00, !dbg !55
|
702 |
+
%659 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %510, float %657) #6, !dbg !56
|
703 |
+
%660 = select i1 %658, float 0.000000e+00, float %659, !dbg !57
|
704 |
+
%661 = fmul float %660, %656, !dbg !58
|
705 |
+
%662 = fadd float %650, %661, !dbg !59
|
706 |
+
%663 = fadd float %431, %655, !dbg !60
|
707 |
+
%664 = fmul float %656, %656, !dbg !61
|
708 |
+
%665 = fmul float %645, %664, !dbg !62
|
709 |
+
%666 = fmul float %660, %665, !dbg !63
|
710 |
+
%667 = fadd float %663, %666, !dbg !64
|
711 |
+
%668 = fsub float %384, %662, !dbg !50
|
712 |
+
%669 = fadd float %511, %657, !dbg !54
|
713 |
+
%670 = fcmp oeq float %669, 0.000000e+00, !dbg !55
|
714 |
+
%671 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %511, float %669) #6, !dbg !56
|
715 |
+
%672 = select i1 %670, float 0.000000e+00, float %671, !dbg !57
|
716 |
+
%673 = fmul float %672, %668, !dbg !58
|
717 |
+
%674 = fadd float %662, %673, !dbg !59
|
718 |
+
%675 = fadd float %432, %667, !dbg !60
|
719 |
+
%676 = fmul float %668, %668, !dbg !61
|
720 |
+
%677 = fmul float %657, %676, !dbg !62
|
721 |
+
%678 = fmul float %672, %677, !dbg !63
|
722 |
+
%679 = fadd float %675, %678, !dbg !64
|
723 |
+
%680 = bitcast float %590 to i32, !dbg !65
|
724 |
+
%681 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %680, i32 4, i32 31), !dbg !65
|
725 |
+
%682 = bitcast i32 %681 to float, !dbg !65
|
726 |
+
%683 = bitcast float %595 to i32, !dbg !65
|
727 |
+
%684 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %683, i32 4, i32 31), !dbg !65
|
728 |
+
%685 = bitcast i32 %684 to float, !dbg !65
|
729 |
+
%686 = bitcast float %585 to i32, !dbg !65
|
730 |
+
%687 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %686, i32 4, i32 31), !dbg !65
|
731 |
+
%688 = bitcast i32 %687 to float, !dbg !65
|
732 |
+
%689 = fsub float %682, %590, !dbg !50
|
733 |
+
%690 = fadd float %585, %688, !dbg !54
|
734 |
+
%691 = fcmp oeq float %690, 0.000000e+00, !dbg !55
|
735 |
+
%692 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %688, float %690) #6, !dbg !56
|
736 |
+
%693 = select i1 %691, float 0.000000e+00, float %692, !dbg !57
|
737 |
+
%694 = fmul float %693, %689, !dbg !58
|
738 |
+
%695 = fadd float %590, %694, !dbg !59
|
739 |
+
%696 = fadd float %595, %685, !dbg !60
|
740 |
+
%697 = fmul float %689, %689, !dbg !61
|
741 |
+
%698 = fmul float %585, %697, !dbg !62
|
742 |
+
%699 = fmul float %693, %698, !dbg !63
|
743 |
+
%700 = fadd float %696, %699, !dbg !64
|
744 |
+
%701 = bitcast float %695 to i32, !dbg !65
|
745 |
+
%702 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %701, i32 2, i32 31), !dbg !65
|
746 |
+
%703 = bitcast i32 %702 to float, !dbg !65
|
747 |
+
%704 = bitcast float %700 to i32, !dbg !65
|
748 |
+
%705 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %704, i32 2, i32 31), !dbg !65
|
749 |
+
%706 = bitcast i32 %705 to float, !dbg !65
|
750 |
+
%707 = bitcast float %690 to i32, !dbg !65
|
751 |
+
%708 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %707, i32 2, i32 31), !dbg !65
|
752 |
+
%709 = bitcast i32 %708 to float, !dbg !65
|
753 |
+
%710 = fsub float %703, %695, !dbg !50
|
754 |
+
%711 = fadd float %690, %709, !dbg !54
|
755 |
+
%712 = fcmp oeq float %711, 0.000000e+00, !dbg !55
|
756 |
+
%713 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %709, float %711) #6, !dbg !56
|
757 |
+
%714 = select i1 %712, float 0.000000e+00, float %713, !dbg !57
|
758 |
+
%715 = fmul float %714, %710, !dbg !58
|
759 |
+
%716 = fadd float %695, %715, !dbg !59
|
760 |
+
%717 = fadd float %700, %706, !dbg !60
|
761 |
+
%718 = fmul float %710, %710, !dbg !61
|
762 |
+
%719 = fmul float %690, %718, !dbg !62
|
763 |
+
%720 = fmul float %714, %719, !dbg !63
|
764 |
+
%721 = fadd float %717, %720, !dbg !64
|
765 |
+
%722 = bitcast float %716 to i32, !dbg !65
|
766 |
+
%723 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %722, i32 1, i32 31), !dbg !65
|
767 |
+
%724 = bitcast i32 %723 to float, !dbg !65
|
768 |
+
%725 = bitcast float %721 to i32, !dbg !65
|
769 |
+
%726 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %725, i32 1, i32 31), !dbg !65
|
770 |
+
%727 = bitcast i32 %726 to float, !dbg !65
|
771 |
+
%728 = bitcast float %711 to i32, !dbg !65
|
772 |
+
%729 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %728, i32 1, i32 31), !dbg !65
|
773 |
+
%730 = bitcast i32 %729 to float, !dbg !65
|
774 |
+
%731 = fsub float %724, %716, !dbg !50
|
775 |
+
%732 = fadd float %711, %730, !dbg !54
|
776 |
+
%733 = fcmp oeq float %732, 0.000000e+00, !dbg !55
|
777 |
+
%734 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %730, float %732) #6, !dbg !56
|
778 |
+
%735 = select i1 %733, float 0.000000e+00, float %734, !dbg !57
|
779 |
+
%736 = fmul float %731, %735, !dbg !58
|
780 |
+
%737 = fadd float %716, %736, !dbg !59
|
781 |
+
%738 = fadd float %721, %727, !dbg !60
|
782 |
+
%739 = fmul float %731, %731, !dbg !61
|
783 |
+
%740 = fmul float %711, %739, !dbg !62
|
784 |
+
%741 = fmul float %735, %740, !dbg !63
|
785 |
+
%742 = fadd float %738, %741, !dbg !64
|
786 |
+
%743 = bitcast float %674 to i32, !dbg !65
|
787 |
+
%744 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %743, i32 4, i32 31), !dbg !65
|
788 |
+
%745 = bitcast i32 %744 to float, !dbg !65
|
789 |
+
%746 = bitcast float %679 to i32, !dbg !65
|
790 |
+
%747 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %746, i32 4, i32 31), !dbg !65
|
791 |
+
%748 = bitcast i32 %747 to float, !dbg !65
|
792 |
+
%749 = bitcast float %669 to i32, !dbg !65
|
793 |
+
%750 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %749, i32 4, i32 31), !dbg !65
|
794 |
+
%751 = bitcast i32 %750 to float, !dbg !65
|
795 |
+
%752 = fsub float %745, %674, !dbg !50
|
796 |
+
%753 = fadd float %669, %751, !dbg !54
|
797 |
+
%754 = fcmp oeq float %753, 0.000000e+00, !dbg !55
|
798 |
+
%755 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %751, float %753) #6, !dbg !56
|
799 |
+
%756 = select i1 %754, float 0.000000e+00, float %755, !dbg !57
|
800 |
+
%757 = fmul float %752, %756, !dbg !58
|
801 |
+
%758 = fadd float %674, %757, !dbg !59
|
802 |
+
%759 = fadd float %679, %748, !dbg !60
|
803 |
+
%760 = fmul float %752, %752, !dbg !61
|
804 |
+
%761 = fmul float %669, %760, !dbg !62
|
805 |
+
%762 = fmul float %761, %756, !dbg !63
|
806 |
+
%763 = fadd float %759, %762, !dbg !64
|
807 |
+
%764 = bitcast float %758 to i32, !dbg !65
|
808 |
+
%765 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %764, i32 2, i32 31), !dbg !65
|
809 |
+
%766 = bitcast i32 %765 to float, !dbg !65
|
810 |
+
%767 = bitcast float %763 to i32, !dbg !65
|
811 |
+
%768 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %767, i32 2, i32 31), !dbg !65
|
812 |
+
%769 = bitcast i32 %768 to float, !dbg !65
|
813 |
+
%770 = bitcast float %753 to i32, !dbg !65
|
814 |
+
%771 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %770, i32 2, i32 31), !dbg !65
|
815 |
+
%772 = bitcast i32 %771 to float, !dbg !65
|
816 |
+
%773 = fsub float %766, %758, !dbg !50
|
817 |
+
%774 = fadd float %753, %772, !dbg !54
|
818 |
+
%775 = fcmp oeq float %774, 0.000000e+00, !dbg !55
|
819 |
+
%776 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %772, float %774) #6, !dbg !56
|
820 |
+
%777 = select i1 %775, float 0.000000e+00, float %776, !dbg !57
|
821 |
+
%778 = fmul float %773, %777, !dbg !58
|
822 |
+
%779 = fadd float %758, %778, !dbg !59
|
823 |
+
%780 = fadd float %763, %769, !dbg !60
|
824 |
+
%781 = fmul float %773, %773, !dbg !61
|
825 |
+
%782 = fmul float %753, %781, !dbg !62
|
826 |
+
%783 = fmul float %777, %782, !dbg !63
|
827 |
+
%784 = fadd float %780, %783, !dbg !64
|
828 |
+
%785 = bitcast float %779 to i32, !dbg !65
|
829 |
+
%786 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %785, i32 1, i32 31), !dbg !65
|
830 |
+
%787 = bitcast i32 %786 to float, !dbg !65
|
831 |
+
%788 = bitcast float %784 to i32, !dbg !65
|
832 |
+
%789 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %788, i32 1, i32 31), !dbg !65
|
833 |
+
%790 = bitcast i32 %789 to float, !dbg !65
|
834 |
+
%791 = bitcast float %774 to i32, !dbg !65
|
835 |
+
%792 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %791, i32 1, i32 31), !dbg !65
|
836 |
+
%793 = bitcast i32 %792 to float, !dbg !65
|
837 |
+
%794 = fsub float %787, %779, !dbg !50
|
838 |
+
%795 = fadd float %774, %793, !dbg !54
|
839 |
+
%796 = fcmp oeq float %795, 0.000000e+00, !dbg !55
|
840 |
+
%797 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %793, float %795) #6, !dbg !56
|
841 |
+
%798 = select i1 %796, float 0.000000e+00, float %797, !dbg !57
|
842 |
+
%799 = fmul float %794, %798, !dbg !58
|
843 |
+
%800 = fadd float %779, %799, !dbg !59
|
844 |
+
%801 = fadd float %784, %790, !dbg !60
|
845 |
+
%802 = fmul float %794, %794, !dbg !61
|
846 |
+
%803 = fmul float %774, %802, !dbg !62
|
847 |
+
%804 = fmul float %798, %803, !dbg !63
|
848 |
+
%805 = fadd float %801, %804, !dbg !64
|
849 |
+
%806 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
850 |
+
%807 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
851 |
+
%808 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
852 |
+
%809 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
853 |
+
%810 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
854 |
+
%811 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
855 |
+
%812 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
856 |
+
%813 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
|
857 |
+
%814 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
858 |
+
%815 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
859 |
+
%816 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
860 |
+
%817 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
861 |
+
%818 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
862 |
+
%819 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
863 |
+
%820 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
864 |
+
%821 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
|
865 |
+
%822 = fadd float %806, 0x3EE4F8B580000000, !dbg !68
|
866 |
+
%823 = fadd float %814, 0x3EE4F8B580000000, !dbg !68
|
867 |
+
br label %824, !dbg !69
|
868 |
+
|
869 |
+
824: ; preds = %435, %__nv_rsqrtf.exit40
|
870 |
+
%825 = phi i32 [ 0, %435 ], [ %1134, %__nv_rsqrtf.exit40 ]
|
871 |
+
%826 = or i32 %825, %14, !dbg !70
|
872 |
+
%827 = or i32 %825, %15, !dbg !70
|
873 |
+
%828 = add i32 %826, %47, !dbg !71
|
874 |
+
%829 = add i32 %827, %47, !dbg !71
|
875 |
+
%830 = add i32 %826, %48, !dbg !71
|
876 |
+
%831 = add i32 %827, %48, !dbg !71
|
877 |
+
%832 = sext i32 %828 to i64, !dbg !72
|
878 |
+
%833 = getelementptr float, ptr addrspace(1) %2, i64 %832, !dbg !72
|
879 |
+
%834 = sext i32 %829 to i64, !dbg !72
|
880 |
+
%835 = getelementptr float, ptr addrspace(1) %2, i64 %834, !dbg !72
|
881 |
+
%836 = sext i32 %830 to i64, !dbg !72
|
882 |
+
%837 = getelementptr float, ptr addrspace(1) %2, i64 %836, !dbg !72
|
883 |
+
%838 = sext i32 %831 to i64, !dbg !72
|
884 |
+
%839 = getelementptr float, ptr addrspace(1) %2, i64 %838, !dbg !72
|
885 |
+
%840 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %833, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
|
886 |
+
%841 = extractvalue { i32, i32, i32, i32 } %840, 0, !dbg !73
|
887 |
+
%842 = extractvalue { i32, i32, i32, i32 } %840, 1, !dbg !73
|
888 |
+
%843 = extractvalue { i32, i32, i32, i32 } %840, 2, !dbg !73
|
889 |
+
%844 = extractvalue { i32, i32, i32, i32 } %840, 3, !dbg !73
|
890 |
+
%845 = bitcast i32 %841 to float, !dbg !73
|
891 |
+
%846 = bitcast i32 %842 to float, !dbg !73
|
892 |
+
%847 = bitcast i32 %843 to float, !dbg !73
|
893 |
+
%848 = bitcast i32 %844 to float, !dbg !73
|
894 |
+
%849 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %835, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
|
895 |
+
%850 = extractvalue { i32, i32, i32, i32 } %849, 0, !dbg !73
|
896 |
+
%851 = extractvalue { i32, i32, i32, i32 } %849, 1, !dbg !73
|
897 |
+
%852 = extractvalue { i32, i32, i32, i32 } %849, 2, !dbg !73
|
898 |
+
%853 = extractvalue { i32, i32, i32, i32 } %849, 3, !dbg !73
|
899 |
+
%854 = bitcast i32 %850 to float, !dbg !73
|
900 |
+
%855 = bitcast i32 %851 to float, !dbg !73
|
901 |
+
%856 = bitcast i32 %852 to float, !dbg !73
|
902 |
+
%857 = bitcast i32 %853 to float, !dbg !73
|
903 |
+
%858 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %837, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
|
904 |
+
%859 = extractvalue { i32, i32, i32, i32 } %858, 0, !dbg !73
|
905 |
+
%860 = extractvalue { i32, i32, i32, i32 } %858, 1, !dbg !73
|
906 |
+
%861 = extractvalue { i32, i32, i32, i32 } %858, 2, !dbg !73
|
907 |
+
%862 = extractvalue { i32, i32, i32, i32 } %858, 3, !dbg !73
|
908 |
+
%863 = bitcast i32 %859 to float, !dbg !73
|
909 |
+
%864 = bitcast i32 %860 to float, !dbg !73
|
910 |
+
%865 = bitcast i32 %861 to float, !dbg !73
|
911 |
+
%866 = bitcast i32 %862 to float, !dbg !73
|
912 |
+
%867 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %839, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
|
913 |
+
%868 = extractvalue { i32, i32, i32, i32 } %867, 0, !dbg !73
|
914 |
+
%869 = extractvalue { i32, i32, i32, i32 } %867, 1, !dbg !73
|
915 |
+
%870 = extractvalue { i32, i32, i32, i32 } %867, 2, !dbg !73
|
916 |
+
%871 = extractvalue { i32, i32, i32, i32 } %867, 3, !dbg !73
|
917 |
+
%872 = bitcast i32 %868 to float, !dbg !73
|
918 |
+
%873 = bitcast i32 %869 to float, !dbg !73
|
919 |
+
%874 = bitcast i32 %870 to float, !dbg !73
|
920 |
+
%875 = bitcast i32 %871 to float, !dbg !73
|
921 |
+
%876 = add i32 %826, %49, !dbg !74
|
922 |
+
%877 = add i32 %826, %50, !dbg !74
|
923 |
+
%878 = sext i32 %876 to i64, !dbg !75
|
924 |
+
%879 = getelementptr i16, ptr addrspace(1) %3, i64 %878, !dbg !75
|
925 |
+
%880 = sext i32 %877 to i64, !dbg !75
|
926 |
+
%881 = getelementptr i16, ptr addrspace(1) %3, i64 %880, !dbg !75
|
927 |
+
%882 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %879, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !76
|
928 |
+
%883 = extractvalue { i32, i32, i32, i32 } %882, 0, !dbg !76
|
929 |
+
%884 = extractvalue { i32, i32, i32, i32 } %882, 1, !dbg !76
|
930 |
+
%885 = extractvalue { i32, i32, i32, i32 } %882, 2, !dbg !76
|
931 |
+
%886 = extractvalue { i32, i32, i32, i32 } %882, 3, !dbg !76
|
932 |
+
%887 = trunc i32 %883 to i16, !dbg !76
|
933 |
+
%extelt.offset = lshr i32 %883, 16, !dbg !76
|
934 |
+
%888 = trunc i32 %extelt.offset to i16, !dbg !76
|
935 |
+
%889 = trunc i32 %884 to i16, !dbg !76
|
936 |
+
%extelt.offset2 = lshr i32 %884, 16, !dbg !76
|
937 |
+
%890 = trunc i32 %extelt.offset2 to i16, !dbg !76
|
938 |
+
%891 = trunc i32 %885 to i16, !dbg !76
|
939 |
+
%extelt.offset3 = lshr i32 %885, 16, !dbg !76
|
940 |
+
%892 = trunc i32 %extelt.offset3 to i16, !dbg !76
|
941 |
+
%893 = trunc i32 %886 to i16, !dbg !76
|
942 |
+
%extelt.offset4 = lshr i32 %886, 16, !dbg !76
|
943 |
+
%894 = trunc i32 %extelt.offset4 to i16, !dbg !76
|
944 |
+
%895 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %881, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !76
|
945 |
+
%896 = extractvalue { i32, i32, i32, i32 } %895, 0, !dbg !76
|
946 |
+
%897 = extractvalue { i32, i32, i32, i32 } %895, 1, !dbg !76
|
947 |
+
%898 = extractvalue { i32, i32, i32, i32 } %895, 2, !dbg !76
|
948 |
+
%899 = extractvalue { i32, i32, i32, i32 } %895, 3, !dbg !76
|
949 |
+
%900 = trunc i32 %896 to i16, !dbg !76
|
950 |
+
%extelt.offset5 = lshr i32 %896, 16, !dbg !76
|
951 |
+
%901 = trunc i32 %extelt.offset5 to i16, !dbg !76
|
952 |
+
%902 = trunc i32 %897 to i16, !dbg !76
|
953 |
+
%extelt.offset6 = lshr i32 %897, 16, !dbg !76
|
954 |
+
%903 = trunc i32 %extelt.offset6 to i16, !dbg !76
|
955 |
+
%904 = trunc i32 %898 to i16, !dbg !76
|
956 |
+
%extelt.offset7 = lshr i32 %898, 16, !dbg !76
|
957 |
+
%905 = trunc i32 %extelt.offset7 to i16, !dbg !76
|
958 |
+
%906 = trunc i32 %899 to i16, !dbg !76
|
959 |
+
%extelt.offset8 = lshr i32 %899, 16, !dbg !76
|
960 |
+
%907 = trunc i32 %extelt.offset8 to i16, !dbg !76
|
961 |
+
%908 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %887) #6, !dbg !77
|
962 |
+
%909 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %888) #6, !dbg !77
|
963 |
+
%910 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %889) #6, !dbg !77
|
964 |
+
%911 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %890) #6, !dbg !77
|
965 |
+
%912 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %891) #6, !dbg !77
|
966 |
+
%913 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %892) #6, !dbg !77
|
967 |
+
%914 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %893) #6, !dbg !77
|
968 |
+
%915 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %894) #6, !dbg !77
|
969 |
+
%916 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %900) #6, !dbg !77
|
970 |
+
%917 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %901) #6, !dbg !77
|
971 |
+
%918 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %902) #6, !dbg !77
|
972 |
+
%919 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %903) #6, !dbg !77
|
973 |
+
%920 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %904) #6, !dbg !77
|
974 |
+
%921 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %905) #6, !dbg !77
|
975 |
+
%922 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %906) #6, !dbg !77
|
976 |
+
%923 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %907) #6, !dbg !77
|
977 |
+
%924 = zext nneg i32 %826 to i64, !dbg !78
|
978 |
+
%925 = getelementptr float, ptr addrspace(1) %4, i64 %924, !dbg !78
|
979 |
+
%926 = zext nneg i32 %827 to i64, !dbg !78
|
980 |
+
%927 = getelementptr float, ptr addrspace(1) %4, i64 %926, !dbg !78
|
981 |
+
%928 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %925, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !79
|
982 |
+
%929 = extractvalue { i32, i32, i32, i32 } %928, 0, !dbg !79
|
983 |
+
%930 = extractvalue { i32, i32, i32, i32 } %928, 1, !dbg !79
|
984 |
+
%931 = extractvalue { i32, i32, i32, i32 } %928, 2, !dbg !79
|
985 |
+
%932 = extractvalue { i32, i32, i32, i32 } %928, 3, !dbg !79
|
986 |
+
%933 = bitcast i32 %929 to float, !dbg !79
|
987 |
+
%934 = bitcast i32 %930 to float, !dbg !79
|
988 |
+
%935 = bitcast i32 %931 to float, !dbg !79
|
989 |
+
%936 = bitcast i32 %932 to float, !dbg !79
|
990 |
+
%937 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %927, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !79
|
991 |
+
%938 = extractvalue { i32, i32, i32, i32 } %937, 0, !dbg !79
|
992 |
+
%939 = extractvalue { i32, i32, i32, i32 } %937, 1, !dbg !79
|
993 |
+
%940 = extractvalue { i32, i32, i32, i32 } %937, 2, !dbg !79
|
994 |
+
%941 = extractvalue { i32, i32, i32, i32 } %937, 3, !dbg !79
|
995 |
+
%942 = bitcast i32 %938 to float, !dbg !79
|
996 |
+
%943 = bitcast i32 %939 to float, !dbg !79
|
997 |
+
%944 = bitcast i32 %940 to float, !dbg !79
|
998 |
+
%945 = bitcast i32 %941 to float, !dbg !79
|
999 |
+
br i1 %56, label %946, label %947, !dbg !80
|
1000 |
+
|
1001 |
+
946: ; preds = %824
|
1002 |
+
tail call void @__assertfail(ptr nonnull @assertMessage_1, ptr nonnull @assertFile_1, i32 883, ptr nonnull @assertFunc_1, i64 1), !dbg !80
|
1003 |
+
br label %947, !dbg !80
|
1004 |
+
|
1005 |
+
947: ; preds = %946, %824
|
1006 |
+
%948 = getelementptr float, ptr addrspace(1) %63, i64 %924, !dbg !81
|
1007 |
+
%949 = getelementptr float, ptr addrspace(1) %63, i64 %926, !dbg !81
|
1008 |
+
%950 = getelementptr float, ptr addrspace(1) %64, i64 %924, !dbg !81
|
1009 |
+
%951 = getelementptr float, ptr addrspace(1) %64, i64 %926, !dbg !81
|
1010 |
+
%952 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %948, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
|
1011 |
+
%953 = extractvalue { i32, i32, i32, i32 } %952, 0, !dbg !82
|
1012 |
+
%954 = extractvalue { i32, i32, i32, i32 } %952, 1, !dbg !82
|
1013 |
+
%955 = extractvalue { i32, i32, i32, i32 } %952, 2, !dbg !82
|
1014 |
+
%956 = extractvalue { i32, i32, i32, i32 } %952, 3, !dbg !82
|
1015 |
+
%957 = bitcast i32 %953 to float, !dbg !82
|
1016 |
+
%958 = bitcast i32 %954 to float, !dbg !82
|
1017 |
+
%959 = bitcast i32 %955 to float, !dbg !82
|
1018 |
+
%960 = bitcast i32 %956 to float, !dbg !82
|
1019 |
+
%961 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %949, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
|
1020 |
+
%962 = extractvalue { i32, i32, i32, i32 } %961, 0, !dbg !82
|
1021 |
+
%963 = extractvalue { i32, i32, i32, i32 } %961, 1, !dbg !82
|
1022 |
+
%964 = extractvalue { i32, i32, i32, i32 } %961, 2, !dbg !82
|
1023 |
+
%965 = extractvalue { i32, i32, i32, i32 } %961, 3, !dbg !82
|
1024 |
+
%966 = bitcast i32 %962 to float, !dbg !82
|
1025 |
+
%967 = bitcast i32 %963 to float, !dbg !82
|
1026 |
+
%968 = bitcast i32 %964 to float, !dbg !82
|
1027 |
+
%969 = bitcast i32 %965 to float, !dbg !82
|
1028 |
+
%970 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %950, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
|
1029 |
+
%971 = extractvalue { i32, i32, i32, i32 } %970, 0, !dbg !82
|
1030 |
+
%972 = extractvalue { i32, i32, i32, i32 } %970, 1, !dbg !82
|
1031 |
+
%973 = extractvalue { i32, i32, i32, i32 } %970, 2, !dbg !82
|
1032 |
+
%974 = extractvalue { i32, i32, i32, i32 } %970, 3, !dbg !82
|
1033 |
+
%975 = bitcast i32 %971 to float, !dbg !82
|
1034 |
+
%976 = bitcast i32 %972 to float, !dbg !82
|
1035 |
+
%977 = bitcast i32 %973 to float, !dbg !82
|
1036 |
+
%978 = bitcast i32 %974 to float, !dbg !82
|
1037 |
+
%979 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %951, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
|
1038 |
+
%980 = extractvalue { i32, i32, i32, i32 } %979, 0, !dbg !82
|
1039 |
+
%981 = extractvalue { i32, i32, i32, i32 } %979, 1, !dbg !82
|
1040 |
+
%982 = extractvalue { i32, i32, i32, i32 } %979, 2, !dbg !82
|
1041 |
+
%983 = extractvalue { i32, i32, i32, i32 } %979, 3, !dbg !82
|
1042 |
+
%984 = bitcast i32 %980 to float, !dbg !82
|
1043 |
+
%985 = bitcast i32 %981 to float, !dbg !82
|
1044 |
+
%986 = bitcast i32 %982 to float, !dbg !82
|
1045 |
+
%987 = bitcast i32 %983 to float, !dbg !82
|
1046 |
+
%988 = fadd float %845, %957, !dbg !83
|
1047 |
+
%989 = fadd float %846, %958, !dbg !83
|
1048 |
+
%990 = fadd float %847, %959, !dbg !83
|
1049 |
+
%991 = fadd float %848, %960, !dbg !83
|
1050 |
+
%992 = fadd float %854, %966, !dbg !83
|
1051 |
+
%993 = fadd float %855, %967, !dbg !83
|
1052 |
+
%994 = fadd float %856, %968, !dbg !83
|
1053 |
+
%995 = fadd float %857, %969, !dbg !83
|
1054 |
+
%996 = fadd float %863, %975, !dbg !83
|
1055 |
+
%997 = fadd float %864, %976, !dbg !83
|
1056 |
+
%998 = fadd float %865, %977, !dbg !83
|
1057 |
+
%999 = fadd float %866, %978, !dbg !83
|
1058 |
+
%1000 = fadd float %872, %984, !dbg !83
|
1059 |
+
%1001 = fadd float %873, %985, !dbg !83
|
1060 |
+
%1002 = fadd float %874, %986, !dbg !83
|
1061 |
+
%1003 = fadd float %875, %987, !dbg !83
|
1062 |
+
%1004 = fadd float %908, %988, !dbg !84
|
1063 |
+
%1005 = fadd float %909, %989, !dbg !84
|
1064 |
+
%1006 = fadd float %910, %990, !dbg !84
|
1065 |
+
%1007 = fadd float %911, %991, !dbg !84
|
1066 |
+
%1008 = fadd float %912, %992, !dbg !84
|
1067 |
+
%1009 = fadd float %913, %993, !dbg !84
|
1068 |
+
%1010 = fadd float %914, %994, !dbg !84
|
1069 |
+
%1011 = fadd float %915, %995, !dbg !84
|
1070 |
+
%1012 = fadd float %916, %996, !dbg !84
|
1071 |
+
%1013 = fadd float %917, %997, !dbg !84
|
1072 |
+
%1014 = fadd float %918, %998, !dbg !84
|
1073 |
+
%1015 = fadd float %919, %999, !dbg !84
|
1074 |
+
%1016 = fadd float %920, %1000, !dbg !84
|
1075 |
+
%1017 = fadd float %921, %1001, !dbg !84
|
1076 |
+
%1018 = fadd float %922, %1002, !dbg !84
|
1077 |
+
%1019 = fadd float %923, %1003, !dbg !84
|
1078 |
+
%1020 = fsub float %1004, %737, !dbg !85
|
1079 |
+
%1021 = fsub float %1005, %737, !dbg !85
|
1080 |
+
%1022 = fsub float %1006, %737, !dbg !85
|
1081 |
+
%1023 = fsub float %1007, %737, !dbg !85
|
1082 |
+
%1024 = fsub float %1008, %737, !dbg !85
|
1083 |
+
%1025 = fsub float %1009, %737, !dbg !85
|
1084 |
+
%1026 = fsub float %1010, %737, !dbg !85
|
1085 |
+
%1027 = fsub float %1011, %737, !dbg !85
|
1086 |
+
%1028 = fsub float %1012, %800, !dbg !85
|
1087 |
+
%1029 = fsub float %1013, %800, !dbg !85
|
1088 |
+
%1030 = fsub float %1014, %800, !dbg !85
|
1089 |
+
%1031 = fsub float %1015, %800, !dbg !85
|
1090 |
+
%1032 = fsub float %1016, %800, !dbg !85
|
1091 |
+
%1033 = fsub float %1017, %800, !dbg !85
|
1092 |
+
%1034 = fsub float %1018, %800, !dbg !85
|
1093 |
+
%1035 = fsub float %1019, %800, !dbg !85
|
1094 |
+
%1036 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1095 |
+
%.not.i = icmp eq i32 %1036, 0, !dbg !86
|
1096 |
+
br i1 %.not.i, label %1039, label %1037, !dbg !86
|
1097 |
+
|
1098 |
+
1037: ; preds = %947
|
1099 |
+
%1038 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %822), !dbg !86
|
1100 |
+
br label %__nv_rsqrtf.exit, !dbg !86
|
1101 |
+
|
1102 |
+
1039: ; preds = %947
|
1103 |
+
%1040 = tail call float @llvm.nvvm.rsqrt.approx.f(float %822), !dbg !86
|
1104 |
+
br label %__nv_rsqrtf.exit, !dbg !86
|
1105 |
+
|
1106 |
+
__nv_rsqrtf.exit: ; preds = %1037, %1039
|
1107 |
+
%.0.i = phi float [ %1038, %1037 ], [ %1040, %1039 ], !dbg !86
|
1108 |
+
%1041 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1109 |
+
%1042 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1110 |
+
%1043 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1111 |
+
%1044 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1112 |
+
%1045 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1113 |
+
%1046 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1114 |
+
%1047 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1115 |
+
%1048 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1116 |
+
%.not.i38 = icmp eq i32 %1048, 0, !dbg !86
|
1117 |
+
br i1 %.not.i38, label %1051, label %1049, !dbg !86
|
1118 |
+
|
1119 |
+
1049: ; preds = %__nv_rsqrtf.exit
|
1120 |
+
%1050 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %823), !dbg !86
|
1121 |
+
br label %__nv_rsqrtf.exit40, !dbg !86
|
1122 |
+
|
1123 |
+
1051: ; preds = %__nv_rsqrtf.exit
|
1124 |
+
%1052 = tail call float @llvm.nvvm.rsqrt.approx.f(float %823), !dbg !86
|
1125 |
+
br label %__nv_rsqrtf.exit40, !dbg !86
|
1126 |
+
|
1127 |
+
__nv_rsqrtf.exit40: ; preds = %1049, %1051
|
1128 |
+
%.0.i39 = phi float [ %1050, %1049 ], [ %1052, %1051 ], !dbg !86
|
1129 |
+
%1053 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1130 |
+
%1054 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1131 |
+
%1055 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1132 |
+
%1056 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1133 |
+
%1057 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1134 |
+
%1058 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1135 |
+
%1059 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
|
1136 |
+
%1060 = fmul float %1020, %.0.i, !dbg !87
|
1137 |
+
%1061 = fmul float %1021, %.0.i, !dbg !87
|
1138 |
+
%1062 = fmul float %1022, %.0.i, !dbg !87
|
1139 |
+
%1063 = fmul float %1023, %.0.i, !dbg !87
|
1140 |
+
%1064 = fmul float %1024, %.0.i, !dbg !87
|
1141 |
+
%1065 = fmul float %1025, %.0.i, !dbg !87
|
1142 |
+
%1066 = fmul float %1026, %.0.i, !dbg !87
|
1143 |
+
%1067 = fmul float %1027, %.0.i, !dbg !87
|
1144 |
+
%1068 = fmul float %1028, %.0.i39, !dbg !87
|
1145 |
+
%1069 = fmul float %1029, %.0.i39, !dbg !87
|
1146 |
+
%1070 = fmul float %1030, %.0.i39, !dbg !87
|
1147 |
+
%1071 = fmul float %1031, %.0.i39, !dbg !87
|
1148 |
+
%1072 = fmul float %1032, %.0.i39, !dbg !87
|
1149 |
+
%1073 = fmul float %1033, %.0.i39, !dbg !87
|
1150 |
+
%1074 = fmul float %1034, %.0.i39, !dbg !87
|
1151 |
+
%1075 = fmul float %1035, %.0.i39, !dbg !87
|
1152 |
+
%1076 = fmul float %1060, %933, !dbg !88
|
1153 |
+
%1077 = fmul float %1061, %934, !dbg !88
|
1154 |
+
%1078 = fmul float %1062, %935, !dbg !88
|
1155 |
+
%1079 = fmul float %1063, %936, !dbg !88
|
1156 |
+
%1080 = fmul float %1064, %942, !dbg !88
|
1157 |
+
%1081 = fmul float %1065, %943, !dbg !88
|
1158 |
+
%1082 = fmul float %1066, %944, !dbg !88
|
1159 |
+
%1083 = fmul float %1067, %945, !dbg !88
|
1160 |
+
%1084 = fmul float %1068, %933, !dbg !88
|
1161 |
+
%1085 = fmul float %1069, %934, !dbg !88
|
1162 |
+
%1086 = fmul float %1070, %935, !dbg !88
|
1163 |
+
%1087 = fmul float %1071, %936, !dbg !88
|
1164 |
+
%1088 = fmul float %1072, %942, !dbg !88
|
1165 |
+
%1089 = fmul float %1073, %943, !dbg !88
|
1166 |
+
%1090 = fmul float %1074, %944, !dbg !88
|
1167 |
+
%1091 = fmul float %1075, %945, !dbg !88
|
1168 |
+
%1092 = getelementptr i16, ptr addrspace(1) %5, i64 %878, !dbg !89
|
1169 |
+
%1093 = getelementptr i16, ptr addrspace(1) %5, i64 %880, !dbg !89
|
1170 |
+
%1094 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1076) #6, !dbg !90
|
1171 |
+
%1095 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1077) #6, !dbg !90
|
1172 |
+
%1096 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1078) #6, !dbg !90
|
1173 |
+
%1097 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1079) #6, !dbg !90
|
1174 |
+
%1098 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1080) #6, !dbg !90
|
1175 |
+
%1099 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1081) #6, !dbg !90
|
1176 |
+
%1100 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1082) #6, !dbg !90
|
1177 |
+
%1101 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1083) #6, !dbg !90
|
1178 |
+
%1102 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1084) #6, !dbg !90
|
1179 |
+
%1103 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1085) #6, !dbg !90
|
1180 |
+
%1104 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1086) #6, !dbg !90
|
1181 |
+
%1105 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1087) #6, !dbg !90
|
1182 |
+
%1106 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1088) #6, !dbg !90
|
1183 |
+
%1107 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1089) #6, !dbg !90
|
1184 |
+
%1108 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1090) #6, !dbg !90
|
1185 |
+
%1109 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1091) #6, !dbg !90
|
1186 |
+
%1110 = insertelement <2 x i16> undef, i16 %1094, i64 0, !dbg !90
|
1187 |
+
%1111 = insertelement <2 x i16> %1110, i16 %1095, i64 1, !dbg !90
|
1188 |
+
%1112 = bitcast <2 x i16> %1111 to i32, !dbg !90
|
1189 |
+
%1113 = insertelement <2 x i16> undef, i16 %1096, i64 0, !dbg !90
|
1190 |
+
%1114 = insertelement <2 x i16> %1113, i16 %1097, i64 1, !dbg !90
|
1191 |
+
%1115 = bitcast <2 x i16> %1114 to i32, !dbg !90
|
1192 |
+
%1116 = insertelement <2 x i16> undef, i16 %1098, i64 0, !dbg !90
|
1193 |
+
%1117 = insertelement <2 x i16> %1116, i16 %1099, i64 1, !dbg !90
|
1194 |
+
%1118 = bitcast <2 x i16> %1117 to i32, !dbg !90
|
1195 |
+
%1119 = insertelement <2 x i16> undef, i16 %1100, i64 0, !dbg !90
|
1196 |
+
%1120 = insertelement <2 x i16> %1119, i16 %1101, i64 1, !dbg !90
|
1197 |
+
%1121 = bitcast <2 x i16> %1120 to i32, !dbg !90
|
1198 |
+
tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %1112, i32 %1115, i32 %1118, i32 %1121, ptr addrspace(1) %1092, i1 true) #6, !dbg !90
|
1199 |
+
%1122 = insertelement <2 x i16> undef, i16 %1102, i64 0, !dbg !90
|
1200 |
+
%1123 = insertelement <2 x i16> %1122, i16 %1103, i64 1, !dbg !90
|
1201 |
+
%1124 = bitcast <2 x i16> %1123 to i32, !dbg !90
|
1202 |
+
%1125 = insertelement <2 x i16> undef, i16 %1104, i64 0, !dbg !90
|
1203 |
+
%1126 = insertelement <2 x i16> %1125, i16 %1105, i64 1, !dbg !90
|
1204 |
+
%1127 = bitcast <2 x i16> %1126 to i32, !dbg !90
|
1205 |
+
%1128 = insertelement <2 x i16> undef, i16 %1106, i64 0, !dbg !90
|
1206 |
+
%1129 = insertelement <2 x i16> %1128, i16 %1107, i64 1, !dbg !90
|
1207 |
+
%1130 = bitcast <2 x i16> %1129 to i32, !dbg !90
|
1208 |
+
%1131 = insertelement <2 x i16> undef, i16 %1108, i64 0, !dbg !90
|
1209 |
+
%1132 = insertelement <2 x i16> %1131, i16 %1109, i64 1, !dbg !90
|
1210 |
+
%1133 = bitcast <2 x i16> %1132 to i32, !dbg !90
|
1211 |
+
tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %1124, i32 %1127, i32 %1130, i32 %1133, ptr addrspace(1) %1093, i1 true) #6, !dbg !90
|
1212 |
+
%1134 = add nuw nsw i32 %825, 64, !dbg !69
|
1213 |
+
%1135 = icmp ult i32 %825, 192, !dbg !69
|
1214 |
+
br i1 %1135, label %824, label %1136, !dbg !69
|
1215 |
+
|
1216 |
+
1136: ; preds = %__nv_rsqrtf.exit40
|
1217 |
+
ret void, !dbg !91
|
1218 |
+
}
|
1219 |
+
|
1220 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
|
1221 |
+
declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
|
1222 |
+
|
1223 |
+
; Function Attrs: convergent nocallback nounwind
|
1224 |
+
declare void @llvm.nvvm.barrier0() #1
|
1225 |
+
|
1226 |
+
; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
|
1227 |
+
declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #2
|
1228 |
+
|
1229 |
+
; Function Attrs: alwaysinline nounwind
|
1230 |
+
define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
|
1231 |
+
%1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
|
1232 |
+
%.not = icmp eq i32 %1, 0
|
1233 |
+
br i1 %.not, label %4, label %2
|
1234 |
+
|
1235 |
+
2: ; preds = %0
|
1236 |
+
%3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
|
1237 |
+
br label %6
|
1238 |
+
|
1239 |
+
4: ; preds = %0
|
1240 |
+
%5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
|
1241 |
+
br label %6
|
1242 |
+
|
1243 |
+
6: ; preds = %4, %2
|
1244 |
+
%.0 = phi float [ %3, %2 ], [ %5, %4 ]
|
1245 |
+
ret float %.0
|
1246 |
+
}
|
1247 |
+
|
1248 |
+
declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
|
1249 |
+
|
1250 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
1251 |
+
declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
|
1252 |
+
|
1253 |
+
; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
|
1254 |
+
declare float @llvm.nvvm.rsqrt.approx.f(float) #5
|
1255 |
+
|
1256 |
+
attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
|
1257 |
+
attributes #1 = { convergent nocallback nounwind }
|
1258 |
+
attributes #2 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
|
1259 |
+
attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
1260 |
+
attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
1261 |
+
attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
|
1262 |
+
attributes #6 = { nounwind }
|
1263 |
+
|
1264 |
+
!llvm.module.flags = !{!0, !1}
|
1265 |
+
!llvm.dbg.cu = !{!2}
|
1266 |
+
!nvvm.annotations = !{!4, !5, !5, !4}
|
1267 |
+
!llvm.ident = !{!6}
|
1268 |
+
|
1269 |
+
!0 = !{i32 2, !"Debug Info Version", i32 3}
|
1270 |
+
!1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
|
1271 |
+
!2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
|
1272 |
+
!3 = !DIFile(filename: "ccig6fki6p4lxrdmgg6eudahiexcvueeol2p4qp532pvve2y463y.py", directory: "/tmp/torchinductor_root/ci")
|
1273 |
+
!4 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"kernel", i32 1}
|
1274 |
+
!5 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"maxntidx", i32 256}
|
1275 |
+
!6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
|
1276 |
+
!7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5d6de7de", linkageName: "triton__0d1d2d3d4d5d6de7de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
|
1277 |
+
!8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
|
1278 |
+
!9 = !{}
|
1279 |
+
!10 = !DILocation(line: 22, column: 44, scope: !7)
|
1280 |
+
!11 = !DILocation(line: 24, column: 33, scope: !7)
|
1281 |
+
!12 = !DILocation(line: 31, column: 36, scope: !7)
|
1282 |
+
!13 = !DILocation(line: 21, column: 28, scope: !7)
|
1283 |
+
!14 = !DILocation(line: 21, column: 33, scope: !7)
|
1284 |
+
!15 = !DILocation(line: 22, column: 23, scope: !7)
|
1285 |
+
!16 = !DILocation(line: 26, column: 30, scope: !7)
|
1286 |
+
!17 = !DILocation(line: 26, column: 35, scope: !7)
|
1287 |
+
!18 = !DILocation(line: 27, column: 18, scope: !7)
|
1288 |
+
!19 = !DILocation(line: 35, column: 44, scope: !7)
|
1289 |
+
!20 = !DILocation(line: 36, column: 44, scope: !7)
|
1290 |
+
!21 = !DILocation(line: 37, column: 22, scope: !7)
|
1291 |
+
!22 = !DILocation(line: 38, column: 22, scope: !7)
|
1292 |
+
!23 = !DILocation(line: 39, column: 36, scope: !7)
|
1293 |
+
!24 = !DILocation(line: 40, column: 40, scope: !7)
|
1294 |
+
!25 = !DILocation(line: 41, column: 44, scope: !7)
|
1295 |
+
!26 = !DILocation(line: 32, column: 27, scope: !7)
|
1296 |
+
!27 = !DILocation(line: 35, column: 40, scope: !7)
|
1297 |
+
!28 = !DILocation(line: 35, column: 34, scope: !7)
|
1298 |
+
!29 = !DILocation(line: 35, column: 50, scope: !7)
|
1299 |
+
!30 = !DILocation(line: 36, column: 40, scope: !7)
|
1300 |
+
!31 = !DILocation(line: 36, column: 34, scope: !7)
|
1301 |
+
!32 = !DILocation(line: 36, column: 50, scope: !7)
|
1302 |
+
!33 = !DILocation(line: 36, column: 101, scope: !7)
|
1303 |
+
!34 = !DILocation(line: 40, column: 55, scope: !7)
|
1304 |
+
!35 = !DILocation(line: 41, column: 40, scope: !7)
|
1305 |
+
!36 = !DILocation(line: 41, column: 34, scope: !7)
|
1306 |
+
!37 = !DILocation(line: 41, column: 52, scope: !7)
|
1307 |
+
!38 = !DILocation(line: 42, column: 22, scope: !7)
|
1308 |
+
!39 = !DILocation(line: 44, column: 22, scope: !7)
|
1309 |
+
!40 = !DILocation(line: 96, column: 20, scope: !41, inlinedAt: !43)
|
1310 |
+
!41 = distinct !DILexicalBlockFile(scope: !7, file: !42, discriminator: 0)
|
1311 |
+
!42 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
|
1312 |
+
!43 = !DILocation(line: 47, column: 41, scope: !41)
|
1313 |
+
!44 = !DILocation(line: 97, column: 26, scope: !41, inlinedAt: !43)
|
1314 |
+
!45 = !DILocation(line: 98, column: 30, scope: !41, inlinedAt: !43)
|
1315 |
+
!46 = !DILocation(line: 98, column: 22, scope: !41, inlinedAt: !43)
|
1316 |
+
!47 = !DILocation(line: 101, column: 30, scope: !41, inlinedAt: !43)
|
1317 |
+
!48 = !DILocation(line: 101, column: 22, scope: !41, inlinedAt: !43)
|
1318 |
+
!49 = !DILocation(line: 50, column: 50, scope: !7)
|
1319 |
+
!50 = !DILocation(line: 108, column: 21, scope: !51, inlinedAt: !52)
|
1320 |
+
!51 = distinct !DILexicalBlockFile(scope: !41, file: !42, discriminator: 0)
|
1321 |
+
!52 = !DILocation(line: 120, column: 46, scope: !51, inlinedAt: !53)
|
1322 |
+
!53 = !DILocation(line: 53, column: 44, scope: !51)
|
1323 |
+
!54 = !DILocation(line: 109, column: 28, scope: !51, inlinedAt: !52)
|
1324 |
+
!55 = !DILocation(line: 110, column: 39, scope: !51, inlinedAt: !52)
|
1325 |
+
!56 = !DILocation(line: 110, column: 60, scope: !51, inlinedAt: !52)
|
1326 |
+
!57 = !DILocation(line: 110, column: 49, scope: !51, inlinedAt: !52)
|
1327 |
+
!58 = !DILocation(line: 112, column: 25, scope: !51, inlinedAt: !52)
|
1328 |
+
!59 = !DILocation(line: 112, column: 17, scope: !51, inlinedAt: !52)
|
1329 |
+
!60 = !DILocation(line: 113, column: 15, scope: !51, inlinedAt: !52)
|
1330 |
+
!61 = !DILocation(line: 113, column: 30, scope: !51, inlinedAt: !52)
|
1331 |
+
!62 = !DILocation(line: 113, column: 38, scope: !51, inlinedAt: !52)
|
1332 |
+
!63 = !DILocation(line: 113, column: 49, scope: !51, inlinedAt: !52)
|
1333 |
+
!64 = !DILocation(line: 113, column: 22, scope: !51, inlinedAt: !52)
|
1334 |
+
!65 = !DILocation(line: 120, column: 46, scope: !41, inlinedAt: !66)
|
1335 |
+
!66 = !DILocation(line: 53, column: 44, scope: !41)
|
1336 |
+
!67 = !DILocation(line: 75, column: 24, scope: !7)
|
1337 |
+
!68 = !DILocation(line: 77, column: 24, scope: !7)
|
1338 |
+
!69 = !DILocation(line: 58, column: 36, scope: !7)
|
1339 |
+
!70 = !DILocation(line: 59, column: 27, scope: !7)
|
1340 |
+
!71 = !DILocation(line: 62, column: 41, scope: !7)
|
1341 |
+
!72 = !DILocation(line: 62, column: 35, scope: !7)
|
1342 |
+
!73 = !DILocation(line: 62, column: 51, scope: !7)
|
1343 |
+
!74 = !DILocation(line: 63, column: 41, scope: !7)
|
1344 |
+
!75 = !DILocation(line: 63, column: 35, scope: !7)
|
1345 |
+
!76 = !DILocation(line: 63, column: 51, scope: !7)
|
1346 |
+
!77 = !DILocation(line: 63, column: 103, scope: !7)
|
1347 |
+
!78 = !DILocation(line: 64, column: 35, scope: !7)
|
1348 |
+
!79 = !DILocation(line: 64, column: 40, scope: !7)
|
1349 |
+
!80 = !DILocation(line: 68, column: 57, scope: !7)
|
1350 |
+
!81 = !DILocation(line: 69, column: 35, scope: !7)
|
1351 |
+
!82 = !DILocation(line: 69, column: 54, scope: !7)
|
1352 |
+
!83 = !DILocation(line: 70, column: 24, scope: !7)
|
1353 |
+
!84 = !DILocation(line: 72, column: 24, scope: !7)
|
1354 |
+
!85 = !DILocation(line: 73, column: 24, scope: !7)
|
1355 |
+
!86 = !DILocation(line: 78, column: 30, scope: !7)
|
1356 |
+
!87 = !DILocation(line: 79, column: 24, scope: !7)
|
1357 |
+
!88 = !DILocation(line: 80, column: 24, scope: !7)
|
1358 |
+
!89 = !DILocation(line: 82, column: 29, scope: !7)
|
1359 |
+
!90 = !DILocation(line: 82, column: 52, scope: !7)
|
1360 |
+
!91 = !DILocation(line: 58, column: 4, scope: !7)
|
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ptx
ADDED
@@ -0,0 +1,2004 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
//
|
2 |
+
// Generated by LLVM NVPTX Back-End
|
3 |
+
//
|
4 |
+
|
5 |
+
.version 8.2
|
6 |
+
.target sm_89
|
7 |
+
.address_size 64
|
8 |
+
|
9 |
+
// .globl triton__0d1d2d3d4d5d6de7de
|
10 |
+
.extern .func __assertfail
|
11 |
+
(
|
12 |
+
.param .b64 __assertfail_param_0,
|
13 |
+
.param .b64 __assertfail_param_1,
|
14 |
+
.param .b32 __assertfail_param_2,
|
15 |
+
.param .b64 __assertfail_param_3,
|
16 |
+
.param .b64 __assertfail_param_4
|
17 |
+
)
|
18 |
+
;
|
19 |
+
.global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
20 |
+
.global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
21 |
+
.global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 54, 32, 60, 32, 53, 48, 50, 53, 55};
|
22 |
+
.global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
|
23 |
+
.global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
|
24 |
+
.global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
|
25 |
+
.extern .shared .align 1 .b8 global_smem[];
|
26 |
+
.global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
|
27 |
+
|
28 |
+
.visible .entry triton__0d1d2d3d4d5d6de7de(
|
29 |
+
.param .u64 triton__0d1d2d3d4d5d6de7de_param_0,
|
30 |
+
.param .u64 triton__0d1d2d3d4d5d6de7de_param_1,
|
31 |
+
.param .u64 triton__0d1d2d3d4d5d6de7de_param_2,
|
32 |
+
.param .u64 triton__0d1d2d3d4d5d6de7de_param_3,
|
33 |
+
.param .u64 triton__0d1d2d3d4d5d6de7de_param_4,
|
34 |
+
.param .u64 triton__0d1d2d3d4d5d6de7de_param_5,
|
35 |
+
.param .u32 triton__0d1d2d3d4d5d6de7de_param_6,
|
36 |
+
.param .u32 triton__0d1d2d3d4d5d6de7de_param_7
|
37 |
+
)
|
38 |
+
.maxntid 256, 1, 1
|
39 |
+
{
|
40 |
+
.reg .pred %p<157>;
|
41 |
+
.reg .b16 %rs<49>;
|
42 |
+
.reg .b32 %r<474>;
|
43 |
+
.reg .f32 %f<678>;
|
44 |
+
.reg .b64 %rd<118>;
|
45 |
+
.loc 1 18 0
|
46 |
+
$L__func_begin0:
|
47 |
+
.loc 1 18 0
|
48 |
+
|
49 |
+
ld.param.u64 %rd16, [triton__0d1d2d3d4d5d6de7de_param_5];
|
50 |
+
ld.param.u64 %rd15, [triton__0d1d2d3d4d5d6de7de_param_4];
|
51 |
+
ld.param.u64 %rd14, [triton__0d1d2d3d4d5d6de7de_param_3];
|
52 |
+
ld.param.u64 %rd52, [triton__0d1d2d3d4d5d6de7de_param_0];
|
53 |
+
ld.param.u64 %rd53, [triton__0d1d2d3d4d5d6de7de_param_1];
|
54 |
+
$L__tmp0:
|
55 |
+
.loc 1 22 44
|
56 |
+
mov.u32 %r12, %tid.x;
|
57 |
+
ld.param.u64 %rd54, [triton__0d1d2d3d4d5d6de7de_param_2];
|
58 |
+
bfe.u32 %r1, %r12, 3, 5;
|
59 |
+
and.b32 %r2, %r12, 63;
|
60 |
+
.loc 1 24 33
|
61 |
+
shl.b32 %r13, %r12, 3;
|
62 |
+
and.b32 %r3, %r13, 56;
|
63 |
+
.loc 1 31 36
|
64 |
+
shr.u32 %r4, %r12, 6;
|
65 |
+
.loc 1 21 28
|
66 |
+
mov.u32 %r10, %ctaid.x;
|
67 |
+
.loc 1 21 33
|
68 |
+
shl.b32 %r14, %r10, 6;
|
69 |
+
.loc 1 22 23
|
70 |
+
or.b32 %r15, %r14, %r1;
|
71 |
+
or.b32 %r16, %r15, 32;
|
72 |
+
or.b32 %r17, %r14, %r2;
|
73 |
+
.loc 1 26 30
|
74 |
+
mul.wide.s32 %rd55, %r15, 8;
|
75 |
+
add.s64 %rd18, %rd52, %rd55;
|
76 |
+
add.s64 %rd34, %rd18, 256;
|
77 |
+
mul.wide.s32 %rd56, %r17, 8;
|
78 |
+
add.s64 %rd50, %rd52, %rd56;
|
79 |
+
mov.pred %p1, -1;
|
80 |
+
.loc 1 26 35
|
81 |
+
mov.u64 %rd17, 0x0;
|
82 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd17 }, [ %rd18 + 0 ];
|
83 |
+
mov.u64 %rd19, 0x0;
|
84 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd19 }, [ %rd18 + 0 ];
|
85 |
+
mov.u64 %rd21, 0x0;
|
86 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd21 }, [ %rd18 + 0 ];
|
87 |
+
mov.u64 %rd23, 0x0;
|
88 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd23 }, [ %rd18 + 0 ];
|
89 |
+
mov.u64 %rd25, 0x0;
|
90 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd25 }, [ %rd18 + 0 ];
|
91 |
+
mov.u64 %rd27, 0x0;
|
92 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd27 }, [ %rd18 + 0 ];
|
93 |
+
mov.u64 %rd29, 0x0;
|
94 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd29 }, [ %rd18 + 0 ];
|
95 |
+
mov.u64 %rd31, 0x0;
|
96 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd31 }, [ %rd18 + 0 ];
|
97 |
+
mov.u64 %rd33, 0x0;
|
98 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd33 }, [ %rd34 + 0 ];
|
99 |
+
mov.u64 %rd35, 0x0;
|
100 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd35 }, [ %rd34 + 0 ];
|
101 |
+
mov.u64 %rd37, 0x0;
|
102 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd37 }, [ %rd34 + 0 ];
|
103 |
+
mov.u64 %rd39, 0x0;
|
104 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd39 }, [ %rd34 + 0 ];
|
105 |
+
mov.u64 %rd41, 0x0;
|
106 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd41 }, [ %rd34 + 0 ];
|
107 |
+
mov.u64 %rd43, 0x0;
|
108 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd43 }, [ %rd34 + 0 ];
|
109 |
+
mov.u64 %rd45, 0x0;
|
110 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd45 }, [ %rd34 + 0 ];
|
111 |
+
mov.u64 %rd47, 0x0;
|
112 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd47 }, [ %rd34 + 0 ];
|
113 |
+
mov.u64 %rd49, 0x0;
|
114 |
+
@%p1 ld.global.L1::evict_last.b64 { %rd49 }, [ %rd50 + 0 ];
|
115 |
+
.loc 1 27 18
|
116 |
+
bfe.s32 %r18, %r10, 25, 1;
|
117 |
+
shr.u32 %r19, %r18, 23;
|
118 |
+
add.s32 %r20, %r15, %r19;
|
119 |
+
and.b32 %r21, %r20, 16776704;
|
120 |
+
sub.s32 %r22, %r15, %r21;
|
121 |
+
add.s32 %r23, %r16, %r19;
|
122 |
+
and.b32 %r24, %r23, 16776704;
|
123 |
+
sub.s32 %r25, %r16, %r24;
|
124 |
+
.loc 1 35 44
|
125 |
+
shl.b32 %r26, %r22, 8;
|
126 |
+
shl.b32 %r27, %r25, 8;
|
127 |
+
.loc 1 37 22
|
128 |
+
add.s64 %rd57, %rd49, 50257;
|
129 |
+
.loc 1 38 22
|
130 |
+
setp.lt.s64 %p18, %rd17, 0;
|
131 |
+
setp.lt.s64 %p19, %rd33, 0;
|
132 |
+
setp.lt.s64 %p20, %rd49, 0;
|
133 |
+
.loc 1 39 36
|
134 |
+
selp.b64 %rd1, %rd57, %rd49, %p20;
|
135 |
+
.loc 1 41 44
|
136 |
+
shl.b64 %rd58, %rd17, 8;
|
137 |
+
add.s64 %rd59, %rd58, 12865792;
|
138 |
+
selp.b64 %rd60, %rd59, %rd58, %p18;
|
139 |
+
shl.b64 %rd61, %rd33, 8;
|
140 |
+
add.s64 %rd62, %rd61, 12865792;
|
141 |
+
selp.b64 %rd63, %rd62, %rd61, %p19;
|
142 |
+
.loc 1 31 36
|
143 |
+
and.b32 %r28, %r12, 7;
|
144 |
+
mul.wide.u32 %rd2, %r28, 32;
|
145 |
+
shl.b64 %rd64, %rd63, 2;
|
146 |
+
or.b64 %rd65, %rd2, %rd64;
|
147 |
+
add.s64 %rd3, %rd53, %rd65;
|
148 |
+
shl.b64 %rd66, %rd60, 2;
|
149 |
+
or.b64 %rd67, %rd2, %rd66;
|
150 |
+
add.s64 %rd4, %rd53, %rd67;
|
151 |
+
or.b32 %r29, %r27, %r3;
|
152 |
+
mul.wide.s32 %rd68, %r29, 4;
|
153 |
+
add.s64 %rd5, %rd54, %rd68;
|
154 |
+
or.b32 %r30, %r26, %r3;
|
155 |
+
mul.wide.s32 %rd69, %r30, 4;
|
156 |
+
add.s64 %rd6, %rd54, %rd69;
|
157 |
+
shl.b32 %r31, %r10, 14;
|
158 |
+
shl.b32 %r32, %r1, 8;
|
159 |
+
or.b32 %r33, %r31, %r32;
|
160 |
+
or.b32 %r5, %r33, %r3;
|
161 |
+
mov.f32 %f614, 0f00000000;
|
162 |
+
mov.u64 %rd116, 0;
|
163 |
+
mov.b32 %r472, -64;
|
164 |
+
mov.f32 %f615, %f614;
|
165 |
+
mov.f32 %f616, %f614;
|
166 |
+
mov.f32 %f617, %f614;
|
167 |
+
mov.f32 %f618, %f614;
|
168 |
+
mov.f32 %f619, %f614;
|
169 |
+
mov.f32 %f620, %f614;
|
170 |
+
mov.f32 %f621, %f614;
|
171 |
+
mov.f32 %f622, %f614;
|
172 |
+
mov.f32 %f623, %f614;
|
173 |
+
mov.f32 %f624, %f614;
|
174 |
+
mov.f32 %f625, %f614;
|
175 |
+
mov.f32 %f626, %f614;
|
176 |
+
mov.f32 %f627, %f614;
|
177 |
+
mov.f32 %f628, %f614;
|
178 |
+
mov.f32 %f629, %f614;
|
179 |
+
mov.f32 %f630, %f614;
|
180 |
+
mov.f32 %f631, %f614;
|
181 |
+
mov.f32 %f632, %f614;
|
182 |
+
mov.f32 %f633, %f614;
|
183 |
+
mov.f32 %f634, %f614;
|
184 |
+
mov.f32 %f635, %f614;
|
185 |
+
mov.f32 %f636, %f614;
|
186 |
+
mov.f32 %f637, %f614;
|
187 |
+
mov.f32 %f638, %f614;
|
188 |
+
mov.f32 %f639, %f614;
|
189 |
+
mov.f32 %f640, %f614;
|
190 |
+
mov.f32 %f641, %f614;
|
191 |
+
mov.f32 %f642, %f614;
|
192 |
+
mov.f32 %f643, %f614;
|
193 |
+
mov.f32 %f644, %f614;
|
194 |
+
mov.f32 %f645, %f614;
|
195 |
+
mov.f32 %f646, %f614;
|
196 |
+
mov.f32 %f647, %f614;
|
197 |
+
mov.f32 %f648, %f614;
|
198 |
+
mov.f32 %f649, %f614;
|
199 |
+
mov.f32 %f650, %f614;
|
200 |
+
mov.f32 %f651, %f614;
|
201 |
+
mov.f32 %f652, %f614;
|
202 |
+
mov.f32 %f653, %f614;
|
203 |
+
mov.f32 %f654, %f614;
|
204 |
+
mov.f32 %f655, %f614;
|
205 |
+
mov.f32 %f656, %f614;
|
206 |
+
mov.f32 %f657, %f614;
|
207 |
+
mov.f32 %f658, %f614;
|
208 |
+
mov.f32 %f659, %f614;
|
209 |
+
mov.f32 %f660, %f614;
|
210 |
+
mov.f32 %f661, %f614;
|
211 |
+
mov.f32 %f662, %f614;
|
212 |
+
mov.f32 %f663, %f614;
|
213 |
+
mov.f32 %f664, %f614;
|
214 |
+
mov.f32 %f665, %f614;
|
215 |
+
mov.f32 %f666, %f614;
|
216 |
+
mov.f32 %f667, %f614;
|
217 |
+
mov.f32 %f668, %f614;
|
218 |
+
mov.f32 %f669, %f614;
|
219 |
+
mov.f32 %f670, %f614;
|
220 |
+
mov.f32 %f671, %f614;
|
221 |
+
mov.f32 %f672, %f614;
|
222 |
+
mov.f32 %f673, %f614;
|
223 |
+
mov.f32 %f674, %f614;
|
224 |
+
mov.f32 %f675, %f614;
|
225 |
+
mov.f32 %f676, %f614;
|
226 |
+
mov.f32 %f677, %f614;
|
227 |
+
bra.uni $L__BB0_1;
|
228 |
+
$L__BB0_3:
|
229 |
+
.loc 1 41 40
|
230 |
+
add.s64 %rd85, %rd4, %rd116;
|
231 |
+
.loc 1 41 34
|
232 |
+
add.s64 %rd86, %rd85, 16;
|
233 |
+
add.s64 %rd87, %rd3, %rd116;
|
234 |
+
.loc 1 41 52
|
235 |
+
add.s64 %rd88, %rd87, 16;
|
236 |
+
mov.u32 %r102, 0x0;
|
237 |
+
mov.u32 %r103, 0x0;
|
238 |
+
mov.u32 %r104, 0x0;
|
239 |
+
mov.u32 %r105, 0x0;
|
240 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r102, %r103, %r104, %r105 }, [ %rd85 + 0 ];
|
241 |
+
@!%p1 mov.u32 %r102, %r411;
|
242 |
+
@!%p1 mov.u32 %r103, %r411;
|
243 |
+
@!%p1 mov.u32 %r104, %r411;
|
244 |
+
@!%p1 mov.u32 %r105, %r411;
|
245 |
+
mov.b32 %f206, %r102;
|
246 |
+
mov.b32 %f207, %r103;
|
247 |
+
mov.b32 %f208, %r104;
|
248 |
+
mov.b32 %f209, %r105;
|
249 |
+
mov.u32 %r110, 0x0;
|
250 |
+
mov.u32 %r111, 0x0;
|
251 |
+
mov.u32 %r112, 0x0;
|
252 |
+
mov.u32 %r113, 0x0;
|
253 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r110, %r111, %r112, %r113 }, [ %rd86 + 0 ];
|
254 |
+
@!%p1 mov.u32 %r110, %r411;
|
255 |
+
@!%p1 mov.u32 %r111, %r411;
|
256 |
+
@!%p1 mov.u32 %r112, %r411;
|
257 |
+
@!%p1 mov.u32 %r113, %r411;
|
258 |
+
mov.b32 %f210, %r110;
|
259 |
+
mov.b32 %f211, %r111;
|
260 |
+
mov.b32 %f212, %r112;
|
261 |
+
mov.b32 %f213, %r113;
|
262 |
+
mov.u32 %r118, 0x0;
|
263 |
+
mov.u32 %r119, 0x0;
|
264 |
+
mov.u32 %r120, 0x0;
|
265 |
+
mov.u32 %r121, 0x0;
|
266 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r118, %r119, %r120, %r121 }, [ %rd87 + 0 ];
|
267 |
+
@!%p1 mov.u32 %r118, %r411;
|
268 |
+
@!%p1 mov.u32 %r119, %r411;
|
269 |
+
@!%p1 mov.u32 %r120, %r411;
|
270 |
+
@!%p1 mov.u32 %r121, %r411;
|
271 |
+
mov.b32 %f214, %r118;
|
272 |
+
mov.b32 %f215, %r119;
|
273 |
+
mov.b32 %f216, %r120;
|
274 |
+
mov.b32 %f217, %r121;
|
275 |
+
mov.u32 %r126, 0x0;
|
276 |
+
mov.u32 %r127, 0x0;
|
277 |
+
mov.u32 %r128, 0x0;
|
278 |
+
mov.u32 %r129, 0x0;
|
279 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r126, %r127, %r128, %r129 }, [ %rd88 + 0 ];
|
280 |
+
@!%p1 mov.u32 %r126, %r411;
|
281 |
+
@!%p1 mov.u32 %r127, %r411;
|
282 |
+
@!%p1 mov.u32 %r128, %r411;
|
283 |
+
@!%p1 mov.u32 %r129, %r411;
|
284 |
+
mov.b32 %f218, %r126;
|
285 |
+
mov.b32 %f219, %r127;
|
286 |
+
mov.b32 %f220, %r128;
|
287 |
+
mov.b32 %f221, %r129;
|
288 |
+
.loc 1 42 22
|
289 |
+
add.f32 %f222, %f65, %f206;
|
290 |
+
add.f32 %f223, %f66, %f207;
|
291 |
+
add.f32 %f224, %f67, %f208;
|
292 |
+
add.f32 %f225, %f68, %f209;
|
293 |
+
add.f32 %f226, %f69, %f210;
|
294 |
+
add.f32 %f227, %f70, %f211;
|
295 |
+
add.f32 %f228, %f71, %f212;
|
296 |
+
add.f32 %f229, %f72, %f213;
|
297 |
+
add.f32 %f230, %f73, %f214;
|
298 |
+
add.f32 %f231, %f74, %f215;
|
299 |
+
add.f32 %f232, %f75, %f216;
|
300 |
+
add.f32 %f233, %f76, %f217;
|
301 |
+
add.f32 %f234, %f77, %f218;
|
302 |
+
add.f32 %f235, %f78, %f219;
|
303 |
+
add.f32 %f236, %f79, %f220;
|
304 |
+
add.f32 %f237, %f80, %f221;
|
305 |
+
.loc 1 44 22
|
306 |
+
add.f32 %f238, %f81, %f222;
|
307 |
+
add.f32 %f239, %f82, %f223;
|
308 |
+
add.f32 %f240, %f83, %f224;
|
309 |
+
add.f32 %f241, %f84, %f225;
|
310 |
+
add.f32 %f242, %f85, %f226;
|
311 |
+
add.f32 %f243, %f86, %f227;
|
312 |
+
add.f32 %f244, %f87, %f228;
|
313 |
+
add.f32 %f245, %f88, %f229;
|
314 |
+
add.f32 %f246, %f89, %f230;
|
315 |
+
add.f32 %f247, %f90, %f231;
|
316 |
+
add.f32 %f248, %f91, %f232;
|
317 |
+
add.f32 %f249, %f92, %f233;
|
318 |
+
add.f32 %f250, %f93, %f234;
|
319 |
+
add.f32 %f251, %f94, %f235;
|
320 |
+
add.f32 %f252, %f95, %f236;
|
321 |
+
add.f32 %f253, %f96, %f237;
|
322 |
+
$L__tmp1:
|
323 |
+
.loc 2 96 20
|
324 |
+
sub.f32 %f254, %f238, %f662;
|
325 |
+
sub.f32 %f255, %f239, %f663;
|
326 |
+
sub.f32 %f256, %f240, %f664;
|
327 |
+
sub.f32 %f257, %f241, %f665;
|
328 |
+
sub.f32 %f258, %f242, %f666;
|
329 |
+
sub.f32 %f259, %f243, %f667;
|
330 |
+
sub.f32 %f260, %f244, %f668;
|
331 |
+
sub.f32 %f261, %f245, %f669;
|
332 |
+
sub.f32 %f262, %f246, %f670;
|
333 |
+
sub.f32 %f263, %f247, %f671;
|
334 |
+
sub.f32 %f264, %f248, %f672;
|
335 |
+
sub.f32 %f265, %f249, %f673;
|
336 |
+
sub.f32 %f266, %f250, %f674;
|
337 |
+
sub.f32 %f267, %f251, %f675;
|
338 |
+
sub.f32 %f268, %f252, %f676;
|
339 |
+
sub.f32 %f269, %f253, %f677;
|
340 |
+
.loc 2 97 26
|
341 |
+
add.f32 %f614, %f614, 0f3F800000;
|
342 |
+
add.f32 %f615, %f615, 0f3F800000;
|
343 |
+
add.f32 %f616, %f616, 0f3F800000;
|
344 |
+
add.f32 %f617, %f617, 0f3F800000;
|
345 |
+
add.f32 %f618, %f618, 0f3F800000;
|
346 |
+
add.f32 %f619, %f619, 0f3F800000;
|
347 |
+
add.f32 %f620, %f620, 0f3F800000;
|
348 |
+
add.f32 %f621, %f621, 0f3F800000;
|
349 |
+
add.f32 %f622, %f622, 0f3F800000;
|
350 |
+
add.f32 %f623, %f623, 0f3F800000;
|
351 |
+
add.f32 %f624, %f624, 0f3F800000;
|
352 |
+
add.f32 %f625, %f625, 0f3F800000;
|
353 |
+
add.f32 %f626, %f626, 0f3F800000;
|
354 |
+
add.f32 %f627, %f627, 0f3F800000;
|
355 |
+
add.f32 %f628, %f628, 0f3F800000;
|
356 |
+
add.f32 %f629, %f629, 0f3F800000;
|
357 |
+
add.f32 %f630, %f630, 0f3F800000;
|
358 |
+
add.f32 %f631, %f631, 0f3F800000;
|
359 |
+
add.f32 %f632, %f632, 0f3F800000;
|
360 |
+
add.f32 %f633, %f633, 0f3F800000;
|
361 |
+
add.f32 %f634, %f634, 0f3F800000;
|
362 |
+
add.f32 %f635, %f635, 0f3F800000;
|
363 |
+
add.f32 %f636, %f636, 0f3F800000;
|
364 |
+
add.f32 %f637, %f637, 0f3F800000;
|
365 |
+
add.f32 %f638, %f638, 0f3F800000;
|
366 |
+
add.f32 %f639, %f639, 0f3F800000;
|
367 |
+
add.f32 %f640, %f640, 0f3F800000;
|
368 |
+
add.f32 %f641, %f641, 0f3F800000;
|
369 |
+
add.f32 %f642, %f642, 0f3F800000;
|
370 |
+
add.f32 %f643, %f643, 0f3F800000;
|
371 |
+
add.f32 %f644, %f644, 0f3F800000;
|
372 |
+
add.f32 %f645, %f645, 0f3F800000;
|
373 |
+
.loc 2 98 30
|
374 |
+
mov.b32 %r135, %f254;
|
375 |
+
mov.b32 %r136, %f614;
|
376 |
+
div.full.f32 %r134, %r135, %r136;
|
377 |
+
mov.b32 %f270, %r134;
|
378 |
+
mov.b32 %r138, %f255;
|
379 |
+
mov.b32 %r139, %f615;
|
380 |
+
div.full.f32 %r137, %r138, %r139;
|
381 |
+
mov.b32 %f271, %r137;
|
382 |
+
mov.b32 %r141, %f256;
|
383 |
+
mov.b32 %r142, %f616;
|
384 |
+
div.full.f32 %r140, %r141, %r142;
|
385 |
+
mov.b32 %f272, %r140;
|
386 |
+
mov.b32 %r144, %f257;
|
387 |
+
mov.b32 %r145, %f617;
|
388 |
+
div.full.f32 %r143, %r144, %r145;
|
389 |
+
mov.b32 %f273, %r143;
|
390 |
+
mov.b32 %r147, %f258;
|
391 |
+
mov.b32 %r148, %f618;
|
392 |
+
div.full.f32 %r146, %r147, %r148;
|
393 |
+
mov.b32 %f274, %r146;
|
394 |
+
mov.b32 %r150, %f259;
|
395 |
+
mov.b32 %r151, %f619;
|
396 |
+
div.full.f32 %r149, %r150, %r151;
|
397 |
+
mov.b32 %f275, %r149;
|
398 |
+
mov.b32 %r153, %f260;
|
399 |
+
mov.b32 %r154, %f620;
|
400 |
+
div.full.f32 %r152, %r153, %r154;
|
401 |
+
mov.b32 %f276, %r152;
|
402 |
+
mov.b32 %r156, %f261;
|
403 |
+
mov.b32 %r157, %f621;
|
404 |
+
div.full.f32 %r155, %r156, %r157;
|
405 |
+
mov.b32 %f277, %r155;
|
406 |
+
mov.b32 %r159, %f262;
|
407 |
+
mov.b32 %r160, %f622;
|
408 |
+
div.full.f32 %r158, %r159, %r160;
|
409 |
+
mov.b32 %f278, %r158;
|
410 |
+
mov.b32 %r162, %f263;
|
411 |
+
mov.b32 %r163, %f623;
|
412 |
+
div.full.f32 %r161, %r162, %r163;
|
413 |
+
mov.b32 %f279, %r161;
|
414 |
+
mov.b32 %r165, %f264;
|
415 |
+
mov.b32 %r166, %f624;
|
416 |
+
div.full.f32 %r164, %r165, %r166;
|
417 |
+
mov.b32 %f280, %r164;
|
418 |
+
mov.b32 %r168, %f265;
|
419 |
+
mov.b32 %r169, %f625;
|
420 |
+
div.full.f32 %r167, %r168, %r169;
|
421 |
+
mov.b32 %f281, %r167;
|
422 |
+
mov.b32 %r171, %f266;
|
423 |
+
mov.b32 %r172, %f626;
|
424 |
+
div.full.f32 %r170, %r171, %r172;
|
425 |
+
mov.b32 %f282, %r170;
|
426 |
+
mov.b32 %r174, %f267;
|
427 |
+
mov.b32 %r175, %f627;
|
428 |
+
div.full.f32 %r173, %r174, %r175;
|
429 |
+
mov.b32 %f283, %r173;
|
430 |
+
mov.b32 %r177, %f268;
|
431 |
+
mov.b32 %r178, %f628;
|
432 |
+
div.full.f32 %r176, %r177, %r178;
|
433 |
+
mov.b32 %f284, %r176;
|
434 |
+
mov.b32 %r180, %f269;
|
435 |
+
mov.b32 %r181, %f629;
|
436 |
+
div.full.f32 %r179, %r180, %r181;
|
437 |
+
mov.b32 %f285, %r179;
|
438 |
+
.loc 2 98 22
|
439 |
+
add.f32 %f662, %f662, %f270;
|
440 |
+
add.f32 %f663, %f663, %f271;
|
441 |
+
add.f32 %f664, %f664, %f272;
|
442 |
+
add.f32 %f665, %f665, %f273;
|
443 |
+
add.f32 %f666, %f666, %f274;
|
444 |
+
add.f32 %f667, %f667, %f275;
|
445 |
+
add.f32 %f668, %f668, %f276;
|
446 |
+
add.f32 %f669, %f669, %f277;
|
447 |
+
add.f32 %f670, %f670, %f278;
|
448 |
+
add.f32 %f671, %f671, %f279;
|
449 |
+
add.f32 %f672, %f672, %f280;
|
450 |
+
add.f32 %f673, %f673, %f281;
|
451 |
+
add.f32 %f674, %f674, %f282;
|
452 |
+
add.f32 %f675, %f675, %f283;
|
453 |
+
add.f32 %f676, %f676, %f284;
|
454 |
+
add.f32 %f677, %f677, %f285;
|
455 |
+
.loc 2 101 30
|
456 |
+
sub.f32 %f286, %f238, %f662;
|
457 |
+
sub.f32 %f287, %f239, %f663;
|
458 |
+
sub.f32 %f288, %f240, %f664;
|
459 |
+
sub.f32 %f289, %f241, %f665;
|
460 |
+
sub.f32 %f290, %f242, %f666;
|
461 |
+
sub.f32 %f291, %f243, %f667;
|
462 |
+
sub.f32 %f292, %f244, %f668;
|
463 |
+
sub.f32 %f293, %f245, %f669;
|
464 |
+
sub.f32 %f294, %f246, %f670;
|
465 |
+
sub.f32 %f295, %f247, %f671;
|
466 |
+
sub.f32 %f296, %f248, %f672;
|
467 |
+
sub.f32 %f297, %f249, %f673;
|
468 |
+
sub.f32 %f298, %f250, %f674;
|
469 |
+
sub.f32 %f299, %f251, %f675;
|
470 |
+
sub.f32 %f300, %f252, %f676;
|
471 |
+
sub.f32 %f301, %f253, %f677;
|
472 |
+
$L__tmp2:
|
473 |
+
.loc 1 50 50
|
474 |
+
fma.rn.f32 %f646, %f254, %f286, %f646;
|
475 |
+
fma.rn.f32 %f647, %f255, %f287, %f647;
|
476 |
+
fma.rn.f32 %f648, %f256, %f288, %f648;
|
477 |
+
fma.rn.f32 %f649, %f257, %f289, %f649;
|
478 |
+
fma.rn.f32 %f650, %f258, %f290, %f650;
|
479 |
+
fma.rn.f32 %f651, %f259, %f291, %f651;
|
480 |
+
fma.rn.f32 %f652, %f260, %f292, %f652;
|
481 |
+
fma.rn.f32 %f653, %f261, %f293, %f653;
|
482 |
+
fma.rn.f32 %f654, %f262, %f294, %f654;
|
483 |
+
fma.rn.f32 %f655, %f263, %f295, %f655;
|
484 |
+
fma.rn.f32 %f656, %f264, %f296, %f656;
|
485 |
+
fma.rn.f32 %f657, %f265, %f297, %f657;
|
486 |
+
fma.rn.f32 %f658, %f266, %f298, %f658;
|
487 |
+
fma.rn.f32 %f659, %f267, %f299, %f659;
|
488 |
+
fma.rn.f32 %f660, %f268, %f300, %f660;
|
489 |
+
fma.rn.f32 %f661, %f269, %f301, %f661;
|
490 |
+
.loc 1 31 36
|
491 |
+
add.s64 %rd116, %rd116, 256;
|
492 |
+
add.s32 %r472, %r472, 64;
|
493 |
+
setp.lt.u32 %p72, %r472, 192;
|
494 |
+
@%p72 bra $L__BB0_1;
|
495 |
+
bra.uni $L__BB0_4;
|
496 |
+
$L__BB0_1:
|
497 |
+
.loc 1 40 40
|
498 |
+
setp.lt.u64 %p51, %rd1, 50257;
|
499 |
+
.loc 1 35 34
|
500 |
+
add.s64 %rd70, %rd6, %rd116;
|
501 |
+
add.s64 %rd71, %rd70, 16;
|
502 |
+
add.s64 %rd72, %rd5, %rd116;
|
503 |
+
.loc 1 35 50
|
504 |
+
add.s64 %rd73, %rd72, 16;
|
505 |
+
mov.b32 %r411, 0;
|
506 |
+
mov.u32 %r34, 0x0;
|
507 |
+
mov.u32 %r35, 0x0;
|
508 |
+
mov.u32 %r36, 0x0;
|
509 |
+
mov.u32 %r37, 0x0;
|
510 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r34, %r35, %r36, %r37 }, [ %rd70 + 0 ];
|
511 |
+
@!%p1 mov.u32 %r34, %r411;
|
512 |
+
@!%p1 mov.u32 %r35, %r411;
|
513 |
+
@!%p1 mov.u32 %r36, %r411;
|
514 |
+
@!%p1 mov.u32 %r37, %r411;
|
515 |
+
mov.b32 %f65, %r34;
|
516 |
+
mov.b32 %f66, %r35;
|
517 |
+
mov.b32 %f67, %r36;
|
518 |
+
mov.b32 %f68, %r37;
|
519 |
+
mov.u32 %r42, 0x0;
|
520 |
+
mov.u32 %r43, 0x0;
|
521 |
+
mov.u32 %r44, 0x0;
|
522 |
+
mov.u32 %r45, 0x0;
|
523 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r42, %r43, %r44, %r45 }, [ %rd71 + 0 ];
|
524 |
+
@!%p1 mov.u32 %r42, %r411;
|
525 |
+
@!%p1 mov.u32 %r43, %r411;
|
526 |
+
@!%p1 mov.u32 %r44, %r411;
|
527 |
+
@!%p1 mov.u32 %r45, %r411;
|
528 |
+
mov.b32 %f69, %r42;
|
529 |
+
mov.b32 %f70, %r43;
|
530 |
+
mov.b32 %f71, %r44;
|
531 |
+
mov.b32 %f72, %r45;
|
532 |
+
mov.u32 %r50, 0x0;
|
533 |
+
mov.u32 %r51, 0x0;
|
534 |
+
mov.u32 %r52, 0x0;
|
535 |
+
mov.u32 %r53, 0x0;
|
536 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r50, %r51, %r52, %r53 }, [ %rd72 + 0 ];
|
537 |
+
@!%p1 mov.u32 %r50, %r411;
|
538 |
+
@!%p1 mov.u32 %r51, %r411;
|
539 |
+
@!%p1 mov.u32 %r52, %r411;
|
540 |
+
@!%p1 mov.u32 %r53, %r411;
|
541 |
+
mov.b32 %f73, %r50;
|
542 |
+
mov.b32 %f74, %r51;
|
543 |
+
mov.b32 %f75, %r52;
|
544 |
+
mov.b32 %f76, %r53;
|
545 |
+
mov.u32 %r58, 0x0;
|
546 |
+
mov.u32 %r59, 0x0;
|
547 |
+
mov.u32 %r60, 0x0;
|
548 |
+
mov.u32 %r61, 0x0;
|
549 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r58, %r59, %r60, %r61 }, [ %rd73 + 0 ];
|
550 |
+
@!%p1 mov.u32 %r58, %r411;
|
551 |
+
@!%p1 mov.u32 %r59, %r411;
|
552 |
+
@!%p1 mov.u32 %r60, %r411;
|
553 |
+
@!%p1 mov.u32 %r61, %r411;
|
554 |
+
mov.b32 %f77, %r58;
|
555 |
+
mov.b32 %f78, %r59;
|
556 |
+
mov.b32 %f79, %r60;
|
557 |
+
mov.b32 %f80, %r61;
|
558 |
+
.loc 1 36 40
|
559 |
+
add.s32 %r98, %r5, %r472;
|
560 |
+
add.s32 %r99, %r98, 64;
|
561 |
+
.loc 1 36 34
|
562 |
+
add.s32 %r100, %r98, 8256;
|
563 |
+
mul.wide.s32 %rd76, %r99, 2;
|
564 |
+
add.s64 %rd74, %rd14, %rd76;
|
565 |
+
mul.wide.s32 %rd77, %r100, 2;
|
566 |
+
add.s64 %rd75, %rd14, %rd77;
|
567 |
+
.loc 1 36 50
|
568 |
+
mov.u32 %r66, 0x0;
|
569 |
+
mov.u32 %r67, 0x0;
|
570 |
+
mov.u32 %r68, 0x0;
|
571 |
+
mov.u32 %r69, 0x0;
|
572 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r66, %r67, %r68, %r69 }, [ %rd74 + 0 ];
|
573 |
+
@!%p1 mov.u32 %r66, %r411;
|
574 |
+
@!%p1 mov.u32 %r67, %r411;
|
575 |
+
@!%p1 mov.u32 %r68, %r411;
|
576 |
+
@!%p1 mov.u32 %r69, %r411;
|
577 |
+
cvt.u16.u32 %rs1, %r66;
|
578 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs2}, %r66; }
|
579 |
+
cvt.u16.u32 %rs3, %r67;
|
580 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs4}, %r67; }
|
581 |
+
cvt.u16.u32 %rs5, %r68;
|
582 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs6}, %r68; }
|
583 |
+
cvt.u16.u32 %rs7, %r69;
|
584 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs8}, %r69; }
|
585 |
+
mov.u32 %r74, 0x0;
|
586 |
+
mov.u32 %r75, 0x0;
|
587 |
+
mov.u32 %r76, 0x0;
|
588 |
+
mov.u32 %r77, 0x0;
|
589 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r74, %r75, %r76, %r77 }, [ %rd75 + 0 ];
|
590 |
+
@!%p1 mov.u32 %r74, %r411;
|
591 |
+
@!%p1 mov.u32 %r75, %r411;
|
592 |
+
@!%p1 mov.u32 %r76, %r411;
|
593 |
+
@!%p1 mov.u32 %r77, %r411;
|
594 |
+
cvt.u16.u32 %rs9, %r74;
|
595 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs10}, %r74; }
|
596 |
+
cvt.u16.u32 %rs11, %r75;
|
597 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs12}, %r75; }
|
598 |
+
cvt.u16.u32 %rs13, %r76;
|
599 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs14}, %r76; }
|
600 |
+
cvt.u16.u32 %rs15, %r77;
|
601 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs16}, %r77; }
|
602 |
+
.loc 1 36 101
|
603 |
+
cvt.f32.bf16 %r82, %rs1;
|
604 |
+
mov.b32 %f81, %r82;
|
605 |
+
cvt.f32.bf16 %r83, %rs2;
|
606 |
+
mov.b32 %f82, %r83;
|
607 |
+
cvt.f32.bf16 %r84, %rs3;
|
608 |
+
mov.b32 %f83, %r84;
|
609 |
+
cvt.f32.bf16 %r85, %rs4;
|
610 |
+
mov.b32 %f84, %r85;
|
611 |
+
cvt.f32.bf16 %r86, %rs5;
|
612 |
+
mov.b32 %f85, %r86;
|
613 |
+
cvt.f32.bf16 %r87, %rs6;
|
614 |
+
mov.b32 %f86, %r87;
|
615 |
+
cvt.f32.bf16 %r88, %rs7;
|
616 |
+
mov.b32 %f87, %r88;
|
617 |
+
cvt.f32.bf16 %r89, %rs8;
|
618 |
+
mov.b32 %f88, %r89;
|
619 |
+
cvt.f32.bf16 %r90, %rs9;
|
620 |
+
mov.b32 %f89, %r90;
|
621 |
+
cvt.f32.bf16 %r91, %rs10;
|
622 |
+
mov.b32 %f90, %r91;
|
623 |
+
cvt.f32.bf16 %r92, %rs11;
|
624 |
+
mov.b32 %f91, %r92;
|
625 |
+
cvt.f32.bf16 %r93, %rs12;
|
626 |
+
mov.b32 %f92, %r93;
|
627 |
+
cvt.f32.bf16 %r94, %rs13;
|
628 |
+
mov.b32 %f93, %r94;
|
629 |
+
cvt.f32.bf16 %r95, %rs14;
|
630 |
+
mov.b32 %f94, %r95;
|
631 |
+
cvt.f32.bf16 %r96, %rs15;
|
632 |
+
mov.b32 %f95, %r96;
|
633 |
+
cvt.f32.bf16 %r97, %rs16;
|
634 |
+
mov.b32 %f96, %r97;
|
635 |
+
mov.b32 %r471, 883;
|
636 |
+
mov.u64 %rd115, 1;
|
637 |
+
.loc 1 40 55
|
638 |
+
@%p51 bra $L__BB0_3;
|
639 |
+
mov.u64 %rd78, assertMessage_0;
|
640 |
+
cvta.global.u64 %rd79, %rd78;
|
641 |
+
mov.u64 %rd80, assertFile_0;
|
642 |
+
cvta.global.u64 %rd81, %rd80;
|
643 |
+
mov.u64 %rd82, assertFunc_0;
|
644 |
+
cvta.global.u64 %rd83, %rd82;
|
645 |
+
{ // callseq 6, 0
|
646 |
+
.reg .b32 temp_param_reg;
|
647 |
+
.param .b64 param0;
|
648 |
+
st.param.b64 [param0+0], %rd79;
|
649 |
+
.param .b64 param1;
|
650 |
+
st.param.b64 [param1+0], %rd81;
|
651 |
+
.param .b32 param2;
|
652 |
+
st.param.b32 [param2+0], %r471;
|
653 |
+
.param .b64 param3;
|
654 |
+
st.param.b64 [param3+0], %rd83;
|
655 |
+
.param .b64 param4;
|
656 |
+
st.param.b64 [param4+0], %rd115;
|
657 |
+
call.uni
|
658 |
+
__assertfail,
|
659 |
+
(
|
660 |
+
param0,
|
661 |
+
param1,
|
662 |
+
param2,
|
663 |
+
param3,
|
664 |
+
param4
|
665 |
+
);
|
666 |
+
} // callseq 6
|
667 |
+
bra.uni $L__BB0_3;
|
668 |
+
$L__BB0_4:
|
669 |
+
.loc 1 31 36
|
670 |
+
and.b32 %r291, %r4, 3;
|
671 |
+
mad.lo.s32 %r292, %r291, 72, %r2;
|
672 |
+
shl.b32 %r293, %r292, 2;
|
673 |
+
mov.u32 %r294, global_smem;
|
674 |
+
add.s32 %r295, %r294, %r293;
|
675 |
+
st.shared.f32 [%r295], %f630;
|
676 |
+
st.shared.f32 [%r295+1152], %f631;
|
677 |
+
st.shared.f32 [%r295+2304], %f632;
|
678 |
+
st.shared.f32 [%r295+3456], %f633;
|
679 |
+
st.shared.f32 [%r295+4608], %f634;
|
680 |
+
st.shared.f32 [%r295+5760], %f635;
|
681 |
+
st.shared.f32 [%r295+6912], %f636;
|
682 |
+
st.shared.f32 [%r295+8064], %f637;
|
683 |
+
bar.sync 0;
|
684 |
+
mad.lo.s32 %r296, %r1, 72, %r3;
|
685 |
+
shl.b32 %r297, %r296, 2;
|
686 |
+
add.s32 %r298, %r294, %r297;
|
687 |
+
ld.shared.v4.f32 {%f302, %f303, %f304, %f305}, [%r298];
|
688 |
+
ld.shared.v4.f32 {%f306, %f307, %f308, %f309}, [%r298+16];
|
689 |
+
bar.sync 0;
|
690 |
+
st.shared.f32 [%r295], %f638;
|
691 |
+
st.shared.f32 [%r295+1152], %f639;
|
692 |
+
st.shared.f32 [%r295+2304], %f640;
|
693 |
+
st.shared.f32 [%r295+3456], %f641;
|
694 |
+
st.shared.f32 [%r295+4608], %f642;
|
695 |
+
st.shared.f32 [%r295+5760], %f643;
|
696 |
+
st.shared.f32 [%r295+6912], %f644;
|
697 |
+
st.shared.f32 [%r295+8064], %f645;
|
698 |
+
bar.sync 0;
|
699 |
+
ld.shared.v4.f32 {%f310, %f311, %f312, %f313}, [%r298];
|
700 |
+
ld.shared.v4.f32 {%f314, %f315, %f316, %f317}, [%r298+16];
|
701 |
+
$L__tmp3:
|
702 |
+
.loc 2 108 21
|
703 |
+
sub.f32 %f318, %f663, %f662;
|
704 |
+
.loc 2 109 28
|
705 |
+
add.f32 %f319, %f302, %f303;
|
706 |
+
.loc 2 110 39
|
707 |
+
setp.eq.f32 %p73, %f319, 0f00000000;
|
708 |
+
.loc 2 110 60
|
709 |
+
mov.b32 %r183, %f303;
|
710 |
+
mov.b32 %r184, %f319;
|
711 |
+
div.full.f32 %r182, %r183, %r184;
|
712 |
+
mov.b32 %f320, %r182;
|
713 |
+
.loc 2 110 49
|
714 |
+
selp.f32 %f321, 0f00000000, %f320, %p73;
|
715 |
+
.loc 2 112 17
|
716 |
+
fma.rn.f32 %f322, %f318, %f321, %f662;
|
717 |
+
.loc 2 113 15
|
718 |
+
add.f32 %f323, %f646, %f647;
|
719 |
+
.loc 2 113 30
|
720 |
+
mul.f32 %f324, %f318, %f318;
|
721 |
+
.loc 2 113 38
|
722 |
+
mul.f32 %f325, %f324, %f302;
|
723 |
+
.loc 2 113 22
|
724 |
+
fma.rn.f32 %f326, %f325, %f321, %f323;
|
725 |
+
.loc 2 108 21
|
726 |
+
sub.f32 %f327, %f664, %f322;
|
727 |
+
.loc 2 109 28
|
728 |
+
add.f32 %f328, %f304, %f319;
|
729 |
+
.loc 2 110 39
|
730 |
+
setp.eq.f32 %p74, %f328, 0f00000000;
|
731 |
+
.loc 2 110 60
|
732 |
+
mov.b32 %r187, %f328;
|
733 |
+
mov.b32 %r186, %f304;
|
734 |
+
div.full.f32 %r185, %r186, %r187;
|
735 |
+
mov.b32 %f329, %r185;
|
736 |
+
.loc 2 110 49
|
737 |
+
selp.f32 %f330, 0f00000000, %f329, %p74;
|
738 |
+
.loc 2 112 17
|
739 |
+
fma.rn.f32 %f331, %f330, %f327, %f322;
|
740 |
+
.loc 2 113 15
|
741 |
+
add.f32 %f332, %f648, %f326;
|
742 |
+
.loc 2 113 30
|
743 |
+
mul.f32 %f333, %f327, %f327;
|
744 |
+
.loc 2 113 38
|
745 |
+
mul.f32 %f334, %f319, %f333;
|
746 |
+
.loc 2 113 22
|
747 |
+
fma.rn.f32 %f335, %f330, %f334, %f332;
|
748 |
+
.loc 2 108 21
|
749 |
+
sub.f32 %f336, %f665, %f331;
|
750 |
+
.loc 2 109 28
|
751 |
+
add.f32 %f337, %f305, %f328;
|
752 |
+
.loc 2 110 39
|
753 |
+
setp.eq.f32 %p75, %f337, 0f00000000;
|
754 |
+
.loc 2 110 60
|
755 |
+
mov.b32 %r190, %f337;
|
756 |
+
mov.b32 %r189, %f305;
|
757 |
+
div.full.f32 %r188, %r189, %r190;
|
758 |
+
mov.b32 %f338, %r188;
|
759 |
+
.loc 2 110 49
|
760 |
+
selp.f32 %f339, 0f00000000, %f338, %p75;
|
761 |
+
.loc 2 112 17
|
762 |
+
fma.rn.f32 %f340, %f339, %f336, %f331;
|
763 |
+
.loc 2 113 15
|
764 |
+
add.f32 %f341, %f649, %f335;
|
765 |
+
.loc 2 113 30
|
766 |
+
mul.f32 %f342, %f336, %f336;
|
767 |
+
.loc 2 113 38
|
768 |
+
mul.f32 %f343, %f328, %f342;
|
769 |
+
.loc 2 113 22
|
770 |
+
fma.rn.f32 %f344, %f339, %f343, %f341;
|
771 |
+
.loc 2 108 21
|
772 |
+
sub.f32 %f345, %f666, %f340;
|
773 |
+
.loc 2 109 28
|
774 |
+
add.f32 %f346, %f306, %f337;
|
775 |
+
.loc 2 110 39
|
776 |
+
setp.eq.f32 %p76, %f346, 0f00000000;
|
777 |
+
.loc 2 110 60
|
778 |
+
mov.b32 %r193, %f346;
|
779 |
+
mov.b32 %r192, %f306;
|
780 |
+
div.full.f32 %r191, %r192, %r193;
|
781 |
+
mov.b32 %f347, %r191;
|
782 |
+
.loc 2 110 49
|
783 |
+
selp.f32 %f348, 0f00000000, %f347, %p76;
|
784 |
+
.loc 2 112 17
|
785 |
+
fma.rn.f32 %f349, %f348, %f345, %f340;
|
786 |
+
.loc 2 113 15
|
787 |
+
add.f32 %f350, %f650, %f344;
|
788 |
+
.loc 2 113 30
|
789 |
+
mul.f32 %f351, %f345, %f345;
|
790 |
+
.loc 2 113 38
|
791 |
+
mul.f32 %f352, %f337, %f351;
|
792 |
+
.loc 2 113 22
|
793 |
+
fma.rn.f32 %f353, %f348, %f352, %f350;
|
794 |
+
.loc 2 108 21
|
795 |
+
sub.f32 %f354, %f667, %f349;
|
796 |
+
.loc 2 109 28
|
797 |
+
add.f32 %f355, %f307, %f346;
|
798 |
+
.loc 2 110 39
|
799 |
+
setp.eq.f32 %p77, %f355, 0f00000000;
|
800 |
+
.loc 2 110 60
|
801 |
+
mov.b32 %r196, %f355;
|
802 |
+
mov.b32 %r195, %f307;
|
803 |
+
div.full.f32 %r194, %r195, %r196;
|
804 |
+
mov.b32 %f356, %r194;
|
805 |
+
.loc 2 110 49
|
806 |
+
selp.f32 %f357, 0f00000000, %f356, %p77;
|
807 |
+
.loc 2 112 17
|
808 |
+
fma.rn.f32 %f358, %f357, %f354, %f349;
|
809 |
+
.loc 2 113 15
|
810 |
+
add.f32 %f359, %f651, %f353;
|
811 |
+
.loc 2 113 30
|
812 |
+
mul.f32 %f360, %f354, %f354;
|
813 |
+
.loc 2 113 38
|
814 |
+
mul.f32 %f361, %f346, %f360;
|
815 |
+
.loc 2 113 22
|
816 |
+
fma.rn.f32 %f362, %f357, %f361, %f359;
|
817 |
+
.loc 2 108 21
|
818 |
+
sub.f32 %f363, %f668, %f358;
|
819 |
+
.loc 2 109 28
|
820 |
+
add.f32 %f364, %f308, %f355;
|
821 |
+
.loc 2 110 39
|
822 |
+
setp.eq.f32 %p78, %f364, 0f00000000;
|
823 |
+
.loc 2 110 60
|
824 |
+
mov.b32 %r199, %f364;
|
825 |
+
mov.b32 %r198, %f308;
|
826 |
+
div.full.f32 %r197, %r198, %r199;
|
827 |
+
mov.b32 %f365, %r197;
|
828 |
+
.loc 2 110 49
|
829 |
+
selp.f32 %f366, 0f00000000, %f365, %p78;
|
830 |
+
.loc 2 112 17
|
831 |
+
fma.rn.f32 %f367, %f366, %f363, %f358;
|
832 |
+
.loc 2 113 15
|
833 |
+
add.f32 %f368, %f652, %f362;
|
834 |
+
.loc 2 113 30
|
835 |
+
mul.f32 %f369, %f363, %f363;
|
836 |
+
.loc 2 113 38
|
837 |
+
mul.f32 %f370, %f355, %f369;
|
838 |
+
.loc 2 113 22
|
839 |
+
fma.rn.f32 %f371, %f366, %f370, %f368;
|
840 |
+
.loc 2 108 21
|
841 |
+
sub.f32 %f372, %f669, %f367;
|
842 |
+
.loc 2 109 28
|
843 |
+
add.f32 %f373, %f309, %f364;
|
844 |
+
.loc 2 110 39
|
845 |
+
setp.eq.f32 %p79, %f373, 0f00000000;
|
846 |
+
.loc 2 110 60
|
847 |
+
mov.b32 %r202, %f373;
|
848 |
+
mov.b32 %r201, %f309;
|
849 |
+
div.full.f32 %r200, %r201, %r202;
|
850 |
+
mov.b32 %f374, %r200;
|
851 |
+
.loc 2 110 49
|
852 |
+
selp.f32 %f375, 0f00000000, %f374, %p79;
|
853 |
+
.loc 2 112 17
|
854 |
+
fma.rn.f32 %f376, %f375, %f372, %f367;
|
855 |
+
.loc 2 113 15
|
856 |
+
add.f32 %f377, %f653, %f371;
|
857 |
+
.loc 2 113 30
|
858 |
+
mul.f32 %f378, %f372, %f372;
|
859 |
+
.loc 2 113 38
|
860 |
+
mul.f32 %f379, %f364, %f378;
|
861 |
+
.loc 2 113 22
|
862 |
+
fma.rn.f32 %f380, %f375, %f379, %f377;
|
863 |
+
.loc 2 108 21
|
864 |
+
sub.f32 %f381, %f671, %f670;
|
865 |
+
.loc 2 109 28
|
866 |
+
add.f32 %f382, %f310, %f311;
|
867 |
+
.loc 2 110 39
|
868 |
+
setp.eq.f32 %p80, %f382, 0f00000000;
|
869 |
+
.loc 2 110 60
|
870 |
+
mov.b32 %r204, %f311;
|
871 |
+
mov.b32 %r205, %f382;
|
872 |
+
div.full.f32 %r203, %r204, %r205;
|
873 |
+
mov.b32 %f383, %r203;
|
874 |
+
.loc 2 110 49
|
875 |
+
selp.f32 %f384, 0f00000000, %f383, %p80;
|
876 |
+
.loc 2 112 17
|
877 |
+
fma.rn.f32 %f385, %f381, %f384, %f670;
|
878 |
+
.loc 2 113 15
|
879 |
+
add.f32 %f386, %f654, %f655;
|
880 |
+
.loc 2 113 30
|
881 |
+
mul.f32 %f387, %f381, %f381;
|
882 |
+
.loc 2 113 38
|
883 |
+
mul.f32 %f388, %f387, %f310;
|
884 |
+
.loc 2 113 22
|
885 |
+
fma.rn.f32 %f389, %f388, %f384, %f386;
|
886 |
+
.loc 2 108 21
|
887 |
+
sub.f32 %f390, %f672, %f385;
|
888 |
+
.loc 2 109 28
|
889 |
+
add.f32 %f391, %f312, %f382;
|
890 |
+
.loc 2 110 39
|
891 |
+
setp.eq.f32 %p81, %f391, 0f00000000;
|
892 |
+
.loc 2 110 60
|
893 |
+
mov.b32 %r208, %f391;
|
894 |
+
mov.b32 %r207, %f312;
|
895 |
+
div.full.f32 %r206, %r207, %r208;
|
896 |
+
mov.b32 %f392, %r206;
|
897 |
+
.loc 2 110 49
|
898 |
+
selp.f32 %f393, 0f00000000, %f392, %p81;
|
899 |
+
.loc 2 112 17
|
900 |
+
fma.rn.f32 %f394, %f393, %f390, %f385;
|
901 |
+
.loc 2 113 15
|
902 |
+
add.f32 %f395, %f656, %f389;
|
903 |
+
.loc 2 113 30
|
904 |
+
mul.f32 %f396, %f390, %f390;
|
905 |
+
.loc 2 113 38
|
906 |
+
mul.f32 %f397, %f382, %f396;
|
907 |
+
.loc 2 113 22
|
908 |
+
fma.rn.f32 %f398, %f393, %f397, %f395;
|
909 |
+
.loc 2 108 21
|
910 |
+
sub.f32 %f399, %f673, %f394;
|
911 |
+
.loc 2 109 28
|
912 |
+
add.f32 %f400, %f313, %f391;
|
913 |
+
.loc 2 110 39
|
914 |
+
setp.eq.f32 %p82, %f400, 0f00000000;
|
915 |
+
.loc 2 110 60
|
916 |
+
mov.b32 %r211, %f400;
|
917 |
+
mov.b32 %r210, %f313;
|
918 |
+
div.full.f32 %r209, %r210, %r211;
|
919 |
+
mov.b32 %f401, %r209;
|
920 |
+
.loc 2 110 49
|
921 |
+
selp.f32 %f402, 0f00000000, %f401, %p82;
|
922 |
+
.loc 2 112 17
|
923 |
+
fma.rn.f32 %f403, %f402, %f399, %f394;
|
924 |
+
.loc 2 113 15
|
925 |
+
add.f32 %f404, %f657, %f398;
|
926 |
+
.loc 2 113 30
|
927 |
+
mul.f32 %f405, %f399, %f399;
|
928 |
+
.loc 2 113 38
|
929 |
+
mul.f32 %f406, %f391, %f405;
|
930 |
+
.loc 2 113 22
|
931 |
+
fma.rn.f32 %f407, %f402, %f406, %f404;
|
932 |
+
.loc 2 108 21
|
933 |
+
sub.f32 %f408, %f674, %f403;
|
934 |
+
.loc 2 109 28
|
935 |
+
add.f32 %f409, %f314, %f400;
|
936 |
+
.loc 2 110 39
|
937 |
+
setp.eq.f32 %p83, %f409, 0f00000000;
|
938 |
+
.loc 2 110 60
|
939 |
+
mov.b32 %r214, %f409;
|
940 |
+
mov.b32 %r213, %f314;
|
941 |
+
div.full.f32 %r212, %r213, %r214;
|
942 |
+
mov.b32 %f410, %r212;
|
943 |
+
.loc 2 110 49
|
944 |
+
selp.f32 %f411, 0f00000000, %f410, %p83;
|
945 |
+
.loc 2 112 17
|
946 |
+
fma.rn.f32 %f412, %f411, %f408, %f403;
|
947 |
+
.loc 2 113 15
|
948 |
+
add.f32 %f413, %f658, %f407;
|
949 |
+
.loc 2 113 30
|
950 |
+
mul.f32 %f414, %f408, %f408;
|
951 |
+
.loc 2 113 38
|
952 |
+
mul.f32 %f415, %f400, %f414;
|
953 |
+
.loc 2 113 22
|
954 |
+
fma.rn.f32 %f416, %f411, %f415, %f413;
|
955 |
+
.loc 2 108 21
|
956 |
+
sub.f32 %f417, %f675, %f412;
|
957 |
+
.loc 2 109 28
|
958 |
+
add.f32 %f418, %f315, %f409;
|
959 |
+
.loc 2 110 39
|
960 |
+
setp.eq.f32 %p84, %f418, 0f00000000;
|
961 |
+
.loc 2 110 60
|
962 |
+
mov.b32 %r217, %f418;
|
963 |
+
mov.b32 %r216, %f315;
|
964 |
+
div.full.f32 %r215, %r216, %r217;
|
965 |
+
mov.b32 %f419, %r215;
|
966 |
+
.loc 2 110 49
|
967 |
+
selp.f32 %f420, 0f00000000, %f419, %p84;
|
968 |
+
.loc 2 112 17
|
969 |
+
fma.rn.f32 %f421, %f420, %f417, %f412;
|
970 |
+
.loc 2 113 15
|
971 |
+
add.f32 %f422, %f659, %f416;
|
972 |
+
.loc 2 113 30
|
973 |
+
mul.f32 %f423, %f417, %f417;
|
974 |
+
.loc 2 113 38
|
975 |
+
mul.f32 %f424, %f409, %f423;
|
976 |
+
.loc 2 113 22
|
977 |
+
fma.rn.f32 %f425, %f420, %f424, %f422;
|
978 |
+
.loc 2 108 21
|
979 |
+
sub.f32 %f426, %f676, %f421;
|
980 |
+
.loc 2 109 28
|
981 |
+
add.f32 %f427, %f316, %f418;
|
982 |
+
.loc 2 110 39
|
983 |
+
setp.eq.f32 %p85, %f427, 0f00000000;
|
984 |
+
.loc 2 110 60
|
985 |
+
mov.b32 %r220, %f427;
|
986 |
+
mov.b32 %r219, %f316;
|
987 |
+
div.full.f32 %r218, %r219, %r220;
|
988 |
+
mov.b32 %f428, %r218;
|
989 |
+
.loc 2 110 49
|
990 |
+
selp.f32 %f429, 0f00000000, %f428, %p85;
|
991 |
+
.loc 2 112 17
|
992 |
+
fma.rn.f32 %f430, %f429, %f426, %f421;
|
993 |
+
.loc 2 113 15
|
994 |
+
add.f32 %f431, %f660, %f425;
|
995 |
+
.loc 2 113 30
|
996 |
+
mul.f32 %f432, %f426, %f426;
|
997 |
+
.loc 2 113 38
|
998 |
+
mul.f32 %f433, %f418, %f432;
|
999 |
+
.loc 2 113 22
|
1000 |
+
fma.rn.f32 %f434, %f429, %f433, %f431;
|
1001 |
+
.loc 2 108 21
|
1002 |
+
sub.f32 %f435, %f677, %f430;
|
1003 |
+
.loc 2 109 28
|
1004 |
+
add.f32 %f436, %f317, %f427;
|
1005 |
+
.loc 2 110 39
|
1006 |
+
setp.eq.f32 %p86, %f436, 0f00000000;
|
1007 |
+
.loc 2 110 60
|
1008 |
+
mov.b32 %r223, %f436;
|
1009 |
+
mov.b32 %r222, %f317;
|
1010 |
+
div.full.f32 %r221, %r222, %r223;
|
1011 |
+
mov.b32 %f437, %r221;
|
1012 |
+
.loc 2 110 49
|
1013 |
+
selp.f32 %f438, 0f00000000, %f437, %p86;
|
1014 |
+
.loc 2 112 17
|
1015 |
+
fma.rn.f32 %f439, %f438, %f435, %f430;
|
1016 |
+
.loc 2 113 15
|
1017 |
+
add.f32 %f440, %f661, %f434;
|
1018 |
+
.loc 2 113 30
|
1019 |
+
mul.f32 %f441, %f435, %f435;
|
1020 |
+
.loc 2 113 38
|
1021 |
+
mul.f32 %f442, %f427, %f441;
|
1022 |
+
.loc 2 113 22
|
1023 |
+
fma.rn.f32 %f443, %f438, %f442, %f440;
|
1024 |
+
$L__tmp4:
|
1025 |
+
.loc 2 120 46
|
1026 |
+
mov.b32 %r299, %f376;
|
1027 |
+
shfl.sync.bfly.b32 %r300, %r299, 4, 31, -1;
|
1028 |
+
mov.b32 %f444, %r300;
|
1029 |
+
mov.b32 %r301, %f380;
|
1030 |
+
shfl.sync.bfly.b32 %r302, %r301, 4, 31, -1;
|
1031 |
+
mov.b32 %f445, %r302;
|
1032 |
+
shfl.sync.bfly.b32 %r225, %r202, 4, 31, -1;
|
1033 |
+
mov.b32 %f446, %r225;
|
1034 |
+
$L__tmp5:
|
1035 |
+
.loc 2 108 21
|
1036 |
+
sub.f32 %f447, %f444, %f376;
|
1037 |
+
.loc 2 109 28
|
1038 |
+
add.f32 %f448, %f373, %f446;
|
1039 |
+
.loc 2 110 39
|
1040 |
+
setp.eq.f32 %p87, %f448, 0f00000000;
|
1041 |
+
.loc 2 110 60
|
1042 |
+
mov.b32 %r226, %f448;
|
1043 |
+
div.full.f32 %r224, %r225, %r226;
|
1044 |
+
mov.b32 %f449, %r224;
|
1045 |
+
.loc 2 110 49
|
1046 |
+
selp.f32 %f450, 0f00000000, %f449, %p87;
|
1047 |
+
.loc 2 112 17
|
1048 |
+
fma.rn.f32 %f451, %f450, %f447, %f376;
|
1049 |
+
.loc 2 113 15
|
1050 |
+
add.f32 %f452, %f380, %f445;
|
1051 |
+
.loc 2 113 30
|
1052 |
+
mul.f32 %f453, %f447, %f447;
|
1053 |
+
.loc 2 113 38
|
1054 |
+
mul.f32 %f454, %f373, %f453;
|
1055 |
+
.loc 2 113 22
|
1056 |
+
fma.rn.f32 %f455, %f450, %f454, %f452;
|
1057 |
+
$L__tmp6:
|
1058 |
+
.loc 2 120 46
|
1059 |
+
mov.b32 %r303, %f451;
|
1060 |
+
shfl.sync.bfly.b32 %r304, %r303, 2, 31, -1;
|
1061 |
+
mov.b32 %f456, %r304;
|
1062 |
+
mov.b32 %r305, %f455;
|
1063 |
+
shfl.sync.bfly.b32 %r306, %r305, 2, 31, -1;
|
1064 |
+
mov.b32 %f457, %r306;
|
1065 |
+
shfl.sync.bfly.b32 %r228, %r226, 2, 31, -1;
|
1066 |
+
mov.b32 %f458, %r228;
|
1067 |
+
$L__tmp7:
|
1068 |
+
.loc 2 108 21
|
1069 |
+
sub.f32 %f459, %f456, %f451;
|
1070 |
+
.loc 2 109 28
|
1071 |
+
add.f32 %f460, %f448, %f458;
|
1072 |
+
.loc 2 110 39
|
1073 |
+
setp.eq.f32 %p88, %f460, 0f00000000;
|
1074 |
+
.loc 2 110 60
|
1075 |
+
mov.b32 %r229, %f460;
|
1076 |
+
div.full.f32 %r227, %r228, %r229;
|
1077 |
+
mov.b32 %f461, %r227;
|
1078 |
+
.loc 2 110 49
|
1079 |
+
selp.f32 %f462, 0f00000000, %f461, %p88;
|
1080 |
+
.loc 2 112 17
|
1081 |
+
fma.rn.f32 %f463, %f462, %f459, %f451;
|
1082 |
+
.loc 2 113 15
|
1083 |
+
add.f32 %f464, %f455, %f457;
|
1084 |
+
.loc 2 113 30
|
1085 |
+
mul.f32 %f465, %f459, %f459;
|
1086 |
+
.loc 2 113 38
|
1087 |
+
mul.f32 %f466, %f448, %f465;
|
1088 |
+
.loc 2 113 22
|
1089 |
+
fma.rn.f32 %f467, %f462, %f466, %f464;
|
1090 |
+
$L__tmp8:
|
1091 |
+
.loc 2 120 46
|
1092 |
+
mov.b32 %r307, %f463;
|
1093 |
+
shfl.sync.bfly.b32 %r308, %r307, 1, 31, -1;
|
1094 |
+
mov.b32 %f468, %r308;
|
1095 |
+
mov.b32 %r309, %f467;
|
1096 |
+
shfl.sync.bfly.b32 %r310, %r309, 1, 31, -1;
|
1097 |
+
mov.b32 %f469, %r310;
|
1098 |
+
shfl.sync.bfly.b32 %r231, %r229, 1, 31, -1;
|
1099 |
+
mov.b32 %f470, %r231;
|
1100 |
+
$L__tmp9:
|
1101 |
+
.loc 2 108 21
|
1102 |
+
sub.f32 %f471, %f468, %f463;
|
1103 |
+
.loc 2 109 28
|
1104 |
+
add.f32 %f472, %f460, %f470;
|
1105 |
+
.loc 2 110 39
|
1106 |
+
setp.eq.f32 %p89, %f472, 0f00000000;
|
1107 |
+
.loc 2 110 60
|
1108 |
+
mov.b32 %r232, %f472;
|
1109 |
+
div.full.f32 %r230, %r231, %r232;
|
1110 |
+
mov.b32 %f473, %r230;
|
1111 |
+
.loc 2 110 49
|
1112 |
+
selp.f32 %f474, 0f00000000, %f473, %p89;
|
1113 |
+
.loc 2 112 17
|
1114 |
+
fma.rn.f32 %f161, %f471, %f474, %f463;
|
1115 |
+
.loc 2 113 15
|
1116 |
+
add.f32 %f475, %f467, %f469;
|
1117 |
+
.loc 2 113 30
|
1118 |
+
mul.f32 %f476, %f471, %f471;
|
1119 |
+
.loc 2 113 38
|
1120 |
+
mul.f32 %f477, %f460, %f476;
|
1121 |
+
.loc 2 113 22
|
1122 |
+
fma.rn.f32 %f478, %f474, %f477, %f475;
|
1123 |
+
$L__tmp10:
|
1124 |
+
.loc 2 120 46
|
1125 |
+
mov.b32 %r311, %f439;
|
1126 |
+
shfl.sync.bfly.b32 %r312, %r311, 4, 31, -1;
|
1127 |
+
mov.b32 %f479, %r312;
|
1128 |
+
mov.b32 %r313, %f443;
|
1129 |
+
shfl.sync.bfly.b32 %r314, %r313, 4, 31, -1;
|
1130 |
+
mov.b32 %f480, %r314;
|
1131 |
+
shfl.sync.bfly.b32 %r234, %r223, 4, 31, -1;
|
1132 |
+
mov.b32 %f481, %r234;
|
1133 |
+
$L__tmp11:
|
1134 |
+
.loc 2 108 21
|
1135 |
+
sub.f32 %f482, %f479, %f439;
|
1136 |
+
.loc 2 109 28
|
1137 |
+
add.f32 %f483, %f436, %f481;
|
1138 |
+
.loc 2 110 39
|
1139 |
+
setp.eq.f32 %p90, %f483, 0f00000000;
|
1140 |
+
.loc 2 110 60
|
1141 |
+
mov.b32 %r235, %f483;
|
1142 |
+
div.full.f32 %r233, %r234, %r235;
|
1143 |
+
mov.b32 %f484, %r233;
|
1144 |
+
.loc 2 110 49
|
1145 |
+
selp.f32 %f485, 0f00000000, %f484, %p90;
|
1146 |
+
.loc 2 112 17
|
1147 |
+
fma.rn.f32 %f486, %f482, %f485, %f439;
|
1148 |
+
.loc 2 113 15
|
1149 |
+
add.f32 %f487, %f443, %f480;
|
1150 |
+
.loc 2 113 30
|
1151 |
+
mul.f32 %f488, %f482, %f482;
|
1152 |
+
.loc 2 113 38
|
1153 |
+
mul.f32 %f489, %f436, %f488;
|
1154 |
+
.loc 2 113 22
|
1155 |
+
fma.rn.f32 %f490, %f489, %f485, %f487;
|
1156 |
+
$L__tmp12:
|
1157 |
+
.loc 2 120 46
|
1158 |
+
mov.b32 %r315, %f486;
|
1159 |
+
shfl.sync.bfly.b32 %r316, %r315, 2, 31, -1;
|
1160 |
+
mov.b32 %f491, %r316;
|
1161 |
+
mov.b32 %r317, %f490;
|
1162 |
+
shfl.sync.bfly.b32 %r318, %r317, 2, 31, -1;
|
1163 |
+
mov.b32 %f492, %r318;
|
1164 |
+
shfl.sync.bfly.b32 %r237, %r235, 2, 31, -1;
|
1165 |
+
mov.b32 %f493, %r237;
|
1166 |
+
$L__tmp13:
|
1167 |
+
.loc 2 108 21
|
1168 |
+
sub.f32 %f494, %f491, %f486;
|
1169 |
+
.loc 2 109 28
|
1170 |
+
add.f32 %f495, %f483, %f493;
|
1171 |
+
.loc 2 110 39
|
1172 |
+
setp.eq.f32 %p91, %f495, 0f00000000;
|
1173 |
+
.loc 2 110 60
|
1174 |
+
mov.b32 %r238, %f495;
|
1175 |
+
div.full.f32 %r236, %r237, %r238;
|
1176 |
+
mov.b32 %f496, %r236;
|
1177 |
+
.loc 2 110 49
|
1178 |
+
selp.f32 %f497, 0f00000000, %f496, %p91;
|
1179 |
+
.loc 2 112 17
|
1180 |
+
fma.rn.f32 %f498, %f494, %f497, %f486;
|
1181 |
+
.loc 2 113 15
|
1182 |
+
add.f32 %f499, %f490, %f492;
|
1183 |
+
.loc 2 113 30
|
1184 |
+
mul.f32 %f500, %f494, %f494;
|
1185 |
+
.loc 2 113 38
|
1186 |
+
mul.f32 %f501, %f483, %f500;
|
1187 |
+
.loc 2 113 22
|
1188 |
+
fma.rn.f32 %f502, %f497, %f501, %f499;
|
1189 |
+
$L__tmp14:
|
1190 |
+
.loc 2 120 46
|
1191 |
+
mov.b32 %r319, %f498;
|
1192 |
+
shfl.sync.bfly.b32 %r320, %r319, 1, 31, -1;
|
1193 |
+
mov.b32 %f503, %r320;
|
1194 |
+
mov.b32 %r321, %f502;
|
1195 |
+
shfl.sync.bfly.b32 %r322, %r321, 1, 31, -1;
|
1196 |
+
mov.b32 %f504, %r322;
|
1197 |
+
shfl.sync.bfly.b32 %r240, %r238, 1, 31, -1;
|
1198 |
+
mov.b32 %f505, %r240;
|
1199 |
+
$L__tmp15:
|
1200 |
+
.loc 2 108 21
|
1201 |
+
sub.f32 %f506, %f503, %f498;
|
1202 |
+
.loc 2 109 28
|
1203 |
+
add.f32 %f507, %f495, %f505;
|
1204 |
+
.loc 2 110 39
|
1205 |
+
setp.eq.f32 %p92, %f507, 0f00000000;
|
1206 |
+
.loc 2 110 60
|
1207 |
+
mov.b32 %r241, %f507;
|
1208 |
+
div.full.f32 %r239, %r240, %r241;
|
1209 |
+
mov.b32 %f508, %r239;
|
1210 |
+
.loc 2 110 49
|
1211 |
+
selp.f32 %f509, 0f00000000, %f508, %p92;
|
1212 |
+
.loc 2 112 17
|
1213 |
+
fma.rn.f32 %f162, %f506, %f509, %f498;
|
1214 |
+
.loc 2 113 15
|
1215 |
+
add.f32 %f510, %f502, %f504;
|
1216 |
+
.loc 2 113 30
|
1217 |
+
mul.f32 %f511, %f506, %f506;
|
1218 |
+
.loc 2 113 38
|
1219 |
+
mul.f32 %f512, %f495, %f511;
|
1220 |
+
.loc 2 113 22
|
1221 |
+
fma.rn.f32 %f513, %f509, %f512, %f510;
|
1222 |
+
$L__tmp16:
|
1223 |
+
.loc 1 75 24
|
1224 |
+
mov.b32 %r243, %f478;
|
1225 |
+
mov.b32 %r244, 1132462080;
|
1226 |
+
div.full.f32 %r242, %r243, %r244;
|
1227 |
+
mov.b32 %f514, %r242;
|
1228 |
+
mov.b32 %r267, %f513;
|
1229 |
+
div.full.f32 %r266, %r267, %r244;
|
1230 |
+
mov.b32 %f515, %r266;
|
1231 |
+
.loc 1 77 24
|
1232 |
+
add.f32 %f163, %f514, 0f3727C5AC;
|
1233 |
+
add.f32 %f164, %f515, 0f3727C5AC;
|
1234 |
+
.loc 1 58 36
|
1235 |
+
add.s64 %rd9, %rd15, %rd2;
|
1236 |
+
mov.u64 %rd117, 0;
|
1237 |
+
mov.b32 %r473, -64;
|
1238 |
+
rsqrt.approx.ftz.f32 %f580, %f163;
|
1239 |
+
rsqrt.approx.ftz.f32 %f581, %f164;
|
1240 |
+
bra.uni $L__BB0_5;
|
1241 |
+
$L__BB0_7:
|
1242 |
+
.loc 1 69 35
|
1243 |
+
add.s64 %rd107, %rd4, %rd117;
|
1244 |
+
add.s64 %rd108, %rd107, 16;
|
1245 |
+
add.s64 %rd109, %rd3, %rd117;
|
1246 |
+
.loc 1 69 54
|
1247 |
+
add.s64 %rd110, %rd109, 16;
|
1248 |
+
mov.u32 %r407, 0x0;
|
1249 |
+
mov.u32 %r408, 0x0;
|
1250 |
+
mov.u32 %r409, 0x0;
|
1251 |
+
mov.u32 %r410, 0x0;
|
1252 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r407, %r408, %r409, %r410 }, [ %rd107 + 0 ];
|
1253 |
+
@!%p1 mov.u32 %r407, %r411;
|
1254 |
+
@!%p1 mov.u32 %r408, %r411;
|
1255 |
+
@!%p1 mov.u32 %r409, %r411;
|
1256 |
+
@!%p1 mov.u32 %r410, %r411;
|
1257 |
+
mov.b32 %f516, %r407;
|
1258 |
+
mov.b32 %f517, %r408;
|
1259 |
+
mov.b32 %f518, %r409;
|
1260 |
+
mov.b32 %f519, %r410;
|
1261 |
+
mov.u32 %r415, 0x0;
|
1262 |
+
mov.u32 %r416, 0x0;
|
1263 |
+
mov.u32 %r417, 0x0;
|
1264 |
+
mov.u32 %r418, 0x0;
|
1265 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r415, %r416, %r417, %r418 }, [ %rd108 + 0 ];
|
1266 |
+
@!%p1 mov.u32 %r415, %r411;
|
1267 |
+
@!%p1 mov.u32 %r416, %r411;
|
1268 |
+
@!%p1 mov.u32 %r417, %r411;
|
1269 |
+
@!%p1 mov.u32 %r418, %r411;
|
1270 |
+
mov.b32 %f520, %r415;
|
1271 |
+
mov.b32 %f521, %r416;
|
1272 |
+
mov.b32 %f522, %r417;
|
1273 |
+
mov.b32 %f523, %r418;
|
1274 |
+
mov.u32 %r423, 0x0;
|
1275 |
+
mov.u32 %r424, 0x0;
|
1276 |
+
mov.u32 %r425, 0x0;
|
1277 |
+
mov.u32 %r426, 0x0;
|
1278 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r423, %r424, %r425, %r426 }, [ %rd109 + 0 ];
|
1279 |
+
@!%p1 mov.u32 %r423, %r411;
|
1280 |
+
@!%p1 mov.u32 %r424, %r411;
|
1281 |
+
@!%p1 mov.u32 %r425, %r411;
|
1282 |
+
@!%p1 mov.u32 %r426, %r411;
|
1283 |
+
mov.b32 %f524, %r423;
|
1284 |
+
mov.b32 %f525, %r424;
|
1285 |
+
mov.b32 %f526, %r425;
|
1286 |
+
mov.b32 %f527, %r426;
|
1287 |
+
mov.u32 %r431, 0x0;
|
1288 |
+
mov.u32 %r432, 0x0;
|
1289 |
+
mov.u32 %r433, 0x0;
|
1290 |
+
mov.u32 %r434, 0x0;
|
1291 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r431, %r432, %r433, %r434 }, [ %rd110 + 0 ];
|
1292 |
+
@!%p1 mov.u32 %r431, %r411;
|
1293 |
+
@!%p1 mov.u32 %r432, %r411;
|
1294 |
+
@!%p1 mov.u32 %r433, %r411;
|
1295 |
+
@!%p1 mov.u32 %r434, %r411;
|
1296 |
+
mov.b32 %f528, %r431;
|
1297 |
+
mov.b32 %f529, %r432;
|
1298 |
+
mov.b32 %f530, %r433;
|
1299 |
+
mov.b32 %f531, %r434;
|
1300 |
+
.loc 1 70 24
|
1301 |
+
add.f32 %f532, %f165, %f516;
|
1302 |
+
add.f32 %f533, %f166, %f517;
|
1303 |
+
add.f32 %f534, %f167, %f518;
|
1304 |
+
add.f32 %f535, %f168, %f519;
|
1305 |
+
add.f32 %f536, %f169, %f520;
|
1306 |
+
add.f32 %f537, %f170, %f521;
|
1307 |
+
add.f32 %f538, %f171, %f522;
|
1308 |
+
add.f32 %f539, %f172, %f523;
|
1309 |
+
add.f32 %f540, %f173, %f524;
|
1310 |
+
add.f32 %f541, %f174, %f525;
|
1311 |
+
add.f32 %f542, %f175, %f526;
|
1312 |
+
add.f32 %f543, %f176, %f527;
|
1313 |
+
add.f32 %f544, %f177, %f528;
|
1314 |
+
add.f32 %f545, %f178, %f529;
|
1315 |
+
add.f32 %f546, %f179, %f530;
|
1316 |
+
add.f32 %f547, %f180, %f531;
|
1317 |
+
.loc 1 72 24
|
1318 |
+
add.f32 %f548, %f181, %f532;
|
1319 |
+
add.f32 %f549, %f182, %f533;
|
1320 |
+
add.f32 %f550, %f183, %f534;
|
1321 |
+
add.f32 %f551, %f184, %f535;
|
1322 |
+
add.f32 %f552, %f185, %f536;
|
1323 |
+
add.f32 %f553, %f186, %f537;
|
1324 |
+
add.f32 %f554, %f187, %f538;
|
1325 |
+
add.f32 %f555, %f188, %f539;
|
1326 |
+
add.f32 %f556, %f189, %f540;
|
1327 |
+
add.f32 %f557, %f190, %f541;
|
1328 |
+
add.f32 %f558, %f191, %f542;
|
1329 |
+
add.f32 %f559, %f192, %f543;
|
1330 |
+
add.f32 %f560, %f193, %f544;
|
1331 |
+
add.f32 %f561, %f194, %f545;
|
1332 |
+
add.f32 %f562, %f195, %f546;
|
1333 |
+
add.f32 %f563, %f196, %f547;
|
1334 |
+
.loc 1 73 24
|
1335 |
+
sub.f32 %f564, %f548, %f161;
|
1336 |
+
sub.f32 %f565, %f549, %f161;
|
1337 |
+
sub.f32 %f566, %f550, %f161;
|
1338 |
+
sub.f32 %f567, %f551, %f161;
|
1339 |
+
sub.f32 %f568, %f552, %f161;
|
1340 |
+
sub.f32 %f569, %f553, %f161;
|
1341 |
+
sub.f32 %f570, %f554, %f161;
|
1342 |
+
sub.f32 %f571, %f555, %f161;
|
1343 |
+
sub.f32 %f572, %f556, %f162;
|
1344 |
+
sub.f32 %f573, %f557, %f162;
|
1345 |
+
sub.f32 %f574, %f558, %f162;
|
1346 |
+
sub.f32 %f575, %f559, %f162;
|
1347 |
+
sub.f32 %f576, %f560, %f162;
|
1348 |
+
sub.f32 %f577, %f561, %f162;
|
1349 |
+
sub.f32 %f578, %f562, %f162;
|
1350 |
+
sub.f32 %f579, %f563, %f162;
|
1351 |
+
.loc 1 79 24
|
1352 |
+
mul.f32 %f582, %f564, %f580;
|
1353 |
+
mul.f32 %f583, %f565, %f580;
|
1354 |
+
mul.f32 %f584, %f566, %f580;
|
1355 |
+
mul.f32 %f585, %f567, %f580;
|
1356 |
+
mul.f32 %f586, %f568, %f580;
|
1357 |
+
mul.f32 %f587, %f569, %f580;
|
1358 |
+
mul.f32 %f588, %f570, %f580;
|
1359 |
+
mul.f32 %f589, %f571, %f580;
|
1360 |
+
mul.f32 %f590, %f572, %f581;
|
1361 |
+
mul.f32 %f591, %f573, %f581;
|
1362 |
+
mul.f32 %f592, %f574, %f581;
|
1363 |
+
mul.f32 %f593, %f575, %f581;
|
1364 |
+
mul.f32 %f594, %f576, %f581;
|
1365 |
+
mul.f32 %f595, %f577, %f581;
|
1366 |
+
mul.f32 %f596, %f578, %f581;
|
1367 |
+
mul.f32 %f597, %f579, %f581;
|
1368 |
+
.loc 1 80 24
|
1369 |
+
mul.f32 %f598, %f582, %f197;
|
1370 |
+
mul.f32 %f599, %f583, %f198;
|
1371 |
+
mul.f32 %f600, %f584, %f199;
|
1372 |
+
mul.f32 %f601, %f585, %f200;
|
1373 |
+
mul.f32 %f602, %f586, %f201;
|
1374 |
+
mul.f32 %f603, %f587, %f202;
|
1375 |
+
mul.f32 %f604, %f588, %f203;
|
1376 |
+
mul.f32 %f605, %f589, %f204;
|
1377 |
+
mul.f32 %f606, %f590, %f197;
|
1378 |
+
mul.f32 %f607, %f591, %f198;
|
1379 |
+
mul.f32 %f608, %f592, %f199;
|
1380 |
+
mul.f32 %f609, %f593, %f200;
|
1381 |
+
mul.f32 %f610, %f594, %f201;
|
1382 |
+
mul.f32 %f611, %f595, %f202;
|
1383 |
+
mul.f32 %f612, %f596, %f203;
|
1384 |
+
mul.f32 %f613, %f597, %f204;
|
1385 |
+
.loc 1 82 29
|
1386 |
+
shl.b64 %rd113, %rd11, 1;
|
1387 |
+
add.s64 %rd111, %rd16, %rd113;
|
1388 |
+
shl.b64 %rd114, %rd12, 1;
|
1389 |
+
add.s64 %rd112, %rd16, %rd114;
|
1390 |
+
.loc 1 82 52
|
1391 |
+
mov.b32 %r439, %f598;
|
1392 |
+
cvt.rn.bf16.f32 %rs33, %r439;
|
1393 |
+
mov.b32 %r440, %f599;
|
1394 |
+
cvt.rn.bf16.f32 %rs34, %r440;
|
1395 |
+
mov.b32 %r441, %f600;
|
1396 |
+
cvt.rn.bf16.f32 %rs35, %r441;
|
1397 |
+
mov.b32 %r442, %f601;
|
1398 |
+
cvt.rn.bf16.f32 %rs36, %r442;
|
1399 |
+
mov.b32 %r443, %f602;
|
1400 |
+
cvt.rn.bf16.f32 %rs37, %r443;
|
1401 |
+
mov.b32 %r444, %f603;
|
1402 |
+
cvt.rn.bf16.f32 %rs38, %r444;
|
1403 |
+
mov.b32 %r445, %f604;
|
1404 |
+
cvt.rn.bf16.f32 %rs39, %r445;
|
1405 |
+
mov.b32 %r446, %f605;
|
1406 |
+
cvt.rn.bf16.f32 %rs40, %r446;
|
1407 |
+
mov.b32 %r447, %f606;
|
1408 |
+
cvt.rn.bf16.f32 %rs41, %r447;
|
1409 |
+
mov.b32 %r448, %f607;
|
1410 |
+
cvt.rn.bf16.f32 %rs42, %r448;
|
1411 |
+
mov.b32 %r449, %f608;
|
1412 |
+
cvt.rn.bf16.f32 %rs43, %r449;
|
1413 |
+
mov.b32 %r450, %f609;
|
1414 |
+
cvt.rn.bf16.f32 %rs44, %r450;
|
1415 |
+
mov.b32 %r451, %f610;
|
1416 |
+
cvt.rn.bf16.f32 %rs45, %r451;
|
1417 |
+
mov.b32 %r452, %f611;
|
1418 |
+
cvt.rn.bf16.f32 %rs46, %r452;
|
1419 |
+
mov.b32 %r453, %f612;
|
1420 |
+
cvt.rn.bf16.f32 %rs47, %r453;
|
1421 |
+
mov.b32 %r454, %f613;
|
1422 |
+
cvt.rn.bf16.f32 %rs48, %r454;
|
1423 |
+
mov.b32 %r463, {%rs33, %rs34};
|
1424 |
+
mov.b32 %r464, {%rs35, %rs36};
|
1425 |
+
mov.b32 %r465, {%rs37, %rs38};
|
1426 |
+
mov.b32 %r466, {%rs39, %rs40};
|
1427 |
+
@%p1 st.global.v4.b32 [ %rd111 + 0 ], { %r463, %r464, %r465, %r466 };
|
1428 |
+
mov.b32 %r467, {%rs41, %rs42};
|
1429 |
+
mov.b32 %r468, {%rs43, %rs44};
|
1430 |
+
mov.b32 %r469, {%rs45, %rs46};
|
1431 |
+
mov.b32 %r470, {%rs47, %rs48};
|
1432 |
+
@%p1 st.global.v4.b32 [ %rd112 + 0 ], { %r467, %r468, %r469, %r470 };
|
1433 |
+
.loc 1 58 36
|
1434 |
+
add.s64 %rd117, %rd117, 256;
|
1435 |
+
add.s32 %r473, %r473, 64;
|
1436 |
+
setp.lt.u32 %p156, %r473, 192;
|
1437 |
+
@%p156 bra $L__BB0_5;
|
1438 |
+
bra.uni $L__BB0_8;
|
1439 |
+
$L__BB0_5:
|
1440 |
+
.loc 1 62 35
|
1441 |
+
add.s64 %rd90, %rd6, %rd117;
|
1442 |
+
add.s64 %rd91, %rd90, 16;
|
1443 |
+
add.s64 %rd92, %rd5, %rd117;
|
1444 |
+
.loc 1 62 51
|
1445 |
+
add.s64 %rd93, %rd92, 16;
|
1446 |
+
mov.u32 %r323, 0x0;
|
1447 |
+
mov.u32 %r324, 0x0;
|
1448 |
+
mov.u32 %r325, 0x0;
|
1449 |
+
mov.u32 %r326, 0x0;
|
1450 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r323, %r324, %r325, %r326 }, [ %rd90 + 0 ];
|
1451 |
+
@!%p1 mov.u32 %r323, %r411;
|
1452 |
+
@!%p1 mov.u32 %r324, %r411;
|
1453 |
+
@!%p1 mov.u32 %r325, %r411;
|
1454 |
+
@!%p1 mov.u32 %r326, %r411;
|
1455 |
+
mov.b32 %f165, %r323;
|
1456 |
+
mov.b32 %f166, %r324;
|
1457 |
+
mov.b32 %f167, %r325;
|
1458 |
+
mov.b32 %f168, %r326;
|
1459 |
+
mov.u32 %r331, 0x0;
|
1460 |
+
mov.u32 %r332, 0x0;
|
1461 |
+
mov.u32 %r333, 0x0;
|
1462 |
+
mov.u32 %r334, 0x0;
|
1463 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r331, %r332, %r333, %r334 }, [ %rd91 + 0 ];
|
1464 |
+
@!%p1 mov.u32 %r331, %r411;
|
1465 |
+
@!%p1 mov.u32 %r332, %r411;
|
1466 |
+
@!%p1 mov.u32 %r333, %r411;
|
1467 |
+
@!%p1 mov.u32 %r334, %r411;
|
1468 |
+
mov.b32 %f169, %r331;
|
1469 |
+
mov.b32 %f170, %r332;
|
1470 |
+
mov.b32 %f171, %r333;
|
1471 |
+
mov.b32 %f172, %r334;
|
1472 |
+
mov.u32 %r339, 0x0;
|
1473 |
+
mov.u32 %r340, 0x0;
|
1474 |
+
mov.u32 %r341, 0x0;
|
1475 |
+
mov.u32 %r342, 0x0;
|
1476 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r339, %r340, %r341, %r342 }, [ %rd92 + 0 ];
|
1477 |
+
@!%p1 mov.u32 %r339, %r411;
|
1478 |
+
@!%p1 mov.u32 %r340, %r411;
|
1479 |
+
@!%p1 mov.u32 %r341, %r411;
|
1480 |
+
@!%p1 mov.u32 %r342, %r411;
|
1481 |
+
mov.b32 %f173, %r339;
|
1482 |
+
mov.b32 %f174, %r340;
|
1483 |
+
mov.b32 %f175, %r341;
|
1484 |
+
mov.b32 %f176, %r342;
|
1485 |
+
mov.u32 %r347, 0x0;
|
1486 |
+
mov.u32 %r348, 0x0;
|
1487 |
+
mov.u32 %r349, 0x0;
|
1488 |
+
mov.u32 %r350, 0x0;
|
1489 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r347, %r348, %r349, %r350 }, [ %rd93 + 0 ];
|
1490 |
+
@!%p1 mov.u32 %r347, %r411;
|
1491 |
+
@!%p1 mov.u32 %r348, %r411;
|
1492 |
+
@!%p1 mov.u32 %r349, %r411;
|
1493 |
+
@!%p1 mov.u32 %r350, %r411;
|
1494 |
+
mov.b32 %f177, %r347;
|
1495 |
+
mov.b32 %f178, %r348;
|
1496 |
+
mov.b32 %f179, %r349;
|
1497 |
+
mov.b32 %f180, %r350;
|
1498 |
+
.loc 1 63 41
|
1499 |
+
add.s32 %r403, %r5, %r473;
|
1500 |
+
add.s32 %r404, %r403, 64;
|
1501 |
+
.loc 1 63 35
|
1502 |
+
add.s32 %r405, %r403, 8256;
|
1503 |
+
cvt.s64.s32 %rd11, %r404;
|
1504 |
+
mul.wide.s32 %rd98, %r404, 2;
|
1505 |
+
add.s64 %rd94, %rd14, %rd98;
|
1506 |
+
cvt.s64.s32 %rd12, %r405;
|
1507 |
+
mul.wide.s32 %rd99, %r405, 2;
|
1508 |
+
add.s64 %rd95, %rd14, %rd99;
|
1509 |
+
.loc 1 63 51
|
1510 |
+
mov.u32 %r355, 0x0;
|
1511 |
+
mov.u32 %r356, 0x0;
|
1512 |
+
mov.u32 %r357, 0x0;
|
1513 |
+
mov.u32 %r358, 0x0;
|
1514 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r355, %r356, %r357, %r358 }, [ %rd94 + 0 ];
|
1515 |
+
@!%p1 mov.u32 %r355, %r411;
|
1516 |
+
@!%p1 mov.u32 %r356, %r411;
|
1517 |
+
@!%p1 mov.u32 %r357, %r411;
|
1518 |
+
@!%p1 mov.u32 %r358, %r411;
|
1519 |
+
cvt.u16.u32 %rs17, %r355;
|
1520 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs18}, %r355; }
|
1521 |
+
cvt.u16.u32 %rs19, %r356;
|
1522 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs20}, %r356; }
|
1523 |
+
cvt.u16.u32 %rs21, %r357;
|
1524 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs22}, %r357; }
|
1525 |
+
cvt.u16.u32 %rs23, %r358;
|
1526 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs24}, %r358; }
|
1527 |
+
mov.u32 %r363, 0x0;
|
1528 |
+
mov.u32 %r364, 0x0;
|
1529 |
+
mov.u32 %r365, 0x0;
|
1530 |
+
mov.u32 %r366, 0x0;
|
1531 |
+
@%p1 ld.global.L1::evict_first.v4.b32 { %r363, %r364, %r365, %r366 }, [ %rd95 + 0 ];
|
1532 |
+
@!%p1 mov.u32 %r363, %r411;
|
1533 |
+
@!%p1 mov.u32 %r364, %r411;
|
1534 |
+
@!%p1 mov.u32 %r365, %r411;
|
1535 |
+
@!%p1 mov.u32 %r366, %r411;
|
1536 |
+
cvt.u16.u32 %rs25, %r363;
|
1537 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs26}, %r363; }
|
1538 |
+
cvt.u16.u32 %rs27, %r364;
|
1539 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs28}, %r364; }
|
1540 |
+
cvt.u16.u32 %rs29, %r365;
|
1541 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs30}, %r365; }
|
1542 |
+
cvt.u16.u32 %rs31, %r366;
|
1543 |
+
{ .reg .b16 tmp; mov.b32 {tmp, %rs32}, %r366; }
|
1544 |
+
.loc 1 63 103
|
1545 |
+
cvt.f32.bf16 %r371, %rs17;
|
1546 |
+
mov.b32 %f181, %r371;
|
1547 |
+
cvt.f32.bf16 %r372, %rs18;
|
1548 |
+
mov.b32 %f182, %r372;
|
1549 |
+
cvt.f32.bf16 %r373, %rs19;
|
1550 |
+
mov.b32 %f183, %r373;
|
1551 |
+
cvt.f32.bf16 %r374, %rs20;
|
1552 |
+
mov.b32 %f184, %r374;
|
1553 |
+
cvt.f32.bf16 %r375, %rs21;
|
1554 |
+
mov.b32 %f185, %r375;
|
1555 |
+
cvt.f32.bf16 %r376, %rs22;
|
1556 |
+
mov.b32 %f186, %r376;
|
1557 |
+
cvt.f32.bf16 %r377, %rs23;
|
1558 |
+
mov.b32 %f187, %r377;
|
1559 |
+
cvt.f32.bf16 %r378, %rs24;
|
1560 |
+
mov.b32 %f188, %r378;
|
1561 |
+
cvt.f32.bf16 %r379, %rs25;
|
1562 |
+
mov.b32 %f189, %r379;
|
1563 |
+
cvt.f32.bf16 %r380, %rs26;
|
1564 |
+
mov.b32 %f190, %r380;
|
1565 |
+
cvt.f32.bf16 %r381, %rs27;
|
1566 |
+
mov.b32 %f191, %r381;
|
1567 |
+
cvt.f32.bf16 %r382, %rs28;
|
1568 |
+
mov.b32 %f192, %r382;
|
1569 |
+
cvt.f32.bf16 %r383, %rs29;
|
1570 |
+
mov.b32 %f193, %r383;
|
1571 |
+
cvt.f32.bf16 %r384, %rs30;
|
1572 |
+
mov.b32 %f194, %r384;
|
1573 |
+
cvt.f32.bf16 %r385, %rs31;
|
1574 |
+
mov.b32 %f195, %r385;
|
1575 |
+
cvt.f32.bf16 %r386, %rs32;
|
1576 |
+
mov.b32 %f196, %r386;
|
1577 |
+
.loc 1 64 35
|
1578 |
+
add.s64 %rd96, %rd9, %rd117;
|
1579 |
+
.loc 1 64 40
|
1580 |
+
add.s64 %rd97, %rd96, 16;
|
1581 |
+
mov.u32 %r387, 0x0;
|
1582 |
+
mov.u32 %r388, 0x0;
|
1583 |
+
mov.u32 %r389, 0x0;
|
1584 |
+
mov.u32 %r390, 0x0;
|
1585 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r387, %r388, %r389, %r390 }, [ %rd96 + 0 ];
|
1586 |
+
@!%p1 mov.u32 %r387, %r411;
|
1587 |
+
@!%p1 mov.u32 %r388, %r411;
|
1588 |
+
@!%p1 mov.u32 %r389, %r411;
|
1589 |
+
@!%p1 mov.u32 %r390, %r411;
|
1590 |
+
mov.b32 %f197, %r387;
|
1591 |
+
mov.b32 %f198, %r388;
|
1592 |
+
mov.b32 %f199, %r389;
|
1593 |
+
mov.b32 %f200, %r390;
|
1594 |
+
mov.u32 %r395, 0x0;
|
1595 |
+
mov.u32 %r396, 0x0;
|
1596 |
+
mov.u32 %r397, 0x0;
|
1597 |
+
mov.u32 %r398, 0x0;
|
1598 |
+
@%p1 ld.global.L1::evict_last.v4.b32 { %r395, %r396, %r397, %r398 }, [ %rd97 + 0 ];
|
1599 |
+
@!%p1 mov.u32 %r395, %r411;
|
1600 |
+
@!%p1 mov.u32 %r396, %r411;
|
1601 |
+
@!%p1 mov.u32 %r397, %r411;
|
1602 |
+
@!%p1 mov.u32 %r398, %r411;
|
1603 |
+
mov.b32 %f201, %r395;
|
1604 |
+
mov.b32 %f202, %r396;
|
1605 |
+
mov.b32 %f203, %r397;
|
1606 |
+
mov.b32 %f204, %r398;
|
1607 |
+
.loc 1 68 57
|
1608 |
+
@%p51 bra $L__BB0_7;
|
1609 |
+
mov.u64 %rd100, assertMessage_1;
|
1610 |
+
cvta.global.u64 %rd101, %rd100;
|
1611 |
+
mov.u64 %rd102, assertFile_1;
|
1612 |
+
cvta.global.u64 %rd103, %rd102;
|
1613 |
+
mov.u64 %rd104, assertFunc_1;
|
1614 |
+
cvta.global.u64 %rd105, %rd104;
|
1615 |
+
{ // callseq 7, 0
|
1616 |
+
.reg .b32 temp_param_reg;
|
1617 |
+
.param .b64 param0;
|
1618 |
+
st.param.b64 [param0+0], %rd101;
|
1619 |
+
.param .b64 param1;
|
1620 |
+
st.param.b64 [param1+0], %rd103;
|
1621 |
+
.param .b32 param2;
|
1622 |
+
st.param.b32 [param2+0], %r471;
|
1623 |
+
.param .b64 param3;
|
1624 |
+
st.param.b64 [param3+0], %rd105;
|
1625 |
+
.param .b64 param4;
|
1626 |
+
st.param.b64 [param4+0], %rd115;
|
1627 |
+
call.uni
|
1628 |
+
__assertfail,
|
1629 |
+
(
|
1630 |
+
param0,
|
1631 |
+
param1,
|
1632 |
+
param2,
|
1633 |
+
param3,
|
1634 |
+
param4
|
1635 |
+
);
|
1636 |
+
} // callseq 7
|
1637 |
+
bra.uni $L__BB0_7;
|
1638 |
+
$L__BB0_8:
|
1639 |
+
.loc 1 58 4
|
1640 |
+
ret;
|
1641 |
+
$L__tmp17:
|
1642 |
+
$L__func_end0:
|
1643 |
+
|
1644 |
+
}
|
1645 |
+
// .globl __nv_rsqrtf
|
1646 |
+
.visible .func (.param .b32 func_retval0) __nv_rsqrtf(
|
1647 |
+
.param .b32 __nv_rsqrtf_param_0
|
1648 |
+
)
|
1649 |
+
{
|
1650 |
+
.reg .f32 %f<3>;
|
1651 |
+
$L__func_begin1:
|
1652 |
+
|
1653 |
+
ld.param.f32 %f1, [__nv_rsqrtf_param_0];
|
1654 |
+
rsqrt.approx.ftz.f32 %f2, %f1;
|
1655 |
+
st.param.f32 [func_retval0+0], %f2;
|
1656 |
+
ret;
|
1657 |
+
$L__func_end1:
|
1658 |
+
|
1659 |
+
}
|
1660 |
+
.file 1 "/tmp/torchinductor_root/ci/ccig6fki6p4lxrdmgg6eudahiexcvueeol2p4qp532pvve2y463y.py"
|
1661 |
+
.file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
|
1662 |
+
.section .debug_abbrev
|
1663 |
+
{
|
1664 |
+
.b8 1
|
1665 |
+
.b8 17
|
1666 |
+
.b8 1
|
1667 |
+
.b8 37
|
1668 |
+
.b8 8
|
1669 |
+
.b8 19
|
1670 |
+
.b8 5
|
1671 |
+
.b8 3
|
1672 |
+
.b8 8
|
1673 |
+
.b8 16
|
1674 |
+
.b8 6
|
1675 |
+
.b8 27
|
1676 |
+
.b8 8
|
1677 |
+
.b8 180
|
1678 |
+
.b8 66
|
1679 |
+
.b8 12
|
1680 |
+
.b8 17
|
1681 |
+
.b8 1
|
1682 |
+
.b8 18
|
1683 |
+
.b8 1
|
1684 |
+
.b8 0
|
1685 |
+
.b8 0
|
1686 |
+
.b8 2
|
1687 |
+
.b8 46
|
1688 |
+
.b8 0
|
1689 |
+
.b8 135
|
1690 |
+
.b8 64
|
1691 |
+
.b8 8
|
1692 |
+
.b8 3
|
1693 |
+
.b8 8
|
1694 |
+
.b8 58
|
1695 |
+
.b8 11
|
1696 |
+
.b8 59
|
1697 |
+
.b8 11
|
1698 |
+
.b8 63
|
1699 |
+
.b8 12
|
1700 |
+
.b8 32
|
1701 |
+
.b8 11
|
1702 |
+
.b8 0
|
1703 |
+
.b8 0
|
1704 |
+
.b8 3
|
1705 |
+
.b8 46
|
1706 |
+
.b8 1
|
1707 |
+
.b8 17
|
1708 |
+
.b8 1
|
1709 |
+
.b8 18
|
1710 |
+
.b8 1
|
1711 |
+
.b8 64
|
1712 |
+
.b8 10
|
1713 |
+
.b8 49
|
1714 |
+
.b8 19
|
1715 |
+
.b8 0
|
1716 |
+
.b8 0
|
1717 |
+
.b8 4
|
1718 |
+
.b8 29
|
1719 |
+
.b8 0
|
1720 |
+
.b8 49
|
1721 |
+
.b8 19
|
1722 |
+
.b8 17
|
1723 |
+
.b8 1
|
1724 |
+
.b8 18
|
1725 |
+
.b8 1
|
1726 |
+
.b8 88
|
1727 |
+
.b8 11
|
1728 |
+
.b8 89
|
1729 |
+
.b8 11
|
1730 |
+
.b8 87
|
1731 |
+
.b8 11
|
1732 |
+
.b8 0
|
1733 |
+
.b8 0
|
1734 |
+
.b8 5
|
1735 |
+
.b8 29
|
1736 |
+
.b8 1
|
1737 |
+
.b8 49
|
1738 |
+
.b8 19
|
1739 |
+
.b8 17
|
1740 |
+
.b8 1
|
1741 |
+
.b8 18
|
1742 |
+
.b8 1
|
1743 |
+
.b8 88
|
1744 |
+
.b8 11
|
1745 |
+
.b8 89
|
1746 |
+
.b8 11
|
1747 |
+
.b8 87
|
1748 |
+
.b8 11
|
1749 |
+
.b8 0
|
1750 |
+
.b8 0
|
1751 |
+
.b8 0
|
1752 |
+
}
|
1753 |
+
.section .debug_info
|
1754 |
+
{
|
1755 |
+
.b32 302
|
1756 |
+
.b8 2
|
1757 |
+
.b8 0
|
1758 |
+
.b32 .debug_abbrev
|
1759 |
+
.b8 8
|
1760 |
+
.b8 1
|
1761 |
+
.b8 116
|
1762 |
+
.b8 114
|
1763 |
+
.b8 105
|
1764 |
+
.b8 116
|
1765 |
+
.b8 111
|
1766 |
+
.b8 110
|
1767 |
+
.b8 0
|
1768 |
+
.b8 2
|
1769 |
+
.b8 0
|
1770 |
+
.b8 99
|
1771 |
+
.b8 99
|
1772 |
+
.b8 105
|
1773 |
+
.b8 103
|
1774 |
+
.b8 54
|
1775 |
+
.b8 102
|
1776 |
+
.b8 107
|
1777 |
+
.b8 105
|
1778 |
+
.b8 54
|
1779 |
+
.b8 112
|
1780 |
+
.b8 52
|
1781 |
+
.b8 108
|
1782 |
+
.b8 120
|
1783 |
+
.b8 114
|
1784 |
+
.b8 100
|
1785 |
+
.b8 109
|
1786 |
+
.b8 103
|
1787 |
+
.b8 103
|
1788 |
+
.b8 54
|
1789 |
+
.b8 101
|
1790 |
+
.b8 117
|
1791 |
+
.b8 100
|
1792 |
+
.b8 97
|
1793 |
+
.b8 104
|
1794 |
+
.b8 105
|
1795 |
+
.b8 101
|
1796 |
+
.b8 120
|
1797 |
+
.b8 99
|
1798 |
+
.b8 118
|
1799 |
+
.b8 117
|
1800 |
+
.b8 101
|
1801 |
+
.b8 101
|
1802 |
+
.b8 111
|
1803 |
+
.b8 108
|
1804 |
+
.b8 50
|
1805 |
+
.b8 112
|
1806 |
+
.b8 52
|
1807 |
+
.b8 113
|
1808 |
+
.b8 112
|
1809 |
+
.b8 53
|
1810 |
+
.b8 51
|
1811 |
+
.b8 50
|
1812 |
+
.b8 112
|
1813 |
+
.b8 118
|
1814 |
+
.b8 118
|
1815 |
+
.b8 101
|
1816 |
+
.b8 50
|
1817 |
+
.b8 121
|
1818 |
+
.b8 52
|
1819 |
+
.b8 54
|
1820 |
+
.b8 51
|
1821 |
+
.b8 121
|
1822 |
+
.b8 46
|
1823 |
+
.b8 112
|
1824 |
+
.b8 121
|
1825 |
+
.b8 0
|
1826 |
+
.b32 .debug_line
|
1827 |
+
.b8 47
|
1828 |
+
.b8 116
|
1829 |
+
.b8 109
|
1830 |
+
.b8 112
|
1831 |
+
.b8 47
|
1832 |
+
.b8 116
|
1833 |
+
.b8 111
|
1834 |
+
.b8 114
|
1835 |
+
.b8 99
|
1836 |
+
.b8 104
|
1837 |
+
.b8 105
|
1838 |
+
.b8 110
|
1839 |
+
.b8 100
|
1840 |
+
.b8 117
|
1841 |
+
.b8 99
|
1842 |
+
.b8 116
|
1843 |
+
.b8 111
|
1844 |
+
.b8 114
|
1845 |
+
.b8 95
|
1846 |
+
.b8 114
|
1847 |
+
.b8 111
|
1848 |
+
.b8 111
|
1849 |
+
.b8 116
|
1850 |
+
.b8 47
|
1851 |
+
.b8 99
|
1852 |
+
.b8 105
|
1853 |
+
.b8 0
|
1854 |
+
.b8 1
|
1855 |
+
.b64 $L__func_begin0
|
1856 |
+
.b64 $L__func_end0
|
1857 |
+
.b8 2
|
1858 |
+
.b8 116
|
1859 |
+
.b8 114
|
1860 |
+
.b8 105
|
1861 |
+
.b8 116
|
1862 |
+
.b8 111
|
1863 |
+
.b8 110
|
1864 |
+
.b8 95
|
1865 |
+
.b8 95
|
1866 |
+
.b8 48
|
1867 |
+
.b8 100
|
1868 |
+
.b8 49
|
1869 |
+
.b8 100
|
1870 |
+
.b8 50
|
1871 |
+
.b8 100
|
1872 |
+
.b8 51
|
1873 |
+
.b8 100
|
1874 |
+
.b8 52
|
1875 |
+
.b8 100
|
1876 |
+
.b8 53
|
1877 |
+
.b8 100
|
1878 |
+
.b8 54
|
1879 |
+
.b8 100
|
1880 |
+
.b8 101
|
1881 |
+
.b8 55
|
1882 |
+
.b8 100
|
1883 |
+
.b8 101
|
1884 |
+
.b8 0
|
1885 |
+
.b8 116
|
1886 |
+
.b8 114
|
1887 |
+
.b8 105
|
1888 |
+
.b8 116
|
1889 |
+
.b8 111
|
1890 |
+
.b8 110
|
1891 |
+
.b8 95
|
1892 |
+
.b8 95
|
1893 |
+
.b8 48
|
1894 |
+
.b8 100
|
1895 |
+
.b8 49
|
1896 |
+
.b8 100
|
1897 |
+
.b8 50
|
1898 |
+
.b8 100
|
1899 |
+
.b8 51
|
1900 |
+
.b8 100
|
1901 |
+
.b8 52
|
1902 |
+
.b8 100
|
1903 |
+
.b8 53
|
1904 |
+
.b8 100
|
1905 |
+
.b8 54
|
1906 |
+
.b8 100
|
1907 |
+
.b8 101
|
1908 |
+
.b8 55
|
1909 |
+
.b8 100
|
1910 |
+
.b8 101
|
1911 |
+
.b8 0
|
1912 |
+
.b8 1
|
1913 |
+
.b8 18
|
1914 |
+
.b8 1
|
1915 |
+
.b8 1
|
1916 |
+
.b8 3
|
1917 |
+
.b64 $L__func_begin0
|
1918 |
+
.b64 $L__func_end0
|
1919 |
+
.b8 1
|
1920 |
+
.b8 156
|
1921 |
+
.b32 125
|
1922 |
+
.b8 4
|
1923 |
+
.b32 125
|
1924 |
+
.b64 $L__tmp1
|
1925 |
+
.b64 $L__tmp2
|
1926 |
+
.b8 2
|
1927 |
+
.b8 47
|
1928 |
+
.b8 41
|
1929 |
+
.b8 5
|
1930 |
+
.b32 125
|
1931 |
+
.b64 $L__tmp3
|
1932 |
+
.b64 $L__tmp16
|
1933 |
+
.b8 2
|
1934 |
+
.b8 53
|
1935 |
+
.b8 44
|
1936 |
+
.b8 4
|
1937 |
+
.b32 125
|
1938 |
+
.b64 $L__tmp3
|
1939 |
+
.b64 $L__tmp16
|
1940 |
+
.b8 2
|
1941 |
+
.b8 120
|
1942 |
+
.b8 46
|
1943 |
+
.b8 0
|
1944 |
+
.b8 4
|
1945 |
+
.b32 125
|
1946 |
+
.b64 $L__tmp4
|
1947 |
+
.b64 $L__tmp15
|
1948 |
+
.b8 2
|
1949 |
+
.b8 53
|
1950 |
+
.b8 44
|
1951 |
+
.b8 0
|
1952 |
+
.b8 0
|
1953 |
+
}
|
1954 |
+
.section .debug_pubnames
|
1955 |
+
{
|
1956 |
+
.b32 $L__pubNames_end0-$L__pubNames_start0
|
1957 |
+
$L__pubNames_start0:
|
1958 |
+
.b8 2
|
1959 |
+
.b8 0
|
1960 |
+
.b32 .debug_info
|
1961 |
+
.b32 306
|
1962 |
+
.b32 125
|
1963 |
+
.b8 116
|
1964 |
+
.b8 114
|
1965 |
+
.b8 105
|
1966 |
+
.b8 116
|
1967 |
+
.b8 111
|
1968 |
+
.b8 110
|
1969 |
+
.b8 95
|
1970 |
+
.b8 95
|
1971 |
+
.b8 48
|
1972 |
+
.b8 100
|
1973 |
+
.b8 49
|
1974 |
+
.b8 100
|
1975 |
+
.b8 50
|
1976 |
+
.b8 100
|
1977 |
+
.b8 51
|
1978 |
+
.b8 100
|
1979 |
+
.b8 52
|
1980 |
+
.b8 100
|
1981 |
+
.b8 53
|
1982 |
+
.b8 100
|
1983 |
+
.b8 54
|
1984 |
+
.b8 100
|
1985 |
+
.b8 101
|
1986 |
+
.b8 55
|
1987 |
+
.b8 100
|
1988 |
+
.b8 101
|
1989 |
+
.b8 0
|
1990 |
+
.b32 0
|
1991 |
+
$L__pubNames_end0:
|
1992 |
+
}
|
1993 |
+
.section .debug_pubtypes
|
1994 |
+
{
|
1995 |
+
.b32 $L__pubTypes_end0-$L__pubTypes_start0
|
1996 |
+
$L__pubTypes_start0:
|
1997 |
+
.b8 2
|
1998 |
+
.b8 0
|
1999 |
+
.b32 .debug_info
|
2000 |
+
.b32 306
|
2001 |
+
.b32 0
|
2002 |
+
$L__pubTypes_end0:
|
2003 |
+
}
|
2004 |
+
.section .debug_loc { }
|
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttgir
ADDED
@@ -0,0 +1,164 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [1, 8], threadsPerWarp = [4, 8], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
2 |
+
#blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
3 |
+
#blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [4, 2], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
4 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 8 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
5 |
+
tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
6 |
+
%cst = arith.constant dense<512> : tensor<64x1xi32, #blocked>
|
7 |
+
%cst_0 = arith.constant dense<256> : tensor<1x64xi32, #blocked>
|
8 |
+
%cst_1 = arith.constant dense<256> : tensor<64x1xi32, #blocked>
|
9 |
+
%cst_2 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked>
|
10 |
+
%cst_3 = arith.constant dense<0.000000e+00> : tensor<1x64xf32, #blocked>
|
11 |
+
%cst_4 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked>
|
12 |
+
%cst_5 = arith.constant dense<256> : tensor<64x1xi64, #blocked>
|
13 |
+
%cst_6 = arith.constant dense<0> : tensor<64x1xi64, #blocked>
|
14 |
+
%cst_7 = arith.constant dense<50257> : tensor<64x1xi64, #blocked>
|
15 |
+
%cst_8 = arith.constant dense<50257> : tensor<64x1xi64, #blocked1>
|
16 |
+
%cst_9 = arith.constant dense<0> : tensor<64x1xi64, #blocked1>
|
17 |
+
%c0_i32 = arith.constant 0 : i32
|
18 |
+
%c64_i32 = arith.constant 64 : i32
|
19 |
+
%c256_i32 = arith.constant 256 : i32
|
20 |
+
%cst_10 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked2>
|
21 |
+
%cst_11 = arith.constant 0.000000e+00 : f32
|
22 |
+
%cst_12 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked2>
|
23 |
+
%cst_13 = arith.constant dense<256> : tensor<1x64xi32, #blocked2>
|
24 |
+
%cst_14 = arith.constant dense<9.99999974E-6> : tensor<64x1xf32, #blocked>
|
25 |
+
%cst_15 = arith.constant dense<2.560000e+02> : tensor<64x1xf32, #blocked>
|
26 |
+
%cst_16 = arith.constant dense<0.000000e+00> : tensor<64x64xbf16, #blocked>
|
27 |
+
%0 = tt.get_program_id x : i32
|
28 |
+
%1 = arith.muli %0, %c64_i32 : i32
|
29 |
+
%2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
|
30 |
+
%3 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
|
31 |
+
%4 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xi32, #blocked>
|
32 |
+
%5 = tt.expand_dims %3 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<64x1xi32, #blocked1>
|
33 |
+
%6 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked>
|
34 |
+
%7 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked1>
|
35 |
+
%8 = arith.addi %6, %4 : tensor<64x1xi32, #blocked>
|
36 |
+
%9 = arith.addi %7, %5 : tensor<64x1xi32, #blocked1>
|
37 |
+
%10 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
|
38 |
+
%11 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>
|
39 |
+
%12 = tt.expand_dims %10 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x64xi32, #blocked>
|
40 |
+
%13 = tt.expand_dims %11 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>) -> tensor<1x64xi32, #blocked2>
|
41 |
+
%14 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked>
|
42 |
+
%15 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked1>
|
43 |
+
%16 = tt.addptr %14, %8 : tensor<64x1x!tt.ptr<i64, 1>, #blocked>, tensor<64x1xi32, #blocked>
|
44 |
+
%17 = tt.addptr %15, %9 : tensor<64x1x!tt.ptr<i64, 1>, #blocked1>, tensor<64x1xi32, #blocked1>
|
45 |
+
%18 = tt.load %16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked>
|
46 |
+
%19 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked1>
|
47 |
+
%20 = arith.remsi %8, %cst : tensor<64x1xi32, #blocked>
|
48 |
+
%21 = arith.muli %20, %cst_1 : tensor<64x1xi32, #blocked>
|
49 |
+
%22 = tt.broadcast %21 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
50 |
+
%23 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
|
51 |
+
%24 = arith.muli %8, %cst_1 : tensor<64x1xi32, #blocked>
|
52 |
+
%25 = tt.broadcast %24 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
53 |
+
%26 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>, #blocked>
|
54 |
+
%27 = arith.addi %18, %cst_7 : tensor<64x1xi64, #blocked>
|
55 |
+
%28 = arith.addi %19, %cst_8 : tensor<64x1xi64, #blocked1>
|
56 |
+
%29 = arith.cmpi slt, %18, %cst_6 : tensor<64x1xi64, #blocked>
|
57 |
+
%30 = arith.cmpi slt, %19, %cst_9 : tensor<64x1xi64, #blocked1>
|
58 |
+
%31 = arith.select %29, %27, %18 : tensor<64x1xi1, #blocked>, tensor<64x1xi64, #blocked>
|
59 |
+
%32 = arith.select %30, %28, %19 : tensor<64x1xi1, #blocked1>, tensor<64x1xi64, #blocked1>
|
60 |
+
%33 = arith.cmpi sge, %32, %cst_9 : tensor<64x1xi64, #blocked1>
|
61 |
+
%34 = arith.cmpi slt, %32, %cst_8 : tensor<64x1xi64, #blocked1>
|
62 |
+
%35 = arith.andi %33, %34 : tensor<64x1xi1, #blocked1>
|
63 |
+
%36 = arith.muli %31, %cst_5 : tensor<64x1xi64, #blocked>
|
64 |
+
%37 = tt.broadcast %36 : (tensor<64x1xi64, #blocked>) -> tensor<64x64xi64, #blocked>
|
65 |
+
%38 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
|
66 |
+
%39:4 = scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 iter_args(%arg9 = %cst_2, %arg10 = %cst_2, %arg11 = %cst_12, %arg12 = %cst_2) -> (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>) : i32 {
|
67 |
+
%49 = tt.splat %arg8 : (i32) -> tensor<1x64xi32, #blocked>
|
68 |
+
%50 = tt.splat %arg8 : (i32) -> tensor<1x64xi32, #blocked2>
|
69 |
+
%51 = arith.addi %49, %12 : tensor<1x64xi32, #blocked>
|
70 |
+
%52 = arith.addi %50, %13 : tensor<1x64xi32, #blocked2>
|
71 |
+
%53 = arith.cmpi slt, %51, %cst_0 : tensor<1x64xi32, #blocked>
|
72 |
+
%54 = arith.cmpi slt, %52, %cst_13 : tensor<1x64xi32, #blocked2>
|
73 |
+
%55 = tt.broadcast %51 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
74 |
+
%56 = arith.addi %55, %22 : tensor<64x64xi32, #blocked>
|
75 |
+
%57 = tt.addptr %23, %56 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
76 |
+
%58 = tt.broadcast %53 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
|
77 |
+
%59 = tt.broadcast %54 : (tensor<1x64xi1, #blocked2>) -> tensor<64x64xi1, #blocked2>
|
78 |
+
%60 = tt.load %57, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
79 |
+
%61 = arith.addi %55, %25 : tensor<64x64xi32, #blocked>
|
80 |
+
%62 = tt.addptr %26, %61 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
81 |
+
%63 = tt.load %62, %58, %cst_16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xbf16, #blocked>
|
82 |
+
%64 = arith.extf %63 : tensor<64x64xbf16, #blocked> to tensor<64x64xf32, #blocked>
|
83 |
+
tt.assert %35, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
|
84 |
+
%65 = arith.extsi %51 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
|
85 |
+
%66 = tt.broadcast %65 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
|
86 |
+
%67 = arith.addi %66, %37 : tensor<64x64xi64, #blocked>
|
87 |
+
%68 = tt.addptr %38, %67 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
|
88 |
+
%69 = tt.load %68, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
89 |
+
%70 = arith.addf %69, %60 : tensor<64x64xf32, #blocked>
|
90 |
+
%71 = arith.addf %70, %64 : tensor<64x64xf32, #blocked>
|
91 |
+
%72 = arith.subf %71, %arg9 : tensor<64x64xf32, #blocked>
|
92 |
+
%73 = arith.addf %arg12, %cst_4 : tensor<64x64xf32, #blocked>
|
93 |
+
%74 = arith.addf %arg11, %cst_10 : tensor<64x64xf32, #blocked2>
|
94 |
+
%75 = arith.divf %72, %73 : tensor<64x64xf32, #blocked>
|
95 |
+
%76 = arith.addf %arg9, %75 : tensor<64x64xf32, #blocked>
|
96 |
+
%77 = arith.subf %71, %76 : tensor<64x64xf32, #blocked>
|
97 |
+
%78 = arith.mulf %72, %77 : tensor<64x64xf32, #blocked>
|
98 |
+
%79 = arith.addf %arg10, %78 : tensor<64x64xf32, #blocked>
|
99 |
+
%80 = arith.select %58, %76, %arg9 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
|
100 |
+
%81 = arith.select %58, %79, %arg10 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
|
101 |
+
%82 = arith.select %58, %73, %arg12 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
|
102 |
+
%83 = arith.select %59, %74, %arg11 : tensor<64x64xi1, #blocked2>, tensor<64x64xf32, #blocked2>
|
103 |
+
scf.yield %80, %81, %83, %82 : tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>
|
104 |
+
}
|
105 |
+
%40 = triton_gpu.convert_layout %39#2 : (tensor<64x64xf32, #blocked2>) -> tensor<64x64xf32, #blocked>
|
106 |
+
%41:3 = "tt.reduce"(%39#0, %39#1, %40) <{axis = 1 : i32}> ({
|
107 |
+
^bb0(%arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32, %arg13: f32):
|
108 |
+
%49 = arith.subf %arg11, %arg8 : f32
|
109 |
+
%50 = arith.addf %arg10, %arg13 : f32
|
110 |
+
%51 = arith.cmpf oeq, %50, %cst_11 : f32
|
111 |
+
%52 = arith.divf %arg13, %50 : f32
|
112 |
+
%53 = arith.select %51, %cst_11, %52 : f32
|
113 |
+
%54 = arith.mulf %49, %53 : f32
|
114 |
+
%55 = arith.addf %arg8, %54 : f32
|
115 |
+
%56 = arith.addf %arg9, %arg12 : f32
|
116 |
+
%57 = arith.mulf %49, %49 : f32
|
117 |
+
%58 = arith.mulf %57, %arg10 : f32
|
118 |
+
%59 = arith.mulf %58, %53 : f32
|
119 |
+
%60 = arith.addf %56, %59 : f32
|
120 |
+
tt.reduce.return %55, %60, %50 : f32, f32, f32
|
121 |
+
}) : (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>) -> (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>)
|
122 |
+
%42 = tt.expand_dims %41#0 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
|
123 |
+
%43 = tt.expand_dims %41#1 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
|
124 |
+
%44 = tt.splat %arg4 : (!tt.ptr<f32, 1>) -> tensor<1x64x!tt.ptr<f32, 1>, #blocked>
|
125 |
+
%45 = tt.broadcast %42 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
|
126 |
+
%46 = arith.divf %43, %cst_15 : tensor<64x1xf32, #blocked>
|
127 |
+
%47 = arith.addf %46, %cst_14 : tensor<64x1xf32, #blocked>
|
128 |
+
%48 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>, #blocked>
|
129 |
+
scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 : i32 {
|
130 |
+
%49 = tt.splat %arg8 : (i32) -> tensor<1x64xi32, #blocked>
|
131 |
+
%50 = arith.addi %49, %12 : tensor<1x64xi32, #blocked>
|
132 |
+
%51 = arith.cmpi slt, %50, %cst_0 : tensor<1x64xi32, #blocked>
|
133 |
+
%52 = tt.broadcast %50 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
|
134 |
+
%53 = arith.addi %52, %22 : tensor<64x64xi32, #blocked>
|
135 |
+
%54 = tt.addptr %23, %53 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
136 |
+
%55 = tt.broadcast %51 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
|
137 |
+
%56 = tt.load %54, %55, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
138 |
+
%57 = arith.addi %52, %25 : tensor<64x64xi32, #blocked>
|
139 |
+
%58 = tt.addptr %26, %57 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
140 |
+
%59 = tt.load %58, %55, %cst_16 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xbf16, #blocked>
|
141 |
+
%60 = arith.extf %59 : tensor<64x64xbf16, #blocked> to tensor<64x64xf32, #blocked>
|
142 |
+
%61 = tt.addptr %44, %50 : tensor<1x64x!tt.ptr<f32, 1>, #blocked>, tensor<1x64xi32, #blocked>
|
143 |
+
%62 = tt.load %61, %51, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x64xf32, #blocked>
|
144 |
+
tt.assert %35, "index out of bounds: 0 <= tmp16 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
|
145 |
+
%63 = arith.extsi %50 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
|
146 |
+
%64 = tt.broadcast %63 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
|
147 |
+
%65 = arith.addi %64, %37 : tensor<64x64xi64, #blocked>
|
148 |
+
%66 = tt.addptr %38, %65 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
|
149 |
+
%67 = tt.load %66, %55, %cst_2 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
|
150 |
+
%68 = arith.addf %67, %56 : tensor<64x64xf32, #blocked>
|
151 |
+
%69 = arith.addf %68, %60 : tensor<64x64xf32, #blocked>
|
152 |
+
%70 = arith.subf %69, %45 : tensor<64x64xf32, #blocked>
|
153 |
+
%71 = tt.extern_elementwise %47 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<64x1xf32, #blocked>) -> tensor<64x1xf32, #blocked>
|
154 |
+
%72 = tt.broadcast %71 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
|
155 |
+
%73 = arith.mulf %70, %72 : tensor<64x64xf32, #blocked>
|
156 |
+
%74 = tt.broadcast %62 : (tensor<1x64xf32, #blocked>) -> tensor<64x64xf32, #blocked>
|
157 |
+
%75 = arith.mulf %73, %74 : tensor<64x64xf32, #blocked>
|
158 |
+
%76 = tt.addptr %48, %57 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
|
159 |
+
%77 = arith.truncf %75 : tensor<64x64xf32, #blocked> to tensor<64x64xbf16, #blocked>
|
160 |
+
tt.store %76, %77, %55 {cache = 1 : i32, evict = 1 : i32} : tensor<64x64xbf16, #blocked>
|
161 |
+
}
|
162 |
+
tt.return
|
163 |
+
}
|
164 |
+
}
|
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttir
ADDED
@@ -0,0 +1,151 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
module {
|
2 |
+
tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
3 |
+
%cst = arith.constant dense<0.000000e+00> : tensor<64x64xbf16>
|
4 |
+
%cst_0 = arith.constant 0.000000e+00 : f32
|
5 |
+
%cst_1 = arith.constant dense<1.000000e+00> : tensor<64x64xf32>
|
6 |
+
%c256_i32 = arith.constant 256 : i32
|
7 |
+
%c64_i32 = arith.constant 64 : i32
|
8 |
+
%c0_i32 = arith.constant 0 : i32
|
9 |
+
%cst_2 = arith.constant dense<256> : tensor<64x1xi64>
|
10 |
+
%cst_3 = arith.constant dense<0> : tensor<64x1xi64>
|
11 |
+
%cst_4 = arith.constant dense<50257> : tensor<64x1xi64>
|
12 |
+
%cst_5 = arith.constant dense<9.99999974E-6> : tensor<64x1xf32>
|
13 |
+
%cst_6 = arith.constant dense<2.560000e+02> : tensor<64x1xf32>
|
14 |
+
%cst_7 = arith.constant dense<0.000000e+00> : tensor<1x64xf32>
|
15 |
+
%cst_8 = arith.constant dense<0.000000e+00> : tensor<64x64xf32>
|
16 |
+
%cst_9 = arith.constant dense<256> : tensor<64x1xi32>
|
17 |
+
%cst_10 = arith.constant dense<256> : tensor<1x64xi32>
|
18 |
+
%cst_11 = arith.constant dense<512> : tensor<64x1xi32>
|
19 |
+
%0 = tt.get_program_id x : i32
|
20 |
+
%1 = arith.muli %0, %c64_i32 : i32
|
21 |
+
%2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32>
|
22 |
+
%3 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<64xi32>) -> tensor<64x1xi32>
|
23 |
+
%4 = tt.splat %1 : (i32) -> tensor<64x1xi32>
|
24 |
+
%5 = arith.addi %4, %3 : tensor<64x1xi32>
|
25 |
+
%6 = tt.expand_dims %2 {axis = 0 : i32} : (tensor<64xi32>) -> tensor<1x64xi32>
|
26 |
+
%7 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>>
|
27 |
+
%8 = tt.addptr %7, %5 : tensor<64x1x!tt.ptr<i64, 1>>, tensor<64x1xi32>
|
28 |
+
%9 = tt.load %8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64>
|
29 |
+
%10 = arith.remsi %5, %cst_11 : tensor<64x1xi32>
|
30 |
+
%11 = arith.muli %10, %cst_9 : tensor<64x1xi32>
|
31 |
+
%12 = tt.broadcast %11 : (tensor<64x1xi32>) -> tensor<64x64xi32>
|
32 |
+
%13 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
|
33 |
+
%14 = arith.muli %5, %cst_9 : tensor<64x1xi32>
|
34 |
+
%15 = tt.broadcast %14 : (tensor<64x1xi32>) -> tensor<64x64xi32>
|
35 |
+
%16 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>>
|
36 |
+
%17 = arith.addi %9, %cst_4 : tensor<64x1xi64>
|
37 |
+
%18 = arith.cmpi slt, %9, %cst_3 : tensor<64x1xi64>
|
38 |
+
%19 = arith.select %18, %17, %9 : tensor<64x1xi1>, tensor<64x1xi64>
|
39 |
+
%20 = arith.cmpi sge, %19, %cst_3 : tensor<64x1xi64>
|
40 |
+
%21 = arith.cmpi slt, %19, %cst_4 : tensor<64x1xi64>
|
41 |
+
%22 = arith.andi %20, %21 : tensor<64x1xi1>
|
42 |
+
%23 = arith.muli %19, %cst_2 : tensor<64x1xi64>
|
43 |
+
%24 = tt.broadcast %23 : (tensor<64x1xi64>) -> tensor<64x64xi64>
|
44 |
+
%25 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
|
45 |
+
%26:3 = scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 iter_args(%arg9 = %cst_8, %arg10 = %cst_8, %arg11 = %cst_8) -> (tensor<64x64xf32>, tensor<64x64xf32>, tensor<64x64xf32>) : i32 {
|
46 |
+
%50 = tt.splat %arg8 : (i32) -> tensor<1x64xi32>
|
47 |
+
%51 = arith.addi %50, %6 : tensor<1x64xi32>
|
48 |
+
%52 = arith.cmpi slt, %51, %cst_10 : tensor<1x64xi32>
|
49 |
+
%53 = tt.broadcast %51 : (tensor<1x64xi32>) -> tensor<64x64xi32>
|
50 |
+
%54 = arith.addi %53, %12 : tensor<64x64xi32>
|
51 |
+
%55 = tt.addptr %13, %54 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi32>
|
52 |
+
%56 = tt.broadcast %52 : (tensor<1x64xi1>) -> tensor<64x64xi1>
|
53 |
+
%57 = tt.load %55, %56, %cst_8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32>
|
54 |
+
%58 = arith.addi %53, %15 : tensor<64x64xi32>
|
55 |
+
%59 = tt.addptr %16, %58 : tensor<64x64x!tt.ptr<bf16, 1>>, tensor<64x64xi32>
|
56 |
+
%60 = tt.load %59, %56, %cst {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xbf16>
|
57 |
+
%61 = arith.extf %60 : tensor<64x64xbf16> to tensor<64x64xf32>
|
58 |
+
tt.assert %22, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1>
|
59 |
+
%62 = arith.extsi %51 : tensor<1x64xi32> to tensor<1x64xi64>
|
60 |
+
%63 = tt.broadcast %62 : (tensor<1x64xi64>) -> tensor<64x64xi64>
|
61 |
+
%64 = arith.addi %63, %24 : tensor<64x64xi64>
|
62 |
+
%65 = tt.addptr %25, %64 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi64>
|
63 |
+
%66 = tt.load %65, %56, %cst_8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32>
|
64 |
+
%67 = arith.addf %66, %57 : tensor<64x64xf32>
|
65 |
+
%68 = arith.addf %67, %61 : tensor<64x64xf32>
|
66 |
+
%69 = arith.subf %68, %arg9 : tensor<64x64xf32>
|
67 |
+
%70 = arith.addf %arg11, %cst_1 : tensor<64x64xf32>
|
68 |
+
%71 = arith.divf %69, %70 : tensor<64x64xf32>
|
69 |
+
%72 = arith.addf %arg9, %71 : tensor<64x64xf32>
|
70 |
+
%73 = arith.subf %68, %72 : tensor<64x64xf32>
|
71 |
+
%74 = arith.mulf %69, %73 : tensor<64x64xf32>
|
72 |
+
%75 = arith.addf %arg10, %74 : tensor<64x64xf32>
|
73 |
+
%76 = arith.select %56, %72, %arg9 : tensor<64x64xi1>, tensor<64x64xf32>
|
74 |
+
%77 = arith.select %56, %75, %arg10 : tensor<64x64xi1>, tensor<64x64xf32>
|
75 |
+
%78 = arith.select %56, %70, %arg11 : tensor<64x64xi1>, tensor<64x64xf32>
|
76 |
+
scf.yield %76, %77, %78 : tensor<64x64xf32>, tensor<64x64xf32>, tensor<64x64xf32>
|
77 |
+
}
|
78 |
+
%27:3 = "tt.reduce"(%26#0, %26#1, %26#2) <{axis = 1 : i32}> ({
|
79 |
+
^bb0(%arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32, %arg13: f32):
|
80 |
+
%50 = arith.subf %arg11, %arg8 : f32
|
81 |
+
%51 = arith.addf %arg10, %arg13 : f32
|
82 |
+
%52 = arith.cmpf oeq, %51, %cst_0 : f32
|
83 |
+
%53 = arith.divf %arg13, %51 : f32
|
84 |
+
%54 = arith.select %52, %cst_0, %53 : f32
|
85 |
+
%55 = arith.mulf %50, %54 : f32
|
86 |
+
%56 = arith.addf %arg8, %55 : f32
|
87 |
+
%57 = arith.addf %arg9, %arg12 : f32
|
88 |
+
%58 = arith.mulf %50, %50 : f32
|
89 |
+
%59 = arith.mulf %58, %arg10 : f32
|
90 |
+
%60 = arith.mulf %59, %54 : f32
|
91 |
+
%61 = arith.addf %57, %60 : f32
|
92 |
+
tt.reduce.return %56, %61, %51 : f32, f32, f32
|
93 |
+
}) : (tensor<64x64xf32>, tensor<64x64xf32>, tensor<64x64xf32>) -> (tensor<64xf32>, tensor<64xf32>, tensor<64xf32>)
|
94 |
+
%28 = tt.expand_dims %27#0 {axis = 1 : i32} : (tensor<64xf32>) -> tensor<64x1xf32>
|
95 |
+
%29 = tt.expand_dims %27#1 {axis = 1 : i32} : (tensor<64xf32>) -> tensor<64x1xf32>
|
96 |
+
%30 = arith.muli %10, %cst_9 : tensor<64x1xi32>
|
97 |
+
%31 = tt.broadcast %30 : (tensor<64x1xi32>) -> tensor<64x64xi32>
|
98 |
+
%32 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
|
99 |
+
%33 = arith.muli %5, %cst_9 : tensor<64x1xi32>
|
100 |
+
%34 = tt.broadcast %33 : (tensor<64x1xi32>) -> tensor<64x64xi32>
|
101 |
+
%35 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>>
|
102 |
+
%36 = tt.splat %arg4 : (!tt.ptr<f32, 1>) -> tensor<1x64x!tt.ptr<f32, 1>>
|
103 |
+
%37 = arith.addi %9, %cst_4 : tensor<64x1xi64>
|
104 |
+
%38 = arith.cmpi slt, %9, %cst_3 : tensor<64x1xi64>
|
105 |
+
%39 = arith.select %38, %37, %9 : tensor<64x1xi1>, tensor<64x1xi64>
|
106 |
+
%40 = arith.cmpi sge, %39, %cst_3 : tensor<64x1xi64>
|
107 |
+
%41 = arith.cmpi slt, %39, %cst_4 : tensor<64x1xi64>
|
108 |
+
%42 = arith.andi %40, %41 : tensor<64x1xi1>
|
109 |
+
%43 = arith.muli %39, %cst_2 : tensor<64x1xi64>
|
110 |
+
%44 = tt.broadcast %43 : (tensor<64x1xi64>) -> tensor<64x64xi64>
|
111 |
+
%45 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
|
112 |
+
%46 = tt.broadcast %28 : (tensor<64x1xf32>) -> tensor<64x64xf32>
|
113 |
+
%47 = arith.divf %29, %cst_6 : tensor<64x1xf32>
|
114 |
+
%48 = arith.addf %47, %cst_5 : tensor<64x1xf32>
|
115 |
+
%49 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>>
|
116 |
+
scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 : i32 {
|
117 |
+
%50 = tt.splat %arg8 : (i32) -> tensor<1x64xi32>
|
118 |
+
%51 = arith.addi %50, %6 : tensor<1x64xi32>
|
119 |
+
%52 = arith.cmpi slt, %51, %cst_10 : tensor<1x64xi32>
|
120 |
+
%53 = tt.broadcast %51 : (tensor<1x64xi32>) -> tensor<64x64xi32>
|
121 |
+
%54 = arith.addi %53, %31 : tensor<64x64xi32>
|
122 |
+
%55 = tt.addptr %32, %54 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi32>
|
123 |
+
%56 = tt.broadcast %52 : (tensor<1x64xi1>) -> tensor<64x64xi1>
|
124 |
+
%57 = tt.load %55, %56, %cst_8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32>
|
125 |
+
%58 = arith.addi %53, %34 : tensor<64x64xi32>
|
126 |
+
%59 = tt.addptr %35, %58 : tensor<64x64x!tt.ptr<bf16, 1>>, tensor<64x64xi32>
|
127 |
+
%60 = tt.load %59, %56, %cst {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xbf16>
|
128 |
+
%61 = arith.extf %60 : tensor<64x64xbf16> to tensor<64x64xf32>
|
129 |
+
%62 = tt.addptr %36, %51 : tensor<1x64x!tt.ptr<f32, 1>>, tensor<1x64xi32>
|
130 |
+
%63 = tt.load %62, %52, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x64xf32>
|
131 |
+
tt.assert %42, "index out of bounds: 0 <= tmp16 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1>
|
132 |
+
%64 = arith.extsi %51 : tensor<1x64xi32> to tensor<1x64xi64>
|
133 |
+
%65 = tt.broadcast %64 : (tensor<1x64xi64>) -> tensor<64x64xi64>
|
134 |
+
%66 = arith.addi %65, %44 : tensor<64x64xi64>
|
135 |
+
%67 = tt.addptr %45, %66 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi64>
|
136 |
+
%68 = tt.load %67, %56, %cst_8 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xf32>
|
137 |
+
%69 = arith.addf %68, %57 : tensor<64x64xf32>
|
138 |
+
%70 = arith.addf %69, %61 : tensor<64x64xf32>
|
139 |
+
%71 = arith.subf %70, %46 : tensor<64x64xf32>
|
140 |
+
%72 = tt.extern_elementwise %48 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<64x1xf32>) -> tensor<64x1xf32>
|
141 |
+
%73 = tt.broadcast %72 : (tensor<64x1xf32>) -> tensor<64x64xf32>
|
142 |
+
%74 = arith.mulf %71, %73 : tensor<64x64xf32>
|
143 |
+
%75 = tt.broadcast %63 : (tensor<1x64xf32>) -> tensor<64x64xf32>
|
144 |
+
%76 = arith.mulf %74, %75 : tensor<64x64xf32>
|
145 |
+
%77 = tt.addptr %49, %58 : tensor<64x64x!tt.ptr<bf16, 1>>, tensor<64x64xi32>
|
146 |
+
%78 = arith.truncf %76 : tensor<64x64xf32> to tensor<64x64xbf16>
|
147 |
+
tt.store %77, %78, %56 {cache = 1 : i32, evict = 1 : i32} : tensor<64x64xbf16>
|
148 |
+
}
|
149 |
+
tt.return
|
150 |
+
}
|
151 |
+
}
|
.triton/dump/76fb48b96c75cb8e388c291a18ef9b02/triton_.ttgir
ADDED
@@ -0,0 +1,169 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [1, 2], threadsPerWarp = [1, 32], warpsPerCTA = [2, 2], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
2 |
+
#blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
3 |
+
#blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 4], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
4 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 4 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
5 |
+
tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
6 |
+
%cst = arith.constant dense<512> : tensor<2x1xi32, #blocked>
|
7 |
+
%cst_0 = arith.constant dense<256> : tensor<1x128xi32, #blocked>
|
8 |
+
%cst_1 = arith.constant dense<256> : tensor<2x1xi32, #blocked>
|
9 |
+
%cst_2 = arith.constant dense<0.000000e+00> : tensor<2x128xf32, #blocked>
|
10 |
+
%cst_3 = arith.constant dense<1.000000e+00> : tensor<2x128xf32, #blocked>
|
11 |
+
%cst_4 = arith.constant dense<256> : tensor<2x1xi64, #blocked>
|
12 |
+
%cst_5 = arith.constant dense<0> : tensor<2x1xi64, #blocked>
|
13 |
+
%cst_6 = arith.constant dense<50257> : tensor<2x1xi64, #blocked>
|
14 |
+
%cst_7 = arith.constant dense<50257> : tensor<2x1xi64, #blocked1>
|
15 |
+
%cst_8 = arith.constant dense<0> : tensor<2x1xi64, #blocked1>
|
16 |
+
%c0_i32 = arith.constant 0 : i32
|
17 |
+
%c128_i32 = arith.constant 128 : i32
|
18 |
+
%c256_i32 = arith.constant 256 : i32
|
19 |
+
%cst_9 = arith.constant dense<1.000000e+00> : tensor<2x128xf32, #blocked2>
|
20 |
+
%cst_10 = arith.constant 0.000000e+00 : f32
|
21 |
+
%cst_11 = arith.constant dense<0.000000e+00> : tensor<1x128xf32, #blocked2>
|
22 |
+
%cst_12 = arith.constant dense<0.000000e+00> : tensor<2x128xf32, #blocked2>
|
23 |
+
%cst_13 = arith.constant dense<256> : tensor<1x128xi32, #blocked2>
|
24 |
+
%cst_14 = arith.constant dense<9.99999974E-6> : tensor<2x1xf32, #blocked>
|
25 |
+
%cst_15 = arith.constant dense<2.560000e+02> : tensor<2x1xf32, #blocked>
|
26 |
+
%cst_16 = arith.constant dense<0.000000e+00> : tensor<2x128xbf16, #blocked>
|
27 |
+
%c2_i32 = arith.constant 2 : i32
|
28 |
+
%0 = tt.get_program_id x : i32
|
29 |
+
%1 = arith.muli %0, %c2_i32 : i32
|
30 |
+
%2 = tt.make_range {end = 2 : i32, start = 0 : i32} : tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
|
31 |
+
%3 = tt.make_range {end = 2 : i32, start = 0 : i32} : tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
|
32 |
+
%4 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<2x1xi32, #blocked>
|
33 |
+
%5 = tt.expand_dims %3 {axis = 1 : i32} : (tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<2x1xi32, #blocked1>
|
34 |
+
%6 = tt.splat %1 : (i32) -> tensor<2x1xi32, #blocked>
|
35 |
+
%7 = tt.splat %1 : (i32) -> tensor<2x1xi32, #blocked1>
|
36 |
+
%8 = arith.addi %6, %4 : tensor<2x1xi32, #blocked>
|
37 |
+
%9 = arith.addi %7, %5 : tensor<2x1xi32, #blocked1>
|
38 |
+
%10 = tt.make_range {end = 128 : i32, start = 0 : i32} : tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
|
39 |
+
%11 = tt.make_range {end = 128 : i32, start = 0 : i32} : tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>
|
40 |
+
%12 = tt.expand_dims %10 {axis = 0 : i32} : (tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x128xi32, #blocked>
|
41 |
+
%13 = tt.expand_dims %11 {axis = 0 : i32} : (tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>) -> tensor<1x128xi32, #blocked2>
|
42 |
+
%14 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<2x1x!tt.ptr<i64, 1>, #blocked>
|
43 |
+
%15 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<2x1x!tt.ptr<i64, 1>, #blocked1>
|
44 |
+
%16 = tt.addptr %14, %8 : tensor<2x1x!tt.ptr<i64, 1>, #blocked>, tensor<2x1xi32, #blocked>
|
45 |
+
%17 = tt.addptr %15, %9 : tensor<2x1x!tt.ptr<i64, 1>, #blocked1>, tensor<2x1xi32, #blocked1>
|
46 |
+
%18 = tt.load %16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x1xi64, #blocked>
|
47 |
+
%19 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x1xi64, #blocked1>
|
48 |
+
%20 = arith.remsi %8, %cst : tensor<2x1xi32, #blocked>
|
49 |
+
%21 = arith.muli %20, %cst_1 : tensor<2x1xi32, #blocked>
|
50 |
+
%22 = tt.broadcast %21 : (tensor<2x1xi32, #blocked>) -> tensor<2x128xi32, #blocked>
|
51 |
+
%23 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<2x128x!tt.ptr<f32, 1>, #blocked>
|
52 |
+
%24 = arith.muli %8, %cst_1 : tensor<2x1xi32, #blocked>
|
53 |
+
%25 = tt.broadcast %24 : (tensor<2x1xi32, #blocked>) -> tensor<2x128xi32, #blocked>
|
54 |
+
%26 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<2x128x!tt.ptr<bf16, 1>, #blocked>
|
55 |
+
%27 = arith.addi %18, %cst_6 : tensor<2x1xi64, #blocked>
|
56 |
+
%28 = arith.addi %19, %cst_7 : tensor<2x1xi64, #blocked1>
|
57 |
+
%29 = arith.cmpi slt, %18, %cst_5 : tensor<2x1xi64, #blocked>
|
58 |
+
%30 = arith.cmpi slt, %19, %cst_8 : tensor<2x1xi64, #blocked1>
|
59 |
+
%31 = arith.select %29, %27, %18 : tensor<2x1xi1, #blocked>, tensor<2x1xi64, #blocked>
|
60 |
+
%32 = arith.select %30, %28, %19 : tensor<2x1xi1, #blocked1>, tensor<2x1xi64, #blocked1>
|
61 |
+
%33 = arith.cmpi sge, %32, %cst_8 : tensor<2x1xi64, #blocked1>
|
62 |
+
%34 = arith.cmpi slt, %32, %cst_7 : tensor<2x1xi64, #blocked1>
|
63 |
+
%35 = arith.andi %33, %34 : tensor<2x1xi1, #blocked1>
|
64 |
+
%36 = arith.muli %31, %cst_4 : tensor<2x1xi64, #blocked>
|
65 |
+
%37 = tt.broadcast %36 : (tensor<2x1xi64, #blocked>) -> tensor<2x128xi64, #blocked>
|
66 |
+
%38 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<2x128x!tt.ptr<f32, 1>, #blocked>
|
67 |
+
%39:4 = scf.for %arg8 = %c0_i32 to %c256_i32 step %c128_i32 iter_args(%arg9 = %cst_2, %arg10 = %cst_2, %arg11 = %cst_12, %arg12 = %cst_2) -> (tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked2>, tensor<2x128xf32, #blocked>) : i32 {
|
68 |
+
%49 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked>
|
69 |
+
%50 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked2>
|
70 |
+
%51 = arith.addi %49, %12 : tensor<1x128xi32, #blocked>
|
71 |
+
%52 = arith.addi %50, %13 : tensor<1x128xi32, #blocked2>
|
72 |
+
%53 = arith.cmpi slt, %51, %cst_0 : tensor<1x128xi32, #blocked>
|
73 |
+
%54 = arith.cmpi slt, %52, %cst_13 : tensor<1x128xi32, #blocked2>
|
74 |
+
%55 = tt.broadcast %51 : (tensor<1x128xi32, #blocked>) -> tensor<2x128xi32, #blocked>
|
75 |
+
%56 = arith.addi %55, %22 : tensor<2x128xi32, #blocked>
|
76 |
+
%57 = tt.addptr %23, %56 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi32, #blocked>
|
77 |
+
%58 = tt.broadcast %53 : (tensor<1x128xi1, #blocked>) -> tensor<2x128xi1, #blocked>
|
78 |
+
%59 = tt.broadcast %54 : (tensor<1x128xi1, #blocked2>) -> tensor<2x128xi1, #blocked2>
|
79 |
+
%60 = tt.load %57, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
|
80 |
+
%61 = arith.addi %55, %25 : tensor<2x128xi32, #blocked>
|
81 |
+
%62 = tt.addptr %26, %61 : tensor<2x128x!tt.ptr<bf16, 1>, #blocked>, tensor<2x128xi32, #blocked>
|
82 |
+
%63 = tt.load %62, %58, %cst_16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xbf16, #blocked>
|
83 |
+
%64 = arith.extf %63 : tensor<2x128xbf16, #blocked> to tensor<2x128xf32, #blocked>
|
84 |
+
tt.assert %35, "index out of bounds: 0 <= tmp3 < 50257", "/usr/local/lib/python3.10/dist-packages/torch/_inductor/codecache.py", "<module>", 1892 : tensor<2x1xi1, #blocked1>
|
85 |
+
%65 = arith.extsi %51 : tensor<1x128xi32, #blocked> to tensor<1x128xi64, #blocked>
|
86 |
+
%66 = tt.broadcast %65 : (tensor<1x128xi64, #blocked>) -> tensor<2x128xi64, #blocked>
|
87 |
+
%67 = arith.addi %66, %37 : tensor<2x128xi64, #blocked>
|
88 |
+
%68 = tt.addptr %38, %67 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi64, #blocked>
|
89 |
+
%69 = tt.load %68, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
|
90 |
+
%70 = arith.addf %69, %60 : tensor<2x128xf32, #blocked>
|
91 |
+
%71 = arith.addf %70, %64 : tensor<2x128xf32, #blocked>
|
92 |
+
%72 = arith.subf %71, %arg9 : tensor<2x128xf32, #blocked>
|
93 |
+
%73 = arith.addf %arg12, %cst_3 : tensor<2x128xf32, #blocked>
|
94 |
+
%74 = arith.addf %arg11, %cst_9 : tensor<2x128xf32, #blocked2>
|
95 |
+
%75 = arith.divf %72, %73 : tensor<2x128xf32, #blocked>
|
96 |
+
%76 = arith.addf %arg9, %75 : tensor<2x128xf32, #blocked>
|
97 |
+
%77 = arith.subf %71, %76 : tensor<2x128xf32, #blocked>
|
98 |
+
%78 = arith.mulf %72, %77 : tensor<2x128xf32, #blocked>
|
99 |
+
%79 = arith.addf %arg10, %78 : tensor<2x128xf32, #blocked>
|
100 |
+
%80 = arith.select %58, %76, %arg9 : tensor<2x128xi1, #blocked>, tensor<2x128xf32, #blocked>
|
101 |
+
%81 = arith.select %58, %79, %arg10 : tensor<2x128xi1, #blocked>, tensor<2x128xf32, #blocked>
|
102 |
+
%82 = arith.select %58, %73, %arg12 : tensor<2x128xi1, #blocked>, tensor<2x128xf32, #blocked>
|
103 |
+
%83 = arith.select %59, %74, %arg11 : tensor<2x128xi1, #blocked2>, tensor<2x128xf32, #blocked2>
|
104 |
+
scf.yield %80, %81, %83, %82 : tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked2>, tensor<2x128xf32, #blocked>
|
105 |
+
}
|
106 |
+
%40 = triton_gpu.convert_layout %39#2 : (tensor<2x128xf32, #blocked2>) -> tensor<2x128xf32, #blocked>
|
107 |
+
%41:3 = "tt.reduce"(%39#0, %39#1, %40) <{axis = 1 : i32}> ({
|
108 |
+
^bb0(%arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32, %arg13: f32):
|
109 |
+
%49 = arith.subf %arg11, %arg8 : f32
|
110 |
+
%50 = arith.addf %arg10, %arg13 : f32
|
111 |
+
%51 = arith.cmpf oeq, %50, %cst_10 : f32
|
112 |
+
%52 = arith.divf %arg13, %50 : f32
|
113 |
+
%53 = arith.select %51, %cst_10, %52 : f32
|
114 |
+
%54 = arith.mulf %49, %53 : f32
|
115 |
+
%55 = arith.addf %arg8, %54 : f32
|
116 |
+
%56 = arith.addf %arg9, %arg12 : f32
|
117 |
+
%57 = arith.mulf %49, %49 : f32
|
118 |
+
%58 = arith.mulf %57, %arg10 : f32
|
119 |
+
%59 = arith.mulf %58, %53 : f32
|
120 |
+
%60 = arith.addf %56, %59 : f32
|
121 |
+
tt.reduce.return %55, %60, %50 : f32, f32, f32
|
122 |
+
}) : (tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>) -> (tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>)
|
123 |
+
%42 = tt.expand_dims %41#0 {axis = 1 : i32} : (tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<2x1xf32, #blocked>
|
124 |
+
%43 = tt.expand_dims %41#1 {axis = 1 : i32} : (tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<2x1xf32, #blocked>
|
125 |
+
%44 = tt.splat %arg4 : (!tt.ptr<f32, 1>) -> tensor<1x128x!tt.ptr<f32, 1>, #blocked2>
|
126 |
+
%45 = tt.broadcast %42 : (tensor<2x1xf32, #blocked>) -> tensor<2x128xf32, #blocked>
|
127 |
+
%46 = arith.divf %43, %cst_15 : tensor<2x1xf32, #blocked>
|
128 |
+
%47 = arith.addf %46, %cst_14 : tensor<2x1xf32, #blocked>
|
129 |
+
%48 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<2x128x!tt.ptr<bf16, 1>, #blocked>
|
130 |
+
scf.for %arg8 = %c0_i32 to %c256_i32 step %c128_i32 : i32 {
|
131 |
+
%49 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked>
|
132 |
+
%50 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked2>
|
133 |
+
%51 = arith.addi %49, %12 : tensor<1x128xi32, #blocked>
|
134 |
+
%52 = arith.addi %50, %13 : tensor<1x128xi32, #blocked2>
|
135 |
+
%53 = arith.cmpi slt, %51, %cst_0 : tensor<1x128xi32, #blocked>
|
136 |
+
%54 = arith.cmpi slt, %52, %cst_13 : tensor<1x128xi32, #blocked2>
|
137 |
+
%55 = tt.broadcast %51 : (tensor<1x128xi32, #blocked>) -> tensor<2x128xi32, #blocked>
|
138 |
+
%56 = arith.addi %55, %22 : tensor<2x128xi32, #blocked>
|
139 |
+
%57 = tt.addptr %23, %56 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi32, #blocked>
|
140 |
+
%58 = tt.broadcast %53 : (tensor<1x128xi1, #blocked>) -> tensor<2x128xi1, #blocked>
|
141 |
+
%59 = tt.load %57, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
|
142 |
+
%60 = arith.addi %55, %25 : tensor<2x128xi32, #blocked>
|
143 |
+
%61 = tt.addptr %26, %60 : tensor<2x128x!tt.ptr<bf16, 1>, #blocked>, tensor<2x128xi32, #blocked>
|
144 |
+
%62 = tt.load %61, %58, %cst_16 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<2x128xbf16, #blocked>
|
145 |
+
%63 = arith.extf %62 : tensor<2x128xbf16, #blocked> to tensor<2x128xf32, #blocked>
|
146 |
+
%64 = tt.addptr %44, %52 : tensor<1x128x!tt.ptr<f32, 1>, #blocked2>, tensor<1x128xi32, #blocked2>
|
147 |
+
%65 = tt.load %64, %54, %cst_11 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x128xf32, #blocked2>
|
148 |
+
tt.assert %35, "index out of bounds: 0 <= tmp16 < 50257", "/usr/local/lib/python3.10/dist-packages/torch/_inductor/codecache.py", "<module>", 1892 : tensor<2x1xi1, #blocked1>
|
149 |
+
%66 = arith.extsi %51 : tensor<1x128xi32, #blocked> to tensor<1x128xi64, #blocked>
|
150 |
+
%67 = tt.broadcast %66 : (tensor<1x128xi64, #blocked>) -> tensor<2x128xi64, #blocked>
|
151 |
+
%68 = arith.addi %67, %37 : tensor<2x128xi64, #blocked>
|
152 |
+
%69 = tt.addptr %38, %68 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi64, #blocked>
|
153 |
+
%70 = tt.load %69, %58, %cst_2 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
|
154 |
+
%71 = arith.addf %70, %59 : tensor<2x128xf32, #blocked>
|
155 |
+
%72 = arith.addf %71, %63 : tensor<2x128xf32, #blocked>
|
156 |
+
%73 = arith.subf %72, %45 : tensor<2x128xf32, #blocked>
|
157 |
+
%74 = tt.extern_elementwise %47 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<2x1xf32, #blocked>) -> tensor<2x1xf32, #blocked>
|
158 |
+
%75 = tt.broadcast %74 : (tensor<2x1xf32, #blocked>) -> tensor<2x128xf32, #blocked>
|
159 |
+
%76 = arith.mulf %73, %75 : tensor<2x128xf32, #blocked>
|
160 |
+
%77 = triton_gpu.convert_layout %65 : (tensor<1x128xf32, #blocked2>) -> tensor<1x128xf32, #blocked>
|
161 |
+
%78 = tt.broadcast %77 : (tensor<1x128xf32, #blocked>) -> tensor<2x128xf32, #blocked>
|
162 |
+
%79 = arith.mulf %76, %78 : tensor<2x128xf32, #blocked>
|
163 |
+
%80 = tt.addptr %48, %60 : tensor<2x128x!tt.ptr<bf16, 1>, #blocked>, tensor<2x128xi32, #blocked>
|
164 |
+
%81 = arith.truncf %79 : tensor<2x128xf32, #blocked> to tensor<2x128xbf16, #blocked>
|
165 |
+
tt.store %80, %81, %58 {cache = 1 : i32, evict = 1 : i32} : tensor<2x128xbf16, #blocked>
|
166 |
+
}
|
167 |
+
tt.return
|
168 |
+
}
|
169 |
+
}
|
.triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttgir
ADDED
@@ -0,0 +1,110 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [1, 32], warpsPerCTA = [1, 8], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
2 |
+
#blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 8], order = [0, 1], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
3 |
+
#blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 8], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
|
4 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 8 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
5 |
+
tt.func public @triton__0d1d2d3d4d5d6e7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg6: i64 {tt.max_divisibility = 8 : i32}, %arg7: i64 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
6 |
+
%cst = arith.constant dense<7680> : tensor<1x2048xi64, #blocked>
|
7 |
+
%cst_0 = arith.constant dense<7680> : tensor<1x2048xi64, #blocked1>
|
8 |
+
%cst_1 = arith.constant dense<50257> : tensor<1x2048xi64, #blocked>
|
9 |
+
%c385973760_i64 = arith.constant 385973760 : i64
|
10 |
+
%c7680_i64 = arith.constant 7680 : i64
|
11 |
+
%c8_i64 = arith.constant 8 : i64
|
12 |
+
%cst_2 = arith.constant dense<-1> : tensor<1x2048xi64, #blocked>
|
13 |
+
%cst_3 = arith.constant dense<0> : tensor<1x2048xi64, #blocked>
|
14 |
+
%cst_4 = arith.constant dense<0.000000e+00> : tensor<1x2048xf32, #blocked1>
|
15 |
+
%cst_5 = arith.constant dense<0.000000e+00> : tensor<1x2048xf32, #blocked>
|
16 |
+
%cst_6 = arith.constant dense<0.000000e+00> : tensor<1x2048xbf16, #blocked1>
|
17 |
+
%c0_i32 = arith.constant 0 : i32
|
18 |
+
%c7680_i32 = arith.constant 7680 : i32
|
19 |
+
%c2048_i32 = arith.constant 2048 : i32
|
20 |
+
%0 = tt.get_program_id x : i32
|
21 |
+
%1 = arith.extsi %0 : i32 to i64
|
22 |
+
%2 = arith.cmpi slt, %1, %c8_i64 : i64
|
23 |
+
%3 = tt.make_range {end = 2048 : i32, start = 0 : i32} : tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
|
24 |
+
%4 = tt.make_range {end = 2048 : i32, start = 0 : i32} : tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked1}>>
|
25 |
+
%5 = tt.expand_dims %3 {axis = 0 : i32} : (tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x2048xi32, #blocked>
|
26 |
+
%6 = tt.expand_dims %4 {axis = 0 : i32} : (tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked1}>>) -> tensor<1x2048xi32, #blocked1>
|
27 |
+
%7 = arith.extsi %5 : tensor<1x2048xi32, #blocked> to tensor<1x2048xi64, #blocked>
|
28 |
+
%8 = arith.extsi %6 : tensor<1x2048xi32, #blocked1> to tensor<1x2048xi64, #blocked1>
|
29 |
+
%9 = arith.muli %1, %c7680_i64 : i64
|
30 |
+
%10 = tt.splat %9 : (i64) -> tensor<1x2048xi64, #blocked>
|
31 |
+
%11 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<1x2048x!tt.ptr<i64, 1>, #blocked>
|
32 |
+
%12 = tt.splat %2 : (i1) -> tensor<1x2048xi1, #blocked>
|
33 |
+
%13 = tt.splat %2 : (i1) -> tensor<1x2048xi1, #blocked1>
|
34 |
+
%14 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>, #blocked>
|
35 |
+
%15 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>, #blocked>
|
36 |
+
%16 = arith.muli %1, %c385973760_i64 : i64
|
37 |
+
%17 = tt.splat %16 : (i64) -> tensor<1x2048xi64, #blocked>
|
38 |
+
%18 = tt.splat %arg1 : (!tt.ptr<bf16, 1>) -> tensor<1x2048x!tt.ptr<bf16, 1>, #blocked>
|
39 |
+
%19:2 = scf.for %arg8 = %c0_i32 to %c7680_i32 step %c2048_i32 iter_args(%arg9 = %cst_4, %arg10 = %cst_3) -> (tensor<1x2048xf32, #blocked1>, tensor<1x2048xi64, #blocked>) : i32 {
|
40 |
+
%30 = arith.extsi %arg8 : i32 to i64
|
41 |
+
%31 = tt.splat %30 : (i64) -> tensor<1x2048xi64, #blocked>
|
42 |
+
%32 = tt.splat %30 : (i64) -> tensor<1x2048xi64, #blocked1>
|
43 |
+
%33 = arith.addi %31, %7 : tensor<1x2048xi64, #blocked>
|
44 |
+
%34 = arith.addi %32, %8 : tensor<1x2048xi64, #blocked1>
|
45 |
+
%35 = arith.cmpi slt, %33, %cst : tensor<1x2048xi64, #blocked>
|
46 |
+
%36 = arith.cmpi slt, %34, %cst_0 : tensor<1x2048xi64, #blocked1>
|
47 |
+
%37 = arith.addi %33, %10 : tensor<1x2048xi64, #blocked>
|
48 |
+
%38 = tt.addptr %11, %37 : tensor<1x2048x!tt.ptr<i64, 1>, #blocked>, tensor<1x2048xi64, #blocked>
|
49 |
+
%39 = arith.andi %35, %12 : tensor<1x2048xi1, #blocked>
|
50 |
+
%40 = arith.andi %36, %13 : tensor<1x2048xi1, #blocked1>
|
51 |
+
%41 = tt.load %38, %39, %cst_3 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xi64, #blocked>
|
52 |
+
%42 = tt.addptr %14, %37 : tensor<1x2048x!tt.ptr<f32, 1>, #blocked>, tensor<1x2048xi64, #blocked>
|
53 |
+
%43 = tt.load %42, %39, %cst_5 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32, #blocked>
|
54 |
+
%44 = triton_gpu.convert_layout %43 : (tensor<1x2048xf32, #blocked>) -> tensor<1x2048xf32, #blocked1>
|
55 |
+
%45 = tt.addptr %15, %37 : tensor<1x2048x!tt.ptr<f32, 1>, #blocked>, tensor<1x2048xi64, #blocked>
|
56 |
+
%46 = tt.load %45, %39, %cst_5 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32, #blocked>
|
57 |
+
%47 = arith.cmpi ne, %41, %cst_2 : tensor<1x2048xi64, #blocked>
|
58 |
+
%48 = triton_gpu.convert_layout %47 : (tensor<1x2048xi1, #blocked>) -> tensor<1x2048xi1, #blocked1>
|
59 |
+
%49 = arith.select %47, %41, %cst_3 : tensor<1x2048xi1, #blocked>, tensor<1x2048xi64, #blocked>
|
60 |
+
%50 = arith.addi %49, %cst_1 : tensor<1x2048xi64, #blocked>
|
61 |
+
%51 = arith.cmpi slt, %49, %cst_3 : tensor<1x2048xi64, #blocked>
|
62 |
+
%52 = arith.select %51, %50, %49 : tensor<1x2048xi1, #blocked>, tensor<1x2048xi64, #blocked>
|
63 |
+
%53 = arith.cmpi sge, %52, %cst_3 : tensor<1x2048xi64, #blocked>
|
64 |
+
%54 = arith.cmpi slt, %52, %cst_1 : tensor<1x2048xi64, #blocked>
|
65 |
+
%55 = arith.andi %53, %54 : tensor<1x2048xi1, #blocked>
|
66 |
+
%56 = triton_gpu.convert_layout %55 : (tensor<1x2048xi1, #blocked>) -> tensor<1x2048xi1, #blocked2>
|
67 |
+
tt.assert %56, "index out of bounds: 0 <= tmp7 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<1x2048xi1, #blocked2>
|
68 |
+
%57 = arith.muli %33, %cst_1 : tensor<1x2048xi64, #blocked>
|
69 |
+
%58 = arith.addi %52, %57 : tensor<1x2048xi64, #blocked>
|
70 |
+
%59 = arith.addi %58, %17 : tensor<1x2048xi64, #blocked>
|
71 |
+
%60 = tt.addptr %18, %59 : tensor<1x2048x!tt.ptr<bf16, 1>, #blocked>, tensor<1x2048xi64, #blocked>
|
72 |
+
%61 = triton_gpu.convert_layout %60 : (tensor<1x2048x!tt.ptr<bf16, 1>, #blocked>) -> tensor<1x2048x!tt.ptr<bf16, 1>, #blocked1>
|
73 |
+
%62 = tt.load %61, %40, %cst_6 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x2048xbf16, #blocked1>
|
74 |
+
%63 = arith.extf %62 : tensor<1x2048xbf16, #blocked1> to tensor<1x2048xf32, #blocked1>
|
75 |
+
%64 = arith.subf %63, %44 : tensor<1x2048xf32, #blocked1>
|
76 |
+
%65 = math.log %46 : tensor<1x2048xf32, #blocked>
|
77 |
+
%66 = triton_gpu.convert_layout %65 : (tensor<1x2048xf32, #blocked>) -> tensor<1x2048xf32, #blocked1>
|
78 |
+
%67 = arith.subf %64, %66 : tensor<1x2048xf32, #blocked1>
|
79 |
+
%68 = arith.subf %cst_4, %67 : tensor<1x2048xf32, #blocked1>
|
80 |
+
%69 = arith.select %48, %68, %cst_4 : tensor<1x2048xi1, #blocked1>, tensor<1x2048xf32, #blocked1>
|
81 |
+
%70 = arith.addf %arg9, %69 : tensor<1x2048xf32, #blocked1>
|
82 |
+
%71 = arith.select %40, %70, %arg9 : tensor<1x2048xi1, #blocked1>, tensor<1x2048xf32, #blocked1>
|
83 |
+
%72 = arith.extui %47 : tensor<1x2048xi1, #blocked> to tensor<1x2048xi64, #blocked>
|
84 |
+
%73 = arith.addi %arg10, %72 : tensor<1x2048xi64, #blocked>
|
85 |
+
%74 = arith.select %39, %73, %arg10 : tensor<1x2048xi1, #blocked>, tensor<1x2048xi64, #blocked>
|
86 |
+
scf.yield %71, %74 : tensor<1x2048xf32, #blocked1>, tensor<1x2048xi64, #blocked>
|
87 |
+
}
|
88 |
+
%20 = "tt.reduce"(%19#0) <{axis = 1 : i32}> ({
|
89 |
+
^bb0(%arg8: f32, %arg9: f32):
|
90 |
+
%30 = arith.addf %arg8, %arg9 : f32
|
91 |
+
tt.reduce.return %30 : f32
|
92 |
+
}) : (tensor<1x2048xf32, #blocked1>) -> tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
|
93 |
+
%21 = tt.expand_dims %20 {axis = 1 : i32} : (tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<1x1xf32, #blocked1>
|
94 |
+
%22 = tt.addptr %arg4, %1 : !tt.ptr<f32, 1>, i64
|
95 |
+
%23 = tt.splat %22 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>, #blocked1>
|
96 |
+
%24 = tt.splat %2 : (i1) -> tensor<1x1xi1, #blocked1>
|
97 |
+
tt.store %23, %21, %24 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32, #blocked1>
|
98 |
+
%25 = "tt.reduce"(%19#1) <{axis = 1 : i32}> ({
|
99 |
+
^bb0(%arg8: i64, %arg9: i64):
|
100 |
+
%30 = arith.addi %arg8, %arg9 : i64
|
101 |
+
tt.reduce.return %30 : i64
|
102 |
+
}) : (tensor<1x2048xi64, #blocked>) -> tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
|
103 |
+
%26 = triton_gpu.convert_layout %25 : (tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
|
104 |
+
%27 = tt.expand_dims %26 {axis = 1 : i32} : (tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<1x1xi64, #blocked1>
|
105 |
+
%28 = tt.addptr %arg5, %1 : !tt.ptr<i64, 1>, i64
|
106 |
+
%29 = tt.splat %28 : (!tt.ptr<i64, 1>) -> tensor<1x1x!tt.ptr<i64, 1>, #blocked1>
|
107 |
+
tt.store %29, %27, %24 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xi64, #blocked1>
|
108 |
+
tt.return
|
109 |
+
}
|
110 |
+
}
|
.triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttir
ADDED
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
module {
|
2 |
+
tt.func public @triton__0d1d2d3d4d5d6e7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg6: i64 {tt.max_divisibility = 8 : i32}, %arg7: i64 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
3 |
+
%c8_i64 = arith.constant 8 : i64
|
4 |
+
%c7680_i64 = arith.constant 7680 : i64
|
5 |
+
%c385973760_i64 = arith.constant 385973760 : i64
|
6 |
+
%cst = arith.constant dense<0.000000e+00> : tensor<1x2048xbf16>
|
7 |
+
%cst_0 = arith.constant dense<50257> : tensor<1x2048xi64>
|
8 |
+
%cst_1 = arith.constant dense<7680> : tensor<1x2048xi64>
|
9 |
+
%c2048_i32 = arith.constant 2048 : i32
|
10 |
+
%c7680_i32 = arith.constant 7680 : i32
|
11 |
+
%c0_i32 = arith.constant 0 : i32
|
12 |
+
%cst_2 = arith.constant dense<-1> : tensor<1x2048xi64>
|
13 |
+
%cst_3 = arith.constant dense<0> : tensor<1x2048xi64>
|
14 |
+
%cst_4 = arith.constant dense<0.000000e+00> : tensor<1x2048xf32>
|
15 |
+
%0 = tt.get_program_id x : i32
|
16 |
+
%1 = arith.extsi %0 : i32 to i64
|
17 |
+
%2 = arith.cmpi slt, %1, %c8_i64 : i64
|
18 |
+
%3 = tt.splat %2 : (i1) -> tensor<1x1xi1>
|
19 |
+
%4 = tt.make_range {end = 2048 : i32, start = 0 : i32} : tensor<2048xi32>
|
20 |
+
%5 = tt.expand_dims %4 {axis = 0 : i32} : (tensor<2048xi32>) -> tensor<1x2048xi32>
|
21 |
+
%6 = arith.extsi %5 : tensor<1x2048xi32> to tensor<1x2048xi64>
|
22 |
+
%7 = arith.muli %1, %c7680_i64 : i64
|
23 |
+
%8 = tt.splat %7 : (i64) -> tensor<1x2048xi64>
|
24 |
+
%9 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<1x2048x!tt.ptr<i64, 1>>
|
25 |
+
%10 = tt.splat %2 : (i1) -> tensor<1x2048xi1>
|
26 |
+
%11 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>>
|
27 |
+
%12 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>>
|
28 |
+
%13 = arith.muli %1, %c385973760_i64 : i64
|
29 |
+
%14 = tt.splat %13 : (i64) -> tensor<1x2048xi64>
|
30 |
+
%15 = tt.splat %arg1 : (!tt.ptr<bf16, 1>) -> tensor<1x2048x!tt.ptr<bf16, 1>>
|
31 |
+
%16:2 = scf.for %arg8 = %c0_i32 to %c7680_i32 step %c2048_i32 iter_args(%arg9 = %cst_4, %arg10 = %cst_3) -> (tensor<1x2048xf32>, tensor<1x2048xi64>) : i32 {
|
32 |
+
%25 = arith.extsi %arg8 : i32 to i64
|
33 |
+
%26 = tt.splat %25 : (i64) -> tensor<1x2048xi64>
|
34 |
+
%27 = arith.addi %26, %6 : tensor<1x2048xi64>
|
35 |
+
%28 = arith.cmpi slt, %27, %cst_1 : tensor<1x2048xi64>
|
36 |
+
%29 = arith.addi %27, %8 : tensor<1x2048xi64>
|
37 |
+
%30 = tt.addptr %9, %29 : tensor<1x2048x!tt.ptr<i64, 1>>, tensor<1x2048xi64>
|
38 |
+
%31 = arith.andi %28, %10 : tensor<1x2048xi1>
|
39 |
+
%32 = tt.load %30, %31, %cst_3 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xi64>
|
40 |
+
%33 = tt.addptr %11, %29 : tensor<1x2048x!tt.ptr<f32, 1>>, tensor<1x2048xi64>
|
41 |
+
%34 = tt.load %33, %31, %cst_4 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32>
|
42 |
+
%35 = tt.addptr %12, %29 : tensor<1x2048x!tt.ptr<f32, 1>>, tensor<1x2048xi64>
|
43 |
+
%36 = tt.load %35, %31, %cst_4 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32>
|
44 |
+
%37 = arith.cmpi ne, %32, %cst_2 : tensor<1x2048xi64>
|
45 |
+
%38 = arith.select %37, %32, %cst_3 : tensor<1x2048xi1>, tensor<1x2048xi64>
|
46 |
+
%39 = arith.addi %38, %cst_0 : tensor<1x2048xi64>
|
47 |
+
%40 = arith.cmpi slt, %38, %cst_3 : tensor<1x2048xi64>
|
48 |
+
%41 = arith.select %40, %39, %38 : tensor<1x2048xi1>, tensor<1x2048xi64>
|
49 |
+
%42 = arith.cmpi sge, %41, %cst_3 : tensor<1x2048xi64>
|
50 |
+
%43 = arith.cmpi slt, %41, %cst_0 : tensor<1x2048xi64>
|
51 |
+
%44 = arith.andi %42, %43 : tensor<1x2048xi1>
|
52 |
+
tt.assert %44, "index out of bounds: 0 <= tmp7 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<1x2048xi1>
|
53 |
+
%45 = arith.muli %27, %cst_0 : tensor<1x2048xi64>
|
54 |
+
%46 = arith.addi %41, %45 : tensor<1x2048xi64>
|
55 |
+
%47 = arith.addi %46, %14 : tensor<1x2048xi64>
|
56 |
+
%48 = tt.addptr %15, %47 : tensor<1x2048x!tt.ptr<bf16, 1>>, tensor<1x2048xi64>
|
57 |
+
%49 = tt.load %48, %31, %cst {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x2048xbf16>
|
58 |
+
%50 = arith.extf %49 : tensor<1x2048xbf16> to tensor<1x2048xf32>
|
59 |
+
%51 = arith.subf %50, %34 : tensor<1x2048xf32>
|
60 |
+
%52 = math.log %36 : tensor<1x2048xf32>
|
61 |
+
%53 = arith.subf %51, %52 : tensor<1x2048xf32>
|
62 |
+
%54 = arith.subf %cst_4, %53 : tensor<1x2048xf32>
|
63 |
+
%55 = arith.select %37, %54, %cst_4 : tensor<1x2048xi1>, tensor<1x2048xf32>
|
64 |
+
%56 = arith.addf %arg9, %55 : tensor<1x2048xf32>
|
65 |
+
%57 = arith.select %31, %56, %arg9 : tensor<1x2048xi1>, tensor<1x2048xf32>
|
66 |
+
%58 = arith.extui %37 : tensor<1x2048xi1> to tensor<1x2048xi64>
|
67 |
+
%59 = arith.addi %arg10, %58 : tensor<1x2048xi64>
|
68 |
+
%60 = arith.select %31, %59, %arg10 : tensor<1x2048xi1>, tensor<1x2048xi64>
|
69 |
+
scf.yield %57, %60 : tensor<1x2048xf32>, tensor<1x2048xi64>
|
70 |
+
}
|
71 |
+
%17 = "tt.reduce"(%16#0) <{axis = 1 : i32}> ({
|
72 |
+
^bb0(%arg8: f32, %arg9: f32):
|
73 |
+
%25 = arith.addf %arg8, %arg9 : f32
|
74 |
+
tt.reduce.return %25 : f32
|
75 |
+
}) : (tensor<1x2048xf32>) -> tensor<1xf32>
|
76 |
+
%18 = tt.expand_dims %17 {axis = 1 : i32} : (tensor<1xf32>) -> tensor<1x1xf32>
|
77 |
+
%19 = tt.addptr %arg4, %1 : !tt.ptr<f32, 1>, i64
|
78 |
+
%20 = tt.splat %19 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>>
|
79 |
+
tt.store %20, %18, %3 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32>
|
80 |
+
%21 = "tt.reduce"(%16#1) <{axis = 1 : i32}> ({
|
81 |
+
^bb0(%arg8: i64, %arg9: i64):
|
82 |
+
%25 = arith.addi %arg8, %arg9 : i64
|
83 |
+
tt.reduce.return %25 : i64
|
84 |
+
}) : (tensor<1x2048xi64>) -> tensor<1xi64>
|
85 |
+
%22 = tt.expand_dims %21 {axis = 1 : i32} : (tensor<1xi64>) -> tensor<1x1xi64>
|
86 |
+
%23 = tt.addptr %arg5, %1 : !tt.ptr<i64, 1>, i64
|
87 |
+
%24 = tt.splat %23 : (!tt.ptr<i64, 1>) -> tensor<1x1x!tt.ptr<i64, 1>>
|
88 |
+
tt.store %24, %22, %3 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xi64>
|
89 |
+
tt.return
|
90 |
+
}
|
91 |
+
}
|
.triton/dump/ab89f85e55d5ddd9676325b49df9419f/triton_.ttgir
ADDED
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#blocked = #triton_gpu.blocked<{sizePerThread = [4], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
|
2 |
+
#blocked1 = #triton_gpu.blocked<{sizePerThread = [1], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
|
3 |
+
module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
|
4 |
+
tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
|
5 |
+
%cst = arith.constant dense<256> : tensor<256xi32, #blocked>
|
6 |
+
%cst_0 = arith.constant 9.99999974E-6 : f32
|
7 |
+
%cst_1 = arith.constant 2.560000e+02 : f32
|
8 |
+
%cst_2 = arith.constant 0.000000e+00 : f32
|
9 |
+
%c256_i32 = arith.constant 256 : i32
|
10 |
+
%cst_3 = arith.constant dense<0.000000e+00> : tensor<256xf32, #blocked>
|
11 |
+
%cst_4 = arith.constant dense<0.000000e+00> : tensor<256xbf16, #blocked>
|
12 |
+
%0 = tt.get_program_id x : i32
|
13 |
+
%1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32, #blocked>
|
14 |
+
%2 = arith.cmpi slt, %1, %cst : tensor<256xi32, #blocked>
|
15 |
+
%3 = arith.muli %0, %c256_i32 : i32
|
16 |
+
%4 = tt.splat %3 : (i32) -> tensor<256xi32, #blocked>
|
17 |
+
%5 = arith.addi %1, %4 : tensor<256xi32, #blocked>
|
18 |
+
%6 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
19 |
+
%7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
20 |
+
%8 = tt.load %7, %2, %cst_3 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
21 |
+
%9 = tt.splat %arg2 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
22 |
+
%10 = tt.addptr %9, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
23 |
+
%11 = tt.load %10, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
|
24 |
+
%12 = arith.extf %11 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
|
25 |
+
%13 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
|
26 |
+
%14 = tt.addptr %13, %1 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
|
27 |
+
%15 = tt.load %14, %2, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32, #blocked>
|
28 |
+
%16 = arith.addf %8, %12 : tensor<256xf32, #blocked>
|
29 |
+
%17 = arith.select %2, %16, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
30 |
+
%18 = "tt.reduce"(%17) <{axis = 0 : i32}> ({
|
31 |
+
^bb0(%arg8: f32, %arg9: f32):
|
32 |
+
%42 = arith.addf %arg8, %arg9 : f32
|
33 |
+
tt.reduce.return %42 : f32
|
34 |
+
}) : (tensor<256xf32, #blocked>) -> f32
|
35 |
+
%19 = arith.addf %18, %cst_2 : f32
|
36 |
+
%20 = arith.divf %19, %cst_1 : f32
|
37 |
+
%21 = tt.splat %20 : (f32) -> tensor<1xf32, #blocked1>
|
38 |
+
%22 = tt.splat %20 : (f32) -> tensor<256xf32, #blocked>
|
39 |
+
%23 = arith.subf %16, %22 : tensor<256xf32, #blocked>
|
40 |
+
%24 = arith.mulf %23, %23 : tensor<256xf32, #blocked>
|
41 |
+
%25 = arith.select %2, %24, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
|
42 |
+
%26 = "tt.reduce"(%25) <{axis = 0 : i32}> ({
|
43 |
+
^bb0(%arg8: f32, %arg9: f32):
|
44 |
+
%42 = arith.addf %arg8, %arg9 : f32
|
45 |
+
tt.reduce.return %42 : f32
|
46 |
+
}) : (tensor<256xf32, #blocked>) -> f32
|
47 |
+
%27 = arith.addf %26, %cst_2 : f32
|
48 |
+
%28 = arith.divf %27, %cst_1 : f32
|
49 |
+
%29 = arith.addf %28, %cst_0 : f32
|
50 |
+
%30 = tt.extern_elementwise %29 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (f32) -> f32
|
51 |
+
%31 = tt.splat %30 : (f32) -> tensor<1xf32, #blocked1>
|
52 |
+
%32 = tt.splat %30 : (f32) -> tensor<256xf32, #blocked>
|
53 |
+
%33 = arith.mulf %23, %32 : tensor<256xf32, #blocked>
|
54 |
+
%34 = arith.mulf %33, %15 : tensor<256xf32, #blocked>
|
55 |
+
gpu.barrier
|
56 |
+
%35 = tt.addptr %arg0, %0 : !tt.ptr<f32, 1>, i32
|
57 |
+
%36 = tt.splat %35 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
|
58 |
+
tt.store %36, %31 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
|
59 |
+
%37 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
|
60 |
+
%38 = tt.addptr %37, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
|
61 |
+
%39 = arith.truncf %34 : tensor<256xf32, #blocked> to tensor<256xbf16, #blocked>
|
62 |
+
tt.store %38, %39, %2 {cache = 1 : i32, evict = 1 : i32} : tensor<256xbf16, #blocked>
|
63 |
+
%40 = tt.addptr %arg4, %0 : !tt.ptr<f32, 1>, i32
|
64 |
+
%41 = tt.splat %40 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
|
65 |
+
tt.store %41, %21 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
|
66 |
+
tt.return
|
67 |
+
}
|
68 |
+
}
|