0-hero commited on
Commit
485133c
·
verified ·
1 Parent(s): d742687

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .cache/pip/http-v2/0/1/f/2/0/01f2082df50502ba9492d64e69db99d1fdb5730707a16c6264b355b8.body +3 -0
  2. .cache/pip/http-v2/0/2/d/4/2/02d4221e858694abc22129c65515f1df2c4c326330eb1a34ceb0b382.body +3 -0
  3. .cache/pip/http-v2/4/0/2/3/b/4023be7b5b37a7a4144c804ce69828082d4fb2a124d9d8aabc855da8.body +3 -0
  4. .cache/pip/http-v2/4/f/d/2/5/4fd254dbd56deb4021e55d22c4b489f6c776c69c316eb7345bc91691.body +3 -0
  5. .cache/pip/http-v2/9/6/e/8/3/96e83221dd149da9a3d38feebc955beb2034effd910108971c5b167b.body +3 -0
  6. .cache/pip/http-v2/9/e/8/c/8/9e8c8c0496d6d3384d616902379ed05e07b6b1dba9673d70b5fef231.body +3 -0
  7. .cache/pip/http-v2/a/e/7/a/2/ae7a241673cf118ca18eca030dc29d2715b1980127dd0e2949514433.body +3 -0
  8. .cache/pip/http-v2/d/3/3/a/b/d33abf9ad709d023fff05902f39da682c1afb233bcd9f2c479487586.body +3 -0
  9. .cache/pip/http-v2/d/b/1/f/6/db1f6b45c0850c8e2ce7d8b47148edeca6e8115413af41f4ecc8ce32.body +3 -0
  10. .cache/pip/http-v2/f/5/2/7/6/f52769e4b4d00542e1e056baf2db3e5ad8f277bff67f2636cace711d.body +3 -0
  11. .cache/pip/wheels/7e/e3/c3/89c7a2f3c4adc07cd1c675f8bb7b9ad4d18f64a72bccdfe826/flash_attn-2.6.3-cp310-cp310-linux_x86_64.whl +3 -0
  12. .gitattributes +20 -0
  13. .launchpadlib/api.launchpad.net/cache/api.launchpad.net,devel,-application,vnd.sun.wadl+xml,2f09acb494bdefdbf8ef0d1396a05e86 +0 -0
  14. .local/share/jupyter/nbextensions/go_to_current_running_cell/auto_focus.gif +3 -0
  15. .local/share/jupyter/nbextensions/nbTranslate/demo1.gif +3 -0
  16. .local/share/jupyter/nbextensions/nbTranslate/demo2.gif +3 -0
  17. .local/share/jupyter/nbextensions/scratchpad/demo.gif +3 -0
  18. .local/share/jupyter/nbextensions/toc2/demo.gif +3 -0
  19. .local/share/jupyter/nbextensions/zenmode/images/back11.jpg +0 -0
  20. .local/share/jupyter/nbextensions/zenmode/images/back21.jpg +0 -0
  21. .local/share/jupyter/nbextensions/zenmode/images/ipynblogo1.png +0 -0
  22. .triton/dump/11759acf26ac56366b171628132485d6/triton_.llir +347 -0
  23. .triton/dump/11759acf26ac56366b171628132485d6/triton_.ttgir +78 -0
  24. .triton/dump/11759acf26ac56366b171628132485d6/triton_.ttir +76 -0
  25. .triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ptx +988 -0
  26. .triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ttir +104 -0
  27. .triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ptx +782 -0
  28. .triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ttgir +88 -0
  29. .triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.llir +980 -0
  30. .triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ptx +1654 -0
  31. .triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ttir +104 -0
  32. .triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ptx +465 -0
  33. .triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttgir +39 -0
  34. .triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttir +38 -0
  35. .triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.cubin +0 -0
  36. .triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.llir +550 -0
  37. .triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.cubin +0 -0
  38. .triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.llir +85 -0
  39. .triton/dump/510522bb05917b836ed253751364fcad/triton_.cubin +0 -0
  40. .triton/dump/510522bb05917b836ed253751364fcad/triton_.ptx +1810 -0
  41. .triton/dump/510522bb05917b836ed253751364fcad/triton_.ttgir +153 -0
  42. .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.cubin +0 -0
  43. .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.llir +1360 -0
  44. .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ptx +2004 -0
  45. .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttgir +164 -0
  46. .triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttir +151 -0
  47. .triton/dump/76fb48b96c75cb8e388c291a18ef9b02/triton_.ttgir +169 -0
  48. .triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttgir +110 -0
  49. .triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttir +91 -0
  50. .triton/dump/ab89f85e55d5ddd9676325b49df9419f/triton_.ttgir +68 -0
.cache/pip/http-v2/0/1/f/2/0/01f2082df50502ba9492d64e69db99d1fdb5730707a16c6264b355b8.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d4511c52caacf3c4981d1ae2df85908bd31853f33d30b345c8b6830763f769c
3
+ size 1080866
.cache/pip/http-v2/0/2/d/4/2/02d4221e858694abc22129c65515f1df2c4c326330eb1a34ceb0b382.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bfae9500ad8e7d2937ebccb4906f3bc464d1bf66eedd0e4adabd520811c7b52
3
+ size 2631958
.cache/pip/http-v2/4/0/2/3/b/4023be7b5b37a7a4144c804ce69828082d4fb2a124d9d8aabc855da8.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd4c97d69242efd604c1a2077c8b56341e236cfaca78c40f59dcef9b95464fdc
3
+ size 9663908
.cache/pip/http-v2/4/f/d/2/5/4fd254dbd56deb4021e55d22c4b489f6c776c69c316eb7345bc91691.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f99e4769b4750076cd4235c044b61232110733322384a94a63791d2e7beacc66
3
+ size 9995162
.cache/pip/http-v2/9/6/e/8/3/96e83221dd149da9a3d38feebc955beb2034effd910108971c5b167b.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e32dced201274bf96899e6491d9ba3e9a5f6b336708656466ad0522d8528f69
3
+ size 41178528
.cache/pip/http-v2/9/e/8/c/8/9e8c8c0496d6d3384d616902379ed05e07b6b1dba9673d70b5fef231.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d2665c5df629eb2f981dab244c01bfa6cdc185f4ffa026639286c4d56fafb54
3
+ size 1221827
.cache/pip/http-v2/a/e/7/a/2/ae7a241673cf118ca18eca030dc29d2715b1980127dd0e2949514433.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dca7c3956b03b7663fac4d150f5e6d4f6f38b2462c1e9afd83bcf7019f17913
3
+ size 1080679
.cache/pip/http-v2/d/3/3/a/b/d33abf9ad709d023fff05902f39da682c1afb233bcd9f2c479487586.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba0d021a166865d2265246961bec0152ff124de910c5cc39f1156ce3fa7c69dc
3
+ size 2110226
.cache/pip/http-v2/d/b/1/f/6/db1f6b45c0850c8e2ce7d8b47148edeca6e8115413af41f4ecc8ce32.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7ae2de664e0b158d1607699a16a488de3d008ba99b3a7aa5de1cbc13574d047
3
+ size 39855626
.cache/pip/http-v2/f/5/2/7/6/f52769e4b4d00542e1e056baf2db3e5ad8f277bff67f2636cace711d.body ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86976a1c5b25ae3f8ccae3a5306e443569ee3c3faf444dfd0f41cda24667ad57
3
+ size 13064210
.cache/pip/wheels/7e/e3/c3/89c7a2f3c4adc07cd1c675f8bb7b9ad4d18f64a72bccdfe826/flash_attn-2.6.3-cp310-cp310-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8991eedb5038a1ee6fc9904f99c12b40213d66753ed91e261a43d085f5aeab8f
3
+ size 187219571
.gitattributes CHANGED
@@ -33,3 +33,23 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ .cache/pip/http-v2/a/e/7/a/2/ae7a241673cf118ca18eca030dc29d2715b1980127dd0e2949514433.body filter=lfs diff=lfs merge=lfs -text
37
+ .cache/pip/http-v2/0/2/d/4/2/02d4221e858694abc22129c65515f1df2c4c326330eb1a34ceb0b382.body filter=lfs diff=lfs merge=lfs -text
38
+ .local/share/jupyter/nbextensions/nbTranslate/demo2.gif filter=lfs diff=lfs merge=lfs -text
39
+ .cache/pip/http-v2/4/f/d/2/5/4fd254dbd56deb4021e55d22c4b489f6c776c69c316eb7345bc91691.body filter=lfs diff=lfs merge=lfs -text
40
+ .cache/pip/http-v2/0/1/f/2/0/01f2082df50502ba9492d64e69db99d1fdb5730707a16c6264b355b8.body filter=lfs diff=lfs merge=lfs -text
41
+ .cache/pip/http-v2/4/0/2/3/b/4023be7b5b37a7a4144c804ce69828082d4fb2a124d9d8aabc855da8.body filter=lfs diff=lfs merge=lfs -text
42
+ .local/share/jupyter/nbextensions/scratchpad/demo.gif filter=lfs diff=lfs merge=lfs -text
43
+ .cache/pip/http-v2/f/5/2/7/6/f52769e4b4d00542e1e056baf2db3e5ad8f277bff67f2636cace711d.body filter=lfs diff=lfs merge=lfs -text
44
+ .local/share/jupyter/nbextensions/go_to_current_running_cell/auto_focus.gif filter=lfs diff=lfs merge=lfs -text
45
+ .cache/pip/http-v2/9/e/8/c/8/9e8c8c0496d6d3384d616902379ed05e07b6b1dba9673d70b5fef231.body filter=lfs diff=lfs merge=lfs -text
46
+ .local/share/jupyter/nbextensions/nbTranslate/demo1.gif filter=lfs diff=lfs merge=lfs -text
47
+ .local/share/jupyter/nbextensions/toc2/demo.gif filter=lfs diff=lfs merge=lfs -text
48
+ .cache/pip/http-v2/d/3/3/a/b/d33abf9ad709d023fff05902f39da682c1afb233bcd9f2c479487586.body filter=lfs diff=lfs merge=lfs -text
49
+ wandb/run-20240927_074011-gzu8f7wl/run-gzu8f7wl.wandb filter=lfs diff=lfs merge=lfs -text
50
+ .cache/pip/http-v2/9/6/e/8/3/96e83221dd149da9a3d38feebc955beb2034effd910108971c5b167b.body filter=lfs diff=lfs merge=lfs -text
51
+ wandb/run-20240927_005424-60260ulk/run-60260ulk.wandb filter=lfs diff=lfs merge=lfs -text
52
+ wandb/run-20240926_180814-1klxtkie/run-1klxtkie.wandb filter=lfs diff=lfs merge=lfs -text
53
+ wandb/run-20240926_111257-7yvor5gh/run-7yvor5gh.wandb filter=lfs diff=lfs merge=lfs -text
54
+ .cache/pip/http-v2/d/b/1/f/6/db1f6b45c0850c8e2ce7d8b47148edeca6e8115413af41f4ecc8ce32.body filter=lfs diff=lfs merge=lfs -text
55
+ .cache/pip/wheels/7e/e3/c3/89c7a2f3c4adc07cd1c675f8bb7b9ad4d18f64a72bccdfe826/flash_attn-2.6.3-cp310-cp310-linux_x86_64.whl filter=lfs diff=lfs merge=lfs -text
.launchpadlib/api.launchpad.net/cache/api.launchpad.net,devel,-application,vnd.sun.wadl+xml,2f09acb494bdefdbf8ef0d1396a05e86 ADDED
The diff for this file is too large to render. See raw diff
 
.local/share/jupyter/nbextensions/go_to_current_running_cell/auto_focus.gif ADDED

Git LFS Details

  • SHA256: 3dc033a545fe3eccdeee6e66932f1a46de4d0cafe084d471165e750ede1dcc4f
  • Pointer size: 132 Bytes
  • Size of remote file: 1.75 MB
.local/share/jupyter/nbextensions/nbTranslate/demo1.gif ADDED

Git LFS Details

  • SHA256: d5a8c90d8375a3a4fb62dee376de91281b96c7616d3763eeaba9cc1d4f9c1f9c
  • Pointer size: 132 Bytes
  • Size of remote file: 2.34 MB
.local/share/jupyter/nbextensions/nbTranslate/demo2.gif ADDED

Git LFS Details

  • SHA256: c27adeaf2ba905566502c3f319f1d01e59a3683c77392d890c24a4532396884c
  • Pointer size: 132 Bytes
  • Size of remote file: 3.1 MB
.local/share/jupyter/nbextensions/scratchpad/demo.gif ADDED

Git LFS Details

  • SHA256: cfbc6359d32c4b072feea49dca4880a75c6f49dd54c6d14225658f73e0d3ae27
  • Pointer size: 132 Bytes
  • Size of remote file: 1.16 MB
.local/share/jupyter/nbextensions/toc2/demo.gif ADDED

Git LFS Details

  • SHA256: 19189620710630c6073bc9fb06464e6b3b7faa85d09ea52fa70d803d9cfba587
  • Pointer size: 132 Bytes
  • Size of remote file: 3.49 MB
.local/share/jupyter/nbextensions/zenmode/images/back11.jpg ADDED
.local/share/jupyter/nbextensions/zenmode/images/back21.jpg ADDED
.local/share/jupyter/nbextensions/zenmode/images/ipynblogo1.png ADDED
.triton/dump/11759acf26ac56366b171628132485d6/triton_.llir ADDED
@@ -0,0 +1,347 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+
4
+ @global_smem = external addrspace(3) global [0 x i8]
5
+ @.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
6
+
7
+ define void @triton__0d1d2d3d4d5d6d7d8de9de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, ptr addrspace(1) %6, ptr addrspace(1) %7, i32 %8, i32 %9) local_unnamed_addr !dbg !7 {
8
+ %11 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
9
+ %12 = and i32 %11, 31, !dbg !10
10
+ %13 = lshr i32 %11, 5, !dbg !10
11
+ %14 = and i32 %13, 1, !dbg !10
12
+ %urem = and i32 %11, 63, !dbg !10
13
+ %15 = shl nuw nsw i32 %urem, 2, !dbg !10
14
+ %16 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !11
15
+ %17 = shl i32 %16, 8, !dbg !12
16
+ %18 = or i32 %17, %15, !dbg !13
17
+ %19 = sext i32 %18 to i64, !dbg !14
18
+ %20 = getelementptr float, ptr addrspace(1) %1, i64 %19, !dbg !14
19
+ %21 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %20, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !15
20
+ %22 = extractvalue { i32, i32, i32, i32 } %21, 0, !dbg !15
21
+ %23 = extractvalue { i32, i32, i32, i32 } %21, 1, !dbg !15
22
+ %24 = extractvalue { i32, i32, i32, i32 } %21, 2, !dbg !15
23
+ %25 = extractvalue { i32, i32, i32, i32 } %21, 3, !dbg !15
24
+ %26 = bitcast i32 %24 to float, !dbg !15
25
+ %27 = bitcast i32 %25 to float, !dbg !15
26
+ %28 = getelementptr i16, ptr addrspace(1) %2, i64 %19, !dbg !16
27
+ %29 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %28, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !17
28
+ %30 = extractvalue { i32, i32 } %29, 0, !dbg !17
29
+ %31 = extractvalue { i32, i32 } %29, 1, !dbg !17
30
+ %32 = trunc i32 %30 to i16, !dbg !17
31
+ %extelt.offset = lshr i32 %30, 16, !dbg !17
32
+ %33 = trunc i32 %extelt.offset to i16, !dbg !17
33
+ %34 = trunc i32 %31 to i16, !dbg !17
34
+ %extelt.offset1 = lshr i32 %31, 16, !dbg !17
35
+ %35 = trunc i32 %extelt.offset1 to i16, !dbg !17
36
+ %36 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %32) #6, !dbg !18
37
+ %37 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %33) #6, !dbg !18
38
+ %38 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %34) #6, !dbg !18
39
+ %39 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %35) #6, !dbg !18
40
+ %40 = getelementptr i16, ptr addrspace(1) %3, i64 %19, !dbg !19
41
+ %41 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %40, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !20
42
+ %42 = extractvalue { i32, i32 } %41, 0, !dbg !20
43
+ %43 = extractvalue { i32, i32 } %41, 1, !dbg !20
44
+ %44 = trunc i32 %42 to i16, !dbg !20
45
+ %extelt.offset2 = lshr i32 %42, 16, !dbg !20
46
+ %45 = trunc i32 %extelt.offset2 to i16, !dbg !20
47
+ %46 = trunc i32 %43 to i16, !dbg !20
48
+ %extelt.offset3 = lshr i32 %43, 16, !dbg !20
49
+ %47 = trunc i32 %extelt.offset3 to i16, !dbg !20
50
+ %48 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %44) #6, !dbg !21
51
+ %49 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %45) #6, !dbg !21
52
+ %50 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %46) #6, !dbg !21
53
+ %51 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %47) #6, !dbg !21
54
+ %52 = getelementptr i16, ptr addrspace(1) %4, i64 %19, !dbg !22
55
+ %53 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %52, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !23
56
+ %54 = extractvalue { i32, i32 } %53, 0, !dbg !23
57
+ %55 = extractvalue { i32, i32 } %53, 1, !dbg !23
58
+ %56 = trunc i32 %54 to i16, !dbg !23
59
+ %extelt.offset4 = lshr i32 %54, 16, !dbg !23
60
+ %57 = trunc i32 %extelt.offset4 to i16, !dbg !23
61
+ %58 = trunc i32 %55 to i16, !dbg !23
62
+ %extelt.offset5 = lshr i32 %55, 16, !dbg !23
63
+ %59 = trunc i32 %extelt.offset5 to i16, !dbg !23
64
+ %60 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %56) #6, !dbg !24
65
+ %61 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %57) #6, !dbg !24
66
+ %62 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %58) #6, !dbg !24
67
+ %63 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %59) #6, !dbg !24
68
+ %64 = zext nneg i32 %15 to i64, !dbg !25
69
+ %65 = getelementptr float, ptr addrspace(1) %5, i64 %64, !dbg !25
70
+ %66 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %65, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !26
71
+ %67 = fadd float %38, %26, !dbg !27
72
+ %68 = fadd float %39, %27, !dbg !27
73
+ %69 = fadd float %67, %50, !dbg !28
74
+ %70 = fadd float %68, %51, !dbg !28
75
+ %71 = insertelement <2 x i32> poison, i32 %22, i64 0, !dbg !15
76
+ %72 = insertelement <2 x i32> %71, i32 %23, i64 1, !dbg !15
77
+ %73 = bitcast <2 x i32> %72 to <2 x float>, !dbg !15
78
+ %74 = insertelement <2 x float> poison, float %36, i64 0, !dbg !27
79
+ %75 = insertelement <2 x float> %74, float %37, i64 1, !dbg !27
80
+ %76 = fadd <2 x float> %75, %73, !dbg !27
81
+ %77 = insertelement <2 x float> poison, float %48, i64 0, !dbg !28
82
+ %78 = insertelement <2 x float> %77, float %49, i64 1, !dbg !28
83
+ %79 = fadd <2 x float> %76, %78, !dbg !28
84
+ %80 = insertelement <2 x float> poison, float %60, i64 0, !dbg !29
85
+ %81 = insertelement <2 x float> %80, float %61, i64 1, !dbg !29
86
+ %82 = fadd <2 x float> %79, %81, !dbg !29
87
+ %83 = fadd float %69, %62, !dbg !29
88
+ %84 = fadd float %70, %63, !dbg !29
89
+ %85 = extractelement <2 x float> %82, i64 0, !dbg !30
90
+ %86 = extractelement <2 x float> %82, i64 1, !dbg !30
91
+ %87 = fadd float %85, %86, !dbg !30
92
+ %88 = fadd float %87, %83, !dbg !30
93
+ %89 = fadd float %88, %84, !dbg !30
94
+ %90 = bitcast float %89 to i32, !dbg !36
95
+ %91 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %90, i32 16, i32 31), !dbg !36
96
+ %92 = bitcast i32 %91 to float, !dbg !36
97
+ %93 = fadd float %89, %92, !dbg !30
98
+ %94 = bitcast float %93 to i32, !dbg !36
99
+ %95 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %94, i32 8, i32 31), !dbg !36
100
+ %96 = bitcast i32 %95 to float, !dbg !36
101
+ %97 = fadd float %93, %96, !dbg !30
102
+ %98 = bitcast float %97 to i32, !dbg !36
103
+ %99 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %98, i32 4, i32 31), !dbg !36
104
+ %100 = bitcast i32 %99 to float, !dbg !36
105
+ %101 = fadd float %97, %100, !dbg !30
106
+ %102 = bitcast float %101 to i32, !dbg !36
107
+ %103 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %102, i32 2, i32 31), !dbg !36
108
+ %104 = bitcast i32 %103 to float, !dbg !36
109
+ %105 = fadd float %101, %104, !dbg !30
110
+ %106 = bitcast float %105 to i32, !dbg !36
111
+ %107 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %106, i32 1, i32 31), !dbg !36
112
+ %108 = bitcast i32 %107 to float, !dbg !36
113
+ %109 = fadd float %105, %108, !dbg !30
114
+ %110 = icmp eq i32 %12, 0, !dbg !36
115
+ %111 = zext nneg i32 %14 to i64, !dbg !36
116
+ %112 = getelementptr float, ptr addrspace(3) @global_smem, i64 %111, !dbg !36
117
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %112, float %109, i1 %110) #6, !dbg !36
118
+ tail call void @llvm.nvvm.barrier0(), !dbg !36
119
+ %113 = icmp slt i32 %11, 2, !dbg !36
120
+ %114 = sext i32 %11 to i64, !dbg !36
121
+ %115 = getelementptr float, ptr addrspace(3) @global_smem, i64 %114, !dbg !36
122
+ %116 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %115, i1 %113) #6, !dbg !36
123
+ %117 = bitcast float %116 to i32, !dbg !36
124
+ %118 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %117, i32 1, i32 31), !dbg !36
125
+ %119 = bitcast i32 %118 to float, !dbg !36
126
+ %120 = fadd float %116, %119, !dbg !30
127
+ %121 = and i32 %11, 1, !dbg !36
128
+ %122 = icmp eq i32 %121, 0, !dbg !36
129
+ %123 = and i1 %113, %122, !dbg !36
130
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %115, float %120, i1 %123) #6, !dbg !36
131
+ tail call void @llvm.nvvm.barrier0(), !dbg !36
132
+ %124 = load float, ptr addrspace(3) @global_smem, align 4, !dbg !36
133
+ %125 = fadd float %124, 0.000000e+00, !dbg !38
134
+ %126 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %125, float 2.560000e+02) #6, !dbg !42
135
+ %127 = fsub float %85, %126, !dbg !43
136
+ %128 = fsub float %86, %126, !dbg !43
137
+ %129 = fsub float %83, %126, !dbg !43
138
+ %130 = fsub float %84, %126, !dbg !43
139
+ %131 = fmul float %127, %127, !dbg !44
140
+ %132 = fmul float %128, %128, !dbg !44
141
+ %133 = fmul float %129, %129, !dbg !44
142
+ %134 = fmul float %130, %130, !dbg !44
143
+ tail call void @llvm.nvvm.barrier0(), !dbg !45
144
+ %135 = fadd float %131, %132, !dbg !47
145
+ %136 = fadd float %133, %135, !dbg !47
146
+ %137 = fadd float %134, %136, !dbg !47
147
+ %138 = bitcast float %137 to i32, !dbg !45
148
+ %139 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %138, i32 16, i32 31), !dbg !45
149
+ %140 = bitcast i32 %139 to float, !dbg !45
150
+ %141 = fadd float %137, %140, !dbg !47
151
+ %142 = bitcast float %141 to i32, !dbg !45
152
+ %143 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %142, i32 8, i32 31), !dbg !45
153
+ %144 = bitcast i32 %143 to float, !dbg !45
154
+ %145 = fadd float %141, %144, !dbg !47
155
+ %146 = bitcast float %145 to i32, !dbg !45
156
+ %147 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %146, i32 4, i32 31), !dbg !45
157
+ %148 = bitcast i32 %147 to float, !dbg !45
158
+ %149 = fadd float %145, %148, !dbg !47
159
+ %150 = bitcast float %149 to i32, !dbg !45
160
+ %151 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %150, i32 2, i32 31), !dbg !45
161
+ %152 = bitcast i32 %151 to float, !dbg !45
162
+ %153 = fadd float %149, %152, !dbg !47
163
+ %154 = bitcast float %153 to i32, !dbg !45
164
+ %155 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %154, i32 1, i32 31), !dbg !45
165
+ %156 = bitcast i32 %155 to float, !dbg !45
166
+ %157 = fadd float %153, %156, !dbg !47
167
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %112, float %157, i1 %110) #6, !dbg !45
168
+ tail call void @llvm.nvvm.barrier0(), !dbg !45
169
+ %158 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %115, i1 %113) #6, !dbg !45
170
+ %159 = bitcast float %158 to i32, !dbg !45
171
+ %160 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %159, i32 1, i32 31), !dbg !45
172
+ %161 = bitcast i32 %160 to float, !dbg !45
173
+ %162 = fadd float %158, %161, !dbg !47
174
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %115, float %162, i1 %123) #6, !dbg !45
175
+ tail call void @llvm.nvvm.barrier0(), !dbg !45
176
+ %163 = load float, ptr addrspace(3) @global_smem, align 4, !dbg !45
177
+ %164 = fadd float %163, 0.000000e+00, !dbg !50
178
+ %165 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %164, float 2.560000e+02) #6, !dbg !52
179
+ %166 = fadd float %165, 0x3EE4F8B580000000, !dbg !53
180
+ %167 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !54
181
+ %.not.i = icmp eq i32 %167, 0, !dbg !54
182
+ br i1 %.not.i, label %170, label %168, !dbg !54
183
+
184
+ 168: ; preds = %10
185
+ %169 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %166), !dbg !54
186
+ br label %__nv_rsqrtf.exit, !dbg !54
187
+
188
+ 170: ; preds = %10
189
+ %171 = tail call float @llvm.nvvm.rsqrt.approx.f(float %166), !dbg !54
190
+ br label %__nv_rsqrtf.exit, !dbg !54
191
+
192
+ __nv_rsqrtf.exit: ; preds = %168, %170
193
+ %.0.i = phi float [ %169, %168 ], [ %171, %170 ], !dbg !54
194
+ %172 = extractvalue { i32, i32, i32, i32 } %66, 3, !dbg !26
195
+ %173 = bitcast i32 %172 to float, !dbg !26
196
+ %174 = extractvalue { i32, i32, i32, i32 } %66, 2, !dbg !26
197
+ %175 = bitcast i32 %174 to float, !dbg !26
198
+ %176 = extractvalue { i32, i32, i32, i32 } %66, 1, !dbg !26
199
+ %177 = bitcast i32 %176 to float, !dbg !26
200
+ %178 = extractvalue { i32, i32, i32, i32 } %66, 0, !dbg !26
201
+ %179 = bitcast i32 %178 to float, !dbg !26
202
+ %180 = fmul float %127, %.0.i, !dbg !55
203
+ %181 = fmul float %128, %.0.i, !dbg !55
204
+ %182 = fmul float %129, %.0.i, !dbg !55
205
+ %183 = fmul float %130, %.0.i, !dbg !55
206
+ %184 = fmul float %180, %179, !dbg !56
207
+ %185 = fmul float %181, %177, !dbg !56
208
+ %186 = fmul float %182, %175, !dbg !56
209
+ %187 = fmul float %183, %173, !dbg !56
210
+ tail call void @llvm.nvvm.barrier0(), !dbg !57
211
+ %188 = sext i32 %16 to i64, !dbg !58
212
+ %189 = getelementptr float, ptr addrspace(1) %0, i64 %188, !dbg !58
213
+ %190 = icmp eq i32 %urem, 0, !dbg !59
214
+ %191 = bitcast float %.0.i to i32, !dbg !59
215
+ tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %191, ptr addrspace(1) %189, i1 %190) #6, !dbg !59
216
+ %192 = getelementptr i16, ptr addrspace(1) %7, i64 %19, !dbg !60
217
+ %193 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %184) #6, !dbg !61
218
+ %194 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %185) #6, !dbg !61
219
+ %195 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %186) #6, !dbg !61
220
+ %196 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %187) #6, !dbg !61
221
+ %197 = insertelement <2 x i16> undef, i16 %193, i64 0, !dbg !61
222
+ %198 = insertelement <2 x i16> %197, i16 %194, i64 1, !dbg !61
223
+ %199 = bitcast <2 x i16> %198 to i32, !dbg !61
224
+ %200 = insertelement <2 x i16> undef, i16 %195, i64 0, !dbg !61
225
+ %201 = insertelement <2 x i16> %200, i16 %196, i64 1, !dbg !61
226
+ %202 = bitcast <2 x i16> %201 to i32, !dbg !61
227
+ tail call void asm sideeffect "@$3 st.global.v2.b32 [ $2 + 0 ], { $0, $1 };", "r,r,l,b"(i32 %199, i32 %202, ptr addrspace(1) %192, i1 true) #6, !dbg !61
228
+ %203 = getelementptr float, ptr addrspace(1) %6, i64 %188, !dbg !62
229
+ %204 = bitcast float %126 to i32, !dbg !63
230
+ tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %204, ptr addrspace(1) %203, i1 %190) #6, !dbg !63
231
+ ret void, !dbg !64
232
+ }
233
+
234
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
235
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
236
+
237
+ ; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
238
+ declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
239
+
240
+ ; Function Attrs: convergent nocallback nounwind
241
+ declare void @llvm.nvvm.barrier0() #2
242
+
243
+ ; Function Attrs: alwaysinline nounwind
244
+ define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
245
+ %1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
246
+ %.not = icmp eq i32 %1, 0
247
+ br i1 %.not, label %4, label %2
248
+
249
+ 2: ; preds = %0
250
+ %3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
251
+ br label %6
252
+
253
+ 4: ; preds = %0
254
+ %5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
255
+ br label %6
256
+
257
+ 6: ; preds = %4, %2
258
+ %.0 = phi float [ %3, %2 ], [ %5, %4 ]
259
+ ret float %.0
260
+ }
261
+
262
+ declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
263
+
264
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
265
+ declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
266
+
267
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
268
+ declare float @llvm.nvvm.rsqrt.approx.f(float) #5
269
+
270
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
271
+ attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
272
+ attributes #2 = { convergent nocallback nounwind }
273
+ attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
274
+ attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
275
+ attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
276
+ attributes #6 = { nounwind }
277
+
278
+ !llvm.module.flags = !{!0, !1}
279
+ !llvm.dbg.cu = !{!2}
280
+ !nvvm.annotations = !{!4, !5, !5, !4}
281
+ !llvm.ident = !{!6}
282
+
283
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
284
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
285
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
286
+ !3 = !DIFile(filename: "cdxa5yqgsimvskocpuiz4ajfrjfcwys3opyrdv53xfphj4576qx7.py", directory: "/tmp/torchinductor_root/dx")
287
+ !4 = !{ptr @triton__0d1d2d3d4d5d6d7d8de9de, !"kernel", i32 1}
288
+ !5 = !{ptr @triton__0d1d2d3d4d5d6d7d8de9de, !"maxntidx", i32 64}
289
+ !6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
290
+ !7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5d6d7d8de9de", linkageName: "triton__0d1d2d3d4d5d6d7d8de9de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
291
+ !8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
292
+ !9 = !{}
293
+ !10 = !DILocation(line: 26, column: 26, scope: !7)
294
+ !11 = !DILocation(line: 23, column: 28, scope: !7)
295
+ !12 = !DILocation(line: 30, column: 40, scope: !7)
296
+ !13 = !DILocation(line: 30, column: 36, scope: !7)
297
+ !14 = !DILocation(line: 30, column: 30, scope: !7)
298
+ !15 = !DILocation(line: 30, column: 46, scope: !7)
299
+ !16 = !DILocation(line: 31, column: 30, scope: !7)
300
+ !17 = !DILocation(line: 31, column: 46, scope: !7)
301
+ !18 = !DILocation(line: 31, column: 67, scope: !7)
302
+ !19 = !DILocation(line: 32, column: 30, scope: !7)
303
+ !20 = !DILocation(line: 32, column: 46, scope: !7)
304
+ !21 = !DILocation(line: 32, column: 67, scope: !7)
305
+ !22 = !DILocation(line: 33, column: 30, scope: !7)
306
+ !23 = !DILocation(line: 33, column: 46, scope: !7)
307
+ !24 = !DILocation(line: 33, column: 67, scope: !7)
308
+ !25 = !DILocation(line: 34, column: 31, scope: !7)
309
+ !26 = !DILocation(line: 34, column: 36, scope: !7)
310
+ !27 = !DILocation(line: 36, column: 18, scope: !7)
311
+ !28 = !DILocation(line: 38, column: 18, scope: !7)
312
+ !29 = !DILocation(line: 40, column: 18, scope: !7)
313
+ !30 = !DILocation(line: 233, column: 15, scope: !31, inlinedAt: !34)
314
+ !31 = distinct !DILexicalBlockFile(scope: !33, file: !32, discriminator: 0)
315
+ !32 = !DIFile(filename: "standard.py", directory: "/usr/local/lib/python3.10/dist-packages/triton/language")
316
+ !33 = distinct !DILexicalBlockFile(scope: !7, file: !32, discriminator: 0)
317
+ !34 = !DILocation(line: 243, column: 36, scope: !31, inlinedAt: !35)
318
+ !35 = !DILocation(line: 45, column: 59, scope: !31)
319
+ !36 = !DILocation(line: 243, column: 36, scope: !33, inlinedAt: !37)
320
+ !37 = !DILocation(line: 45, column: 59, scope: !33)
321
+ !38 = !DILocation(line: 8, column: 15, scope: !39, inlinedAt: !41)
322
+ !39 = distinct !DILexicalBlockFile(scope: !7, file: !40, discriminator: 0)
323
+ !40 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
324
+ !41 = !DILocation(line: 45, column: 45, scope: !39)
325
+ !42 = !DILocation(line: 48, column: 20, scope: !7)
326
+ !43 = !DILocation(line: 49, column: 20, scope: !7)
327
+ !44 = !DILocation(line: 50, column: 20, scope: !7)
328
+ !45 = !DILocation(line: 243, column: 36, scope: !33, inlinedAt: !46)
329
+ !46 = !DILocation(line: 53, column: 59, scope: !33)
330
+ !47 = !DILocation(line: 233, column: 15, scope: !31, inlinedAt: !48)
331
+ !48 = !DILocation(line: 243, column: 36, scope: !31, inlinedAt: !49)
332
+ !49 = !DILocation(line: 53, column: 59, scope: !31)
333
+ !50 = !DILocation(line: 8, column: 15, scope: !39, inlinedAt: !51)
334
+ !51 = !DILocation(line: 53, column: 45, scope: !39)
335
+ !52 = !DILocation(line: 55, column: 20, scope: !7)
336
+ !53 = !DILocation(line: 57, column: 20, scope: !7)
337
+ !54 = !DILocation(line: 58, column: 26, scope: !7)
338
+ !55 = !DILocation(line: 60, column: 20, scope: !7)
339
+ !56 = !DILocation(line: 61, column: 20, scope: !7)
340
+ !57 = !DILocation(line: 63, column: 4, scope: !7)
341
+ !58 = !DILocation(line: 64, column: 28, scope: !7)
342
+ !59 = !DILocation(line: 64, column: 40, scope: !7)
343
+ !60 = !DILocation(line: 65, column: 25, scope: !7)
344
+ !61 = !DILocation(line: 65, column: 48, scope: !7)
345
+ !62 = !DILocation(line: 66, column: 25, scope: !7)
346
+ !63 = !DILocation(line: 66, column: 37, scope: !7)
347
+ !64 = !DILocation(line: 66, column: 4, scope: !7)
.triton/dump/11759acf26ac56366b171628132485d6/triton_.ttgir ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [4], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
2
+ #blocked1 = #triton_gpu.blocked<{sizePerThread = [1], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
3
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
4
+ tt.func public @triton__0d1d2d3d4d5d6d7d8de9de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg6: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg7: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg8: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg9: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
5
+ %cst = arith.constant dense<256> : tensor<256xi32, #blocked>
6
+ %cst_0 = arith.constant 9.99999974E-6 : f32
7
+ %cst_1 = arith.constant 2.560000e+02 : f32
8
+ %cst_2 = arith.constant 0.000000e+00 : f32
9
+ %c256_i32 = arith.constant 256 : i32
10
+ %cst_3 = arith.constant dense<0.000000e+00> : tensor<256xf32, #blocked>
11
+ %cst_4 = arith.constant dense<0.000000e+00> : tensor<256xbf16, #blocked>
12
+ %0 = tt.get_program_id x : i32
13
+ %1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32, #blocked>
14
+ %2 = arith.cmpi slt, %1, %cst : tensor<256xi32, #blocked>
15
+ %3 = arith.muli %0, %c256_i32 : i32
16
+ %4 = tt.splat %3 : (i32) -> tensor<256xi32, #blocked>
17
+ %5 = arith.addi %1, %4 : tensor<256xi32, #blocked>
18
+ %6 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
19
+ %7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
20
+ %8 = tt.load %7, %2, %cst_3 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
21
+ %9 = tt.splat %arg2 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
22
+ %10 = tt.addptr %9, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
23
+ %11 = tt.load %10, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
24
+ %12 = arith.extf %11 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
25
+ %13 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
26
+ %14 = tt.addptr %13, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
27
+ %15 = tt.load %14, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
28
+ %16 = arith.extf %15 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
29
+ %17 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
30
+ %18 = tt.addptr %17, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
31
+ %19 = tt.load %18, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
32
+ %20 = arith.extf %19 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
33
+ %21 = tt.splat %arg5 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
34
+ %22 = tt.addptr %21, %1 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
35
+ %23 = tt.load %22, %2, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32, #blocked>
36
+ %24 = arith.addf %8, %12 : tensor<256xf32, #blocked>
37
+ %25 = arith.addf %24, %16 : tensor<256xf32, #blocked>
38
+ %26 = arith.addf %25, %20 : tensor<256xf32, #blocked>
39
+ %27 = arith.select %2, %26, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
40
+ %28 = "tt.reduce"(%27) <{axis = 0 : i32}> ({
41
+ ^bb0(%arg10: f32, %arg11: f32):
42
+ %52 = arith.addf %arg10, %arg11 : f32
43
+ tt.reduce.return %52 : f32
44
+ }) : (tensor<256xf32, #blocked>) -> f32
45
+ %29 = arith.addf %28, %cst_2 : f32
46
+ %30 = arith.divf %29, %cst_1 : f32
47
+ %31 = tt.splat %30 : (f32) -> tensor<1xf32, #blocked1>
48
+ %32 = tt.splat %30 : (f32) -> tensor<256xf32, #blocked>
49
+ %33 = arith.subf %26, %32 : tensor<256xf32, #blocked>
50
+ %34 = arith.mulf %33, %33 : tensor<256xf32, #blocked>
51
+ %35 = arith.select %2, %34, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
52
+ %36 = "tt.reduce"(%35) <{axis = 0 : i32}> ({
53
+ ^bb0(%arg10: f32, %arg11: f32):
54
+ %52 = arith.addf %arg10, %arg11 : f32
55
+ tt.reduce.return %52 : f32
56
+ }) : (tensor<256xf32, #blocked>) -> f32
57
+ %37 = arith.addf %36, %cst_2 : f32
58
+ %38 = arith.divf %37, %cst_1 : f32
59
+ %39 = arith.addf %38, %cst_0 : f32
60
+ %40 = tt.extern_elementwise %39 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (f32) -> f32
61
+ %41 = tt.splat %40 : (f32) -> tensor<1xf32, #blocked1>
62
+ %42 = tt.splat %40 : (f32) -> tensor<256xf32, #blocked>
63
+ %43 = arith.mulf %33, %42 : tensor<256xf32, #blocked>
64
+ %44 = arith.mulf %43, %23 : tensor<256xf32, #blocked>
65
+ gpu.barrier
66
+ %45 = tt.addptr %arg0, %0 : !tt.ptr<f32, 1>, i32
67
+ %46 = tt.splat %45 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
68
+ tt.store %46, %41 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
69
+ %47 = tt.splat %arg7 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
70
+ %48 = tt.addptr %47, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
71
+ %49 = arith.truncf %44 : tensor<256xf32, #blocked> to tensor<256xbf16, #blocked>
72
+ tt.store %48, %49, %2 {cache = 1 : i32, evict = 1 : i32} : tensor<256xbf16, #blocked>
73
+ %50 = tt.addptr %arg6, %0 : !tt.ptr<f32, 1>, i32
74
+ %51 = tt.splat %50 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
75
+ tt.store %51, %31 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
76
+ tt.return
77
+ }
78
+ }
.triton/dump/11759acf26ac56366b171628132485d6/triton_.ttir ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ module {
2
+ tt.func public @triton__0d1d2d3d4d5d6d7d8de9de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg6: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg7: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg8: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg9: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
3
+ %c256_i32 = arith.constant 256 : i32
4
+ %cst = arith.constant dense<0.000000e+00> : tensor<256xbf16>
5
+ %cst_0 = arith.constant 0.000000e+00 : f32
6
+ %cst_1 = arith.constant 2.560000e+02 : f32
7
+ %cst_2 = arith.constant 9.99999974E-6 : f32
8
+ %cst_3 = arith.constant dense<0.000000e+00> : tensor<256xf32>
9
+ %cst_4 = arith.constant dense<256> : tensor<256xi32>
10
+ %0 = tt.get_program_id x : i32
11
+ %1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32>
12
+ %2 = arith.cmpi slt, %1, %cst_4 : tensor<256xi32>
13
+ %3 = arith.muli %0, %c256_i32 : i32
14
+ %4 = tt.splat %3 : (i32) -> tensor<256xi32>
15
+ %5 = arith.addi %1, %4 : tensor<256xi32>
16
+ %6 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>>
17
+ %7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<f32, 1>>, tensor<256xi32>
18
+ %8 = tt.load %7, %2, %cst_3 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32>
19
+ %9 = tt.splat %arg2 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
20
+ %10 = tt.addptr %9, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
21
+ %11 = tt.load %10, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16>
22
+ %12 = arith.extf %11 : tensor<256xbf16> to tensor<256xf32>
23
+ %13 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
24
+ %14 = tt.addptr %13, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
25
+ %15 = tt.load %14, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16>
26
+ %16 = arith.extf %15 : tensor<256xbf16> to tensor<256xf32>
27
+ %17 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
28
+ %18 = tt.addptr %17, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
29
+ %19 = tt.load %18, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16>
30
+ %20 = arith.extf %19 : tensor<256xbf16> to tensor<256xf32>
31
+ %21 = tt.splat %arg5 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>>
32
+ %22 = tt.addptr %21, %1 : tensor<256x!tt.ptr<f32, 1>>, tensor<256xi32>
33
+ %23 = tt.load %22, %2, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32>
34
+ %24 = arith.addf %8, %12 : tensor<256xf32>
35
+ %25 = arith.addf %24, %16 : tensor<256xf32>
36
+ %26 = arith.addf %25, %20 : tensor<256xf32>
37
+ %27 = arith.select %2, %26, %cst_3 : tensor<256xi1>, tensor<256xf32>
38
+ %28 = "tt.reduce"(%27) <{axis = 0 : i32}> ({
39
+ ^bb0(%arg10: f32, %arg11: f32):
40
+ %52 = arith.addf %arg10, %arg11 : f32
41
+ tt.reduce.return %52 : f32
42
+ }) : (tensor<256xf32>) -> f32
43
+ %29 = arith.addf %28, %cst_0 : f32
44
+ %30 = arith.divf %29, %cst_1 : f32
45
+ %31 = tt.splat %30 : (f32) -> tensor<1xf32>
46
+ %32 = tt.splat %30 : (f32) -> tensor<256xf32>
47
+ %33 = arith.subf %26, %32 : tensor<256xf32>
48
+ %34 = arith.mulf %33, %33 : tensor<256xf32>
49
+ %35 = arith.select %2, %34, %cst_3 : tensor<256xi1>, tensor<256xf32>
50
+ %36 = "tt.reduce"(%35) <{axis = 0 : i32}> ({
51
+ ^bb0(%arg10: f32, %arg11: f32):
52
+ %52 = arith.addf %arg10, %arg11 : f32
53
+ tt.reduce.return %52 : f32
54
+ }) : (tensor<256xf32>) -> f32
55
+ %37 = arith.addf %36, %cst_0 : f32
56
+ %38 = arith.divf %37, %cst_1 : f32
57
+ %39 = arith.addf %38, %cst_2 : f32
58
+ %40 = tt.extern_elementwise %39 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (f32) -> f32
59
+ %41 = tt.splat %40 : (f32) -> tensor<1xf32>
60
+ %42 = tt.splat %40 : (f32) -> tensor<256xf32>
61
+ %43 = arith.mulf %33, %42 : tensor<256xf32>
62
+ %44 = arith.mulf %43, %23 : tensor<256xf32>
63
+ gpu.barrier
64
+ %45 = tt.addptr %arg0, %0 : !tt.ptr<f32, 1>, i32
65
+ %46 = tt.splat %45 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>>
66
+ tt.store %46, %41 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32>
67
+ %47 = tt.splat %arg7 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>>
68
+ %48 = tt.addptr %47, %5 : tensor<256x!tt.ptr<bf16, 1>>, tensor<256xi32>
69
+ %49 = arith.truncf %44 : tensor<256xf32> to tensor<256xbf16>
70
+ tt.store %48, %49, %2 {cache = 1 : i32, evict = 1 : i32} : tensor<256xbf16>
71
+ %50 = tt.addptr %arg6, %0 : !tt.ptr<f32, 1>, i32
72
+ %51 = tt.splat %50 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>>
73
+ tt.store %51, %31 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32>
74
+ tt.return
75
+ }
76
+ }
.triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ptx ADDED
@@ -0,0 +1,988 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.2
6
+ .target sm_89
7
+ .address_size 64
8
+
9
+ // .globl triton__0d1d2d3d4d5de6de
10
+ .extern .func __assertfail
11
+ (
12
+ .param .b64 __assertfail_param_0,
13
+ .param .b64 __assertfail_param_1,
14
+ .param .b32 __assertfail_param_2,
15
+ .param .b64 __assertfail_param_3,
16
+ .param .b64 __assertfail_param_4
17
+ )
18
+ ;
19
+ .global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
20
+ .global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
21
+ .global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 51, 32, 60, 32, 53, 48, 50, 53, 55};
22
+ .global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
23
+ .global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
24
+ .global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
25
+ .extern .shared .align 1 .b8 global_smem[];
26
+ .global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
27
+
28
+ .visible .entry triton__0d1d2d3d4d5de6de(
29
+ .param .u64 triton__0d1d2d3d4d5de6de_param_0,
30
+ .param .u64 triton__0d1d2d3d4d5de6de_param_1,
31
+ .param .u64 triton__0d1d2d3d4d5de6de_param_2,
32
+ .param .u64 triton__0d1d2d3d4d5de6de_param_3,
33
+ .param .u64 triton__0d1d2d3d4d5de6de_param_4,
34
+ .param .u32 triton__0d1d2d3d4d5de6de_param_5,
35
+ .param .u32 triton__0d1d2d3d4d5de6de_param_6
36
+ )
37
+ .maxntid 128, 1, 1
38
+ {
39
+ .reg .pred %p<50>;
40
+ .reg .b16 %rs<5>;
41
+ .reg .b32 %r<169>;
42
+ .reg .f32 %f<153>;
43
+ .reg .b64 %rd<53>;
44
+ .loc 1 18 0
45
+ $L__func_begin0:
46
+ .loc 1 18 0
47
+
48
+ ld.param.u64 %rd6, [triton__0d1d2d3d4d5de6de_param_3];
49
+ ld.param.u64 %rd5, [triton__0d1d2d3d4d5de6de_param_1];
50
+ ld.param.u64 %rd19, [triton__0d1d2d3d4d5de6de_param_0];
51
+ $L__tmp0:
52
+ .loc 1 22 44
53
+ mov.u32 %r1, %tid.x;
54
+ and.b32 %r2, %r1, 31;
55
+ ld.param.u64 %rd20, [triton__0d1d2d3d4d5de6de_param_2];
56
+ bfe.u32 %r3, %r1, 6, 1;
57
+ and.b32 %r4, %r1, 1;
58
+ .loc 1 24 33
59
+ bfe.u32 %r5, %r1, 5, 1;
60
+ shl.b32 %r24, %r1, 2;
61
+ and.b32 %r6, %r24, 252;
62
+ shl.b32 %r25, %r1, 1;
63
+ and.b32 %r7, %r25, 254;
64
+ .loc 1 21 28
65
+ mov.u32 %r15, %ctaid.x;
66
+ .loc 1 21 33
67
+ shl.b32 %r26, %r15, 1;
68
+ .loc 1 22 23
69
+ or.b32 %r8, %r26, %r3;
70
+ or.b32 %r27, %r26, %r4;
71
+ .loc 1 26 30
72
+ mul.wide.s32 %rd21, %r8, 8;
73
+ add.s64 %rd9, %rd19, %rd21;
74
+ mul.wide.s32 %rd22, %r27, 8;
75
+ add.s64 %rd17, %rd19, %rd22;
76
+ mov.pred %p44, -1;
77
+ .loc 1 26 35
78
+ mov.u64 %rd8, 0x0;
79
+ @%p44 ld.global.L1::evict_last.b64 { %rd8 }, [ %rd9 + 0 ];
80
+ mov.u64 %rd10, 0x0;
81
+ @%p44 ld.global.L1::evict_last.b64 { %rd10 }, [ %rd9 + 0 ];
82
+ mov.u64 %rd12, 0x0;
83
+ @%p44 ld.global.L1::evict_last.b64 { %rd12 }, [ %rd9 + 0 ];
84
+ mov.u64 %rd14, 0x0;
85
+ @%p44 ld.global.L1::evict_last.b64 { %rd14 }, [ %rd9 + 0 ];
86
+ mov.u64 %rd16, 0x0;
87
+ @%p44 ld.global.L1::evict_last.b64 { %rd16 }, [ %rd17 + 0 ];
88
+ .loc 1 27 18
89
+ bfe.s32 %r28, %r15, 30, 1;
90
+ shr.u32 %r29, %r28, 23;
91
+ add.s32 %r30, %r8, %r29;
92
+ and.b32 %r31, %r30, 16776704;
93
+ sub.s32 %r32, %r8, %r31;
94
+ .loc 1 35 44
95
+ shl.b32 %r33, %r32, 8;
96
+ .loc 1 35 40
97
+ or.b32 %r34, %r33, %r6;
98
+ .loc 1 35 34
99
+ mul.wide.s32 %rd23, %r34, 4;
100
+ add.s64 %rd33, %rd20, %rd23;
101
+ mov.b32 %r137, 0;
102
+ .loc 1 35 50
103
+ mov.u32 %r16, 0x0;
104
+ mov.u32 %r17, 0x0;
105
+ mov.u32 %r18, 0x0;
106
+ mov.u32 %r19, 0x0;
107
+ @%p44 ld.global.L1::evict_last.v4.b32 { %r16, %r17, %r18, %r19 }, [ %rd33 + 0 ];
108
+ @!%p44 mov.u32 %r16, %r137;
109
+ @!%p44 mov.u32 %r17, %r137;
110
+ @!%p44 mov.u32 %r18, %r137;
111
+ @!%p44 mov.u32 %r19, %r137;
112
+ mov.b32 %f1, %r16;
113
+ mov.b32 %f2, %r17;
114
+ mov.b32 %f3, %r18;
115
+ mov.b32 %f4, %r19;
116
+ .loc 1 36 22
117
+ add.s64 %rd24, %rd16, 50257;
118
+ .loc 1 37 22
119
+ setp.lt.s64 %p11, %rd16, 0;
120
+ .loc 1 38 36
121
+ selp.b64 %rd3, %rd24, %rd16, %p11;
122
+ .loc 1 39 40
123
+ setp.lt.u64 %p12, %rd3, 50257;
124
+ mov.b32 %r168, 883;
125
+ mov.u64 %rd52, 1;
126
+ .loc 1 39 55
127
+ @%p12 bra $L__BB0_2;
128
+ mov.u64 %rd25, assertMessage_0;
129
+ cvta.global.u64 %rd26, %rd25;
130
+ mov.u64 %rd27, assertFile_0;
131
+ cvta.global.u64 %rd28, %rd27;
132
+ mov.u64 %rd29, assertFunc_0;
133
+ cvta.global.u64 %rd30, %rd29;
134
+ { // callseq 4, 0
135
+ .reg .b32 temp_param_reg;
136
+ .param .b64 param0;
137
+ st.param.b64 [param0+0], %rd26;
138
+ .param .b64 param1;
139
+ st.param.b64 [param1+0], %rd28;
140
+ .param .b32 param2;
141
+ st.param.b32 [param2+0], %r168;
142
+ .param .b64 param3;
143
+ st.param.b64 [param3+0], %rd30;
144
+ .param .b64 param4;
145
+ st.param.b64 [param4+0], %rd52;
146
+ call.uni
147
+ __assertfail,
148
+ (
149
+ param0,
150
+ param1,
151
+ param2,
152
+ param3,
153
+ param4
154
+ );
155
+ } // callseq 4
156
+ $L__BB0_2:
157
+ .loc 1 0 55
158
+ ld.param.u64 %rd7, [triton__0d1d2d3d4d5de6de_param_4];
159
+ .loc 1 37 22
160
+ setp.lt.s64 %p36, %rd8, 0;
161
+ .loc 1 40 44
162
+ shl.b64 %rd35, %rd8, 8;
163
+ add.s64 %rd36, %rd35, 12865792;
164
+ selp.b64 %rd37, %rd36, %rd35, %p36;
165
+ cvt.u64.u32 %rd38, %r6;
166
+ .loc 1 40 40
167
+ or.b64 %rd39, %rd37, %rd38;
168
+ .loc 1 40 34
169
+ shl.b64 %rd40, %rd39, 2;
170
+ add.s64 %rd49, %rd5, %rd40;
171
+ .loc 1 40 52
172
+ mov.u32 %r36, 0x0;
173
+ mov.u32 %r37, 0x0;
174
+ mov.u32 %r38, 0x0;
175
+ mov.u32 %r39, 0x0;
176
+ @%p44 ld.global.L1::evict_last.v4.b32 { %r36, %r37, %r38, %r39 }, [ %rd49 + 0 ];
177
+ @!%p44 mov.u32 %r36, %r137;
178
+ @!%p44 mov.u32 %r37, %r137;
179
+ @!%p44 mov.u32 %r38, %r137;
180
+ @!%p44 mov.u32 %r39, %r137;
181
+ mov.b32 %f7, %r36;
182
+ mov.b32 %f8, %r37;
183
+ mov.b32 %f9, %r38;
184
+ mov.b32 %f10, %r39;
185
+ .loc 1 41 22
186
+ add.f32 %f11, %f1, %f7;
187
+ add.f32 %f12, %f2, %f8;
188
+ add.f32 %f13, %f3, %f9;
189
+ add.f32 %f14, %f4, %f10;
190
+ $L__tmp1:
191
+ .loc 2 98 22
192
+ add.f32 %f15, %f11, 0f00000000;
193
+ add.f32 %f16, %f12, 0f00000000;
194
+ add.f32 %f17, %f13, 0f00000000;
195
+ add.f32 %f18, %f14, 0f00000000;
196
+ .loc 2 101 30
197
+ sub.f32 %f19, %f11, %f15;
198
+ sub.f32 %f20, %f12, %f16;
199
+ sub.f32 %f21, %f13, %f17;
200
+ sub.f32 %f22, %f14, %f18;
201
+ .loc 2 101 13
202
+ fma.rn.f32 %f23, %f11, %f19, 0f00000000;
203
+ fma.rn.f32 %f24, %f12, %f20, 0f00000000;
204
+ fma.rn.f32 %f25, %f13, %f21, 0f00000000;
205
+ fma.rn.f32 %f26, %f14, %f22, 0f00000000;
206
+ $L__tmp2:
207
+ .loc 2 108 21
208
+ sub.f32 %f27, %f16, %f15;
209
+ mov.b32 %r45, 1065353216;
210
+ mov.b32 %r46, 1073741824;
211
+ .loc 2 110 60
212
+ div.full.f32 %r44, %r45, %r46;
213
+ mov.b32 %f28, %r44;
214
+ .loc 2 112 17
215
+ fma.rn.f32 %f29, %f28, %f27, %f15;
216
+ .loc 2 113 15
217
+ add.f32 %f30, %f23, %f24;
218
+ .loc 2 113 30
219
+ mul.f32 %f31, %f27, %f27;
220
+ .loc 2 113 22
221
+ fma.rn.f32 %f32, %f28, %f31, %f30;
222
+ .loc 2 108 21
223
+ sub.f32 %f33, %f17, %f29;
224
+ mov.b32 %r49, 1077936128;
225
+ .loc 2 110 60
226
+ div.full.f32 %r47, %r45, %r49;
227
+ mov.b32 %f34, %r47;
228
+ .loc 2 112 17
229
+ fma.rn.f32 %f35, %f34, %f33, %f29;
230
+ .loc 2 113 15
231
+ add.f32 %f36, %f25, %f32;
232
+ .loc 2 113 30
233
+ mul.f32 %f37, %f33, %f33;
234
+ .loc 2 113 38
235
+ fma.rn.f32 %f38, %f33, %f33, %f37;
236
+ .loc 2 113 22
237
+ fma.rn.f32 %f39, %f34, %f38, %f36;
238
+ .loc 2 108 21
239
+ sub.f32 %f40, %f18, %f35;
240
+ mov.b32 %r52, 1082130432;
241
+ .loc 2 110 60
242
+ div.full.f32 %r50, %r45, %r52;
243
+ mov.b32 %f41, %r50;
244
+ .loc 2 112 17
245
+ fma.rn.f32 %f42, %f41, %f40, %f35;
246
+ .loc 2 113 15
247
+ add.f32 %f43, %f26, %f39;
248
+ .loc 2 113 30
249
+ mul.f32 %f44, %f40, %f40;
250
+ .loc 2 113 38
251
+ mul.f32 %f45, %f44, 0f40400000;
252
+ .loc 2 113 22
253
+ fma.rn.f32 %f46, %f41, %f45, %f43;
254
+ $L__tmp3:
255
+ .loc 2 120 46
256
+ mov.b32 %r101, %f42;
257
+ shfl.sync.bfly.b32 %r102, %r101, 16, 31, -1;
258
+ mov.b32 %f47, %r102;
259
+ mov.b32 %r103, %f46;
260
+ shfl.sync.bfly.b32 %r104, %r103, 16, 31, -1;
261
+ mov.b32 %f48, %r104;
262
+ shfl.sync.bfly.b32 %r54, %r52, 16, 31, -1;
263
+ mov.b32 %f49, %r54;
264
+ $L__tmp4:
265
+ .loc 2 108 21
266
+ sub.f32 %f50, %f47, %f42;
267
+ .loc 2 109 28
268
+ add.f32 %f51, %f49, 0f40800000;
269
+ .loc 2 110 39
270
+ setp.eq.f32 %p37, %f51, 0f00000000;
271
+ .loc 2 110 60
272
+ mov.b32 %r55, %f51;
273
+ div.full.f32 %r53, %r54, %r55;
274
+ mov.b32 %f52, %r53;
275
+ .loc 2 110 49
276
+ selp.f32 %f53, 0f00000000, %f52, %p37;
277
+ .loc 2 112 17
278
+ fma.rn.f32 %f54, %f53, %f50, %f42;
279
+ .loc 2 113 15
280
+ add.f32 %f55, %f46, %f48;
281
+ .loc 2 113 30
282
+ mul.f32 %f56, %f50, %f50;
283
+ .loc 2 113 38
284
+ mul.f32 %f57, %f56, 0f40800000;
285
+ .loc 2 113 22
286
+ fma.rn.f32 %f58, %f53, %f57, %f55;
287
+ $L__tmp5:
288
+ .loc 2 120 46
289
+ mov.b32 %r105, %f54;
290
+ shfl.sync.bfly.b32 %r106, %r105, 8, 31, -1;
291
+ mov.b32 %f59, %r106;
292
+ mov.b32 %r107, %f58;
293
+ shfl.sync.bfly.b32 %r108, %r107, 8, 31, -1;
294
+ mov.b32 %f60, %r108;
295
+ shfl.sync.bfly.b32 %r57, %r55, 8, 31, -1;
296
+ mov.b32 %f61, %r57;
297
+ $L__tmp6:
298
+ .loc 2 108 21
299
+ sub.f32 %f62, %f59, %f54;
300
+ .loc 2 109 28
301
+ add.f32 %f63, %f51, %f61;
302
+ .loc 2 110 39
303
+ setp.eq.f32 %p38, %f63, 0f00000000;
304
+ .loc 2 110 60
305
+ mov.b32 %r58, %f63;
306
+ div.full.f32 %r56, %r57, %r58;
307
+ mov.b32 %f64, %r56;
308
+ .loc 2 110 49
309
+ selp.f32 %f65, 0f00000000, %f64, %p38;
310
+ .loc 2 112 17
311
+ fma.rn.f32 %f66, %f65, %f62, %f54;
312
+ .loc 2 113 15
313
+ add.f32 %f67, %f58, %f60;
314
+ .loc 2 113 30
315
+ mul.f32 %f68, %f62, %f62;
316
+ .loc 2 113 38
317
+ mul.f32 %f69, %f51, %f68;
318
+ .loc 2 113 22
319
+ fma.rn.f32 %f70, %f65, %f69, %f67;
320
+ $L__tmp7:
321
+ .loc 2 120 46
322
+ mov.b32 %r109, %f66;
323
+ shfl.sync.bfly.b32 %r110, %r109, 4, 31, -1;
324
+ mov.b32 %f71, %r110;
325
+ mov.b32 %r111, %f70;
326
+ shfl.sync.bfly.b32 %r112, %r111, 4, 31, -1;
327
+ mov.b32 %f72, %r112;
328
+ shfl.sync.bfly.b32 %r60, %r58, 4, 31, -1;
329
+ mov.b32 %f73, %r60;
330
+ $L__tmp8:
331
+ .loc 2 108 21
332
+ sub.f32 %f74, %f71, %f66;
333
+ .loc 2 109 28
334
+ add.f32 %f75, %f63, %f73;
335
+ .loc 2 110 39
336
+ setp.eq.f32 %p39, %f75, 0f00000000;
337
+ .loc 2 110 60
338
+ mov.b32 %r61, %f75;
339
+ div.full.f32 %r59, %r60, %r61;
340
+ mov.b32 %f76, %r59;
341
+ .loc 2 110 49
342
+ selp.f32 %f77, 0f00000000, %f76, %p39;
343
+ .loc 2 112 17
344
+ fma.rn.f32 %f78, %f77, %f74, %f66;
345
+ .loc 2 113 15
346
+ add.f32 %f79, %f70, %f72;
347
+ .loc 2 113 30
348
+ mul.f32 %f80, %f74, %f74;
349
+ .loc 2 113 38
350
+ mul.f32 %f81, %f63, %f80;
351
+ .loc 2 113 22
352
+ fma.rn.f32 %f82, %f77, %f81, %f79;
353
+ $L__tmp9:
354
+ .loc 2 120 46
355
+ mov.b32 %r113, %f78;
356
+ shfl.sync.bfly.b32 %r114, %r113, 2, 31, -1;
357
+ mov.b32 %f83, %r114;
358
+ mov.b32 %r115, %f82;
359
+ shfl.sync.bfly.b32 %r116, %r115, 2, 31, -1;
360
+ mov.b32 %f84, %r116;
361
+ shfl.sync.bfly.b32 %r63, %r61, 2, 31, -1;
362
+ mov.b32 %f85, %r63;
363
+ $L__tmp10:
364
+ .loc 2 108 21
365
+ sub.f32 %f86, %f83, %f78;
366
+ .loc 2 109 28
367
+ add.f32 %f87, %f75, %f85;
368
+ .loc 2 110 39
369
+ setp.eq.f32 %p40, %f87, 0f00000000;
370
+ .loc 2 110 60
371
+ mov.b32 %r64, %f87;
372
+ div.full.f32 %r62, %r63, %r64;
373
+ mov.b32 %f88, %r62;
374
+ .loc 2 110 49
375
+ selp.f32 %f89, 0f00000000, %f88, %p40;
376
+ .loc 2 112 17
377
+ fma.rn.f32 %f90, %f89, %f86, %f78;
378
+ .loc 2 113 15
379
+ add.f32 %f91, %f82, %f84;
380
+ .loc 2 113 30
381
+ mul.f32 %f92, %f86, %f86;
382
+ .loc 2 113 38
383
+ mul.f32 %f93, %f75, %f92;
384
+ .loc 2 113 22
385
+ fma.rn.f32 %f94, %f89, %f93, %f91;
386
+ $L__tmp11:
387
+ .loc 2 120 46
388
+ mov.b32 %r117, %f90;
389
+ shfl.sync.bfly.b32 %r118, %r117, 1, 31, -1;
390
+ mov.b32 %f95, %r118;
391
+ mov.b32 %r119, %f94;
392
+ shfl.sync.bfly.b32 %r120, %r119, 1, 31, -1;
393
+ mov.b32 %f96, %r120;
394
+ shfl.sync.bfly.b32 %r66, %r64, 1, 31, -1;
395
+ mov.b32 %f97, %r66;
396
+ $L__tmp12:
397
+ .loc 2 108 21
398
+ sub.f32 %f98, %f95, %f90;
399
+ .loc 2 109 28
400
+ add.f32 %f99, %f87, %f97;
401
+ .loc 2 110 39
402
+ setp.eq.f32 %p41, %f99, 0f00000000;
403
+ .loc 2 110 60
404
+ mov.b32 %r67, %f99;
405
+ div.full.f32 %r65, %r66, %r67;
406
+ mov.b32 %f100, %r65;
407
+ .loc 2 110 49
408
+ selp.f32 %f101, 0f00000000, %f100, %p41;
409
+ .loc 2 112 17
410
+ fma.rn.f32 %f102, %f98, %f101, %f90;
411
+ .loc 2 113 15
412
+ add.f32 %f103, %f94, %f96;
413
+ .loc 2 113 30
414
+ mul.f32 %f104, %f98, %f98;
415
+ .loc 2 113 38
416
+ mul.f32 %f105, %f87, %f104;
417
+ .loc 2 113 22
418
+ fma.rn.f32 %f106, %f101, %f105, %f103;
419
+ $L__tmp13:
420
+ .loc 2 120 46
421
+ setp.eq.s32 %p18, %r2, 0;
422
+ shl.b32 %r121, %r5, 2;
423
+ shl.b32 %r122, %r3, 3;
424
+ or.b32 %r123, %r122, %r121;
425
+ mov.u32 %r124, global_smem;
426
+ add.s32 %r68, %r124, %r123;
427
+ mov.b32 %r69, %f102;
428
+ @%p18 st.shared.b32 [ %r68 + 0 ], %r69;
429
+ add.s32 %r125, %r124, 16;
430
+ add.s32 %r70, %r125, %r123;
431
+ mov.b32 %r71, %f106;
432
+ @%p18 st.shared.b32 [ %r70 + 0 ], %r71;
433
+ add.s32 %r126, %r124, 32;
434
+ add.s32 %r72, %r126, %r123;
435
+ @%p18 st.shared.b32 [ %r72 + 0 ], %r67;
436
+ bar.sync 0;
437
+ setp.lt.s32 %p21, %r1, 4;
438
+ add.s32 %r75, %r124, %r24;
439
+ @%p21 ld.shared.b32 %r74, [ %r75 + 0 ];
440
+ mov.b32 %f107, %r74;
441
+ add.s32 %r77, %r125, %r24;
442
+ @%p21 ld.shared.b32 %r76, [ %r77 + 0 ];
443
+ mov.b32 %f108, %r76;
444
+ add.s32 %r79, %r126, %r24;
445
+ @%p21 ld.shared.b32 %r78, [ %r79 + 0 ];
446
+ mov.b32 %f109, %r78;
447
+ shfl.sync.bfly.b32 %r128, %r74, 1, 31, -1;
448
+ mov.b32 %f110, %r128;
449
+ shfl.sync.bfly.b32 %r129, %r76, 1, 31, -1;
450
+ mov.b32 %f111, %r129;
451
+ shfl.sync.bfly.b32 %r81, %r78, 1, 31, -1;
452
+ mov.b32 %f112, %r81;
453
+ $L__tmp14:
454
+ .loc 2 108 21
455
+ sub.f32 %f113, %f110, %f107;
456
+ .loc 2 109 28
457
+ add.f32 %f114, %f109, %f112;
458
+ .loc 2 110 39
459
+ setp.eq.f32 %p42, %f114, 0f00000000;
460
+ .loc 2 110 60
461
+ mov.b32 %r82, %f114;
462
+ div.full.f32 %r80, %r81, %r82;
463
+ mov.b32 %f115, %r80;
464
+ .loc 2 110 49
465
+ selp.f32 %f116, 0f00000000, %f115, %p42;
466
+ .loc 2 112 17
467
+ fma.rn.f32 %f117, %f113, %f116, %f107;
468
+ .loc 2 113 15
469
+ add.f32 %f118, %f108, %f111;
470
+ .loc 2 113 30
471
+ mul.f32 %f119, %f113, %f113;
472
+ .loc 2 113 38
473
+ mul.f32 %f120, %f109, %f119;
474
+ .loc 2 113 22
475
+ fma.rn.f32 %f121, %f120, %f116, %f118;
476
+ $L__tmp15:
477
+ .loc 2 120 46
478
+ setp.eq.s32 %p43, %r4, 0;
479
+ and.pred %p24, %p21, %p43;
480
+ mov.b32 %r84, %f117;
481
+ @%p24 st.shared.b32 [ %r75 + 0 ], %r84;
482
+ mov.b32 %r86, %f121;
483
+ @%p24 st.shared.b32 [ %r77 + 0 ], %r86;
484
+ @%p24 st.shared.b32 [ %r79 + 0 ], %r82;
485
+ bar.sync 0;
486
+ add.s32 %r130, %r124, %r122;
487
+ ld.shared.f32 %f5, [%r130];
488
+ add.s32 %r131, %r125, %r122;
489
+ ld.shared.f32 %f6, [%r131];
490
+ $L__tmp16:
491
+ .loc 1 59 51
492
+ mov.u32 %r89, 0x0;
493
+ mov.u32 %r90, 0x0;
494
+ mov.u32 %r91, 0x0;
495
+ mov.u32 %r92, 0x0;
496
+ @%p44 ld.global.L1::evict_last.v4.b32 { %r89, %r90, %r91, %r92 }, [ %rd33 + 0 ];
497
+ @!%p44 mov.u32 %r89, %r137;
498
+ @!%p44 mov.u32 %r90, %r137;
499
+ @!%p44 mov.u32 %r91, %r137;
500
+ @!%p44 mov.u32 %r92, %r137;
501
+ .loc 1 60 35
502
+ mul.wide.u32 %rd41, %r7, 4;
503
+ add.s64 %rd34, %rd6, %rd41;
504
+ .loc 1 60 40
505
+ mov.u32 %r97, 0x0;
506
+ mov.u32 %r98, 0x0;
507
+ @%p44 ld.global.L1::evict_last.v2.b32 { %r97, %r98 }, [ %rd34 + 0 ];
508
+ @!%p44 mov.u32 %r97, %r137;
509
+ @!%p44 mov.u32 %r98, %r137;
510
+ .loc 1 64 57
511
+ @%p12 bra $L__BB0_4;
512
+ mov.u64 %rd42, assertMessage_1;
513
+ cvta.global.u64 %rd43, %rd42;
514
+ mov.u64 %rd44, assertFile_1;
515
+ cvta.global.u64 %rd45, %rd44;
516
+ mov.u64 %rd46, assertFunc_1;
517
+ cvta.global.u64 %rd47, %rd46;
518
+ { // callseq 5, 0
519
+ .reg .b32 temp_param_reg;
520
+ .param .b64 param0;
521
+ st.param.b64 [param0+0], %rd43;
522
+ .param .b64 param1;
523
+ st.param.b64 [param1+0], %rd45;
524
+ .param .b32 param2;
525
+ st.param.b32 [param2+0], %r168;
526
+ .param .b64 param3;
527
+ st.param.b64 [param3+0], %rd47;
528
+ .param .b64 param4;
529
+ st.param.b64 [param4+0], %rd52;
530
+ call.uni
531
+ __assertfail,
532
+ (
533
+ param0,
534
+ param1,
535
+ param2,
536
+ param3,
537
+ param4
538
+ );
539
+ } // callseq 5
540
+ $L__BB0_4:
541
+ .loc 1 65 54
542
+ mov.u32 %r133, 0x0;
543
+ mov.u32 %r134, 0x0;
544
+ mov.u32 %r135, 0x0;
545
+ mov.u32 %r136, 0x0;
546
+ @%p44 ld.global.L1::evict_first.v4.b32 { %r133, %r134, %r135, %r136 }, [ %rd49 + 0 ];
547
+ @!%p44 mov.u32 %r133, %r137;
548
+ @!%p44 mov.u32 %r134, %r137;
549
+ @!%p44 mov.u32 %r135, %r137;
550
+ @!%p44 mov.u32 %r136, %r137;
551
+ .loc 1 69 23
552
+ mov.b32 %r142, %f6;
553
+ mov.b32 %r143, 1132462080;
554
+ div.full.f32 %r141, %r142, %r143;
555
+ mov.b32 %f122, %r141;
556
+ .loc 1 71 24
557
+ add.f32 %f123, %f122, 0f3727C5AC;
558
+ .loc 1 72 30
559
+ rsqrt.approx.ftz.f32 %f124, %f123;
560
+ .loc 1 65 54
561
+ mov.b32 %f125, %r136;
562
+ .loc 1 59 51
563
+ mov.b32 %f126, %r92;
564
+ .loc 1 66 24
565
+ add.f32 %f127, %f126, %f125;
566
+ .loc 1 67 24
567
+ sub.f32 %f128, %f127, %f5;
568
+ .loc 1 65 54
569
+ mov.b32 %f129, %r135;
570
+ .loc 1 59 51
571
+ mov.b32 %f130, %r91;
572
+ .loc 1 66 24
573
+ add.f32 %f131, %f130, %f129;
574
+ .loc 1 67 24
575
+ sub.f32 %f132, %f131, %f5;
576
+ .loc 1 65 54
577
+ mov.b32 %f133, %r134;
578
+ .loc 1 59 51
579
+ mov.b32 %f134, %r90;
580
+ .loc 1 66 24
581
+ add.f32 %f135, %f134, %f133;
582
+ .loc 1 67 24
583
+ sub.f32 %f136, %f135, %f5;
584
+ .loc 1 65 54
585
+ mov.b32 %f137, %r133;
586
+ .loc 1 59 51
587
+ mov.b32 %f138, %r89;
588
+ .loc 1 66 24
589
+ add.f32 %f139, %f138, %f137;
590
+ .loc 1 67 24
591
+ sub.f32 %f140, %f139, %f5;
592
+ .loc 1 73 24
593
+ mul.f32 %f141, %f140, %f124;
594
+ mul.f32 %f142, %f136, %f124;
595
+ mul.f32 %f143, %f132, %f124;
596
+ mul.f32 %f144, %f128, %f124;
597
+ .loc 1 74 24
598
+ bar.sync 0;
599
+ shl.b32 %r159, %r7, 2;
600
+ add.s32 %r161, %r124, %r159;
601
+ st.shared.v2.u32 [%r161], {%r97, %r98};
602
+ bar.sync 0;
603
+ shl.b32 %r162, %r6, 2;
604
+ add.s32 %r163, %r124, %r162;
605
+ ld.shared.v4.f32 {%f145, %f146, %f147, %f148}, [%r163];
606
+ mul.f32 %f149, %f141, %f145;
607
+ mul.f32 %f150, %f142, %f146;
608
+ mul.f32 %f151, %f143, %f147;
609
+ mul.f32 %f152, %f144, %f148;
610
+ .loc 1 76 39
611
+ shl.b32 %r164, %r8, 8;
612
+ .loc 1 76 35
613
+ or.b32 %r165, %r164, %r6;
614
+ .loc 1 76 29
615
+ mul.wide.s32 %rd51, %r165, 2;
616
+ add.s64 %rd50, %rd7, %rd51;
617
+ .loc 1 76 52
618
+ mov.b32 %r153, %f149;
619
+ cvt.rn.bf16.f32 %rs1, %r153;
620
+ mov.b32 %r154, %f150;
621
+ cvt.rn.bf16.f32 %rs2, %r154;
622
+ mov.b32 %r155, %f151;
623
+ cvt.rn.bf16.f32 %rs3, %r155;
624
+ mov.b32 %r156, %f152;
625
+ cvt.rn.bf16.f32 %rs4, %r156;
626
+ mov.b32 %r166, {%rs1, %rs2};
627
+ mov.b32 %r167, {%rs3, %rs4};
628
+ @%p44 st.global.v2.b32 [ %rd50 + 0 ], { %r166, %r167 };
629
+ .loc 1 55 4
630
+ ret;
631
+ $L__tmp17:
632
+ $L__func_end0:
633
+
634
+ }
635
+ // .globl __nv_rsqrtf
636
+ .visible .func (.param .b32 func_retval0) __nv_rsqrtf(
637
+ .param .b32 __nv_rsqrtf_param_0
638
+ )
639
+ {
640
+ .reg .f32 %f<3>;
641
+ $L__func_begin1:
642
+
643
+ ld.param.f32 %f1, [__nv_rsqrtf_param_0];
644
+ rsqrt.approx.ftz.f32 %f2, %f1;
645
+ st.param.f32 [func_retval0+0], %f2;
646
+ ret;
647
+ $L__func_end1:
648
+
649
+ }
650
+ .file 1 "/tmp/torchinductor_root/lh/clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py"
651
+ .file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
652
+ .section .debug_abbrev
653
+ {
654
+ .b8 1
655
+ .b8 17
656
+ .b8 1
657
+ .b8 37
658
+ .b8 8
659
+ .b8 19
660
+ .b8 5
661
+ .b8 3
662
+ .b8 8
663
+ .b8 16
664
+ .b8 6
665
+ .b8 27
666
+ .b8 8
667
+ .b8 180
668
+ .b8 66
669
+ .b8 12
670
+ .b8 17
671
+ .b8 1
672
+ .b8 18
673
+ .b8 1
674
+ .b8 0
675
+ .b8 0
676
+ .b8 2
677
+ .b8 46
678
+ .b8 0
679
+ .b8 135
680
+ .b8 64
681
+ .b8 8
682
+ .b8 3
683
+ .b8 8
684
+ .b8 58
685
+ .b8 11
686
+ .b8 59
687
+ .b8 11
688
+ .b8 63
689
+ .b8 12
690
+ .b8 32
691
+ .b8 11
692
+ .b8 0
693
+ .b8 0
694
+ .b8 3
695
+ .b8 46
696
+ .b8 1
697
+ .b8 17
698
+ .b8 1
699
+ .b8 18
700
+ .b8 1
701
+ .b8 64
702
+ .b8 10
703
+ .b8 49
704
+ .b8 19
705
+ .b8 0
706
+ .b8 0
707
+ .b8 4
708
+ .b8 29
709
+ .b8 0
710
+ .b8 49
711
+ .b8 19
712
+ .b8 17
713
+ .b8 1
714
+ .b8 18
715
+ .b8 1
716
+ .b8 88
717
+ .b8 11
718
+ .b8 89
719
+ .b8 11
720
+ .b8 87
721
+ .b8 11
722
+ .b8 0
723
+ .b8 0
724
+ .b8 5
725
+ .b8 29
726
+ .b8 1
727
+ .b8 49
728
+ .b8 19
729
+ .b8 17
730
+ .b8 1
731
+ .b8 18
732
+ .b8 1
733
+ .b8 88
734
+ .b8 11
735
+ .b8 89
736
+ .b8 11
737
+ .b8 87
738
+ .b8 11
739
+ .b8 0
740
+ .b8 0
741
+ .b8 0
742
+ }
743
+ .section .debug_info
744
+ {
745
+ .b32 298
746
+ .b8 2
747
+ .b8 0
748
+ .b32 .debug_abbrev
749
+ .b8 8
750
+ .b8 1
751
+ .b8 116
752
+ .b8 114
753
+ .b8 105
754
+ .b8 116
755
+ .b8 111
756
+ .b8 110
757
+ .b8 0
758
+ .b8 2
759
+ .b8 0
760
+ .b8 99
761
+ .b8 108
762
+ .b8 104
763
+ .b8 101
764
+ .b8 52
765
+ .b8 97
766
+ .b8 51
767
+ .b8 115
768
+ .b8 116
769
+ .b8 118
770
+ .b8 117
771
+ .b8 102
772
+ .b8 120
773
+ .b8 97
774
+ .b8 102
775
+ .b8 109
776
+ .b8 113
777
+ .b8 51
778
+ .b8 107
779
+ .b8 107
780
+ .b8 53
781
+ .b8 104
782
+ .b8 111
783
+ .b8 100
784
+ .b8 97
785
+ .b8 122
786
+ .b8 122
787
+ .b8 50
788
+ .b8 101
789
+ .b8 102
790
+ .b8 99
791
+ .b8 116
792
+ .b8 102
793
+ .b8 102
794
+ .b8 116
795
+ .b8 101
796
+ .b8 54
797
+ .b8 52
798
+ .b8 54
799
+ .b8 122
800
+ .b8 110
801
+ .b8 106
802
+ .b8 100
803
+ .b8 110
804
+ .b8 118
805
+ .b8 51
806
+ .b8 108
807
+ .b8 113
808
+ .b8 105
809
+ .b8 53
810
+ .b8 111
811
+ .b8 97
812
+ .b8 46
813
+ .b8 112
814
+ .b8 121
815
+ .b8 0
816
+ .b32 .debug_line
817
+ .b8 47
818
+ .b8 116
819
+ .b8 109
820
+ .b8 112
821
+ .b8 47
822
+ .b8 116
823
+ .b8 111
824
+ .b8 114
825
+ .b8 99
826
+ .b8 104
827
+ .b8 105
828
+ .b8 110
829
+ .b8 100
830
+ .b8 117
831
+ .b8 99
832
+ .b8 116
833
+ .b8 111
834
+ .b8 114
835
+ .b8 95
836
+ .b8 114
837
+ .b8 111
838
+ .b8 111
839
+ .b8 116
840
+ .b8 47
841
+ .b8 108
842
+ .b8 104
843
+ .b8 0
844
+ .b8 1
845
+ .b64 $L__func_begin0
846
+ .b64 $L__func_end0
847
+ .b8 2
848
+ .b8 116
849
+ .b8 114
850
+ .b8 105
851
+ .b8 116
852
+ .b8 111
853
+ .b8 110
854
+ .b8 95
855
+ .b8 95
856
+ .b8 48
857
+ .b8 100
858
+ .b8 49
859
+ .b8 100
860
+ .b8 50
861
+ .b8 100
862
+ .b8 51
863
+ .b8 100
864
+ .b8 52
865
+ .b8 100
866
+ .b8 53
867
+ .b8 100
868
+ .b8 101
869
+ .b8 54
870
+ .b8 100
871
+ .b8 101
872
+ .b8 0
873
+ .b8 116
874
+ .b8 114
875
+ .b8 105
876
+ .b8 116
877
+ .b8 111
878
+ .b8 110
879
+ .b8 95
880
+ .b8 95
881
+ .b8 48
882
+ .b8 100
883
+ .b8 49
884
+ .b8 100
885
+ .b8 50
886
+ .b8 100
887
+ .b8 51
888
+ .b8 100
889
+ .b8 52
890
+ .b8 100
891
+ .b8 53
892
+ .b8 100
893
+ .b8 101
894
+ .b8 54
895
+ .b8 100
896
+ .b8 101
897
+ .b8 0
898
+ .b8 1
899
+ .b8 18
900
+ .b8 1
901
+ .b8 1
902
+ .b8 3
903
+ .b64 $L__func_begin0
904
+ .b64 $L__func_end0
905
+ .b8 1
906
+ .b8 156
907
+ .b32 125
908
+ .b8 4
909
+ .b32 125
910
+ .b64 $L__tmp1
911
+ .b64 $L__tmp2
912
+ .b8 2
913
+ .b8 44
914
+ .b8 38
915
+ .b8 5
916
+ .b32 125
917
+ .b64 $L__tmp2
918
+ .b64 $L__tmp15
919
+ .b8 2
920
+ .b8 50
921
+ .b8 41
922
+ .b8 4
923
+ .b32 125
924
+ .b64 $L__tmp2
925
+ .b64 $L__tmp15
926
+ .b8 2
927
+ .b8 120
928
+ .b8 46
929
+ .b8 0
930
+ .b8 4
931
+ .b32 125
932
+ .b64 $L__tmp3
933
+ .b64 $L__tmp16
934
+ .b8 2
935
+ .b8 50
936
+ .b8 41
937
+ .b8 0
938
+ .b8 0
939
+ }
940
+ .section .debug_pubnames
941
+ {
942
+ .b32 $L__pubNames_end0-$L__pubNames_start0
943
+ $L__pubNames_start0:
944
+ .b8 2
945
+ .b8 0
946
+ .b32 .debug_info
947
+ .b32 302
948
+ .b32 125
949
+ .b8 116
950
+ .b8 114
951
+ .b8 105
952
+ .b8 116
953
+ .b8 111
954
+ .b8 110
955
+ .b8 95
956
+ .b8 95
957
+ .b8 48
958
+ .b8 100
959
+ .b8 49
960
+ .b8 100
961
+ .b8 50
962
+ .b8 100
963
+ .b8 51
964
+ .b8 100
965
+ .b8 52
966
+ .b8 100
967
+ .b8 53
968
+ .b8 100
969
+ .b8 101
970
+ .b8 54
971
+ .b8 100
972
+ .b8 101
973
+ .b8 0
974
+ .b32 0
975
+ $L__pubNames_end0:
976
+ }
977
+ .section .debug_pubtypes
978
+ {
979
+ .b32 $L__pubTypes_end0-$L__pubTypes_start0
980
+ $L__pubTypes_start0:
981
+ .b8 2
982
+ .b8 0
983
+ .b32 .debug_info
984
+ .b32 302
985
+ .b32 0
986
+ $L__pubTypes_end0:
987
+ }
988
+ .section .debug_loc { }
.triton/dump/1ed98b0d136db679153ca6a42fff755c/triton_.ttir ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ module {
2
+ tt.func public @triton__0d1d2d3d4d5de6de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
3
+ %cst = arith.constant dense<1.000000e+00> : tensor<1x256xf32>
4
+ %cst_0 = arith.constant dense<0.000000e+00> : tensor<1x256xf32>
5
+ %cst_1 = arith.constant 0.000000e+00 : f32
6
+ %cst_2 = arith.constant dense<256> : tensor<2x1xi64>
7
+ %cst_3 = arith.constant dense<50257> : tensor<2x1xi64>
8
+ %cst_4 = arith.constant dense<0> : tensor<2x1xi64>
9
+ %cst_5 = arith.constant dense<9.99999974E-6> : tensor<2x1xf32>
10
+ %cst_6 = arith.constant dense<2.560000e+02> : tensor<2x1xf32>
11
+ %cst_7 = arith.constant dense<0.000000e+00> : tensor<2x256xf32>
12
+ %cst_8 = arith.constant dense<256> : tensor<2x1xi32>
13
+ %cst_9 = arith.constant dense<256> : tensor<1x256xi32>
14
+ %cst_10 = arith.constant dense<512> : tensor<2x1xi32>
15
+ %c2_i32 = arith.constant 2 : i32
16
+ %0 = tt.get_program_id x : i32
17
+ %1 = arith.muli %0, %c2_i32 : i32
18
+ %2 = tt.make_range {end = 2 : i32, start = 0 : i32} : tensor<2xi32>
19
+ %3 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<2xi32>) -> tensor<2x1xi32>
20
+ %4 = tt.splat %1 : (i32) -> tensor<2x1xi32>
21
+ %5 = arith.addi %4, %3 : tensor<2x1xi32>
22
+ %6 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32>
23
+ %7 = tt.expand_dims %6 {axis = 0 : i32} : (tensor<256xi32>) -> tensor<1x256xi32>
24
+ %8 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<2x1x!tt.ptr<i64, 1>>
25
+ %9 = tt.addptr %8, %5 : tensor<2x1x!tt.ptr<i64, 1>>, tensor<2x1xi32>
26
+ %10 = tt.load %9 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x1xi64>
27
+ %11 = arith.remsi %5, %cst_10 : tensor<2x1xi32>
28
+ %12 = arith.cmpi slt, %7, %cst_9 : tensor<1x256xi32>
29
+ %13 = arith.muli %11, %cst_8 : tensor<2x1xi32>
30
+ %14 = tt.broadcast %7 : (tensor<1x256xi32>) -> tensor<2x256xi32>
31
+ %15 = tt.broadcast %13 : (tensor<2x1xi32>) -> tensor<2x256xi32>
32
+ %16 = arith.addi %14, %15 : tensor<2x256xi32>
33
+ %17 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<2x256x!tt.ptr<f32, 1>>
34
+ %18 = tt.addptr %17, %16 : tensor<2x256x!tt.ptr<f32, 1>>, tensor<2x256xi32>
35
+ %19 = tt.broadcast %12 : (tensor<1x256xi1>) -> tensor<2x256xi1>
36
+ %20 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x256xf32>
37
+ %21 = arith.addi %10, %cst_3 : tensor<2x1xi64>
38
+ %22 = arith.cmpi slt, %10, %cst_4 : tensor<2x1xi64>
39
+ %23 = arith.select %22, %21, %10 : tensor<2x1xi1>, tensor<2x1xi64>
40
+ %24 = arith.cmpi sge, %23, %cst_4 : tensor<2x1xi64>
41
+ %25 = arith.cmpi slt, %23, %cst_3 : tensor<2x1xi64>
42
+ %26 = arith.andi %24, %25 : tensor<2x1xi1>
43
+ tt.assert %26, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<2x1xi1>
44
+ %27 = arith.muli %23, %cst_2 : tensor<2x1xi64>
45
+ %28 = tt.broadcast %27 : (tensor<2x1xi64>) -> tensor<2x256xi64>
46
+ %29 = arith.extsi %7 : tensor<1x256xi32> to tensor<1x256xi64>
47
+ %30 = tt.broadcast %29 : (tensor<1x256xi64>) -> tensor<2x256xi64>
48
+ %31 = arith.addi %30, %28 : tensor<2x256xi64>
49
+ %32 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<2x256x!tt.ptr<f32, 1>>
50
+ %33 = tt.addptr %32, %31 : tensor<2x256x!tt.ptr<f32, 1>>, tensor<2x256xi64>
51
+ %34 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x256xf32>
52
+ %35 = arith.addf %34, %20 : tensor<2x256xf32>
53
+ %36 = arith.addf %35, %cst_7 : tensor<2x256xf32>
54
+ %37 = arith.subf %35, %36 : tensor<2x256xf32>
55
+ %38 = arith.mulf %35, %37 : tensor<2x256xf32>
56
+ %39 = arith.addf %38, %cst_7 : tensor<2x256xf32>
57
+ %40 = arith.select %19, %36, %cst_7 : tensor<2x256xi1>, tensor<2x256xf32>
58
+ %41 = arith.select %19, %39, %cst_7 : tensor<2x256xi1>, tensor<2x256xf32>
59
+ %42 = arith.select %12, %cst, %cst_0 : tensor<1x256xi1>, tensor<1x256xf32>
60
+ %43 = tt.broadcast %42 : (tensor<1x256xf32>) -> tensor<2x256xf32>
61
+ %44:3 = "tt.reduce"(%40, %41, %43) <{axis = 1 : i32}> ({
62
+ ^bb0(%arg7: f32, %arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32):
63
+ %68 = arith.subf %arg10, %arg7 : f32
64
+ %69 = arith.addf %arg9, %arg12 : f32
65
+ %70 = arith.cmpf oeq, %69, %cst_1 : f32
66
+ %71 = arith.divf %arg12, %69 : f32
67
+ %72 = arith.select %70, %cst_1, %71 : f32
68
+ %73 = arith.mulf %68, %72 : f32
69
+ %74 = arith.addf %arg7, %73 : f32
70
+ %75 = arith.addf %arg8, %arg11 : f32
71
+ %76 = arith.mulf %68, %68 : f32
72
+ %77 = arith.mulf %76, %arg9 : f32
73
+ %78 = arith.mulf %77, %72 : f32
74
+ %79 = arith.addf %75, %78 : f32
75
+ tt.reduce.return %74, %79, %69 : f32, f32, f32
76
+ }) : (tensor<2x256xf32>, tensor<2x256xf32>, tensor<2x256xf32>) -> (tensor<2xf32>, tensor<2xf32>, tensor<2xf32>)
77
+ %45 = tt.expand_dims %44#0 {axis = 1 : i32} : (tensor<2xf32>) -> tensor<2x1xf32>
78
+ %46 = tt.expand_dims %44#1 {axis = 1 : i32} : (tensor<2xf32>) -> tensor<2x1xf32>
79
+ %47 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x256xf32>
80
+ %48 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x256x!tt.ptr<f32, 1>>
81
+ %49 = tt.addptr %48, %7 : tensor<1x256x!tt.ptr<f32, 1>>, tensor<1x256xi32>
82
+ %50 = tt.load %49, %12, %cst_0 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x256xf32>
83
+ tt.assert %26, "index out of bounds: 0 <= tmp13 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<2x1xi1>
84
+ %51 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<2x256xf32>
85
+ %52 = arith.addf %51, %47 : tensor<2x256xf32>
86
+ %53 = tt.broadcast %45 : (tensor<2x1xf32>) -> tensor<2x256xf32>
87
+ %54 = arith.subf %52, %53 : tensor<2x256xf32>
88
+ %55 = arith.divf %46, %cst_6 : tensor<2x1xf32>
89
+ %56 = arith.addf %55, %cst_5 : tensor<2x1xf32>
90
+ %57 = tt.extern_elementwise %56 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<2x1xf32>) -> tensor<2x1xf32>
91
+ %58 = tt.broadcast %57 : (tensor<2x1xf32>) -> tensor<2x256xf32>
92
+ %59 = arith.mulf %54, %58 : tensor<2x256xf32>
93
+ %60 = tt.broadcast %50 : (tensor<1x256xf32>) -> tensor<2x256xf32>
94
+ %61 = arith.mulf %59, %60 : tensor<2x256xf32>
95
+ %62 = arith.muli %5, %cst_8 : tensor<2x1xi32>
96
+ %63 = tt.broadcast %62 : (tensor<2x1xi32>) -> tensor<2x256xi32>
97
+ %64 = arith.addi %14, %63 : tensor<2x256xi32>
98
+ %65 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<2x256x!tt.ptr<bf16, 1>>
99
+ %66 = tt.addptr %65, %64 : tensor<2x256x!tt.ptr<bf16, 1>>, tensor<2x256xi32>
100
+ %67 = arith.truncf %61 : tensor<2x256xf32> to tensor<2x256xbf16>
101
+ tt.store %66, %67, %19 {cache = 1 : i32, evict = 1 : i32} : tensor<2x256xbf16>
102
+ tt.return
103
+ }
104
+ }
.triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ptx ADDED
@@ -0,0 +1,782 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.2
6
+ .target sm_89
7
+ .address_size 64
8
+
9
+ // .globl triton__0d1d2d3d4d5d6d7d8de9de
10
+ .extern .shared .align 1 .b8 global_smem[];
11
+
12
+ .visible .entry triton__0d1d2d3d4d5d6d7d8de9de(
13
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_0,
14
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_1,
15
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_2,
16
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_3,
17
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_4,
18
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_5,
19
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_6,
20
+ .param .u64 triton__0d1d2d3d4d5d6d7d8de9de_param_7,
21
+ .param .u32 triton__0d1d2d3d4d5d6d7d8de9de_param_8,
22
+ .param .u32 triton__0d1d2d3d4d5d6d7d8de9de_param_9
23
+ )
24
+ .maxntid 64, 1, 1
25
+ {
26
+ .reg .pred %p<45>;
27
+ .reg .b16 %rs<5>;
28
+ .reg .b32 %r<106>;
29
+ .reg .f32 %f<90>;
30
+ .reg .b64 %rd<44>;
31
+ .loc 1 18 0
32
+ $L__func_begin0:
33
+ .loc 1 18 0
34
+
35
+ ld.param.u64 %rd25, [triton__0d1d2d3d4d5d6d7d8de9de_param_0];
36
+ ld.param.u64 %rd26, [triton__0d1d2d3d4d5d6d7d8de9de_param_1];
37
+ $L__tmp0:
38
+ .loc 1 26 26
39
+ mov.u32 %r74, %tid.x;
40
+ and.b32 %r75, %r74, 31;
41
+ ld.param.u64 %rd27, [triton__0d1d2d3d4d5d6d7d8de9de_param_2];
42
+ ld.param.u64 %rd28, [triton__0d1d2d3d4d5d6d7d8de9de_param_3];
43
+ ld.param.u64 %rd29, [triton__0d1d2d3d4d5d6d7d8de9de_param_4];
44
+ shl.b32 %r76, %r74, 2;
45
+ ld.param.u64 %rd30, [triton__0d1d2d3d4d5d6d7d8de9de_param_5];
46
+ and.b32 %r77, %r76, 252;
47
+ ld.param.u64 %rd31, [triton__0d1d2d3d4d5d6d7d8de9de_param_6];
48
+ ld.param.u64 %rd32, [triton__0d1d2d3d4d5d6d7d8de9de_param_7];
49
+ .loc 1 23 28
50
+ mov.u32 %r1, %ctaid.x;
51
+ .loc 1 30 40
52
+ shl.b32 %r78, %r1, 8;
53
+ .loc 1 30 36
54
+ or.b32 %r79, %r78, %r77;
55
+ .loc 1 30 30
56
+ mul.wide.s32 %rd33, %r79, 2;
57
+ add.s64 %rd1, %rd26, %rd33;
58
+ mov.b32 %r4, 0;
59
+ mov.pred %p1, -1;
60
+ .loc 1 30 46
61
+ mov.u32 %r2, 0x0;
62
+ mov.u32 %r3, 0x0;
63
+ @%p1 ld.global.v2.b32 { %r2, %r3 }, [ %rd1 + 0 ];
64
+ @!%p1 mov.u32 %r2, %r4;
65
+ @!%p1 mov.u32 %r3, %r4;
66
+ cvt.u16.u32 %rs1, %r2;
67
+ { .reg .b16 tmp; mov.b32 {tmp, %rs2}, %r2; }
68
+ cvt.u16.u32 %rs3, %r3;
69
+ { .reg .b16 tmp; mov.b32 {tmp, %rs4}, %r3; }
70
+ .loc 1 30 67
71
+ cvt.f32.bf16 %r6, %rs1;
72
+ mov.b32 %f1, %r6;
73
+ cvt.f32.bf16 %r7, %rs2;
74
+ mov.b32 %f2, %r7;
75
+ cvt.f32.bf16 %r8, %rs3;
76
+ mov.b32 %f3, %r8;
77
+ cvt.f32.bf16 %r9, %rs4;
78
+ mov.b32 %f4, %r9;
79
+ .loc 1 31 30
80
+ cvt.u64.u32 %rd34, %r77;
81
+ mul.wide.u32 %rd35, %r77, 4;
82
+ add.s64 %rd2, %rd27, %rd35;
83
+ .loc 1 31 35
84
+ mov.u32 %r10, 0x0;
85
+ mov.u32 %r11, 0x0;
86
+ mov.u32 %r12, 0x0;
87
+ mov.u32 %r13, 0x0;
88
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r10, %r11, %r12, %r13 }, [ %rd2 + 0 ];
89
+ @!%p1 mov.u32 %r10, %r4;
90
+ @!%p1 mov.u32 %r11, %r4;
91
+ @!%p1 mov.u32 %r12, %r4;
92
+ @!%p1 mov.u32 %r13, %r4;
93
+ mov.b32 %f5, %r10;
94
+ mov.b32 %f6, %r11;
95
+ mov.b32 %f7, %r12;
96
+ mov.b32 %f8, %r13;
97
+ .loc 1 32 30
98
+ mul.wide.s32 %rd36, %r79, 4;
99
+ add.s64 %rd3, %rd28, %rd36;
100
+ .loc 1 32 46
101
+ mov.u32 %r18, 0x0;
102
+ mov.u32 %r19, 0x0;
103
+ mov.u32 %r20, 0x0;
104
+ mov.u32 %r21, 0x0;
105
+ @%p1 ld.global.v4.b32 { %r18, %r19, %r20, %r21 }, [ %rd3 + 0 ];
106
+ @!%p1 mov.u32 %r18, %r4;
107
+ @!%p1 mov.u32 %r19, %r4;
108
+ @!%p1 mov.u32 %r20, %r4;
109
+ @!%p1 mov.u32 %r21, %r4;
110
+ mov.b32 %f9, %r18;
111
+ mov.b32 %f10, %r19;
112
+ mov.b32 %f11, %r20;
113
+ mov.b32 %f12, %r21;
114
+ .loc 1 33 30
115
+ mul.wide.s32 %rd37, %r1, 4;
116
+ add.s64 %rd4, %rd29, %rd37;
117
+ .loc 1 33 35
118
+ mov.u32 %r26, 0x0;
119
+ @%p1 ld.global.L1::evict_last.b32 { %r26 }, [ %rd4 + 0 ];
120
+ mov.b32 %f13, %r26;
121
+ mov.u32 %r27, 0x0;
122
+ @%p1 ld.global.L1::evict_last.b32 { %r27 }, [ %rd4 + 0 ];
123
+ mov.u32 %r28, 0x0;
124
+ @%p1 ld.global.L1::evict_last.b32 { %r28 }, [ %rd4 + 0 ];
125
+ mov.u32 %r29, 0x0;
126
+ @%p1 ld.global.L1::evict_last.b32 { %r29 }, [ %rd4 + 0 ];
127
+ .loc 1 34 31
128
+ add.s64 %rd8, %rd30, %rd37;
129
+ .loc 1 34 36
130
+ mov.u32 %r55, 0x0;
131
+ @%p1 ld.global.L1::evict_last.b32 { %r55 }, [ %rd8 + 0 ];
132
+ mov.b32 %f14, %r55;
133
+ mov.u32 %r31, 0x0;
134
+ @%p1 ld.global.L1::evict_last.b32 { %r31 }, [ %rd8 + 0 ];
135
+ mov.u32 %r32, 0x0;
136
+ @%p1 ld.global.L1::evict_last.b32 { %r32 }, [ %rd8 + 0 ];
137
+ mov.u32 %r33, 0x0;
138
+ @%p1 ld.global.L1::evict_last.b32 { %r33 }, [ %rd8 + 0 ];
139
+ .loc 1 35 31
140
+ mul.wide.s32 %rd38, %r1, 8;
141
+ add.s64 %rd13, %rd31, %rd38;
142
+ .loc 1 35 36
143
+ mov.u64 %rd12, 0x0;
144
+ @%p1 ld.global.L1::evict_last.b64 { %rd12 }, [ %rd13 + 0 ];
145
+ mov.u64 %rd14, 0x0;
146
+ @%p1 ld.global.L1::evict_last.b64 { %rd14 }, [ %rd13 + 0 ];
147
+ mov.u64 %rd16, 0x0;
148
+ @%p1 ld.global.L1::evict_last.b64 { %rd16 }, [ %rd13 + 0 ];
149
+ mov.u64 %rd18, 0x0;
150
+ @%p1 ld.global.L1::evict_last.b64 { %rd18 }, [ %rd13 + 0 ];
151
+ .loc 1 36 35
152
+ add.s64 %rd20, %rd25, %rd36;
153
+ .loc 1 36 51
154
+ mov.u32 %r34, 0x0;
155
+ mov.u32 %r35, 0x0;
156
+ mov.u32 %r36, 0x0;
157
+ mov.u32 %r37, 0x0;
158
+ @%p1 ld.global.v4.b32 { %r34, %r35, %r36, %r37 }, [ %rd20 + 0 ];
159
+ @!%p1 mov.u32 %r34, %r4;
160
+ @!%p1 mov.u32 %r35, %r4;
161
+ @!%p1 mov.u32 %r36, %r4;
162
+ @!%p1 mov.u32 %r37, %r4;
163
+ mov.b32 %f15, %r34;
164
+ mov.b32 %f16, %r35;
165
+ mov.b32 %f17, %r36;
166
+ mov.b32 %f18, %r37;
167
+ .loc 1 38 18
168
+ mul.f32 %f19, %f1, %f5;
169
+ mul.f32 %f20, %f2, %f6;
170
+ mul.f32 %f21, %f3, %f7;
171
+ mul.f32 %f22, %f4, %f8;
172
+ $L__tmp1:
173
+ .loc 2 233 15
174
+ fma.rn.f32 %f23, %f1, %f5, %f20;
175
+ fma.rn.f32 %f24, %f3, %f7, %f23;
176
+ fma.rn.f32 %f25, %f4, %f8, %f24;
177
+ $L__tmp2:
178
+ .loc 2 243 36
179
+ mov.b32 %r80, %f25;
180
+ shfl.sync.bfly.b32 %r81, %r80, 16, 31, -1;
181
+ mov.b32 %f26, %r81;
182
+ $L__tmp3:
183
+ .loc 2 233 15
184
+ add.f32 %f27, %f25, %f26;
185
+ $L__tmp4:
186
+ .loc 2 243 36
187
+ mov.b32 %r82, %f27;
188
+ shfl.sync.bfly.b32 %r83, %r82, 8, 31, -1;
189
+ mov.b32 %f28, %r83;
190
+ $L__tmp5:
191
+ .loc 2 233 15
192
+ add.f32 %f29, %f27, %f28;
193
+ $L__tmp6:
194
+ .loc 2 243 36
195
+ mov.b32 %r84, %f29;
196
+ shfl.sync.bfly.b32 %r85, %r84, 4, 31, -1;
197
+ mov.b32 %f30, %r85;
198
+ $L__tmp7:
199
+ .loc 2 233 15
200
+ add.f32 %f31, %f29, %f30;
201
+ $L__tmp8:
202
+ .loc 2 243 36
203
+ mov.b32 %r86, %f31;
204
+ shfl.sync.bfly.b32 %r87, %r86, 2, 31, -1;
205
+ mov.b32 %f32, %r87;
206
+ $L__tmp9:
207
+ .loc 2 233 15
208
+ add.f32 %f33, %f31, %f32;
209
+ $L__tmp10:
210
+ .loc 2 243 36
211
+ mov.b32 %r88, %f33;
212
+ shfl.sync.bfly.b32 %r89, %r88, 1, 31, -1;
213
+ mov.b32 %f34, %r89;
214
+ $L__tmp11:
215
+ .loc 2 233 15
216
+ add.f32 %f35, %f33, %f34;
217
+ $L__tmp12:
218
+ .loc 2 243 36
219
+ setp.eq.s32 %p31, %r75, 0;
220
+ shr.u32 %r90, %r74, 3;
221
+ and.b32 %r91, %r90, 4;
222
+ mov.u32 %r92, global_smem;
223
+ add.s32 %r42, %r92, %r91;
224
+ mov.b32 %r43, %f35;
225
+ @%p31 st.shared.b32 [ %r42 + 0 ], %r43;
226
+ bar.sync 0;
227
+ setp.lt.s32 %p32, %r74, 2;
228
+ add.s32 %r45, %r92, %r76;
229
+ @%p32 ld.shared.b32 %r44, [ %r45 + 0 ];
230
+ mov.b32 %f36, %r44;
231
+ shfl.sync.bfly.b32 %r93, %r44, 1, 31, -1;
232
+ mov.b32 %f37, %r93;
233
+ $L__tmp13:
234
+ .loc 2 233 15
235
+ add.f32 %f38, %f36, %f37;
236
+ $L__tmp14:
237
+ .loc 2 243 36
238
+ and.b32 %r94, %r74, 1;
239
+ setp.eq.b32 %p41, %r94, 1;
240
+ not.pred %p42, %p41;
241
+ and.pred %p33, %p32, %p42;
242
+ mov.b32 %r47, %f38;
243
+ @%p33 st.shared.b32 [ %r45 + 0 ], %r47;
244
+ bar.sync 0;
245
+ ld.shared.f32 %f39, [global_smem];
246
+ $L__tmp15:
247
+ .loc 3 8 15
248
+ add.f32 %f40, %f39, 0f00000000;
249
+ $L__tmp16:
250
+ .loc 1 42 19
251
+ sub.f32 %f41, %f9, %f13;
252
+ sub.f32 %f42, %f10, %f13;
253
+ sub.f32 %f43, %f11, %f13;
254
+ sub.f32 %f44, %f12, %f13;
255
+ .loc 1 43 20
256
+ mul.f32 %f45, %f41, %f14;
257
+ mul.f32 %f46, %f42, %f14;
258
+ mul.f32 %f47, %f43, %f14;
259
+ mul.f32 %f48, %f44, %f14;
260
+ .loc 1 44 19
261
+ mul.f32 %f49, %f20, %f46;
262
+ $L__tmp17:
263
+ .loc 2 243 36
264
+ bar.sync 0;
265
+ $L__tmp18:
266
+ .loc 2 233 15
267
+ fma.rn.f32 %f50, %f19, %f45, %f49;
268
+ fma.rn.f32 %f51, %f21, %f47, %f50;
269
+ fma.rn.f32 %f52, %f22, %f48, %f51;
270
+ $L__tmp19:
271
+ .loc 2 243 36
272
+ mov.b32 %r95, %f52;
273
+ shfl.sync.bfly.b32 %r96, %r95, 16, 31, -1;
274
+ mov.b32 %f53, %r96;
275
+ $L__tmp20:
276
+ .loc 2 233 15
277
+ add.f32 %f54, %f52, %f53;
278
+ $L__tmp21:
279
+ .loc 2 243 36
280
+ mov.b32 %r97, %f54;
281
+ shfl.sync.bfly.b32 %r98, %r97, 8, 31, -1;
282
+ mov.b32 %f55, %r98;
283
+ $L__tmp22:
284
+ .loc 2 233 15
285
+ add.f32 %f56, %f54, %f55;
286
+ $L__tmp23:
287
+ .loc 2 243 36
288
+ mov.b32 %r99, %f56;
289
+ shfl.sync.bfly.b32 %r100, %r99, 4, 31, -1;
290
+ mov.b32 %f57, %r100;
291
+ $L__tmp24:
292
+ .loc 2 233 15
293
+ add.f32 %f58, %f56, %f57;
294
+ $L__tmp25:
295
+ .loc 2 243 36
296
+ mov.b32 %r101, %f58;
297
+ shfl.sync.bfly.b32 %r102, %r101, 2, 31, -1;
298
+ mov.b32 %f59, %r102;
299
+ $L__tmp26:
300
+ .loc 2 233 15
301
+ add.f32 %f60, %f58, %f59;
302
+ $L__tmp27:
303
+ .loc 2 243 36
304
+ mov.b32 %r103, %f60;
305
+ shfl.sync.bfly.b32 %r104, %r103, 1, 31, -1;
306
+ mov.b32 %f61, %r104;
307
+ $L__tmp28:
308
+ .loc 2 233 15
309
+ add.f32 %f62, %f60, %f61;
310
+ $L__tmp29:
311
+ .loc 2 243 36
312
+ mov.b32 %r49, %f62;
313
+ @%p31 st.shared.b32 [ %r42 + 0 ], %r49;
314
+ bar.sync 0;
315
+ @%p32 ld.shared.b32 %r50, [ %r45 + 0 ];
316
+ mov.b32 %f63, %r50;
317
+ shfl.sync.bfly.b32 %r105, %r50, 1, 31, -1;
318
+ mov.b32 %f64, %r105;
319
+ $L__tmp30:
320
+ .loc 2 233 15
321
+ add.f32 %f65, %f63, %f64;
322
+ $L__tmp31:
323
+ .loc 2 243 36
324
+ mov.b32 %r53, %f65;
325
+ @%p33 st.shared.b32 [ %r45 + 0 ], %r53;
326
+ bar.sync 0;
327
+ ld.shared.f32 %f66, [global_smem];
328
+ $L__tmp32:
329
+ .loc 3 8 15
330
+ add.f32 %f67, %f66, 0f00000000;
331
+ $L__tmp33:
332
+ .loc 1 49 21
333
+ setp.eq.s64 %p43, %rd12, -1;
334
+ mov.b32 %r56, 1132462080;
335
+ .loc 1 51 20
336
+ div.full.f32 %r54, %r55, %r56;
337
+ mov.b32 %f68, %r54;
338
+ .loc 1 53 20
339
+ neg.f32 %f69, %f40;
340
+ fma.rn.f32 %f70, %f19, 0f43800000, %f69;
341
+ fma.rn.f32 %f71, %f20, 0f43800000, %f69;
342
+ fma.rn.f32 %f72, %f21, 0f43800000, %f69;
343
+ fma.rn.f32 %f73, %f22, 0f43800000, %f69;
344
+ .loc 1 55 20
345
+ neg.f32 %f74, %f45;
346
+ fma.rn.f32 %f75, %f74, %f67, %f70;
347
+ neg.f32 %f76, %f46;
348
+ fma.rn.f32 %f77, %f76, %f67, %f71;
349
+ neg.f32 %f78, %f47;
350
+ fma.rn.f32 %f79, %f78, %f67, %f72;
351
+ neg.f32 %f80, %f48;
352
+ fma.rn.f32 %f81, %f80, %f67, %f73;
353
+ .loc 1 57 20
354
+ fma.rn.f32 %f82, %f68, %f75, %f15;
355
+ fma.rn.f32 %f83, %f68, %f77, %f16;
356
+ fma.rn.f32 %f84, %f68, %f79, %f17;
357
+ fma.rn.f32 %f85, %f68, %f81, %f18;
358
+ .loc 1 59 35
359
+ selp.f32 %f86, 0f00000000, %f82, %p43;
360
+ selp.f32 %f87, 0f00000000, %f83, %p43;
361
+ selp.f32 %f88, 0f00000000, %f84, %p43;
362
+ selp.f32 %f89, 0f00000000, %f85, %p43;
363
+ .loc 1 61 20
364
+ setp.lt.s64 %p44, %rd12, 0;
365
+ .loc 1 63 56
366
+ shl.b64 %rd39, %rd12, 8;
367
+ add.s64 %rd40, %rd39, 12865792;
368
+ selp.b64 %rd41, %rd40, %rd39, %p44;
369
+ .loc 1 63 52
370
+ or.b64 %rd42, %rd41, %rd34;
371
+ .loc 1 63 30
372
+ shl.b64 %rd43, %rd42, 2;
373
+ add.s64 %rd21, %rd32, %rd43;
374
+ add.s64 %rd22, %rd21, 4;
375
+ add.s64 %rd23, %rd21, 8;
376
+ add.s64 %rd24, %rd21, 12;
377
+ .loc 1 63 83
378
+ mov.b32 %r67, %f86;
379
+ mov.u32 %r66, 0x0;
380
+ @%p1 atom.global.gpu.acq_rel.add.f32 %r66, [ %rd21 + 0 ], %r67;
381
+ mov.b32 %r69, %f87;
382
+ mov.u32 %r68, 0x0;
383
+ @%p1 atom.global.gpu.acq_rel.add.f32 %r68, [ %rd22 + 0 ], %r69;
384
+ mov.b32 %r71, %f88;
385
+ mov.u32 %r70, 0x0;
386
+ @%p1 atom.global.gpu.acq_rel.add.f32 %r70, [ %rd23 + 0 ], %r71;
387
+ mov.b32 %r73, %f89;
388
+ mov.u32 %r72, 0x0;
389
+ @%p1 atom.global.gpu.acq_rel.add.f32 %r72, [ %rd24 + 0 ], %r73;
390
+ .loc 1 63 4
391
+ ret;
392
+ $L__tmp34:
393
+ $L__func_end0:
394
+
395
+ }
396
+ .file 1 "/tmp/torchinductor_root/qr/cqryxm46jcxyr3qdktqirn53eap7h3pjjqiqavyqqyvflabjpvmd.py"
397
+ .file 2 "/usr/local/lib/python3.10/dist-packages/triton/language/standard.py"
398
+ .file 3 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
399
+ .section .debug_abbrev
400
+ {
401
+ .b8 1
402
+ .b8 17
403
+ .b8 1
404
+ .b8 37
405
+ .b8 8
406
+ .b8 19
407
+ .b8 5
408
+ .b8 3
409
+ .b8 8
410
+ .b8 16
411
+ .b8 6
412
+ .b8 27
413
+ .b8 8
414
+ .b8 180
415
+ .b8 66
416
+ .b8 12
417
+ .b8 17
418
+ .b8 1
419
+ .b8 18
420
+ .b8 1
421
+ .b8 0
422
+ .b8 0
423
+ .b8 2
424
+ .b8 46
425
+ .b8 0
426
+ .b8 135
427
+ .b8 64
428
+ .b8 8
429
+ .b8 3
430
+ .b8 8
431
+ .b8 58
432
+ .b8 11
433
+ .b8 59
434
+ .b8 11
435
+ .b8 63
436
+ .b8 12
437
+ .b8 32
438
+ .b8 11
439
+ .b8 0
440
+ .b8 0
441
+ .b8 3
442
+ .b8 46
443
+ .b8 1
444
+ .b8 17
445
+ .b8 1
446
+ .b8 18
447
+ .b8 1
448
+ .b8 64
449
+ .b8 10
450
+ .b8 49
451
+ .b8 19
452
+ .b8 0
453
+ .b8 0
454
+ .b8 4
455
+ .b8 29
456
+ .b8 1
457
+ .b8 49
458
+ .b8 19
459
+ .b8 17
460
+ .b8 1
461
+ .b8 18
462
+ .b8 1
463
+ .b8 88
464
+ .b8 11
465
+ .b8 89
466
+ .b8 11
467
+ .b8 87
468
+ .b8 11
469
+ .b8 0
470
+ .b8 0
471
+ .b8 5
472
+ .b8 29
473
+ .b8 0
474
+ .b8 49
475
+ .b8 19
476
+ .b8 17
477
+ .b8 1
478
+ .b8 18
479
+ .b8 1
480
+ .b8 88
481
+ .b8 11
482
+ .b8 89
483
+ .b8 11
484
+ .b8 87
485
+ .b8 11
486
+ .b8 0
487
+ .b8 0
488
+ .b8 0
489
+ }
490
+ .section .debug_info
491
+ {
492
+ .b32 407
493
+ .b8 2
494
+ .b8 0
495
+ .b32 .debug_abbrev
496
+ .b8 8
497
+ .b8 1
498
+ .b8 116
499
+ .b8 114
500
+ .b8 105
501
+ .b8 116
502
+ .b8 111
503
+ .b8 110
504
+ .b8 0
505
+ .b8 2
506
+ .b8 0
507
+ .b8 99
508
+ .b8 113
509
+ .b8 114
510
+ .b8 121
511
+ .b8 120
512
+ .b8 109
513
+ .b8 52
514
+ .b8 54
515
+ .b8 106
516
+ .b8 99
517
+ .b8 120
518
+ .b8 121
519
+ .b8 114
520
+ .b8 51
521
+ .b8 113
522
+ .b8 100
523
+ .b8 107
524
+ .b8 116
525
+ .b8 113
526
+ .b8 105
527
+ .b8 114
528
+ .b8 110
529
+ .b8 53
530
+ .b8 51
531
+ .b8 101
532
+ .b8 97
533
+ .b8 112
534
+ .b8 55
535
+ .b8 104
536
+ .b8 51
537
+ .b8 112
538
+ .b8 106
539
+ .b8 106
540
+ .b8 113
541
+ .b8 105
542
+ .b8 113
543
+ .b8 97
544
+ .b8 118
545
+ .b8 121
546
+ .b8 113
547
+ .b8 113
548
+ .b8 121
549
+ .b8 118
550
+ .b8 102
551
+ .b8 108
552
+ .b8 97
553
+ .b8 98
554
+ .b8 106
555
+ .b8 112
556
+ .b8 118
557
+ .b8 109
558
+ .b8 100
559
+ .b8 46
560
+ .b8 112
561
+ .b8 121
562
+ .b8 0
563
+ .b32 .debug_line
564
+ .b8 47
565
+ .b8 116
566
+ .b8 109
567
+ .b8 112
568
+ .b8 47
569
+ .b8 116
570
+ .b8 111
571
+ .b8 114
572
+ .b8 99
573
+ .b8 104
574
+ .b8 105
575
+ .b8 110
576
+ .b8 100
577
+ .b8 117
578
+ .b8 99
579
+ .b8 116
580
+ .b8 111
581
+ .b8 114
582
+ .b8 95
583
+ .b8 114
584
+ .b8 111
585
+ .b8 111
586
+ .b8 116
587
+ .b8 47
588
+ .b8 113
589
+ .b8 114
590
+ .b8 0
591
+ .b8 1
592
+ .b64 $L__func_begin0
593
+ .b64 $L__func_end0
594
+ .b8 2
595
+ .b8 116
596
+ .b8 114
597
+ .b8 105
598
+ .b8 116
599
+ .b8 111
600
+ .b8 110
601
+ .b8 95
602
+ .b8 95
603
+ .b8 48
604
+ .b8 100
605
+ .b8 49
606
+ .b8 100
607
+ .b8 50
608
+ .b8 100
609
+ .b8 51
610
+ .b8 100
611
+ .b8 52
612
+ .b8 100
613
+ .b8 53
614
+ .b8 100
615
+ .b8 54
616
+ .b8 100
617
+ .b8 55
618
+ .b8 100
619
+ .b8 56
620
+ .b8 100
621
+ .b8 101
622
+ .b8 57
623
+ .b8 100
624
+ .b8 101
625
+ .b8 0
626
+ .b8 116
627
+ .b8 114
628
+ .b8 105
629
+ .b8 116
630
+ .b8 111
631
+ .b8 110
632
+ .b8 95
633
+ .b8 95
634
+ .b8 48
635
+ .b8 100
636
+ .b8 49
637
+ .b8 100
638
+ .b8 50
639
+ .b8 100
640
+ .b8 51
641
+ .b8 100
642
+ .b8 52
643
+ .b8 100
644
+ .b8 53
645
+ .b8 100
646
+ .b8 54
647
+ .b8 100
648
+ .b8 55
649
+ .b8 100
650
+ .b8 56
651
+ .b8 100
652
+ .b8 101
653
+ .b8 57
654
+ .b8 100
655
+ .b8 101
656
+ .b8 0
657
+ .b8 1
658
+ .b8 18
659
+ .b8 1
660
+ .b8 1
661
+ .b8 3
662
+ .b64 $L__func_begin0
663
+ .b64 $L__func_end0
664
+ .b8 1
665
+ .b8 156
666
+ .b32 125
667
+ .b8 4
668
+ .b32 125
669
+ .b64 $L__tmp1
670
+ .b64 $L__tmp14
671
+ .b8 2
672
+ .b8 41
673
+ .b8 57
674
+ .b8 5
675
+ .b32 125
676
+ .b64 $L__tmp1
677
+ .b64 $L__tmp14
678
+ .b8 2
679
+ .b8 243
680
+ .b8 36
681
+ .b8 0
682
+ .b8 5
683
+ .b32 125
684
+ .b64 $L__tmp2
685
+ .b64 $L__tmp15
686
+ .b8 2
687
+ .b8 41
688
+ .b8 57
689
+ .b8 5
690
+ .b32 125
691
+ .b64 $L__tmp15
692
+ .b64 $L__tmp16
693
+ .b8 3
694
+ .b8 41
695
+ .b8 44
696
+ .b8 5
697
+ .b32 125
698
+ .b64 $L__tmp17
699
+ .b64 $L__tmp32
700
+ .b8 2
701
+ .b8 47
702
+ .b8 59
703
+ .b8 4
704
+ .b32 125
705
+ .b64 $L__tmp18
706
+ .b64 $L__tmp31
707
+ .b8 2
708
+ .b8 47
709
+ .b8 59
710
+ .b8 5
711
+ .b32 125
712
+ .b64 $L__tmp18
713
+ .b64 $L__tmp31
714
+ .b8 2
715
+ .b8 243
716
+ .b8 36
717
+ .b8 0
718
+ .b8 5
719
+ .b32 125
720
+ .b64 $L__tmp32
721
+ .b64 $L__tmp33
722
+ .b8 3
723
+ .b8 47
724
+ .b8 45
725
+ .b8 0
726
+ .b8 0
727
+ }
728
+ .section .debug_pubnames
729
+ {
730
+ .b32 $L__pubNames_end0-$L__pubNames_start0
731
+ $L__pubNames_start0:
732
+ .b8 2
733
+ .b8 0
734
+ .b32 .debug_info
735
+ .b32 411
736
+ .b32 125
737
+ .b8 116
738
+ .b8 114
739
+ .b8 105
740
+ .b8 116
741
+ .b8 111
742
+ .b8 110
743
+ .b8 95
744
+ .b8 95
745
+ .b8 48
746
+ .b8 100
747
+ .b8 49
748
+ .b8 100
749
+ .b8 50
750
+ .b8 100
751
+ .b8 51
752
+ .b8 100
753
+ .b8 52
754
+ .b8 100
755
+ .b8 53
756
+ .b8 100
757
+ .b8 54
758
+ .b8 100
759
+ .b8 55
760
+ .b8 100
761
+ .b8 56
762
+ .b8 100
763
+ .b8 101
764
+ .b8 57
765
+ .b8 100
766
+ .b8 101
767
+ .b8 0
768
+ .b32 0
769
+ $L__pubNames_end0:
770
+ }
771
+ .section .debug_pubtypes
772
+ {
773
+ .b32 $L__pubTypes_end0-$L__pubTypes_start0
774
+ $L__pubTypes_start0:
775
+ .b8 2
776
+ .b8 0
777
+ .b32 .debug_info
778
+ .b32 411
779
+ .b32 0
780
+ $L__pubTypes_end0:
781
+ }
782
+ .section .debug_loc { }
.triton/dump/21d0195c63fb062bfc567b79c9bb2771/triton_.ttgir ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [4], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
2
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
3
+ tt.func public @triton__0d1d2d3d4d5d6d7d8de9de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg6: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg7: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg8: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg9: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
4
+ %cst = arith.constant dense<256> : tensor<256xi32, #blocked>
5
+ %cst_0 = arith.constant dense<-1> : tensor<1xi64, #blocked>
6
+ %cst_1 = arith.constant dense<2.560000e+02> : tensor<1xf32, #blocked>
7
+ %cst_2 = arith.constant dense<256> : tensor<1xi64, #blocked>
8
+ %cst_3 = arith.constant dense<0> : tensor<1xi64, #blocked>
9
+ %cst_4 = arith.constant dense<50257> : tensor<1xi64, #blocked>
10
+ %cst_5 = arith.constant 0.000000e+00 : f32
11
+ %c256_i32 = arith.constant 256 : i32
12
+ %cst_6 = arith.constant dense<0.000000e+00> : tensor<256xf32, #blocked>
13
+ %cst_7 = arith.constant dense<2.560000e+02> : tensor<256xf32, #blocked>
14
+ %cst_8 = arith.constant dense<0.000000e+00> : tensor<256xbf16, #blocked>
15
+ %0 = tt.get_program_id x : i32
16
+ %1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32, #blocked>
17
+ %2 = arith.cmpi slt, %1, %cst : tensor<256xi32, #blocked>
18
+ %3 = arith.muli %0, %c256_i32 : i32
19
+ %4 = tt.splat %3 : (i32) -> tensor<256xi32, #blocked>
20
+ %5 = arith.addi %1, %4 : tensor<256xi32, #blocked>
21
+ %6 = tt.splat %arg1 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
22
+ %7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
23
+ %8 = tt.load %7, %2, %cst_8 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
24
+ %9 = arith.extf %8 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
25
+ %10 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
26
+ %11 = tt.addptr %10, %1 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
27
+ %12 = tt.load %11, %2, %cst_6 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32, #blocked>
28
+ %13 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
29
+ %14 = tt.addptr %13, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
30
+ %15 = tt.load %14, %2, %cst_6 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
31
+ %16 = tt.addptr %arg4, %0 : !tt.ptr<f32, 1>, i32
32
+ %17 = tt.splat %16 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked>
33
+ %18 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1xf32, #blocked>
34
+ %19 = tt.addptr %arg5, %0 : !tt.ptr<f32, 1>, i32
35
+ %20 = tt.splat %19 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked>
36
+ %21 = tt.load %20 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1xf32, #blocked>
37
+ %22 = tt.addptr %arg6, %0 : !tt.ptr<i64, 1>, i32
38
+ %23 = tt.splat %22 : (!tt.ptr<i64, 1>) -> tensor<1x!tt.ptr<i64, 1>, #blocked>
39
+ %24 = tt.load %23 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1xi64, #blocked>
40
+ %25 = tt.splat %arg0 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
41
+ %26 = tt.addptr %25, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
42
+ %27 = tt.load %26, %2, %cst_6 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
43
+ %28 = arith.mulf %9, %12 : tensor<256xf32, #blocked>
44
+ %29 = arith.select %2, %28, %cst_6 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
45
+ %30 = "tt.reduce"(%29) <{axis = 0 : i32}> ({
46
+ ^bb0(%arg10: f32, %arg11: f32):
47
+ %63 = arith.addf %arg10, %arg11 : f32
48
+ tt.reduce.return %63 : f32
49
+ }) : (tensor<256xf32, #blocked>) -> f32
50
+ %31 = arith.addf %30, %cst_5 : f32
51
+ %32 = tt.broadcast %18 : (tensor<1xf32, #blocked>) -> tensor<256xf32, #blocked>
52
+ %33 = arith.subf %15, %32 : tensor<256xf32, #blocked>
53
+ %34 = tt.broadcast %21 : (tensor<1xf32, #blocked>) -> tensor<256xf32, #blocked>
54
+ %35 = arith.mulf %33, %34 : tensor<256xf32, #blocked>
55
+ %36 = arith.mulf %28, %35 : tensor<256xf32, #blocked>
56
+ %37 = arith.select %2, %36, %cst_6 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
57
+ %38 = "tt.reduce"(%37) <{axis = 0 : i32}> ({
58
+ ^bb0(%arg10: f32, %arg11: f32):
59
+ %63 = arith.addf %arg10, %arg11 : f32
60
+ tt.reduce.return %63 : f32
61
+ }) : (tensor<256xf32, #blocked>) -> f32
62
+ %39 = arith.addf %38, %cst_5 : f32
63
+ %40 = arith.cmpi eq, %24, %cst_0 : tensor<1xi64, #blocked>
64
+ %41 = arith.divf %21, %cst_1 : tensor<1xf32, #blocked>
65
+ %42 = arith.mulf %28, %cst_7 : tensor<256xf32, #blocked>
66
+ %43 = tt.splat %31 : (f32) -> tensor<256xf32, #blocked>
67
+ %44 = arith.subf %42, %43 : tensor<256xf32, #blocked>
68
+ %45 = tt.splat %39 : (f32) -> tensor<256xf32, #blocked>
69
+ %46 = arith.mulf %35, %45 : tensor<256xf32, #blocked>
70
+ %47 = arith.subf %44, %46 : tensor<256xf32, #blocked>
71
+ %48 = tt.broadcast %41 : (tensor<1xf32, #blocked>) -> tensor<256xf32, #blocked>
72
+ %49 = arith.mulf %48, %47 : tensor<256xf32, #blocked>
73
+ %50 = arith.addf %27, %49 : tensor<256xf32, #blocked>
74
+ %51 = tt.broadcast %40 : (tensor<1xi1, #blocked>) -> tensor<256xi1, #blocked>
75
+ %52 = arith.select %51, %cst_6, %50 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
76
+ %53 = arith.addi %24, %cst_4 : tensor<1xi64, #blocked>
77
+ %54 = arith.cmpi slt, %24, %cst_3 : tensor<1xi64, #blocked>
78
+ %55 = arith.select %54, %53, %24 : tensor<1xi1, #blocked>, tensor<1xi64, #blocked>
79
+ %56 = arith.muli %55, %cst_2 : tensor<1xi64, #blocked>
80
+ %57 = tt.broadcast %56 : (tensor<1xi64, #blocked>) -> tensor<256xi64, #blocked>
81
+ %58 = arith.extsi %1 : tensor<256xi32, #blocked> to tensor<256xi64, #blocked>
82
+ %59 = arith.addi %58, %57 : tensor<256xi64, #blocked>
83
+ %60 = tt.splat %arg7 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
84
+ %61 = tt.addptr %60, %59 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi64, #blocked>
85
+ %62 = "tt.atomic_rmw"(%61, %52, %2) <{atomic_rmw_op = 5 : i32, scope = 1 : i32, sem = 4 : i32}> : (tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xf32, #blocked>, tensor<256xi1, #blocked>) -> tensor<256xf32, #blocked>
86
+ tt.return
87
+ }
88
+ }
.triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.llir ADDED
@@ -0,0 +1,980 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+
4
+ @assertFunc_1 = internal constant [25 x i8] c"_call_with_frames_removed"
5
+ @assertFile_1 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
6
+ @assertMessage_1 = internal constant [39 x i8] c"index out of bounds: 0 <= tmp13 < 50257"
7
+ @assertFunc_0 = internal constant [25 x i8] c"_call_with_frames_removed"
8
+ @assertFile_0 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
9
+ @assertMessage_0 = internal constant [38 x i8] c"index out of bounds: 0 <= tmp3 < 50257"
10
+ @global_smem = external local_unnamed_addr addrspace(3) global [0 x i8]
11
+ @.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
12
+
13
+ declare void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr
14
+
15
+ define void @triton__0d1d2d3d4d5de6de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, i32 %5, i32 %6) local_unnamed_addr !dbg !7 {
16
+ %8 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
17
+ %9 = lshr i32 %8, 5, !dbg !10
18
+ %10 = and i32 %9, 7, !dbg !10
19
+ %11 = and i32 %8, 15, !dbg !10
20
+ %12 = shl i32 %8, 3, !dbg !11
21
+ %13 = and i32 %12, 248, !dbg !11
22
+ %14 = or i32 %13, 4, !dbg !11
23
+ %urem = and i32 %8, 255, !dbg !11
24
+ %15 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !12
25
+ %16 = shl i32 %15, 4, !dbg !13
26
+ %17 = or i32 %16, %10, !dbg !14
27
+ %18 = or i32 %17, 8, !dbg !14
28
+ %19 = or i32 %16, %11, !dbg !14
29
+ %20 = sext i32 %17 to i64, !dbg !15
30
+ %21 = getelementptr i64, ptr addrspace(1) %0, i64 %20, !dbg !15
31
+ %22 = sext i32 %18 to i64, !dbg !15
32
+ %23 = getelementptr i64, ptr addrspace(1) %0, i64 %22, !dbg !15
33
+ %24 = sext i32 %19 to i64, !dbg !15
34
+ %25 = getelementptr i64, ptr addrspace(1) %0, i64 %24, !dbg !15
35
+ %26 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
36
+ %27 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
37
+ %28 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
38
+ %29 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
39
+ %30 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
40
+ %31 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
41
+ %32 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
42
+ %33 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %21, i1 true) #6, !dbg !16
43
+ %34 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
44
+ %35 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
45
+ %36 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
46
+ %37 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
47
+ %38 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
48
+ %39 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
49
+ %40 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
50
+ %41 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !16
51
+ %42 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !16
52
+ %43 = srem i32 %17, 512, !dbg !17
53
+ %44 = srem i32 %18, 512, !dbg !17
54
+ %45 = shl nsw i32 %43, 8, !dbg !18
55
+ %46 = shl nsw i32 %44, 8, !dbg !18
56
+ %47 = or i32 %45, %13, !dbg !19
57
+ %48 = or i32 %45, %14, !dbg !19
58
+ %49 = or i32 %46, %13, !dbg !19
59
+ %50 = or i32 %46, %14, !dbg !19
60
+ %51 = sext i32 %47 to i64, !dbg !20
61
+ %52 = getelementptr float, ptr addrspace(1) %2, i64 %51, !dbg !20
62
+ %53 = sext i32 %48 to i64, !dbg !20
63
+ %54 = getelementptr float, ptr addrspace(1) %2, i64 %53, !dbg !20
64
+ %55 = sext i32 %49 to i64, !dbg !20
65
+ %56 = getelementptr float, ptr addrspace(1) %2, i64 %55, !dbg !20
66
+ %57 = sext i32 %50 to i64, !dbg !20
67
+ %58 = getelementptr float, ptr addrspace(1) %2, i64 %57, !dbg !20
68
+ %59 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %52, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
69
+ %60 = extractvalue { i32, i32, i32, i32 } %59, 0, !dbg !21
70
+ %61 = extractvalue { i32, i32, i32, i32 } %59, 1, !dbg !21
71
+ %62 = extractvalue { i32, i32, i32, i32 } %59, 2, !dbg !21
72
+ %63 = extractvalue { i32, i32, i32, i32 } %59, 3, !dbg !21
73
+ %64 = bitcast i32 %60 to float, !dbg !21
74
+ %65 = bitcast i32 %61 to float, !dbg !21
75
+ %66 = bitcast i32 %62 to float, !dbg !21
76
+ %67 = bitcast i32 %63 to float, !dbg !21
77
+ %68 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %54, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
78
+ %69 = extractvalue { i32, i32, i32, i32 } %68, 0, !dbg !21
79
+ %70 = extractvalue { i32, i32, i32, i32 } %68, 1, !dbg !21
80
+ %71 = extractvalue { i32, i32, i32, i32 } %68, 2, !dbg !21
81
+ %72 = extractvalue { i32, i32, i32, i32 } %68, 3, !dbg !21
82
+ %73 = bitcast i32 %69 to float, !dbg !21
83
+ %74 = bitcast i32 %70 to float, !dbg !21
84
+ %75 = bitcast i32 %71 to float, !dbg !21
85
+ %76 = bitcast i32 %72 to float, !dbg !21
86
+ %77 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %56, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
87
+ %78 = extractvalue { i32, i32, i32, i32 } %77, 0, !dbg !21
88
+ %79 = extractvalue { i32, i32, i32, i32 } %77, 1, !dbg !21
89
+ %80 = extractvalue { i32, i32, i32, i32 } %77, 2, !dbg !21
90
+ %81 = extractvalue { i32, i32, i32, i32 } %77, 3, !dbg !21
91
+ %82 = bitcast i32 %78 to float, !dbg !21
92
+ %83 = bitcast i32 %79 to float, !dbg !21
93
+ %84 = bitcast i32 %80 to float, !dbg !21
94
+ %85 = bitcast i32 %81 to float, !dbg !21
95
+ %86 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %58, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
96
+ %87 = extractvalue { i32, i32, i32, i32 } %86, 0, !dbg !21
97
+ %88 = extractvalue { i32, i32, i32, i32 } %86, 1, !dbg !21
98
+ %89 = extractvalue { i32, i32, i32, i32 } %86, 2, !dbg !21
99
+ %90 = extractvalue { i32, i32, i32, i32 } %86, 3, !dbg !21
100
+ %91 = bitcast i32 %87 to float, !dbg !21
101
+ %92 = bitcast i32 %88 to float, !dbg !21
102
+ %93 = bitcast i32 %89 to float, !dbg !21
103
+ %94 = bitcast i32 %90 to float, !dbg !21
104
+ %95 = add i64 %42, 50257, !dbg !22
105
+ %96 = icmp slt i64 %26, 0, !dbg !23
106
+ %97 = icmp slt i64 %34, 0, !dbg !23
107
+ %98 = icmp slt i64 %42, 0, !dbg !23
108
+ %99 = select i1 %98, i64 %95, i64 %42, !dbg !24
109
+ %100 = icmp ugt i64 %99, 50256, !dbg !25
110
+ br i1 %100, label %101, label %102, !dbg !26
111
+
112
+ 101: ; preds = %7
113
+ tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 883, ptr nonnull @assertFunc_0, i64 1), !dbg !26
114
+ br label %102, !dbg !26
115
+
116
+ 102: ; preds = %101, %7
117
+ %103 = shl i64 %26, 8, !dbg !27
118
+ %104 = add i64 %103, 12865792, !dbg !27
119
+ %105 = select i1 %96, i64 %104, i64 %103, !dbg !27
120
+ %106 = shl i64 %34, 8, !dbg !27
121
+ %107 = add i64 %106, 12865792, !dbg !27
122
+ %108 = select i1 %97, i64 %107, i64 %106, !dbg !27
123
+ %109 = zext nneg i32 %13 to i64
124
+ %110 = zext nneg i32 %14 to i64
125
+ %111 = or i64 %105, %109, !dbg !28
126
+ %112 = or i64 %105, %110, !dbg !28
127
+ %113 = or i64 %108, %109, !dbg !28
128
+ %114 = or i64 %108, %110, !dbg !28
129
+ %115 = getelementptr float, ptr addrspace(1) %1, i64 %111, !dbg !29
130
+ %116 = getelementptr float, ptr addrspace(1) %1, i64 %112, !dbg !29
131
+ %117 = getelementptr float, ptr addrspace(1) %1, i64 %113, !dbg !29
132
+ %118 = getelementptr float, ptr addrspace(1) %1, i64 %114, !dbg !29
133
+ %119 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %115, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
134
+ %120 = extractvalue { i32, i32, i32, i32 } %119, 0, !dbg !30
135
+ %121 = extractvalue { i32, i32, i32, i32 } %119, 1, !dbg !30
136
+ %122 = extractvalue { i32, i32, i32, i32 } %119, 2, !dbg !30
137
+ %123 = extractvalue { i32, i32, i32, i32 } %119, 3, !dbg !30
138
+ %124 = bitcast i32 %120 to float, !dbg !30
139
+ %125 = bitcast i32 %121 to float, !dbg !30
140
+ %126 = bitcast i32 %122 to float, !dbg !30
141
+ %127 = bitcast i32 %123 to float, !dbg !30
142
+ %128 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %116, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
143
+ %129 = extractvalue { i32, i32, i32, i32 } %128, 0, !dbg !30
144
+ %130 = extractvalue { i32, i32, i32, i32 } %128, 1, !dbg !30
145
+ %131 = extractvalue { i32, i32, i32, i32 } %128, 2, !dbg !30
146
+ %132 = extractvalue { i32, i32, i32, i32 } %128, 3, !dbg !30
147
+ %133 = bitcast i32 %129 to float, !dbg !30
148
+ %134 = bitcast i32 %130 to float, !dbg !30
149
+ %135 = bitcast i32 %131 to float, !dbg !30
150
+ %136 = bitcast i32 %132 to float, !dbg !30
151
+ %137 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %117, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
152
+ %138 = extractvalue { i32, i32, i32, i32 } %137, 0, !dbg !30
153
+ %139 = extractvalue { i32, i32, i32, i32 } %137, 1, !dbg !30
154
+ %140 = extractvalue { i32, i32, i32, i32 } %137, 2, !dbg !30
155
+ %141 = extractvalue { i32, i32, i32, i32 } %137, 3, !dbg !30
156
+ %142 = bitcast i32 %138 to float, !dbg !30
157
+ %143 = bitcast i32 %139 to float, !dbg !30
158
+ %144 = bitcast i32 %140 to float, !dbg !30
159
+ %145 = bitcast i32 %141 to float, !dbg !30
160
+ %146 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %118, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !30
161
+ %147 = extractvalue { i32, i32, i32, i32 } %146, 0, !dbg !30
162
+ %148 = extractvalue { i32, i32, i32, i32 } %146, 1, !dbg !30
163
+ %149 = extractvalue { i32, i32, i32, i32 } %146, 2, !dbg !30
164
+ %150 = extractvalue { i32, i32, i32, i32 } %146, 3, !dbg !30
165
+ %151 = bitcast i32 %147 to float, !dbg !30
166
+ %152 = bitcast i32 %148 to float, !dbg !30
167
+ %153 = bitcast i32 %149 to float, !dbg !30
168
+ %154 = bitcast i32 %150 to float, !dbg !30
169
+ %155 = fadd float %64, %124, !dbg !31
170
+ %156 = fadd float %65, %125, !dbg !31
171
+ %157 = fadd float %66, %126, !dbg !31
172
+ %158 = fadd float %67, %127, !dbg !31
173
+ %159 = fadd float %73, %133, !dbg !31
174
+ %160 = fadd float %74, %134, !dbg !31
175
+ %161 = fadd float %75, %135, !dbg !31
176
+ %162 = fadd float %76, %136, !dbg !31
177
+ %163 = fadd float %82, %142, !dbg !31
178
+ %164 = fadd float %83, %143, !dbg !31
179
+ %165 = fadd float %84, %144, !dbg !31
180
+ %166 = fadd float %85, %145, !dbg !31
181
+ %167 = fadd float %91, %151, !dbg !31
182
+ %168 = fadd float %92, %152, !dbg !31
183
+ %169 = fadd float %93, %153, !dbg !31
184
+ %170 = fadd float %94, %154, !dbg !31
185
+ %171 = fadd float %155, 0.000000e+00, !dbg !32
186
+ %172 = fadd float %156, 0.000000e+00, !dbg !32
187
+ %173 = fadd float %157, 0.000000e+00, !dbg !32
188
+ %174 = fadd float %158, 0.000000e+00, !dbg !32
189
+ %175 = fadd float %159, 0.000000e+00, !dbg !32
190
+ %176 = fadd float %160, 0.000000e+00, !dbg !32
191
+ %177 = fadd float %161, 0.000000e+00, !dbg !32
192
+ %178 = fadd float %162, 0.000000e+00, !dbg !32
193
+ %179 = fadd float %163, 0.000000e+00, !dbg !32
194
+ %180 = fadd float %164, 0.000000e+00, !dbg !32
195
+ %181 = fadd float %165, 0.000000e+00, !dbg !32
196
+ %182 = fadd float %166, 0.000000e+00, !dbg !32
197
+ %183 = fadd float %167, 0.000000e+00, !dbg !32
198
+ %184 = fadd float %168, 0.000000e+00, !dbg !32
199
+ %185 = fadd float %169, 0.000000e+00, !dbg !32
200
+ %186 = fadd float %170, 0.000000e+00, !dbg !32
201
+ %187 = fsub float %155, %171, !dbg !36
202
+ %188 = fsub float %156, %172, !dbg !36
203
+ %189 = fsub float %157, %173, !dbg !36
204
+ %190 = fsub float %158, %174, !dbg !36
205
+ %191 = fsub float %159, %175, !dbg !36
206
+ %192 = fsub float %160, %176, !dbg !36
207
+ %193 = fsub float %161, %177, !dbg !36
208
+ %194 = fsub float %162, %178, !dbg !36
209
+ %195 = fsub float %163, %179, !dbg !36
210
+ %196 = fsub float %164, %180, !dbg !36
211
+ %197 = fsub float %165, %181, !dbg !36
212
+ %198 = fsub float %166, %182, !dbg !36
213
+ %199 = fsub float %167, %183, !dbg !36
214
+ %200 = fsub float %168, %184, !dbg !36
215
+ %201 = fsub float %169, %185, !dbg !36
216
+ %202 = fsub float %170, %186, !dbg !36
217
+ %203 = fmul float %155, %187, !dbg !37
218
+ %204 = fmul float %156, %188, !dbg !37
219
+ %205 = fmul float %157, %189, !dbg !37
220
+ %206 = fmul float %158, %190, !dbg !37
221
+ %207 = fmul float %159, %191, !dbg !37
222
+ %208 = fmul float %160, %192, !dbg !37
223
+ %209 = fmul float %161, %193, !dbg !37
224
+ %210 = fmul float %162, %194, !dbg !37
225
+ %211 = fmul float %163, %195, !dbg !37
226
+ %212 = fmul float %164, %196, !dbg !37
227
+ %213 = fmul float %165, %197, !dbg !37
228
+ %214 = fmul float %166, %198, !dbg !37
229
+ %215 = fmul float %167, %199, !dbg !37
230
+ %216 = fmul float %168, %200, !dbg !37
231
+ %217 = fmul float %169, %201, !dbg !37
232
+ %218 = fmul float %170, %202, !dbg !37
233
+ %219 = fadd float %203, 0.000000e+00, !dbg !38
234
+ %220 = fadd float %204, 0.000000e+00, !dbg !38
235
+ %221 = fadd float %205, 0.000000e+00, !dbg !38
236
+ %222 = fadd float %206, 0.000000e+00, !dbg !38
237
+ %223 = fadd float %207, 0.000000e+00, !dbg !38
238
+ %224 = fadd float %208, 0.000000e+00, !dbg !38
239
+ %225 = fadd float %209, 0.000000e+00, !dbg !38
240
+ %226 = fadd float %210, 0.000000e+00, !dbg !38
241
+ %227 = fadd float %211, 0.000000e+00, !dbg !38
242
+ %228 = fadd float %212, 0.000000e+00, !dbg !38
243
+ %229 = fadd float %213, 0.000000e+00, !dbg !38
244
+ %230 = fadd float %214, 0.000000e+00, !dbg !38
245
+ %231 = fadd float %215, 0.000000e+00, !dbg !38
246
+ %232 = fadd float %216, 0.000000e+00, !dbg !38
247
+ %233 = fadd float %217, 0.000000e+00, !dbg !38
248
+ %234 = fadd float %218, 0.000000e+00, !dbg !38
249
+ %235 = fsub float %172, %171, !dbg !39
250
+ %236 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 2.000000e+00) #6, !dbg !43
251
+ %237 = fmul float %236, %235, !dbg !44
252
+ %238 = fadd float %171, %237, !dbg !45
253
+ %239 = fadd float %219, %220, !dbg !46
254
+ %240 = fmul float %235, %235, !dbg !47
255
+ %241 = fmul float %236, %240, !dbg !48
256
+ %242 = fadd float %241, %239, !dbg !49
257
+ %243 = fsub float %173, %238, !dbg !39
258
+ %244 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 3.000000e+00) #6, !dbg !43
259
+ %245 = fmul float %244, %243, !dbg !44
260
+ %246 = fadd float %238, %245, !dbg !45
261
+ %247 = fadd float %221, %242, !dbg !46
262
+ %248 = fmul float %243, %243, !dbg !47
263
+ %249 = fmul float %248, 2.000000e+00, !dbg !50
264
+ %250 = fmul float %244, %249, !dbg !48
265
+ %251 = fadd float %247, %250, !dbg !49
266
+ %252 = fsub float %174, %246, !dbg !39
267
+ %253 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 4.000000e+00) #6, !dbg !43
268
+ %254 = fmul float %253, %252, !dbg !44
269
+ %255 = fadd float %246, %254, !dbg !45
270
+ %256 = fadd float %222, %251, !dbg !46
271
+ %257 = fmul float %252, %252, !dbg !47
272
+ %258 = fmul float %257, 3.000000e+00, !dbg !50
273
+ %259 = fmul float %253, %258, !dbg !48
274
+ %260 = fadd float %256, %259, !dbg !49
275
+ %261 = fsub float %175, %255, !dbg !39
276
+ %262 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 5.000000e+00) #6, !dbg !43
277
+ %263 = fmul float %262, %261, !dbg !44
278
+ %264 = fadd float %255, %263, !dbg !45
279
+ %265 = fadd float %223, %260, !dbg !46
280
+ %266 = fmul float %261, %261, !dbg !47
281
+ %267 = fmul float %266, 4.000000e+00, !dbg !50
282
+ %268 = fmul float %262, %267, !dbg !48
283
+ %269 = fadd float %265, %268, !dbg !49
284
+ %270 = fsub float %176, %264, !dbg !39
285
+ %271 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 6.000000e+00) #6, !dbg !43
286
+ %272 = fmul float %271, %270, !dbg !44
287
+ %273 = fadd float %264, %272, !dbg !45
288
+ %274 = fadd float %224, %269, !dbg !46
289
+ %275 = fmul float %270, %270, !dbg !47
290
+ %276 = fmul float %275, 5.000000e+00, !dbg !50
291
+ %277 = fmul float %271, %276, !dbg !48
292
+ %278 = fadd float %274, %277, !dbg !49
293
+ %279 = fsub float %177, %273, !dbg !39
294
+ %280 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 7.000000e+00) #6, !dbg !43
295
+ %281 = fmul float %280, %279, !dbg !44
296
+ %282 = fadd float %273, %281, !dbg !45
297
+ %283 = fadd float %225, %278, !dbg !46
298
+ %284 = fmul float %279, %279, !dbg !47
299
+ %285 = fmul float %284, 6.000000e+00, !dbg !50
300
+ %286 = fmul float %280, %285, !dbg !48
301
+ %287 = fadd float %283, %286, !dbg !49
302
+ %288 = fsub float %178, %282, !dbg !39
303
+ %289 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 8.000000e+00) #6, !dbg !43
304
+ %290 = fmul float %289, %288, !dbg !44
305
+ %291 = fadd float %282, %290, !dbg !45
306
+ %292 = fadd float %226, %287, !dbg !46
307
+ %293 = fmul float %288, %288, !dbg !47
308
+ %294 = fmul float %293, 7.000000e+00, !dbg !50
309
+ %295 = fmul float %289, %294, !dbg !48
310
+ %296 = fadd float %292, %295, !dbg !49
311
+ %297 = fsub float %180, %179, !dbg !39
312
+ %298 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 2.000000e+00) #6, !dbg !43
313
+ %299 = fmul float %297, %298, !dbg !44
314
+ %300 = fadd float %179, %299, !dbg !45
315
+ %301 = fadd float %227, %228, !dbg !46
316
+ %302 = fmul float %297, %297, !dbg !47
317
+ %303 = fmul float %302, %298, !dbg !48
318
+ %304 = fadd float %301, %303, !dbg !49
319
+ %305 = fsub float %181, %300, !dbg !39
320
+ %306 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 3.000000e+00) #6, !dbg !43
321
+ %307 = fmul float %306, %305, !dbg !44
322
+ %308 = fadd float %300, %307, !dbg !45
323
+ %309 = fadd float %229, %304, !dbg !46
324
+ %310 = fmul float %305, %305, !dbg !47
325
+ %311 = fmul float %310, 2.000000e+00, !dbg !50
326
+ %312 = fmul float %306, %311, !dbg !48
327
+ %313 = fadd float %309, %312, !dbg !49
328
+ %314 = fsub float %182, %308, !dbg !39
329
+ %315 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 4.000000e+00) #6, !dbg !43
330
+ %316 = fmul float %315, %314, !dbg !44
331
+ %317 = fadd float %308, %316, !dbg !45
332
+ %318 = fadd float %230, %313, !dbg !46
333
+ %319 = fmul float %314, %314, !dbg !47
334
+ %320 = fmul float %319, 3.000000e+00, !dbg !50
335
+ %321 = fmul float %315, %320, !dbg !48
336
+ %322 = fadd float %318, %321, !dbg !49
337
+ %323 = fsub float %183, %317, !dbg !39
338
+ %324 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 5.000000e+00) #6, !dbg !43
339
+ %325 = fmul float %324, %323, !dbg !44
340
+ %326 = fadd float %317, %325, !dbg !45
341
+ %327 = fadd float %231, %322, !dbg !46
342
+ %328 = fmul float %323, %323, !dbg !47
343
+ %329 = fmul float %328, 4.000000e+00, !dbg !50
344
+ %330 = fmul float %324, %329, !dbg !48
345
+ %331 = fadd float %327, %330, !dbg !49
346
+ %332 = fsub float %184, %326, !dbg !39
347
+ %333 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 6.000000e+00) #6, !dbg !43
348
+ %334 = fmul float %333, %332, !dbg !44
349
+ %335 = fadd float %326, %334, !dbg !45
350
+ %336 = fadd float %232, %331, !dbg !46
351
+ %337 = fmul float %332, %332, !dbg !47
352
+ %338 = fmul float %337, 5.000000e+00, !dbg !50
353
+ %339 = fmul float %333, %338, !dbg !48
354
+ %340 = fadd float %336, %339, !dbg !49
355
+ %341 = fsub float %185, %335, !dbg !39
356
+ %342 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 7.000000e+00) #6, !dbg !43
357
+ %343 = fmul float %342, %341, !dbg !44
358
+ %344 = fadd float %335, %343, !dbg !45
359
+ %345 = fadd float %233, %340, !dbg !46
360
+ %346 = fmul float %341, %341, !dbg !47
361
+ %347 = fmul float %346, 6.000000e+00, !dbg !50
362
+ %348 = fmul float %342, %347, !dbg !48
363
+ %349 = fadd float %345, %348, !dbg !49
364
+ %350 = fsub float %186, %344, !dbg !39
365
+ %351 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 8.000000e+00) #6, !dbg !43
366
+ %352 = fmul float %351, %350, !dbg !44
367
+ %353 = fadd float %344, %352, !dbg !45
368
+ %354 = fadd float %234, %349, !dbg !46
369
+ %355 = fmul float %350, %350, !dbg !47
370
+ %356 = fmul float %355, 7.000000e+00, !dbg !50
371
+ %357 = fmul float %351, %356, !dbg !48
372
+ %358 = fadd float %354, %357, !dbg !49
373
+ %359 = bitcast float %291 to i32, !dbg !51
374
+ %360 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %359, i32 16, i32 31), !dbg !51
375
+ %361 = bitcast i32 %360 to float, !dbg !51
376
+ %362 = bitcast float %296 to i32, !dbg !51
377
+ %363 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %362, i32 16, i32 31), !dbg !51
378
+ %364 = bitcast i32 %363 to float, !dbg !51
379
+ %365 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 1090519040, i32 16, i32 31), !dbg !51
380
+ %366 = bitcast i32 %365 to float, !dbg !51
381
+ %367 = fsub float %361, %291, !dbg !39
382
+ %368 = fadd float %366, 8.000000e+00, !dbg !53
383
+ %369 = fcmp oeq float %368, 0.000000e+00, !dbg !54
384
+ %370 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %366, float %368) #6, !dbg !43
385
+ %371 = select i1 %369, float 0.000000e+00, float %370, !dbg !55
386
+ %372 = fmul float %371, %367, !dbg !44
387
+ %373 = fadd float %291, %372, !dbg !45
388
+ %374 = fadd float %296, %364, !dbg !46
389
+ %375 = fmul float %367, %367, !dbg !47
390
+ %376 = fmul float %375, 8.000000e+00, !dbg !50
391
+ %377 = fmul float %371, %376, !dbg !48
392
+ %378 = fadd float %374, %377, !dbg !49
393
+ %379 = bitcast float %373 to i32, !dbg !51
394
+ %380 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %379, i32 8, i32 31), !dbg !51
395
+ %381 = bitcast i32 %380 to float, !dbg !51
396
+ %382 = bitcast float %378 to i32, !dbg !51
397
+ %383 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %382, i32 8, i32 31), !dbg !51
398
+ %384 = bitcast i32 %383 to float, !dbg !51
399
+ %385 = bitcast float %368 to i32, !dbg !51
400
+ %386 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %385, i32 8, i32 31), !dbg !51
401
+ %387 = bitcast i32 %386 to float, !dbg !51
402
+ %388 = fsub float %381, %373, !dbg !39
403
+ %389 = fadd float %368, %387, !dbg !53
404
+ %390 = fcmp oeq float %389, 0.000000e+00, !dbg !54
405
+ %391 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %387, float %389) #6, !dbg !43
406
+ %392 = select i1 %390, float 0.000000e+00, float %391, !dbg !55
407
+ %393 = fmul float %392, %388, !dbg !44
408
+ %394 = fadd float %373, %393, !dbg !45
409
+ %395 = fadd float %378, %384, !dbg !46
410
+ %396 = fmul float %388, %388, !dbg !47
411
+ %397 = fmul float %368, %396, !dbg !50
412
+ %398 = fmul float %392, %397, !dbg !48
413
+ %399 = fadd float %395, %398, !dbg !49
414
+ %400 = bitcast float %394 to i32, !dbg !51
415
+ %401 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %400, i32 4, i32 31), !dbg !51
416
+ %402 = bitcast i32 %401 to float, !dbg !51
417
+ %403 = bitcast float %399 to i32, !dbg !51
418
+ %404 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %403, i32 4, i32 31), !dbg !51
419
+ %405 = bitcast i32 %404 to float, !dbg !51
420
+ %406 = bitcast float %389 to i32, !dbg !51
421
+ %407 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %406, i32 4, i32 31), !dbg !51
422
+ %408 = bitcast i32 %407 to float, !dbg !51
423
+ %409 = fsub float %402, %394, !dbg !39
424
+ %410 = fadd float %389, %408, !dbg !53
425
+ %411 = fcmp oeq float %410, 0.000000e+00, !dbg !54
426
+ %412 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %408, float %410) #6, !dbg !43
427
+ %413 = select i1 %411, float 0.000000e+00, float %412, !dbg !55
428
+ %414 = fmul float %409, %413, !dbg !44
429
+ %415 = fadd float %394, %414, !dbg !45
430
+ %416 = fadd float %399, %405, !dbg !46
431
+ %417 = fmul float %409, %409, !dbg !47
432
+ %418 = fmul float %389, %417, !dbg !50
433
+ %419 = fmul float %413, %418, !dbg !48
434
+ %420 = fadd float %416, %419, !dbg !49
435
+ %421 = bitcast float %415 to i32, !dbg !51
436
+ %422 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %421, i32 2, i32 31), !dbg !51
437
+ %423 = bitcast i32 %422 to float, !dbg !51
438
+ %424 = bitcast float %420 to i32, !dbg !51
439
+ %425 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %424, i32 2, i32 31), !dbg !51
440
+ %426 = bitcast i32 %425 to float, !dbg !51
441
+ %427 = bitcast float %410 to i32, !dbg !51
442
+ %428 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %427, i32 2, i32 31), !dbg !51
443
+ %429 = bitcast i32 %428 to float, !dbg !51
444
+ %430 = fsub float %423, %415, !dbg !39
445
+ %431 = fadd float %410, %429, !dbg !53
446
+ %432 = fcmp oeq float %431, 0.000000e+00, !dbg !54
447
+ %433 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %429, float %431) #6, !dbg !43
448
+ %434 = select i1 %432, float 0.000000e+00, float %433, !dbg !55
449
+ %435 = fmul float %430, %434, !dbg !44
450
+ %436 = fadd float %415, %435, !dbg !45
451
+ %437 = fadd float %420, %426, !dbg !46
452
+ %438 = fmul float %430, %430, !dbg !47
453
+ %439 = fmul float %410, %438, !dbg !50
454
+ %440 = fmul float %434, %439, !dbg !48
455
+ %441 = fadd float %437, %440, !dbg !49
456
+ %442 = bitcast float %436 to i32, !dbg !51
457
+ %443 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %442, i32 1, i32 31), !dbg !51
458
+ %444 = bitcast float %441 to i32, !dbg !51
459
+ %445 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %444, i32 1, i32 31), !dbg !51
460
+ %446 = bitcast float %431 to i32, !dbg !51
461
+ %447 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %446, i32 1, i32 31), !dbg !51
462
+ %448 = bitcast i32 %447 to float, !dbg !51
463
+ %449 = fadd float %431, %448, !dbg !53
464
+ %450 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %448, float %449) #6, !dbg !43
465
+ %451 = bitcast float %353 to i32, !dbg !51
466
+ %452 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %451, i32 16, i32 31), !dbg !51
467
+ %453 = bitcast i32 %452 to float, !dbg !51
468
+ %454 = bitcast float %358 to i32, !dbg !51
469
+ %455 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %454, i32 16, i32 31), !dbg !51
470
+ %456 = bitcast i32 %455 to float, !dbg !51
471
+ %457 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 1090519040, i32 16, i32 31), !dbg !51
472
+ %458 = bitcast i32 %457 to float, !dbg !51
473
+ %459 = fsub float %453, %353, !dbg !39
474
+ %460 = fadd float %458, 8.000000e+00, !dbg !53
475
+ %461 = fcmp oeq float %460, 0.000000e+00, !dbg !54
476
+ %462 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %458, float %460) #6, !dbg !43
477
+ %463 = select i1 %461, float 0.000000e+00, float %462, !dbg !55
478
+ %464 = fmul float %459, %463, !dbg !44
479
+ %465 = fadd float %353, %464, !dbg !45
480
+ %466 = fadd float %358, %456, !dbg !46
481
+ %467 = fmul float %459, %459, !dbg !47
482
+ %468 = fmul float %467, 8.000000e+00, !dbg !50
483
+ %469 = fmul float %468, %463, !dbg !48
484
+ %470 = fadd float %466, %469, !dbg !49
485
+ %471 = bitcast float %465 to i32, !dbg !51
486
+ %472 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %471, i32 8, i32 31), !dbg !51
487
+ %473 = bitcast i32 %472 to float, !dbg !51
488
+ %474 = bitcast float %470 to i32, !dbg !51
489
+ %475 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %474, i32 8, i32 31), !dbg !51
490
+ %476 = bitcast i32 %475 to float, !dbg !51
491
+ %477 = bitcast float %460 to i32, !dbg !51
492
+ %478 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %477, i32 8, i32 31), !dbg !51
493
+ %479 = bitcast i32 %478 to float, !dbg !51
494
+ %480 = fsub float %473, %465, !dbg !39
495
+ %481 = fadd float %460, %479, !dbg !53
496
+ %482 = fcmp oeq float %481, 0.000000e+00, !dbg !54
497
+ %483 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %479, float %481) #6, !dbg !43
498
+ %484 = select i1 %482, float 0.000000e+00, float %483, !dbg !55
499
+ %485 = fmul float %480, %484, !dbg !44
500
+ %486 = fadd float %465, %485, !dbg !45
501
+ %487 = fadd float %470, %476, !dbg !46
502
+ %488 = fmul float %480, %480, !dbg !47
503
+ %489 = fmul float %460, %488, !dbg !50
504
+ %490 = fmul float %484, %489, !dbg !48
505
+ %491 = fadd float %487, %490, !dbg !49
506
+ %492 = bitcast float %486 to i32, !dbg !51
507
+ %493 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %492, i32 4, i32 31), !dbg !51
508
+ %494 = bitcast i32 %493 to float, !dbg !51
509
+ %495 = bitcast float %491 to i32, !dbg !51
510
+ %496 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %495, i32 4, i32 31), !dbg !51
511
+ %497 = bitcast i32 %496 to float, !dbg !51
512
+ %498 = bitcast float %481 to i32, !dbg !51
513
+ %499 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %498, i32 4, i32 31), !dbg !51
514
+ %500 = bitcast i32 %499 to float, !dbg !51
515
+ %501 = fsub float %494, %486, !dbg !39
516
+ %502 = fadd float %481, %500, !dbg !53
517
+ %503 = fcmp oeq float %502, 0.000000e+00, !dbg !54
518
+ %504 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %500, float %502) #6, !dbg !43
519
+ %505 = select i1 %503, float 0.000000e+00, float %504, !dbg !55
520
+ %506 = fmul float %501, %505, !dbg !44
521
+ %507 = fadd float %486, %506, !dbg !45
522
+ %508 = fadd float %491, %497, !dbg !46
523
+ %509 = fmul float %501, %501, !dbg !47
524
+ %510 = fmul float %481, %509, !dbg !50
525
+ %511 = fmul float %505, %510, !dbg !48
526
+ %512 = fadd float %508, %511, !dbg !49
527
+ %513 = bitcast float %507 to i32, !dbg !51
528
+ %514 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %513, i32 2, i32 31), !dbg !51
529
+ %515 = bitcast i32 %514 to float, !dbg !51
530
+ %516 = bitcast float %512 to i32, !dbg !51
531
+ %517 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %516, i32 2, i32 31), !dbg !51
532
+ %518 = bitcast i32 %517 to float, !dbg !51
533
+ %519 = bitcast float %502 to i32, !dbg !51
534
+ %520 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %519, i32 2, i32 31), !dbg !51
535
+ %521 = bitcast i32 %520 to float, !dbg !51
536
+ %522 = fsub float %515, %507, !dbg !39
537
+ %523 = fadd float %502, %521, !dbg !53
538
+ %524 = fcmp oeq float %523, 0.000000e+00, !dbg !54
539
+ %525 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %521, float %523) #6, !dbg !43
540
+ %526 = select i1 %524, float 0.000000e+00, float %525, !dbg !55
541
+ %527 = fmul float %522, %526, !dbg !44
542
+ %528 = fadd float %507, %527, !dbg !45
543
+ %529 = fadd float %512, %518, !dbg !46
544
+ %530 = fmul float %522, %522, !dbg !47
545
+ %531 = fmul float %502, %530, !dbg !50
546
+ %532 = fmul float %526, %531, !dbg !48
547
+ %533 = fadd float %529, %532, !dbg !49
548
+ %534 = bitcast float %528 to i32, !dbg !51
549
+ %535 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %534, i32 1, i32 31), !dbg !51
550
+ %536 = bitcast float %533 to i32, !dbg !51
551
+ %537 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %536, i32 1, i32 31), !dbg !51
552
+ %538 = bitcast float %523 to i32, !dbg !51
553
+ %539 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %538, i32 1, i32 31), !dbg !51
554
+ %540 = bitcast i32 %539 to float, !dbg !51
555
+ %541 = fadd float %523, %540, !dbg !53
556
+ %542 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %540, float %541) #6, !dbg !43
557
+ %543 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %52, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
558
+ %544 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %54, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
559
+ %545 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %56, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
560
+ %546 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %58, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !56
561
+ %547 = zext nneg i32 %urem to i64, !dbg !57
562
+ %548 = getelementptr float, ptr addrspace(1) %3, i64 %547, !dbg !57
563
+ %549 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];\0A\09@!$4 mov.u32 $0, $3;", "=r,l,b,r,b"(ptr addrspace(1) %548, i1 true, i32 0, i1 true) #6, !dbg !58
564
+ br i1 %100, label %550, label %551, !dbg !59
565
+
566
+ 550: ; preds = %102
567
+ tail call void @__assertfail(ptr nonnull @assertMessage_1, ptr nonnull @assertFile_1, i32 883, ptr nonnull @assertFunc_1, i64 1), !dbg !59
568
+ br label %551, !dbg !59
569
+
570
+ 551: ; preds = %550, %102
571
+ %552 = bitcast i32 %537 to float, !dbg !51
572
+ %553 = fadd float %533, %552, !dbg !46
573
+ %554 = bitcast i32 %535 to float, !dbg !51
574
+ %555 = fsub float %554, %528, !dbg !39
575
+ %556 = fmul float %555, %555, !dbg !47
576
+ %557 = fmul float %523, %556, !dbg !50
577
+ %558 = fcmp oeq float %541, 0.000000e+00, !dbg !54
578
+ %559 = select i1 %558, float 0.000000e+00, float %542, !dbg !55
579
+ %560 = fmul float %559, %557, !dbg !48
580
+ %561 = fadd float %553, %560, !dbg !49
581
+ %562 = bitcast i32 %445 to float, !dbg !51
582
+ %563 = fadd float %441, %562, !dbg !46
583
+ %564 = bitcast i32 %443 to float, !dbg !51
584
+ %565 = fsub float %564, %436, !dbg !39
585
+ %566 = fmul float %565, %565, !dbg !47
586
+ %567 = fmul float %431, %566, !dbg !50
587
+ %568 = fcmp oeq float %449, 0.000000e+00, !dbg !54
588
+ %569 = select i1 %568, float 0.000000e+00, float %450, !dbg !55
589
+ %570 = fmul float %569, %567, !dbg !48
590
+ %571 = fadd float %563, %570, !dbg !49
591
+ %572 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %115, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
592
+ %573 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %116, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
593
+ %574 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %117, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
594
+ %575 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %118, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !60
595
+ %576 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
596
+ %577 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
597
+ %578 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
598
+ %579 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
599
+ %580 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
600
+ %581 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
601
+ %582 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
602
+ %583 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %571, float 2.560000e+02) #6, !dbg !61
603
+ %584 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
604
+ %585 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
605
+ %586 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
606
+ %587 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
607
+ %588 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
608
+ %589 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
609
+ %590 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
610
+ %591 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %561, float 2.560000e+02) #6, !dbg !61
611
+ %592 = fadd float %576, 0x3EE4F8B580000000, !dbg !62
612
+ %593 = fadd float %584, 0x3EE4F8B580000000, !dbg !62
613
+ %594 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
614
+ %.not.i = icmp eq i32 %594, 0, !dbg !63
615
+ br i1 %.not.i, label %597, label %595, !dbg !63
616
+
617
+ 595: ; preds = %551
618
+ %596 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %592), !dbg !63
619
+ br label %__nv_rsqrtf.exit, !dbg !63
620
+
621
+ 597: ; preds = %551
622
+ %598 = tail call float @llvm.nvvm.rsqrt.approx.f(float %592), !dbg !63
623
+ br label %__nv_rsqrtf.exit, !dbg !63
624
+
625
+ __nv_rsqrtf.exit: ; preds = %595, %597
626
+ %.0.i = phi float [ %596, %595 ], [ %598, %597 ], !dbg !63
627
+ %599 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
628
+ %600 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
629
+ %601 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
630
+ %602 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
631
+ %603 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
632
+ %604 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
633
+ %605 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
634
+ %606 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
635
+ %.not.i22 = icmp eq i32 %606, 0, !dbg !63
636
+ br i1 %.not.i22, label %609, label %607, !dbg !63
637
+
638
+ 607: ; preds = %__nv_rsqrtf.exit
639
+ %608 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %593), !dbg !63
640
+ br label %__nv_rsqrtf.exit24, !dbg !63
641
+
642
+ 609: ; preds = %__nv_rsqrtf.exit
643
+ %610 = tail call float @llvm.nvvm.rsqrt.approx.f(float %593), !dbg !63
644
+ br label %__nv_rsqrtf.exit24, !dbg !63
645
+
646
+ __nv_rsqrtf.exit24: ; preds = %607, %609
647
+ %.0.i23 = phi float [ %608, %607 ], [ %610, %609 ], !dbg !63
648
+ %611 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
649
+ %612 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
650
+ %613 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
651
+ %614 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
652
+ %615 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
653
+ %616 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
654
+ %617 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !63
655
+ %618 = extractvalue { i32, i32, i32, i32 } %575, 3, !dbg !60
656
+ %619 = bitcast i32 %618 to float, !dbg !60
657
+ %620 = extractvalue { i32, i32, i32, i32 } %546, 3, !dbg !56
658
+ %621 = bitcast i32 %620 to float, !dbg !56
659
+ %622 = fadd float %621, %619, !dbg !64
660
+ %623 = fmul float %555, %559, !dbg !44
661
+ %624 = fadd float %528, %623, !dbg !45
662
+ %625 = fsub float %622, %624, !dbg !65
663
+ %626 = extractvalue { i32, i32, i32, i32 } %575, 2, !dbg !60
664
+ %627 = bitcast i32 %626 to float, !dbg !60
665
+ %628 = extractvalue { i32, i32, i32, i32 } %546, 2, !dbg !56
666
+ %629 = bitcast i32 %628 to float, !dbg !56
667
+ %630 = fadd float %629, %627, !dbg !64
668
+ %631 = fsub float %630, %624, !dbg !65
669
+ %632 = extractvalue { i32, i32, i32, i32 } %575, 1, !dbg !60
670
+ %633 = bitcast i32 %632 to float, !dbg !60
671
+ %634 = extractvalue { i32, i32, i32, i32 } %546, 1, !dbg !56
672
+ %635 = bitcast i32 %634 to float, !dbg !56
673
+ %636 = fadd float %635, %633, !dbg !64
674
+ %637 = fsub float %636, %624, !dbg !65
675
+ %638 = extractvalue { i32, i32, i32, i32 } %575, 0, !dbg !60
676
+ %639 = bitcast i32 %638 to float, !dbg !60
677
+ %640 = extractvalue { i32, i32, i32, i32 } %546, 0, !dbg !56
678
+ %641 = bitcast i32 %640 to float, !dbg !56
679
+ %642 = fadd float %641, %639, !dbg !64
680
+ %643 = fsub float %642, %624, !dbg !65
681
+ %644 = extractvalue { i32, i32, i32, i32 } %574, 3, !dbg !60
682
+ %645 = bitcast i32 %644 to float, !dbg !60
683
+ %646 = extractvalue { i32, i32, i32, i32 } %545, 3, !dbg !56
684
+ %647 = bitcast i32 %646 to float, !dbg !56
685
+ %648 = fadd float %647, %645, !dbg !64
686
+ %649 = fsub float %648, %624, !dbg !65
687
+ %650 = extractvalue { i32, i32, i32, i32 } %574, 2, !dbg !60
688
+ %651 = bitcast i32 %650 to float, !dbg !60
689
+ %652 = extractvalue { i32, i32, i32, i32 } %545, 2, !dbg !56
690
+ %653 = bitcast i32 %652 to float, !dbg !56
691
+ %654 = fadd float %653, %651, !dbg !64
692
+ %655 = fsub float %654, %624, !dbg !65
693
+ %656 = extractvalue { i32, i32, i32, i32 } %574, 1, !dbg !60
694
+ %657 = bitcast i32 %656 to float, !dbg !60
695
+ %658 = extractvalue { i32, i32, i32, i32 } %545, 1, !dbg !56
696
+ %659 = bitcast i32 %658 to float, !dbg !56
697
+ %660 = fadd float %659, %657, !dbg !64
698
+ %661 = fsub float %660, %624, !dbg !65
699
+ %662 = extractvalue { i32, i32, i32, i32 } %574, 0, !dbg !60
700
+ %663 = bitcast i32 %662 to float, !dbg !60
701
+ %664 = extractvalue { i32, i32, i32, i32 } %545, 0, !dbg !56
702
+ %665 = bitcast i32 %664 to float, !dbg !56
703
+ %666 = fadd float %665, %663, !dbg !64
704
+ %667 = fsub float %666, %624, !dbg !65
705
+ %668 = extractvalue { i32, i32, i32, i32 } %573, 3, !dbg !60
706
+ %669 = bitcast i32 %668 to float, !dbg !60
707
+ %670 = extractvalue { i32, i32, i32, i32 } %544, 3, !dbg !56
708
+ %671 = bitcast i32 %670 to float, !dbg !56
709
+ %672 = fadd float %671, %669, !dbg !64
710
+ %673 = fmul float %565, %569, !dbg !44
711
+ %674 = fadd float %436, %673, !dbg !45
712
+ %675 = fsub float %672, %674, !dbg !65
713
+ %676 = extractvalue { i32, i32, i32, i32 } %573, 2, !dbg !60
714
+ %677 = bitcast i32 %676 to float, !dbg !60
715
+ %678 = extractvalue { i32, i32, i32, i32 } %544, 2, !dbg !56
716
+ %679 = bitcast i32 %678 to float, !dbg !56
717
+ %680 = fadd float %679, %677, !dbg !64
718
+ %681 = fsub float %680, %674, !dbg !65
719
+ %682 = extractvalue { i32, i32, i32, i32 } %573, 1, !dbg !60
720
+ %683 = bitcast i32 %682 to float, !dbg !60
721
+ %684 = extractvalue { i32, i32, i32, i32 } %544, 1, !dbg !56
722
+ %685 = bitcast i32 %684 to float, !dbg !56
723
+ %686 = fadd float %685, %683, !dbg !64
724
+ %687 = fsub float %686, %674, !dbg !65
725
+ %688 = extractvalue { i32, i32, i32, i32 } %573, 0, !dbg !60
726
+ %689 = bitcast i32 %688 to float, !dbg !60
727
+ %690 = extractvalue { i32, i32, i32, i32 } %544, 0, !dbg !56
728
+ %691 = bitcast i32 %690 to float, !dbg !56
729
+ %692 = fadd float %691, %689, !dbg !64
730
+ %693 = fsub float %692, %674, !dbg !65
731
+ %694 = extractvalue { i32, i32, i32, i32 } %572, 3, !dbg !60
732
+ %695 = bitcast i32 %694 to float, !dbg !60
733
+ %696 = extractvalue { i32, i32, i32, i32 } %543, 3, !dbg !56
734
+ %697 = bitcast i32 %696 to float, !dbg !56
735
+ %698 = fadd float %697, %695, !dbg !64
736
+ %699 = fsub float %698, %674, !dbg !65
737
+ %700 = extractvalue { i32, i32, i32, i32 } %572, 2, !dbg !60
738
+ %701 = bitcast i32 %700 to float, !dbg !60
739
+ %702 = extractvalue { i32, i32, i32, i32 } %543, 2, !dbg !56
740
+ %703 = bitcast i32 %702 to float, !dbg !56
741
+ %704 = fadd float %703, %701, !dbg !64
742
+ %705 = fsub float %704, %674, !dbg !65
743
+ %706 = extractvalue { i32, i32, i32, i32 } %572, 1, !dbg !60
744
+ %707 = bitcast i32 %706 to float, !dbg !60
745
+ %708 = extractvalue { i32, i32, i32, i32 } %543, 1, !dbg !56
746
+ %709 = bitcast i32 %708 to float, !dbg !56
747
+ %710 = fadd float %709, %707, !dbg !64
748
+ %711 = fsub float %710, %674, !dbg !65
749
+ %712 = extractvalue { i32, i32, i32, i32 } %572, 0, !dbg !60
750
+ %713 = bitcast i32 %712 to float, !dbg !60
751
+ %714 = extractvalue { i32, i32, i32, i32 } %543, 0, !dbg !56
752
+ %715 = bitcast i32 %714 to float, !dbg !56
753
+ %716 = fadd float %715, %713, !dbg !64
754
+ %717 = fsub float %716, %674, !dbg !65
755
+ %718 = fmul float %717, %.0.i, !dbg !66
756
+ %719 = fmul float %711, %.0.i, !dbg !66
757
+ %720 = fmul float %705, %.0.i, !dbg !66
758
+ %721 = fmul float %699, %.0.i, !dbg !66
759
+ %722 = fmul float %693, %.0.i, !dbg !66
760
+ %723 = fmul float %687, %.0.i, !dbg !66
761
+ %724 = fmul float %681, %.0.i, !dbg !66
762
+ %725 = fmul float %675, %.0.i, !dbg !66
763
+ %726 = fmul float %667, %.0.i23, !dbg !66
764
+ %727 = fmul float %661, %.0.i23, !dbg !66
765
+ %728 = fmul float %655, %.0.i23, !dbg !66
766
+ %729 = fmul float %649, %.0.i23, !dbg !66
767
+ %730 = fmul float %643, %.0.i23, !dbg !66
768
+ %731 = fmul float %637, %.0.i23, !dbg !66
769
+ %732 = fmul float %631, %.0.i23, !dbg !66
770
+ %733 = fmul float %625, %.0.i23, !dbg !66
771
+ %734 = getelementptr float, ptr addrspace(3) @global_smem, i64 %547, !dbg !67
772
+ store i32 %549, ptr addrspace(3) %734, align 4, !dbg !67
773
+ tail call void @llvm.nvvm.barrier0(), !dbg !67
774
+ %735 = getelementptr float, ptr addrspace(3) @global_smem, i64 %109, !dbg !67
775
+ %736 = load float, ptr addrspace(3) %735, align 32, !dbg !67
776
+ %737 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 1, !dbg !67
777
+ %738 = load float, ptr addrspace(3) %737, align 4, !dbg !67
778
+ %739 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 2, !dbg !67
779
+ %740 = load float, ptr addrspace(3) %739, align 8, !dbg !67
780
+ %741 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 3, !dbg !67
781
+ %742 = load float, ptr addrspace(3) %741, align 4, !dbg !67
782
+ %743 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 4, !dbg !67
783
+ %744 = load float, ptr addrspace(3) %743, align 16, !dbg !67
784
+ %745 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 5, !dbg !67
785
+ %746 = load float, ptr addrspace(3) %745, align 4, !dbg !67
786
+ %747 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 6, !dbg !67
787
+ %748 = load float, ptr addrspace(3) %747, align 8, !dbg !67
788
+ %749 = getelementptr inbounds <8 x float>, ptr addrspace(3) %735, i64 0, i64 7, !dbg !67
789
+ %750 = load float, ptr addrspace(3) %749, align 4, !dbg !67
790
+ %751 = fmul float %718, %736, !dbg !67
791
+ %752 = fmul float %719, %738, !dbg !67
792
+ %753 = fmul float %720, %740, !dbg !67
793
+ %754 = fmul float %721, %742, !dbg !67
794
+ %755 = fmul float %722, %744, !dbg !67
795
+ %756 = fmul float %723, %746, !dbg !67
796
+ %757 = fmul float %724, %748, !dbg !67
797
+ %758 = fmul float %725, %750, !dbg !67
798
+ %759 = fmul float %726, %736, !dbg !67
799
+ %760 = fmul float %727, %738, !dbg !67
800
+ %761 = fmul float %728, %740, !dbg !67
801
+ %762 = fmul float %729, %742, !dbg !67
802
+ %763 = fmul float %730, %744, !dbg !67
803
+ %764 = fmul float %731, %746, !dbg !67
804
+ %765 = fmul float %732, %748, !dbg !67
805
+ %766 = fmul float %733, %750, !dbg !67
806
+ %767 = shl i32 %17, 8, !dbg !68
807
+ %768 = shl i32 %18, 8, !dbg !68
808
+ %769 = or i32 %767, %13, !dbg !69
809
+ %770 = or i32 %768, %13, !dbg !69
810
+ %771 = sext i32 %769 to i64, !dbg !70
811
+ %772 = getelementptr i16, ptr addrspace(1) %4, i64 %771, !dbg !70
812
+ %773 = sext i32 %770 to i64, !dbg !70
813
+ %774 = getelementptr i16, ptr addrspace(1) %4, i64 %773, !dbg !70
814
+ %775 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %751) #6, !dbg !71
815
+ %776 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %752) #6, !dbg !71
816
+ %777 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %753) #6, !dbg !71
817
+ %778 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %754) #6, !dbg !71
818
+ %779 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %755) #6, !dbg !71
819
+ %780 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %756) #6, !dbg !71
820
+ %781 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %757) #6, !dbg !71
821
+ %782 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %758) #6, !dbg !71
822
+ %783 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %759) #6, !dbg !71
823
+ %784 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %760) #6, !dbg !71
824
+ %785 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %761) #6, !dbg !71
825
+ %786 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %762) #6, !dbg !71
826
+ %787 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %763) #6, !dbg !71
827
+ %788 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %764) #6, !dbg !71
828
+ %789 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %765) #6, !dbg !71
829
+ %790 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %766) #6, !dbg !71
830
+ %791 = insertelement <2 x i16> undef, i16 %775, i64 0, !dbg !71
831
+ %792 = insertelement <2 x i16> %791, i16 %776, i64 1, !dbg !71
832
+ %793 = bitcast <2 x i16> %792 to i32, !dbg !71
833
+ %794 = insertelement <2 x i16> undef, i16 %777, i64 0, !dbg !71
834
+ %795 = insertelement <2 x i16> %794, i16 %778, i64 1, !dbg !71
835
+ %796 = bitcast <2 x i16> %795 to i32, !dbg !71
836
+ %797 = insertelement <2 x i16> undef, i16 %779, i64 0, !dbg !71
837
+ %798 = insertelement <2 x i16> %797, i16 %780, i64 1, !dbg !71
838
+ %799 = bitcast <2 x i16> %798 to i32, !dbg !71
839
+ %800 = insertelement <2 x i16> undef, i16 %781, i64 0, !dbg !71
840
+ %801 = insertelement <2 x i16> %800, i16 %782, i64 1, !dbg !71
841
+ %802 = bitcast <2 x i16> %801 to i32, !dbg !71
842
+ tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %793, i32 %796, i32 %799, i32 %802, ptr addrspace(1) %772, i1 true) #6, !dbg !71
843
+ %803 = insertelement <2 x i16> undef, i16 %783, i64 0, !dbg !71
844
+ %804 = insertelement <2 x i16> %803, i16 %784, i64 1, !dbg !71
845
+ %805 = bitcast <2 x i16> %804 to i32, !dbg !71
846
+ %806 = insertelement <2 x i16> undef, i16 %785, i64 0, !dbg !71
847
+ %807 = insertelement <2 x i16> %806, i16 %786, i64 1, !dbg !71
848
+ %808 = bitcast <2 x i16> %807 to i32, !dbg !71
849
+ %809 = insertelement <2 x i16> undef, i16 %787, i64 0, !dbg !71
850
+ %810 = insertelement <2 x i16> %809, i16 %788, i64 1, !dbg !71
851
+ %811 = bitcast <2 x i16> %810 to i32, !dbg !71
852
+ %812 = insertelement <2 x i16> undef, i16 %789, i64 0, !dbg !71
853
+ %813 = insertelement <2 x i16> %812, i16 %790, i64 1, !dbg !71
854
+ %814 = bitcast <2 x i16> %813 to i32, !dbg !71
855
+ tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %805, i32 %808, i32 %811, i32 %814, ptr addrspace(1) %774, i1 true) #6, !dbg !71
856
+ ret void, !dbg !72
857
+ }
858
+
859
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
860
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
861
+
862
+ ; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
863
+ declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
864
+
865
+ ; Function Attrs: convergent nocallback nounwind
866
+ declare void @llvm.nvvm.barrier0() #2
867
+
868
+ ; Function Attrs: alwaysinline nounwind
869
+ define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
870
+ %1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
871
+ %.not = icmp eq i32 %1, 0
872
+ br i1 %.not, label %4, label %2
873
+
874
+ 2: ; preds = %0
875
+ %3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
876
+ br label %6
877
+
878
+ 4: ; preds = %0
879
+ %5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
880
+ br label %6
881
+
882
+ 6: ; preds = %4, %2
883
+ %.0 = phi float [ %3, %2 ], [ %5, %4 ]
884
+ ret float %.0
885
+ }
886
+
887
+ declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
888
+
889
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
890
+ declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
891
+
892
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
893
+ declare float @llvm.nvvm.rsqrt.approx.f(float) #5
894
+
895
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
896
+ attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
897
+ attributes #2 = { convergent nocallback nounwind }
898
+ attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
899
+ attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
900
+ attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
901
+ attributes #6 = { nounwind }
902
+
903
+ !llvm.module.flags = !{!0, !1}
904
+ !llvm.dbg.cu = !{!2}
905
+ !nvvm.annotations = !{!4, !5, !5, !4}
906
+ !llvm.ident = !{!6}
907
+
908
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
909
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
910
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
911
+ !3 = !DIFile(filename: "clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py", directory: "/tmp/torchinductor_root/lh")
912
+ !4 = !{ptr @triton__0d1d2d3d4d5de6de, !"kernel", i32 1}
913
+ !5 = !{ptr @triton__0d1d2d3d4d5de6de, !"maxntidx", i32 256}
914
+ !6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
915
+ !7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5de6de", linkageName: "triton__0d1d2d3d4d5de6de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
916
+ !8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
917
+ !9 = !{}
918
+ !10 = !DILocation(line: 22, column: 44, scope: !7)
919
+ !11 = !DILocation(line: 24, column: 33, scope: !7)
920
+ !12 = !DILocation(line: 21, column: 28, scope: !7)
921
+ !13 = !DILocation(line: 21, column: 33, scope: !7)
922
+ !14 = !DILocation(line: 22, column: 23, scope: !7)
923
+ !15 = !DILocation(line: 26, column: 30, scope: !7)
924
+ !16 = !DILocation(line: 26, column: 35, scope: !7)
925
+ !17 = !DILocation(line: 27, column: 18, scope: !7)
926
+ !18 = !DILocation(line: 35, column: 44, scope: !7)
927
+ !19 = !DILocation(line: 35, column: 40, scope: !7)
928
+ !20 = !DILocation(line: 35, column: 34, scope: !7)
929
+ !21 = !DILocation(line: 35, column: 50, scope: !7)
930
+ !22 = !DILocation(line: 36, column: 22, scope: !7)
931
+ !23 = !DILocation(line: 37, column: 22, scope: !7)
932
+ !24 = !DILocation(line: 38, column: 36, scope: !7)
933
+ !25 = !DILocation(line: 39, column: 40, scope: !7)
934
+ !26 = !DILocation(line: 39, column: 55, scope: !7)
935
+ !27 = !DILocation(line: 40, column: 44, scope: !7)
936
+ !28 = !DILocation(line: 40, column: 40, scope: !7)
937
+ !29 = !DILocation(line: 40, column: 34, scope: !7)
938
+ !30 = !DILocation(line: 40, column: 52, scope: !7)
939
+ !31 = !DILocation(line: 41, column: 22, scope: !7)
940
+ !32 = !DILocation(line: 98, column: 22, scope: !33, inlinedAt: !35)
941
+ !33 = distinct !DILexicalBlockFile(scope: !7, file: !34, discriminator: 0)
942
+ !34 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
943
+ !35 = !DILocation(line: 44, column: 38, scope: !33)
944
+ !36 = !DILocation(line: 101, column: 30, scope: !33, inlinedAt: !35)
945
+ !37 = !DILocation(line: 101, column: 22, scope: !33, inlinedAt: !35)
946
+ !38 = !DILocation(line: 101, column: 13, scope: !33, inlinedAt: !35)
947
+ !39 = !DILocation(line: 108, column: 21, scope: !40, inlinedAt: !41)
948
+ !40 = distinct !DILexicalBlockFile(scope: !33, file: !34, discriminator: 0)
949
+ !41 = !DILocation(line: 120, column: 46, scope: !40, inlinedAt: !42)
950
+ !42 = !DILocation(line: 50, column: 41, scope: !40)
951
+ !43 = !DILocation(line: 110, column: 60, scope: !40, inlinedAt: !41)
952
+ !44 = !DILocation(line: 112, column: 25, scope: !40, inlinedAt: !41)
953
+ !45 = !DILocation(line: 112, column: 17, scope: !40, inlinedAt: !41)
954
+ !46 = !DILocation(line: 113, column: 15, scope: !40, inlinedAt: !41)
955
+ !47 = !DILocation(line: 113, column: 30, scope: !40, inlinedAt: !41)
956
+ !48 = !DILocation(line: 113, column: 49, scope: !40, inlinedAt: !41)
957
+ !49 = !DILocation(line: 113, column: 22, scope: !40, inlinedAt: !41)
958
+ !50 = !DILocation(line: 113, column: 38, scope: !40, inlinedAt: !41)
959
+ !51 = !DILocation(line: 120, column: 46, scope: !33, inlinedAt: !52)
960
+ !52 = !DILocation(line: 50, column: 41, scope: !33)
961
+ !53 = !DILocation(line: 109, column: 28, scope: !40, inlinedAt: !41)
962
+ !54 = !DILocation(line: 110, column: 39, scope: !40, inlinedAt: !41)
963
+ !55 = !DILocation(line: 110, column: 49, scope: !40, inlinedAt: !41)
964
+ !56 = !DILocation(line: 59, column: 51, scope: !7)
965
+ !57 = !DILocation(line: 60, column: 35, scope: !7)
966
+ !58 = !DILocation(line: 60, column: 40, scope: !7)
967
+ !59 = !DILocation(line: 64, column: 57, scope: !7)
968
+ !60 = !DILocation(line: 65, column: 54, scope: !7)
969
+ !61 = !DILocation(line: 69, column: 23, scope: !7)
970
+ !62 = !DILocation(line: 71, column: 24, scope: !7)
971
+ !63 = !DILocation(line: 72, column: 30, scope: !7)
972
+ !64 = !DILocation(line: 66, column: 24, scope: !7)
973
+ !65 = !DILocation(line: 67, column: 24, scope: !7)
974
+ !66 = !DILocation(line: 73, column: 24, scope: !7)
975
+ !67 = !DILocation(line: 74, column: 24, scope: !7)
976
+ !68 = !DILocation(line: 76, column: 39, scope: !7)
977
+ !69 = !DILocation(line: 76, column: 35, scope: !7)
978
+ !70 = !DILocation(line: 76, column: 29, scope: !7)
979
+ !71 = !DILocation(line: 76, column: 52, scope: !7)
980
+ !72 = !DILocation(line: 55, column: 4, scope: !7)
.triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ptx ADDED
@@ -0,0 +1,1654 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.2
6
+ .target sm_89
7
+ .address_size 64
8
+
9
+ // .globl triton__0d1d2d3d4d5de6de
10
+ .extern .func __assertfail
11
+ (
12
+ .param .b64 __assertfail_param_0,
13
+ .param .b64 __assertfail_param_1,
14
+ .param .b32 __assertfail_param_2,
15
+ .param .b64 __assertfail_param_3,
16
+ .param .b64 __assertfail_param_4
17
+ )
18
+ ;
19
+ .global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
20
+ .global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
21
+ .global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 51, 32, 60, 32, 53, 48, 50, 53, 55};
22
+ .global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
23
+ .global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
24
+ .global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
25
+ .extern .shared .align 1 .b8 global_smem[];
26
+ .global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
27
+
28
+ .visible .entry triton__0d1d2d3d4d5de6de(
29
+ .param .u64 triton__0d1d2d3d4d5de6de_param_0,
30
+ .param .u64 triton__0d1d2d3d4d5de6de_param_1,
31
+ .param .u64 triton__0d1d2d3d4d5de6de_param_2,
32
+ .param .u64 triton__0d1d2d3d4d5de6de_param_3,
33
+ .param .u64 triton__0d1d2d3d4d5de6de_param_4,
34
+ .param .u32 triton__0d1d2d3d4d5de6de_param_5,
35
+ .param .u32 triton__0d1d2d3d4d5de6de_param_6
36
+ )
37
+ .maxntid 256, 1, 1
38
+ {
39
+ .reg .pred %p<117>;
40
+ .reg .b16 %rs<17>;
41
+ .reg .b32 %r<375>;
42
+ .reg .f32 %f<423>;
43
+ .reg .b64 %rd<113>;
44
+ .loc 1 18 0
45
+ $L__func_begin0:
46
+ .loc 1 18 0
47
+
48
+ ld.param.u64 %rd13, [triton__0d1d2d3d4d5de6de_param_3];
49
+ ld.param.u64 %rd12, [triton__0d1d2d3d4d5de6de_param_1];
50
+ ld.param.u64 %rd53, [triton__0d1d2d3d4d5de6de_param_0];
51
+ $L__tmp0:
52
+ .loc 1 22 44
53
+ mov.u32 %r59, %tid.x;
54
+ ld.param.u64 %rd54, [triton__0d1d2d3d4d5de6de_param_2];
55
+ bfe.u32 %r60, %r59, 5, 3;
56
+ and.b32 %r61, %r59, 15;
57
+ .loc 1 24 33
58
+ shl.b32 %r62, %r59, 3;
59
+ and.b32 %r1, %r62, 248;
60
+ and.b32 %r2, %r59, 255;
61
+ .loc 1 21 28
62
+ mov.u32 %r26, %ctaid.x;
63
+ .loc 1 21 33
64
+ shl.b32 %r63, %r26, 4;
65
+ .loc 1 22 23
66
+ or.b32 %r3, %r63, %r60;
67
+ or.b32 %r4, %r3, 8;
68
+ or.b32 %r64, %r63, %r61;
69
+ .loc 1 26 30
70
+ mul.wide.s32 %rd55, %r3, 8;
71
+ add.s64 %rd16, %rd53, %rd55;
72
+ add.s64 %rd32, %rd16, 64;
73
+ mul.wide.s32 %rd56, %r64, 8;
74
+ add.s64 %rd48, %rd53, %rd56;
75
+ mov.pred %p93, -1;
76
+ .loc 1 26 35
77
+ mov.u64 %rd15, 0x0;
78
+ @%p93 ld.global.L1::evict_last.b64 { %rd15 }, [ %rd16 + 0 ];
79
+ mov.u64 %rd17, 0x0;
80
+ @%p93 ld.global.L1::evict_last.b64 { %rd17 }, [ %rd16 + 0 ];
81
+ mov.u64 %rd19, 0x0;
82
+ @%p93 ld.global.L1::evict_last.b64 { %rd19 }, [ %rd16 + 0 ];
83
+ mov.u64 %rd21, 0x0;
84
+ @%p93 ld.global.L1::evict_last.b64 { %rd21 }, [ %rd16 + 0 ];
85
+ mov.u64 %rd23, 0x0;
86
+ @%p93 ld.global.L1::evict_last.b64 { %rd23 }, [ %rd16 + 0 ];
87
+ mov.u64 %rd25, 0x0;
88
+ @%p93 ld.global.L1::evict_last.b64 { %rd25 }, [ %rd16 + 0 ];
89
+ mov.u64 %rd27, 0x0;
90
+ @%p93 ld.global.L1::evict_last.b64 { %rd27 }, [ %rd16 + 0 ];
91
+ mov.u64 %rd29, 0x0;
92
+ @%p93 ld.global.L1::evict_last.b64 { %rd29 }, [ %rd16 + 0 ];
93
+ mov.u64 %rd31, 0x0;
94
+ @%p93 ld.global.L1::evict_last.b64 { %rd31 }, [ %rd32 + 0 ];
95
+ mov.u64 %rd33, 0x0;
96
+ @%p93 ld.global.L1::evict_last.b64 { %rd33 }, [ %rd32 + 0 ];
97
+ mov.u64 %rd35, 0x0;
98
+ @%p93 ld.global.L1::evict_last.b64 { %rd35 }, [ %rd32 + 0 ];
99
+ mov.u64 %rd37, 0x0;
100
+ @%p93 ld.global.L1::evict_last.b64 { %rd37 }, [ %rd32 + 0 ];
101
+ mov.u64 %rd39, 0x0;
102
+ @%p93 ld.global.L1::evict_last.b64 { %rd39 }, [ %rd32 + 0 ];
103
+ mov.u64 %rd41, 0x0;
104
+ @%p93 ld.global.L1::evict_last.b64 { %rd41 }, [ %rd32 + 0 ];
105
+ mov.u64 %rd43, 0x0;
106
+ @%p93 ld.global.L1::evict_last.b64 { %rd43 }, [ %rd32 + 0 ];
107
+ mov.u64 %rd45, 0x0;
108
+ @%p93 ld.global.L1::evict_last.b64 { %rd45 }, [ %rd32 + 0 ];
109
+ mov.u64 %rd47, 0x0;
110
+ @%p93 ld.global.L1::evict_last.b64 { %rd47 }, [ %rd48 + 0 ];
111
+ .loc 1 27 18
112
+ bfe.s32 %r65, %r26, 27, 1;
113
+ shr.u32 %r66, %r65, 23;
114
+ add.s32 %r67, %r3, %r66;
115
+ and.b32 %r68, %r67, 16776704;
116
+ sub.s32 %r69, %r3, %r68;
117
+ add.s32 %r70, %r4, %r66;
118
+ and.b32 %r71, %r70, 16776704;
119
+ sub.s32 %r72, %r4, %r71;
120
+ .loc 1 35 44
121
+ shl.b32 %r73, %r69, 8;
122
+ shl.b32 %r74, %r72, 8;
123
+ .loc 1 35 40
124
+ or.b32 %r75, %r73, %r1;
125
+ or.b32 %r76, %r74, %r1;
126
+ .loc 1 35 34
127
+ mul.wide.s32 %rd57, %r75, 4;
128
+ add.s64 %rd80, %rd54, %rd57;
129
+ cvt.s64.s32 %rd58, %r73;
130
+ cvt.u64.u32 %rd59, %r1;
131
+ or.b64 %rd60, %rd58, %rd59;
132
+ shl.b64 %rd61, %rd60, 2;
133
+ add.s64 %rd62, %rd54, %rd61;
134
+ add.s64 %rd81, %rd62, 16;
135
+ mul.wide.s32 %rd63, %r76, 4;
136
+ add.s64 %rd82, %rd54, %rd63;
137
+ cvt.s64.s32 %rd64, %r74;
138
+ or.b64 %rd65, %rd64, %rd59;
139
+ shl.b64 %rd66, %rd65, 2;
140
+ add.s64 %rd67, %rd54, %rd66;
141
+ add.s64 %rd83, %rd67, 16;
142
+ mov.b32 %r257, 0;
143
+ .loc 1 35 50
144
+ mov.u32 %r27, 0x0;
145
+ mov.u32 %r28, 0x0;
146
+ mov.u32 %r29, 0x0;
147
+ mov.u32 %r30, 0x0;
148
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r27, %r28, %r29, %r30 }, [ %rd80 + 0 ];
149
+ @!%p93 mov.u32 %r27, %r257;
150
+ @!%p93 mov.u32 %r28, %r257;
151
+ @!%p93 mov.u32 %r29, %r257;
152
+ @!%p93 mov.u32 %r30, %r257;
153
+ mov.b32 %f1, %r27;
154
+ mov.b32 %f2, %r28;
155
+ mov.b32 %f3, %r29;
156
+ mov.b32 %f4, %r30;
157
+ mov.u32 %r35, 0x0;
158
+ mov.u32 %r36, 0x0;
159
+ mov.u32 %r37, 0x0;
160
+ mov.u32 %r38, 0x0;
161
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r35, %r36, %r37, %r38 }, [ %rd81 + 0 ];
162
+ @!%p93 mov.u32 %r35, %r257;
163
+ @!%p93 mov.u32 %r36, %r257;
164
+ @!%p93 mov.u32 %r37, %r257;
165
+ @!%p93 mov.u32 %r38, %r257;
166
+ mov.b32 %f5, %r35;
167
+ mov.b32 %f6, %r36;
168
+ mov.b32 %f7, %r37;
169
+ mov.b32 %f8, %r38;
170
+ mov.u32 %r43, 0x0;
171
+ mov.u32 %r44, 0x0;
172
+ mov.u32 %r45, 0x0;
173
+ mov.u32 %r46, 0x0;
174
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r43, %r44, %r45, %r46 }, [ %rd82 + 0 ];
175
+ @!%p93 mov.u32 %r43, %r257;
176
+ @!%p93 mov.u32 %r44, %r257;
177
+ @!%p93 mov.u32 %r45, %r257;
178
+ @!%p93 mov.u32 %r46, %r257;
179
+ mov.b32 %f9, %r43;
180
+ mov.b32 %f10, %r44;
181
+ mov.b32 %f11, %r45;
182
+ mov.b32 %f12, %r46;
183
+ mov.u32 %r51, 0x0;
184
+ mov.u32 %r52, 0x0;
185
+ mov.u32 %r53, 0x0;
186
+ mov.u32 %r54, 0x0;
187
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r51, %r52, %r53, %r54 }, [ %rd83 + 0 ];
188
+ @!%p93 mov.u32 %r51, %r257;
189
+ @!%p93 mov.u32 %r52, %r257;
190
+ @!%p93 mov.u32 %r53, %r257;
191
+ @!%p93 mov.u32 %r54, %r257;
192
+ mov.b32 %f13, %r51;
193
+ mov.b32 %f14, %r52;
194
+ mov.b32 %f15, %r53;
195
+ mov.b32 %f16, %r54;
196
+ .loc 1 36 22
197
+ add.s64 %rd68, %rd47, 50257;
198
+ .loc 1 37 22
199
+ setp.lt.s64 %p38, %rd47, 0;
200
+ .loc 1 38 36
201
+ selp.b64 %rd7, %rd68, %rd47, %p38;
202
+ .loc 1 39 40
203
+ setp.lt.u64 %p39, %rd7, 50257;
204
+ mov.b32 %r374, 883;
205
+ mov.u64 %rd112, 1;
206
+ .loc 1 39 55
207
+ @%p39 bra $L__BB0_2;
208
+ mov.u64 %rd69, assertMessage_0;
209
+ cvta.global.u64 %rd70, %rd69;
210
+ mov.u64 %rd71, assertFile_0;
211
+ cvta.global.u64 %rd72, %rd71;
212
+ mov.u64 %rd73, assertFunc_0;
213
+ cvta.global.u64 %rd74, %rd73;
214
+ { // callseq 8, 0
215
+ .reg .b32 temp_param_reg;
216
+ .param .b64 param0;
217
+ st.param.b64 [param0+0], %rd70;
218
+ .param .b64 param1;
219
+ st.param.b64 [param1+0], %rd72;
220
+ .param .b32 param2;
221
+ st.param.b32 [param2+0], %r374;
222
+ .param .b64 param3;
223
+ st.param.b64 [param3+0], %rd74;
224
+ .param .b64 param4;
225
+ st.param.b64 [param4+0], %rd112;
226
+ call.uni
227
+ __assertfail,
228
+ (
229
+ param0,
230
+ param1,
231
+ param2,
232
+ param3,
233
+ param4
234
+ );
235
+ } // callseq 8
236
+ $L__BB0_2:
237
+ .loc 1 0 55
238
+ ld.param.u64 %rd14, [triton__0d1d2d3d4d5de6de_param_4];
239
+ .loc 1 37 22
240
+ setp.lt.s64 %p83, %rd31, 0;
241
+ setp.lt.s64 %p84, %rd15, 0;
242
+ .loc 1 40 44
243
+ shl.b64 %rd85, %rd15, 8;
244
+ add.s64 %rd86, %rd85, 12865792;
245
+ selp.b64 %rd87, %rd86, %rd85, %p84;
246
+ shl.b64 %rd88, %rd31, 8;
247
+ add.s64 %rd89, %rd88, 12865792;
248
+ selp.b64 %rd90, %rd89, %rd88, %p83;
249
+ .loc 1 40 40
250
+ or.b64 %rd92, %rd87, %rd59;
251
+ or.b64 %rd93, %rd90, %rd59;
252
+ .loc 1 40 34
253
+ shl.b64 %rd94, %rd92, 2;
254
+ add.s64 %rd104, %rd12, %rd94;
255
+ add.s64 %rd105, %rd104, 16;
256
+ shl.b64 %rd95, %rd93, 2;
257
+ add.s64 %rd106, %rd12, %rd95;
258
+ add.s64 %rd107, %rd106, 16;
259
+ .loc 1 40 52
260
+ mov.u32 %r78, 0x0;
261
+ mov.u32 %r79, 0x0;
262
+ mov.u32 %r80, 0x0;
263
+ mov.u32 %r81, 0x0;
264
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r78, %r79, %r80, %r81 }, [ %rd104 + 0 ];
265
+ @!%p93 mov.u32 %r78, %r257;
266
+ @!%p93 mov.u32 %r79, %r257;
267
+ @!%p93 mov.u32 %r80, %r257;
268
+ @!%p93 mov.u32 %r81, %r257;
269
+ mov.b32 %f27, %r78;
270
+ mov.b32 %f28, %r79;
271
+ mov.b32 %f29, %r80;
272
+ mov.b32 %f30, %r81;
273
+ mov.u32 %r86, 0x0;
274
+ mov.u32 %r87, 0x0;
275
+ mov.u32 %r88, 0x0;
276
+ mov.u32 %r89, 0x0;
277
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r86, %r87, %r88, %r89 }, [ %rd105 + 0 ];
278
+ @!%p93 mov.u32 %r86, %r257;
279
+ @!%p93 mov.u32 %r87, %r257;
280
+ @!%p93 mov.u32 %r88, %r257;
281
+ @!%p93 mov.u32 %r89, %r257;
282
+ mov.b32 %f31, %r86;
283
+ mov.b32 %f32, %r87;
284
+ mov.b32 %f33, %r88;
285
+ mov.b32 %f34, %r89;
286
+ mov.u32 %r94, 0x0;
287
+ mov.u32 %r95, 0x0;
288
+ mov.u32 %r96, 0x0;
289
+ mov.u32 %r97, 0x0;
290
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r94, %r95, %r96, %r97 }, [ %rd106 + 0 ];
291
+ @!%p93 mov.u32 %r94, %r257;
292
+ @!%p93 mov.u32 %r95, %r257;
293
+ @!%p93 mov.u32 %r96, %r257;
294
+ @!%p93 mov.u32 %r97, %r257;
295
+ mov.b32 %f35, %r94;
296
+ mov.b32 %f36, %r95;
297
+ mov.b32 %f37, %r96;
298
+ mov.b32 %f38, %r97;
299
+ mov.u32 %r102, 0x0;
300
+ mov.u32 %r103, 0x0;
301
+ mov.u32 %r104, 0x0;
302
+ mov.u32 %r105, 0x0;
303
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r102, %r103, %r104, %r105 }, [ %rd107 + 0 ];
304
+ @!%p93 mov.u32 %r102, %r257;
305
+ @!%p93 mov.u32 %r103, %r257;
306
+ @!%p93 mov.u32 %r104, %r257;
307
+ @!%p93 mov.u32 %r105, %r257;
308
+ mov.b32 %f39, %r102;
309
+ mov.b32 %f40, %r103;
310
+ mov.b32 %f41, %r104;
311
+ mov.b32 %f42, %r105;
312
+ .loc 1 41 22
313
+ add.f32 %f43, %f1, %f27;
314
+ add.f32 %f44, %f2, %f28;
315
+ add.f32 %f45, %f3, %f29;
316
+ add.f32 %f46, %f4, %f30;
317
+ add.f32 %f47, %f5, %f31;
318
+ add.f32 %f48, %f6, %f32;
319
+ add.f32 %f49, %f7, %f33;
320
+ add.f32 %f50, %f8, %f34;
321
+ add.f32 %f51, %f9, %f35;
322
+ add.f32 %f52, %f10, %f36;
323
+ add.f32 %f53, %f11, %f37;
324
+ add.f32 %f54, %f12, %f38;
325
+ add.f32 %f55, %f13, %f39;
326
+ add.f32 %f56, %f14, %f40;
327
+ add.f32 %f57, %f15, %f41;
328
+ add.f32 %f58, %f16, %f42;
329
+ $L__tmp1:
330
+ .loc 2 98 22
331
+ add.f32 %f59, %f43, 0f00000000;
332
+ add.f32 %f60, %f44, 0f00000000;
333
+ add.f32 %f61, %f45, 0f00000000;
334
+ add.f32 %f62, %f46, 0f00000000;
335
+ add.f32 %f63, %f47, 0f00000000;
336
+ add.f32 %f64, %f48, 0f00000000;
337
+ add.f32 %f65, %f49, 0f00000000;
338
+ add.f32 %f66, %f50, 0f00000000;
339
+ add.f32 %f67, %f51, 0f00000000;
340
+ add.f32 %f68, %f52, 0f00000000;
341
+ add.f32 %f69, %f53, 0f00000000;
342
+ add.f32 %f70, %f54, 0f00000000;
343
+ add.f32 %f71, %f55, 0f00000000;
344
+ add.f32 %f72, %f56, 0f00000000;
345
+ add.f32 %f73, %f57, 0f00000000;
346
+ add.f32 %f74, %f58, 0f00000000;
347
+ .loc 2 101 30
348
+ sub.f32 %f75, %f43, %f59;
349
+ sub.f32 %f76, %f44, %f60;
350
+ sub.f32 %f77, %f45, %f61;
351
+ sub.f32 %f78, %f46, %f62;
352
+ sub.f32 %f79, %f47, %f63;
353
+ sub.f32 %f80, %f48, %f64;
354
+ sub.f32 %f81, %f49, %f65;
355
+ sub.f32 %f82, %f50, %f66;
356
+ sub.f32 %f83, %f51, %f67;
357
+ sub.f32 %f84, %f52, %f68;
358
+ sub.f32 %f85, %f53, %f69;
359
+ sub.f32 %f86, %f54, %f70;
360
+ sub.f32 %f87, %f55, %f71;
361
+ sub.f32 %f88, %f56, %f72;
362
+ sub.f32 %f89, %f57, %f73;
363
+ sub.f32 %f90, %f58, %f74;
364
+ .loc 2 101 13
365
+ fma.rn.f32 %f91, %f43, %f75, 0f00000000;
366
+ fma.rn.f32 %f92, %f44, %f76, 0f00000000;
367
+ fma.rn.f32 %f93, %f45, %f77, 0f00000000;
368
+ fma.rn.f32 %f94, %f46, %f78, 0f00000000;
369
+ fma.rn.f32 %f95, %f47, %f79, 0f00000000;
370
+ fma.rn.f32 %f96, %f48, %f80, 0f00000000;
371
+ fma.rn.f32 %f97, %f49, %f81, 0f00000000;
372
+ fma.rn.f32 %f98, %f50, %f82, 0f00000000;
373
+ fma.rn.f32 %f99, %f51, %f83, 0f00000000;
374
+ fma.rn.f32 %f100, %f52, %f84, 0f00000000;
375
+ fma.rn.f32 %f101, %f53, %f85, 0f00000000;
376
+ fma.rn.f32 %f102, %f54, %f86, 0f00000000;
377
+ fma.rn.f32 %f103, %f55, %f87, 0f00000000;
378
+ fma.rn.f32 %f104, %f56, %f88, 0f00000000;
379
+ fma.rn.f32 %f105, %f57, %f89, 0f00000000;
380
+ fma.rn.f32 %f106, %f58, %f90, 0f00000000;
381
+ $L__tmp2:
382
+ .loc 2 108 21
383
+ sub.f32 %f107, %f60, %f59;
384
+ mov.b32 %r111, 1065353216;
385
+ mov.b32 %r112, 1073741824;
386
+ .loc 2 110 60
387
+ div.full.f32 %r110, %r111, %r112;
388
+ mov.b32 %f108, %r110;
389
+ .loc 2 112 17
390
+ fma.rn.f32 %f109, %f108, %f107, %f59;
391
+ .loc 2 113 15
392
+ add.f32 %f110, %f91, %f92;
393
+ .loc 2 113 30
394
+ mul.f32 %f111, %f107, %f107;
395
+ .loc 2 113 22
396
+ fma.rn.f32 %f112, %f108, %f111, %f110;
397
+ .loc 2 108 21
398
+ sub.f32 %f113, %f61, %f109;
399
+ mov.b32 %r115, 1077936128;
400
+ .loc 2 110 60
401
+ div.full.f32 %r113, %r111, %r115;
402
+ mov.b32 %f114, %r113;
403
+ .loc 2 112 17
404
+ fma.rn.f32 %f115, %f114, %f113, %f109;
405
+ .loc 2 113 15
406
+ add.f32 %f116, %f93, %f112;
407
+ .loc 2 113 30
408
+ mul.f32 %f117, %f113, %f113;
409
+ .loc 2 113 38
410
+ fma.rn.f32 %f118, %f113, %f113, %f117;
411
+ .loc 2 113 22
412
+ fma.rn.f32 %f119, %f114, %f118, %f116;
413
+ .loc 2 108 21
414
+ sub.f32 %f120, %f62, %f115;
415
+ mov.b32 %r118, 1082130432;
416
+ .loc 2 110 60
417
+ div.full.f32 %r116, %r111, %r118;
418
+ mov.b32 %f121, %r116;
419
+ .loc 2 112 17
420
+ fma.rn.f32 %f122, %f121, %f120, %f115;
421
+ .loc 2 113 15
422
+ add.f32 %f123, %f94, %f119;
423
+ .loc 2 113 30
424
+ mul.f32 %f124, %f120, %f120;
425
+ .loc 2 113 38
426
+ mul.f32 %f125, %f124, 0f40400000;
427
+ .loc 2 113 22
428
+ fma.rn.f32 %f126, %f121, %f125, %f123;
429
+ .loc 2 108 21
430
+ sub.f32 %f127, %f63, %f122;
431
+ mov.b32 %r121, 1084227584;
432
+ .loc 2 110 60
433
+ div.full.f32 %r119, %r111, %r121;
434
+ mov.b32 %f128, %r119;
435
+ .loc 2 112 17
436
+ fma.rn.f32 %f129, %f128, %f127, %f122;
437
+ .loc 2 113 15
438
+ add.f32 %f130, %f95, %f126;
439
+ .loc 2 113 30
440
+ mul.f32 %f131, %f127, %f127;
441
+ .loc 2 113 38
442
+ mul.f32 %f132, %f131, 0f40800000;
443
+ .loc 2 113 22
444
+ fma.rn.f32 %f133, %f128, %f132, %f130;
445
+ .loc 2 108 21
446
+ sub.f32 %f134, %f64, %f129;
447
+ mov.b32 %r124, 1086324736;
448
+ .loc 2 110 60
449
+ div.full.f32 %r122, %r111, %r124;
450
+ mov.b32 %f135, %r122;
451
+ .loc 2 112 17
452
+ fma.rn.f32 %f136, %f135, %f134, %f129;
453
+ .loc 2 113 15
454
+ add.f32 %f137, %f96, %f133;
455
+ .loc 2 113 30
456
+ mul.f32 %f138, %f134, %f134;
457
+ .loc 2 113 38
458
+ mul.f32 %f139, %f138, 0f40A00000;
459
+ .loc 2 113 22
460
+ fma.rn.f32 %f140, %f135, %f139, %f137;
461
+ .loc 2 108 21
462
+ sub.f32 %f141, %f65, %f136;
463
+ mov.b32 %r127, 1088421888;
464
+ .loc 2 110 60
465
+ div.full.f32 %r125, %r111, %r127;
466
+ mov.b32 %f142, %r125;
467
+ .loc 2 112 17
468
+ fma.rn.f32 %f143, %f142, %f141, %f136;
469
+ .loc 2 113 15
470
+ add.f32 %f144, %f97, %f140;
471
+ .loc 2 113 30
472
+ mul.f32 %f145, %f141, %f141;
473
+ .loc 2 113 38
474
+ mul.f32 %f146, %f145, 0f40C00000;
475
+ .loc 2 113 22
476
+ fma.rn.f32 %f147, %f142, %f146, %f144;
477
+ .loc 2 108 21
478
+ sub.f32 %f148, %f66, %f143;
479
+ mov.b32 %r130, 1090519040;
480
+ .loc 2 110 60
481
+ div.full.f32 %r128, %r111, %r130;
482
+ mov.b32 %f149, %r128;
483
+ .loc 2 112 17
484
+ fma.rn.f32 %f150, %f149, %f148, %f143;
485
+ .loc 2 113 15
486
+ add.f32 %f151, %f98, %f147;
487
+ .loc 2 113 30
488
+ mul.f32 %f152, %f148, %f148;
489
+ .loc 2 113 38
490
+ mul.f32 %f153, %f152, 0f40E00000;
491
+ .loc 2 113 22
492
+ fma.rn.f32 %f154, %f149, %f153, %f151;
493
+ .loc 2 108 21
494
+ sub.f32 %f155, %f68, %f67;
495
+ .loc 2 110 60
496
+ div.full.f32 %r131, %r111, %r112;
497
+ mov.b32 %f156, %r131;
498
+ .loc 2 112 17
499
+ fma.rn.f32 %f157, %f155, %f156, %f67;
500
+ .loc 2 113 15
501
+ add.f32 %f158, %f99, %f100;
502
+ .loc 2 113 30
503
+ mul.f32 %f159, %f155, %f155;
504
+ .loc 2 113 22
505
+ fma.rn.f32 %f160, %f159, %f156, %f158;
506
+ .loc 2 108 21
507
+ sub.f32 %f161, %f69, %f157;
508
+ .loc 2 110 60
509
+ div.full.f32 %r134, %r111, %r115;
510
+ mov.b32 %f162, %r134;
511
+ .loc 2 112 17
512
+ fma.rn.f32 %f163, %f162, %f161, %f157;
513
+ .loc 2 113 15
514
+ add.f32 %f164, %f101, %f160;
515
+ .loc 2 113 30
516
+ mul.f32 %f165, %f161, %f161;
517
+ .loc 2 113 38
518
+ fma.rn.f32 %f166, %f161, %f161, %f165;
519
+ .loc 2 113 22
520
+ fma.rn.f32 %f167, %f162, %f166, %f164;
521
+ .loc 2 108 21
522
+ sub.f32 %f168, %f70, %f163;
523
+ .loc 2 110 60
524
+ div.full.f32 %r137, %r111, %r118;
525
+ mov.b32 %f169, %r137;
526
+ .loc 2 112 17
527
+ fma.rn.f32 %f170, %f169, %f168, %f163;
528
+ .loc 2 113 15
529
+ add.f32 %f171, %f102, %f167;
530
+ .loc 2 113 30
531
+ mul.f32 %f172, %f168, %f168;
532
+ .loc 2 113 38
533
+ mul.f32 %f173, %f172, 0f40400000;
534
+ .loc 2 113 22
535
+ fma.rn.f32 %f174, %f169, %f173, %f171;
536
+ .loc 2 108 21
537
+ sub.f32 %f175, %f71, %f170;
538
+ .loc 2 110 60
539
+ div.full.f32 %r140, %r111, %r121;
540
+ mov.b32 %f176, %r140;
541
+ .loc 2 112 17
542
+ fma.rn.f32 %f177, %f176, %f175, %f170;
543
+ .loc 2 113 15
544
+ add.f32 %f178, %f103, %f174;
545
+ .loc 2 113 30
546
+ mul.f32 %f179, %f175, %f175;
547
+ .loc 2 113 38
548
+ mul.f32 %f180, %f179, 0f40800000;
549
+ .loc 2 113 22
550
+ fma.rn.f32 %f181, %f176, %f180, %f178;
551
+ .loc 2 108 21
552
+ sub.f32 %f182, %f72, %f177;
553
+ .loc 2 110 60
554
+ div.full.f32 %r143, %r111, %r124;
555
+ mov.b32 %f183, %r143;
556
+ .loc 2 112 17
557
+ fma.rn.f32 %f184, %f183, %f182, %f177;
558
+ .loc 2 113 15
559
+ add.f32 %f185, %f104, %f181;
560
+ .loc 2 113 30
561
+ mul.f32 %f186, %f182, %f182;
562
+ .loc 2 113 38
563
+ mul.f32 %f187, %f186, 0f40A00000;
564
+ .loc 2 113 22
565
+ fma.rn.f32 %f188, %f183, %f187, %f185;
566
+ .loc 2 108 21
567
+ sub.f32 %f189, %f73, %f184;
568
+ .loc 2 110 60
569
+ div.full.f32 %r146, %r111, %r127;
570
+ mov.b32 %f190, %r146;
571
+ .loc 2 112 17
572
+ fma.rn.f32 %f191, %f190, %f189, %f184;
573
+ .loc 2 113 15
574
+ add.f32 %f192, %f105, %f188;
575
+ .loc 2 113 30
576
+ mul.f32 %f193, %f189, %f189;
577
+ .loc 2 113 38
578
+ mul.f32 %f194, %f193, 0f40C00000;
579
+ .loc 2 113 22
580
+ fma.rn.f32 %f195, %f190, %f194, %f192;
581
+ .loc 2 108 21
582
+ sub.f32 %f196, %f74, %f191;
583
+ .loc 2 110 60
584
+ div.full.f32 %r149, %r111, %r130;
585
+ mov.b32 %f197, %r149;
586
+ .loc 2 112 17
587
+ fma.rn.f32 %f198, %f197, %f196, %f191;
588
+ .loc 2 113 15
589
+ add.f32 %f199, %f106, %f195;
590
+ .loc 2 113 30
591
+ mul.f32 %f200, %f196, %f196;
592
+ .loc 2 113 38
593
+ mul.f32 %f201, %f200, 0f40E00000;
594
+ .loc 2 113 22
595
+ fma.rn.f32 %f202, %f197, %f201, %f199;
596
+ $L__tmp3:
597
+ .loc 2 120 46
598
+ mov.b32 %r216, %f150;
599
+ shfl.sync.bfly.b32 %r217, %r216, 16, 31, -1;
600
+ mov.b32 %f203, %r217;
601
+ mov.b32 %r218, %f154;
602
+ shfl.sync.bfly.b32 %r219, %r218, 16, 31, -1;
603
+ mov.b32 %f204, %r219;
604
+ shfl.sync.bfly.b32 %r153, %r130, 16, 31, -1;
605
+ mov.b32 %f205, %r153;
606
+ $L__tmp4:
607
+ .loc 2 108 21
608
+ sub.f32 %f206, %f203, %f150;
609
+ .loc 2 109 28
610
+ add.f32 %f207, %f205, 0f41000000;
611
+ .loc 2 110 39
612
+ setp.eq.f32 %p85, %f207, 0f00000000;
613
+ .loc 2 110 60
614
+ mov.b32 %r154, %f207;
615
+ div.full.f32 %r152, %r153, %r154;
616
+ mov.b32 %f208, %r152;
617
+ .loc 2 110 49
618
+ selp.f32 %f209, 0f00000000, %f208, %p85;
619
+ .loc 2 112 17
620
+ fma.rn.f32 %f210, %f209, %f206, %f150;
621
+ .loc 2 113 15
622
+ add.f32 %f211, %f154, %f204;
623
+ .loc 2 113 30
624
+ mul.f32 %f212, %f206, %f206;
625
+ .loc 2 113 38
626
+ mul.f32 %f213, %f212, 0f41000000;
627
+ .loc 2 113 22
628
+ fma.rn.f32 %f214, %f209, %f213, %f211;
629
+ $L__tmp5:
630
+ .loc 2 120 46
631
+ mov.b32 %r220, %f210;
632
+ shfl.sync.bfly.b32 %r221, %r220, 8, 31, -1;
633
+ mov.b32 %f215, %r221;
634
+ mov.b32 %r222, %f214;
635
+ shfl.sync.bfly.b32 %r223, %r222, 8, 31, -1;
636
+ mov.b32 %f216, %r223;
637
+ shfl.sync.bfly.b32 %r156, %r154, 8, 31, -1;
638
+ mov.b32 %f217, %r156;
639
+ $L__tmp6:
640
+ .loc 2 108 21
641
+ sub.f32 %f218, %f215, %f210;
642
+ .loc 2 109 28
643
+ add.f32 %f219, %f207, %f217;
644
+ .loc 2 110 39
645
+ setp.eq.f32 %p86, %f219, 0f00000000;
646
+ .loc 2 110 60
647
+ mov.b32 %r157, %f219;
648
+ div.full.f32 %r155, %r156, %r157;
649
+ mov.b32 %f220, %r155;
650
+ .loc 2 110 49
651
+ selp.f32 %f221, 0f00000000, %f220, %p86;
652
+ .loc 2 112 17
653
+ fma.rn.f32 %f222, %f221, %f218, %f210;
654
+ .loc 2 113 15
655
+ add.f32 %f223, %f214, %f216;
656
+ .loc 2 113 30
657
+ mul.f32 %f224, %f218, %f218;
658
+ .loc 2 113 38
659
+ mul.f32 %f225, %f207, %f224;
660
+ .loc 2 113 22
661
+ fma.rn.f32 %f226, %f221, %f225, %f223;
662
+ $L__tmp7:
663
+ .loc 2 120 46
664
+ mov.b32 %r224, %f222;
665
+ shfl.sync.bfly.b32 %r225, %r224, 4, 31, -1;
666
+ mov.b32 %f227, %r225;
667
+ mov.b32 %r226, %f226;
668
+ shfl.sync.bfly.b32 %r227, %r226, 4, 31, -1;
669
+ mov.b32 %f228, %r227;
670
+ shfl.sync.bfly.b32 %r159, %r157, 4, 31, -1;
671
+ mov.b32 %f229, %r159;
672
+ $L__tmp8:
673
+ .loc 2 108 21
674
+ sub.f32 %f230, %f227, %f222;
675
+ .loc 2 109 28
676
+ add.f32 %f231, %f219, %f229;
677
+ .loc 2 110 39
678
+ setp.eq.f32 %p87, %f231, 0f00000000;
679
+ .loc 2 110 60
680
+ mov.b32 %r160, %f231;
681
+ div.full.f32 %r158, %r159, %r160;
682
+ mov.b32 %f232, %r158;
683
+ .loc 2 110 49
684
+ selp.f32 %f233, 0f00000000, %f232, %p87;
685
+ .loc 2 112 17
686
+ fma.rn.f32 %f234, %f230, %f233, %f222;
687
+ .loc 2 113 15
688
+ add.f32 %f235, %f226, %f228;
689
+ .loc 2 113 30
690
+ mul.f32 %f236, %f230, %f230;
691
+ .loc 2 113 38
692
+ mul.f32 %f237, %f219, %f236;
693
+ .loc 2 113 22
694
+ fma.rn.f32 %f238, %f233, %f237, %f235;
695
+ $L__tmp9:
696
+ .loc 2 120 46
697
+ mov.b32 %r228, %f234;
698
+ shfl.sync.bfly.b32 %r229, %r228, 2, 31, -1;
699
+ mov.b32 %f239, %r229;
700
+ mov.b32 %r230, %f238;
701
+ shfl.sync.bfly.b32 %r231, %r230, 2, 31, -1;
702
+ mov.b32 %f240, %r231;
703
+ shfl.sync.bfly.b32 %r162, %r160, 2, 31, -1;
704
+ mov.b32 %f241, %r162;
705
+ $L__tmp10:
706
+ .loc 2 108 21
707
+ sub.f32 %f242, %f239, %f234;
708
+ .loc 2 109 28
709
+ add.f32 %f17, %f231, %f241;
710
+ .loc 2 110 39
711
+ setp.eq.f32 %p88, %f17, 0f00000000;
712
+ .loc 2 110 60
713
+ mov.b32 %r163, %f17;
714
+ div.full.f32 %r161, %r162, %r163;
715
+ mov.b32 %f243, %r161;
716
+ .loc 2 110 49
717
+ selp.f32 %f244, 0f00000000, %f243, %p88;
718
+ .loc 2 112 17
719
+ fma.rn.f32 %f18, %f242, %f244, %f234;
720
+ .loc 2 113 15
721
+ add.f32 %f245, %f238, %f240;
722
+ .loc 2 113 30
723
+ mul.f32 %f246, %f242, %f242;
724
+ .loc 2 113 38
725
+ mul.f32 %f247, %f231, %f246;
726
+ .loc 2 113 22
727
+ fma.rn.f32 %f19, %f244, %f247, %f245;
728
+ $L__tmp11:
729
+ .loc 2 120 46
730
+ mov.b32 %r232, %f18;
731
+ shfl.sync.bfly.b32 %r5, %r232, 1, 31, -1;
732
+ mov.b32 %r233, %f19;
733
+ shfl.sync.bfly.b32 %r6, %r233, 1, 31, -1;
734
+ shfl.sync.bfly.b32 %r165, %r163, 1, 31, -1;
735
+ mov.b32 %f248, %r165;
736
+ $L__tmp12:
737
+ .loc 2 109 28
738
+ add.f32 %f20, %f17, %f248;
739
+ .loc 2 110 60
740
+ mov.b32 %r166, %f20;
741
+ div.full.f32 %r164, %r165, %r166;
742
+ mov.b32 %f21, %r164;
743
+ $L__tmp13:
744
+ .loc 2 120 46
745
+ mov.b32 %r234, %f198;
746
+ shfl.sync.bfly.b32 %r235, %r234, 16, 31, -1;
747
+ mov.b32 %f249, %r235;
748
+ mov.b32 %r236, %f202;
749
+ shfl.sync.bfly.b32 %r237, %r236, 16, 31, -1;
750
+ mov.b32 %f250, %r237;
751
+ shfl.sync.bfly.b32 %r168, %r130, 16, 31, -1;
752
+ mov.b32 %f251, %r168;
753
+ $L__tmp14:
754
+ .loc 2 108 21
755
+ sub.f32 %f252, %f249, %f198;
756
+ .loc 2 109 28
757
+ add.f32 %f253, %f251, 0f41000000;
758
+ .loc 2 110 39
759
+ setp.eq.f32 %p89, %f253, 0f00000000;
760
+ .loc 2 110 60
761
+ mov.b32 %r169, %f253;
762
+ div.full.f32 %r167, %r168, %r169;
763
+ mov.b32 %f254, %r167;
764
+ .loc 2 110 49
765
+ selp.f32 %f255, 0f00000000, %f254, %p89;
766
+ .loc 2 112 17
767
+ fma.rn.f32 %f256, %f252, %f255, %f198;
768
+ .loc 2 113 15
769
+ add.f32 %f257, %f202, %f250;
770
+ .loc 2 113 30
771
+ mul.f32 %f258, %f252, %f252;
772
+ .loc 2 113 38
773
+ mul.f32 %f259, %f258, 0f41000000;
774
+ .loc 2 113 22
775
+ fma.rn.f32 %f260, %f259, %f255, %f257;
776
+ $L__tmp15:
777
+ .loc 2 120 46
778
+ mov.b32 %r238, %f256;
779
+ shfl.sync.bfly.b32 %r239, %r238, 8, 31, -1;
780
+ mov.b32 %f261, %r239;
781
+ mov.b32 %r240, %f260;
782
+ shfl.sync.bfly.b32 %r241, %r240, 8, 31, -1;
783
+ mov.b32 %f262, %r241;
784
+ shfl.sync.bfly.b32 %r171, %r169, 8, 31, -1;
785
+ mov.b32 %f263, %r171;
786
+ $L__tmp16:
787
+ .loc 2 108 21
788
+ sub.f32 %f264, %f261, %f256;
789
+ .loc 2 109 28
790
+ add.f32 %f265, %f253, %f263;
791
+ .loc 2 110 39
792
+ setp.eq.f32 %p90, %f265, 0f00000000;
793
+ .loc 2 110 60
794
+ mov.b32 %r172, %f265;
795
+ div.full.f32 %r170, %r171, %r172;
796
+ mov.b32 %f266, %r170;
797
+ .loc 2 110 49
798
+ selp.f32 %f267, 0f00000000, %f266, %p90;
799
+ .loc 2 112 17
800
+ fma.rn.f32 %f268, %f264, %f267, %f256;
801
+ .loc 2 113 15
802
+ add.f32 %f269, %f260, %f262;
803
+ .loc 2 113 30
804
+ mul.f32 %f270, %f264, %f264;
805
+ .loc 2 113 38
806
+ mul.f32 %f271, %f253, %f270;
807
+ .loc 2 113 22
808
+ fma.rn.f32 %f272, %f267, %f271, %f269;
809
+ $L__tmp17:
810
+ .loc 2 120 46
811
+ mov.b32 %r242, %f268;
812
+ shfl.sync.bfly.b32 %r243, %r242, 4, 31, -1;
813
+ mov.b32 %f273, %r243;
814
+ mov.b32 %r244, %f272;
815
+ shfl.sync.bfly.b32 %r245, %r244, 4, 31, -1;
816
+ mov.b32 %f274, %r245;
817
+ shfl.sync.bfly.b32 %r174, %r172, 4, 31, -1;
818
+ mov.b32 %f275, %r174;
819
+ $L__tmp18:
820
+ .loc 2 108 21
821
+ sub.f32 %f276, %f273, %f268;
822
+ .loc 2 109 28
823
+ add.f32 %f277, %f265, %f275;
824
+ .loc 2 110 39
825
+ setp.eq.f32 %p91, %f277, 0f00000000;
826
+ .loc 2 110 60
827
+ mov.b32 %r175, %f277;
828
+ div.full.f32 %r173, %r174, %r175;
829
+ mov.b32 %f278, %r173;
830
+ .loc 2 110 49
831
+ selp.f32 %f279, 0f00000000, %f278, %p91;
832
+ .loc 2 112 17
833
+ fma.rn.f32 %f280, %f276, %f279, %f268;
834
+ .loc 2 113 15
835
+ add.f32 %f281, %f272, %f274;
836
+ .loc 2 113 30
837
+ mul.f32 %f282, %f276, %f276;
838
+ .loc 2 113 38
839
+ mul.f32 %f283, %f265, %f282;
840
+ .loc 2 113 22
841
+ fma.rn.f32 %f284, %f279, %f283, %f281;
842
+ $L__tmp19:
843
+ .loc 2 120 46
844
+ mov.b32 %r246, %f280;
845
+ shfl.sync.bfly.b32 %r247, %r246, 2, 31, -1;
846
+ mov.b32 %f285, %r247;
847
+ mov.b32 %r248, %f284;
848
+ shfl.sync.bfly.b32 %r249, %r248, 2, 31, -1;
849
+ mov.b32 %f286, %r249;
850
+ shfl.sync.bfly.b32 %r177, %r175, 2, 31, -1;
851
+ mov.b32 %f287, %r177;
852
+ $L__tmp20:
853
+ .loc 2 108 21
854
+ sub.f32 %f288, %f285, %f280;
855
+ .loc 2 109 28
856
+ add.f32 %f22, %f277, %f287;
857
+ .loc 2 110 39
858
+ setp.eq.f32 %p92, %f22, 0f00000000;
859
+ .loc 2 110 60
860
+ mov.b32 %r178, %f22;
861
+ div.full.f32 %r176, %r177, %r178;
862
+ mov.b32 %f289, %r176;
863
+ .loc 2 110 49
864
+ selp.f32 %f290, 0f00000000, %f289, %p92;
865
+ .loc 2 112 17
866
+ fma.rn.f32 %f23, %f288, %f290, %f280;
867
+ .loc 2 113 15
868
+ add.f32 %f291, %f284, %f286;
869
+ .loc 2 113 30
870
+ mul.f32 %f292, %f288, %f288;
871
+ .loc 2 113 38
872
+ mul.f32 %f293, %f277, %f292;
873
+ .loc 2 113 22
874
+ fma.rn.f32 %f24, %f290, %f293, %f291;
875
+ $L__tmp21:
876
+ .loc 2 120 46
877
+ mov.b32 %r250, %f23;
878
+ shfl.sync.bfly.b32 %r7, %r250, 1, 31, -1;
879
+ mov.b32 %r251, %f24;
880
+ shfl.sync.bfly.b32 %r8, %r251, 1, 31, -1;
881
+ shfl.sync.bfly.b32 %r180, %r178, 1, 31, -1;
882
+ mov.b32 %f294, %r180;
883
+ $L__tmp22:
884
+ .loc 2 109 28
885
+ add.f32 %f25, %f22, %f294;
886
+ .loc 2 110 60
887
+ mov.b32 %r181, %f25;
888
+ div.full.f32 %r179, %r180, %r181;
889
+ mov.b32 %f26, %r179;
890
+ $L__tmp23:
891
+ .loc 1 59 51
892
+ mov.u32 %r182, 0x0;
893
+ mov.u32 %r183, 0x0;
894
+ mov.u32 %r184, 0x0;
895
+ mov.u32 %r185, 0x0;
896
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r182, %r183, %r184, %r185 }, [ %rd80 + 0 ];
897
+ @!%p93 mov.u32 %r182, %r257;
898
+ @!%p93 mov.u32 %r183, %r257;
899
+ @!%p93 mov.u32 %r184, %r257;
900
+ @!%p93 mov.u32 %r185, %r257;
901
+ mov.u32 %r190, 0x0;
902
+ mov.u32 %r191, 0x0;
903
+ mov.u32 %r192, 0x0;
904
+ mov.u32 %r193, 0x0;
905
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r190, %r191, %r192, %r193 }, [ %rd81 + 0 ];
906
+ @!%p93 mov.u32 %r190, %r257;
907
+ @!%p93 mov.u32 %r191, %r257;
908
+ @!%p93 mov.u32 %r192, %r257;
909
+ @!%p93 mov.u32 %r193, %r257;
910
+ mov.u32 %r198, 0x0;
911
+ mov.u32 %r199, 0x0;
912
+ mov.u32 %r200, 0x0;
913
+ mov.u32 %r201, 0x0;
914
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r198, %r199, %r200, %r201 }, [ %rd82 + 0 ];
915
+ @!%p93 mov.u32 %r198, %r257;
916
+ @!%p93 mov.u32 %r199, %r257;
917
+ @!%p93 mov.u32 %r200, %r257;
918
+ @!%p93 mov.u32 %r201, %r257;
919
+ mov.u32 %r206, 0x0;
920
+ mov.u32 %r207, 0x0;
921
+ mov.u32 %r208, 0x0;
922
+ mov.u32 %r209, 0x0;
923
+ @%p93 ld.global.L1::evict_last.v4.b32 { %r206, %r207, %r208, %r209 }, [ %rd83 + 0 ];
924
+ @!%p93 mov.u32 %r206, %r257;
925
+ @!%p93 mov.u32 %r207, %r257;
926
+ @!%p93 mov.u32 %r208, %r257;
927
+ @!%p93 mov.u32 %r209, %r257;
928
+ .loc 1 60 35
929
+ mul.wide.u32 %rd96, %r2, 4;
930
+ add.s64 %rd84, %rd13, %rd96;
931
+ .loc 1 60 40
932
+ mov.u32 %r214, 0x0;
933
+ @%p93 ld.global.L1::evict_last.b32 { %r214 }, [ %rd84 + 0 ];
934
+ @!%p93 mov.u32 %r214, %r257;
935
+ .loc 1 64 57
936
+ @%p39 bra $L__BB0_4;
937
+ mov.u64 %rd97, assertMessage_1;
938
+ cvta.global.u64 %rd98, %rd97;
939
+ mov.u64 %rd99, assertFile_1;
940
+ cvta.global.u64 %rd100, %rd99;
941
+ mov.u64 %rd101, assertFunc_1;
942
+ cvta.global.u64 %rd102, %rd101;
943
+ { // callseq 9, 0
944
+ .reg .b32 temp_param_reg;
945
+ .param .b64 param0;
946
+ st.param.b64 [param0+0], %rd98;
947
+ .param .b64 param1;
948
+ st.param.b64 [param1+0], %rd100;
949
+ .param .b32 param2;
950
+ st.param.b32 [param2+0], %r374;
951
+ .param .b64 param3;
952
+ st.param.b64 [param3+0], %rd102;
953
+ .param .b64 param4;
954
+ st.param.b64 [param4+0], %rd112;
955
+ call.uni
956
+ __assertfail,
957
+ (
958
+ param0,
959
+ param1,
960
+ param2,
961
+ param3,
962
+ param4
963
+ );
964
+ } // callseq 9
965
+ $L__BB0_4:
966
+ $L__tmp24:
967
+ .loc 2 120 46
968
+ mov.b32 %f295, %r8;
969
+ $L__tmp25:
970
+ .loc 2 113 15
971
+ add.f32 %f296, %f24, %f295;
972
+ $L__tmp26:
973
+ .loc 2 120 46
974
+ mov.b32 %f297, %r7;
975
+ $L__tmp27:
976
+ .loc 2 108 21
977
+ sub.f32 %f298, %f297, %f23;
978
+ .loc 2 113 30
979
+ mul.f32 %f299, %f298, %f298;
980
+ .loc 2 113 38
981
+ mul.f32 %f300, %f22, %f299;
982
+ .loc 2 110 39
983
+ setp.eq.f32 %p115, %f25, 0f00000000;
984
+ .loc 2 110 49
985
+ selp.f32 %f301, 0f00000000, %f26, %p115;
986
+ .loc 2 113 22
987
+ fma.rn.f32 %f302, %f301, %f300, %f296;
988
+ $L__tmp28:
989
+ .loc 2 120 46
990
+ mov.b32 %f303, %r6;
991
+ $L__tmp29:
992
+ .loc 2 113 15
993
+ add.f32 %f304, %f19, %f303;
994
+ $L__tmp30:
995
+ .loc 2 120 46
996
+ mov.b32 %f305, %r5;
997
+ $L__tmp31:
998
+ .loc 2 108 21
999
+ sub.f32 %f306, %f305, %f18;
1000
+ .loc 2 113 30
1001
+ mul.f32 %f307, %f306, %f306;
1002
+ .loc 2 113 38
1003
+ mul.f32 %f308, %f17, %f307;
1004
+ .loc 2 110 39
1005
+ setp.eq.f32 %p116, %f20, 0f00000000;
1006
+ .loc 2 110 49
1007
+ selp.f32 %f309, 0f00000000, %f21, %p116;
1008
+ .loc 2 113 22
1009
+ fma.rn.f32 %f310, %f309, %f308, %f304;
1010
+ $L__tmp32:
1011
+ .loc 1 65 54
1012
+ mov.u32 %r253, 0x0;
1013
+ mov.u32 %r254, 0x0;
1014
+ mov.u32 %r255, 0x0;
1015
+ mov.u32 %r256, 0x0;
1016
+ @%p93 ld.global.L1::evict_first.v4.b32 { %r253, %r254, %r255, %r256 }, [ %rd104 + 0 ];
1017
+ @!%p93 mov.u32 %r253, %r257;
1018
+ @!%p93 mov.u32 %r254, %r257;
1019
+ @!%p93 mov.u32 %r255, %r257;
1020
+ @!%p93 mov.u32 %r256, %r257;
1021
+ mov.u32 %r261, 0x0;
1022
+ mov.u32 %r262, 0x0;
1023
+ mov.u32 %r263, 0x0;
1024
+ mov.u32 %r264, 0x0;
1025
+ @%p93 ld.global.L1::evict_first.v4.b32 { %r261, %r262, %r263, %r264 }, [ %rd105 + 0 ];
1026
+ @!%p93 mov.u32 %r261, %r257;
1027
+ @!%p93 mov.u32 %r262, %r257;
1028
+ @!%p93 mov.u32 %r263, %r257;
1029
+ @!%p93 mov.u32 %r264, %r257;
1030
+ mov.u32 %r269, 0x0;
1031
+ mov.u32 %r270, 0x0;
1032
+ mov.u32 %r271, 0x0;
1033
+ mov.u32 %r272, 0x0;
1034
+ @%p93 ld.global.L1::evict_first.v4.b32 { %r269, %r270, %r271, %r272 }, [ %rd106 + 0 ];
1035
+ @!%p93 mov.u32 %r269, %r257;
1036
+ @!%p93 mov.u32 %r270, %r257;
1037
+ @!%p93 mov.u32 %r271, %r257;
1038
+ @!%p93 mov.u32 %r272, %r257;
1039
+ mov.u32 %r277, 0x0;
1040
+ mov.u32 %r278, 0x0;
1041
+ mov.u32 %r279, 0x0;
1042
+ mov.u32 %r280, 0x0;
1043
+ @%p93 ld.global.L1::evict_first.v4.b32 { %r277, %r278, %r279, %r280 }, [ %rd107 + 0 ];
1044
+ @!%p93 mov.u32 %r277, %r257;
1045
+ @!%p93 mov.u32 %r278, %r257;
1046
+ @!%p93 mov.u32 %r279, %r257;
1047
+ @!%p93 mov.u32 %r280, %r257;
1048
+ .loc 1 69 23
1049
+ mov.b32 %r286, %f310;
1050
+ mov.b32 %r287, 1132462080;
1051
+ div.full.f32 %r285, %r286, %r287;
1052
+ mov.b32 %f311, %r285;
1053
+ mov.b32 %r310, %f302;
1054
+ div.full.f32 %r309, %r310, %r287;
1055
+ mov.b32 %f312, %r309;
1056
+ .loc 1 71 24
1057
+ add.f32 %f313, %f311, 0f3727C5AC;
1058
+ add.f32 %f314, %f312, 0f3727C5AC;
1059
+ .loc 1 72 30
1060
+ rsqrt.approx.ftz.f32 %f315, %f313;
1061
+ rsqrt.approx.ftz.f32 %f316, %f314;
1062
+ .loc 1 65 54
1063
+ mov.b32 %f317, %r280;
1064
+ .loc 1 59 51
1065
+ mov.b32 %f318, %r209;
1066
+ .loc 1 66 24
1067
+ add.f32 %f319, %f318, %f317;
1068
+ $L__tmp33:
1069
+ .loc 2 112 17
1070
+ fma.rn.f32 %f320, %f298, %f301, %f23;
1071
+ $L__tmp34:
1072
+ .loc 1 67 24
1073
+ sub.f32 %f321, %f319, %f320;
1074
+ .loc 1 65 54
1075
+ mov.b32 %f322, %r279;
1076
+ .loc 1 59 51
1077
+ mov.b32 %f323, %r208;
1078
+ .loc 1 66 24
1079
+ add.f32 %f324, %f323, %f322;
1080
+ .loc 1 67 24
1081
+ sub.f32 %f325, %f324, %f320;
1082
+ .loc 1 65 54
1083
+ mov.b32 %f326, %r278;
1084
+ .loc 1 59 51
1085
+ mov.b32 %f327, %r207;
1086
+ .loc 1 66 24
1087
+ add.f32 %f328, %f327, %f326;
1088
+ .loc 1 67 24
1089
+ sub.f32 %f329, %f328, %f320;
1090
+ .loc 1 65 54
1091
+ mov.b32 %f330, %r277;
1092
+ .loc 1 59 51
1093
+ mov.b32 %f331, %r206;
1094
+ .loc 1 66 24
1095
+ add.f32 %f332, %f331, %f330;
1096
+ .loc 1 67 24
1097
+ sub.f32 %f333, %f332, %f320;
1098
+ .loc 1 65 54
1099
+ mov.b32 %f334, %r272;
1100
+ .loc 1 59 51
1101
+ mov.b32 %f335, %r201;
1102
+ .loc 1 66 24
1103
+ add.f32 %f336, %f335, %f334;
1104
+ .loc 1 67 24
1105
+ sub.f32 %f337, %f336, %f320;
1106
+ .loc 1 65 54
1107
+ mov.b32 %f338, %r271;
1108
+ .loc 1 59 51
1109
+ mov.b32 %f339, %r200;
1110
+ .loc 1 66 24
1111
+ add.f32 %f340, %f339, %f338;
1112
+ .loc 1 67 24
1113
+ sub.f32 %f341, %f340, %f320;
1114
+ .loc 1 65 54
1115
+ mov.b32 %f342, %r270;
1116
+ .loc 1 59 51
1117
+ mov.b32 %f343, %r199;
1118
+ .loc 1 66 24
1119
+ add.f32 %f344, %f343, %f342;
1120
+ .loc 1 67 24
1121
+ sub.f32 %f345, %f344, %f320;
1122
+ .loc 1 65 54
1123
+ mov.b32 %f346, %r269;
1124
+ .loc 1 59 51
1125
+ mov.b32 %f347, %r198;
1126
+ .loc 1 66 24
1127
+ add.f32 %f348, %f347, %f346;
1128
+ .loc 1 67 24
1129
+ sub.f32 %f349, %f348, %f320;
1130
+ .loc 1 65 54
1131
+ mov.b32 %f350, %r264;
1132
+ .loc 1 59 51
1133
+ mov.b32 %f351, %r193;
1134
+ .loc 1 66 24
1135
+ add.f32 %f352, %f351, %f350;
1136
+ $L__tmp35:
1137
+ .loc 2 112 17
1138
+ fma.rn.f32 %f353, %f306, %f309, %f18;
1139
+ $L__tmp36:
1140
+ .loc 1 67 24
1141
+ sub.f32 %f354, %f352, %f353;
1142
+ .loc 1 65 54
1143
+ mov.b32 %f355, %r263;
1144
+ .loc 1 59 51
1145
+ mov.b32 %f356, %r192;
1146
+ .loc 1 66 24
1147
+ add.f32 %f357, %f356, %f355;
1148
+ .loc 1 67 24
1149
+ sub.f32 %f358, %f357, %f353;
1150
+ .loc 1 65 54
1151
+ mov.b32 %f359, %r262;
1152
+ .loc 1 59 51
1153
+ mov.b32 %f360, %r191;
1154
+ .loc 1 66 24
1155
+ add.f32 %f361, %f360, %f359;
1156
+ .loc 1 67 24
1157
+ sub.f32 %f362, %f361, %f353;
1158
+ .loc 1 65 54
1159
+ mov.b32 %f363, %r261;
1160
+ .loc 1 59 51
1161
+ mov.b32 %f364, %r190;
1162
+ .loc 1 66 24
1163
+ add.f32 %f365, %f364, %f363;
1164
+ .loc 1 67 24
1165
+ sub.f32 %f366, %f365, %f353;
1166
+ .loc 1 65 54
1167
+ mov.b32 %f367, %r256;
1168
+ .loc 1 59 51
1169
+ mov.b32 %f368, %r185;
1170
+ .loc 1 66 24
1171
+ add.f32 %f369, %f368, %f367;
1172
+ .loc 1 67 24
1173
+ sub.f32 %f370, %f369, %f353;
1174
+ .loc 1 65 54
1175
+ mov.b32 %f371, %r255;
1176
+ .loc 1 59 51
1177
+ mov.b32 %f372, %r184;
1178
+ .loc 1 66 24
1179
+ add.f32 %f373, %f372, %f371;
1180
+ .loc 1 67 24
1181
+ sub.f32 %f374, %f373, %f353;
1182
+ .loc 1 65 54
1183
+ mov.b32 %f375, %r254;
1184
+ .loc 1 59 51
1185
+ mov.b32 %f376, %r183;
1186
+ .loc 1 66 24
1187
+ add.f32 %f377, %f376, %f375;
1188
+ .loc 1 67 24
1189
+ sub.f32 %f378, %f377, %f353;
1190
+ .loc 1 65 54
1191
+ mov.b32 %f379, %r253;
1192
+ .loc 1 59 51
1193
+ mov.b32 %f380, %r182;
1194
+ .loc 1 66 24
1195
+ add.f32 %f381, %f380, %f379;
1196
+ .loc 1 67 24
1197
+ sub.f32 %f382, %f381, %f353;
1198
+ .loc 1 73 24
1199
+ mul.f32 %f383, %f382, %f315;
1200
+ mul.f32 %f384, %f378, %f315;
1201
+ mul.f32 %f385, %f374, %f315;
1202
+ mul.f32 %f386, %f370, %f315;
1203
+ mul.f32 %f387, %f366, %f315;
1204
+ mul.f32 %f388, %f362, %f315;
1205
+ mul.f32 %f389, %f358, %f315;
1206
+ mul.f32 %f390, %f354, %f315;
1207
+ mul.f32 %f391, %f349, %f316;
1208
+ mul.f32 %f392, %f345, %f316;
1209
+ mul.f32 %f393, %f341, %f316;
1210
+ mul.f32 %f394, %f337, %f316;
1211
+ mul.f32 %f395, %f333, %f316;
1212
+ mul.f32 %f396, %f329, %f316;
1213
+ mul.f32 %f397, %f325, %f316;
1214
+ mul.f32 %f398, %f321, %f316;
1215
+ .loc 1 74 24
1216
+ shl.b32 %r357, %r2, 2;
1217
+ mov.u32 %r358, global_smem;
1218
+ add.s32 %r359, %r358, %r357;
1219
+ st.shared.u32 [%r359], %r214;
1220
+ bar.sync 0;
1221
+ shl.b32 %r360, %r1, 2;
1222
+ add.s32 %r361, %r358, %r360;
1223
+ ld.shared.v4.f32 {%f399, %f400, %f401, %f402}, [%r361];
1224
+ ld.shared.v4.f32 {%f403, %f404, %f405, %f406}, [%r361+16];
1225
+ mul.f32 %f407, %f383, %f399;
1226
+ mul.f32 %f408, %f384, %f400;
1227
+ mul.f32 %f409, %f385, %f401;
1228
+ mul.f32 %f410, %f386, %f402;
1229
+ mul.f32 %f411, %f387, %f403;
1230
+ mul.f32 %f412, %f388, %f404;
1231
+ mul.f32 %f413, %f389, %f405;
1232
+ mul.f32 %f414, %f390, %f406;
1233
+ mul.f32 %f415, %f391, %f399;
1234
+ mul.f32 %f416, %f392, %f400;
1235
+ mul.f32 %f417, %f393, %f401;
1236
+ mul.f32 %f418, %f394, %f402;
1237
+ mul.f32 %f419, %f395, %f403;
1238
+ mul.f32 %f420, %f396, %f404;
1239
+ mul.f32 %f421, %f397, %f405;
1240
+ mul.f32 %f422, %f398, %f406;
1241
+ .loc 1 76 39
1242
+ shl.b32 %r362, %r3, 8;
1243
+ shl.b32 %r363, %r4, 8;
1244
+ .loc 1 76 35
1245
+ or.b32 %r364, %r362, %r1;
1246
+ or.b32 %r365, %r363, %r1;
1247
+ .loc 1 76 29
1248
+ mul.wide.s32 %rd110, %r364, 2;
1249
+ add.s64 %rd108, %rd14, %rd110;
1250
+ mul.wide.s32 %rd111, %r365, 2;
1251
+ add.s64 %rd109, %rd14, %rd111;
1252
+ .loc 1 76 52
1253
+ mov.b32 %r333, %f407;
1254
+ cvt.rn.bf16.f32 %rs1, %r333;
1255
+ mov.b32 %r334, %f408;
1256
+ cvt.rn.bf16.f32 %rs2, %r334;
1257
+ mov.b32 %r335, %f409;
1258
+ cvt.rn.bf16.f32 %rs3, %r335;
1259
+ mov.b32 %r336, %f410;
1260
+ cvt.rn.bf16.f32 %rs4, %r336;
1261
+ mov.b32 %r337, %f411;
1262
+ cvt.rn.bf16.f32 %rs5, %r337;
1263
+ mov.b32 %r338, %f412;
1264
+ cvt.rn.bf16.f32 %rs6, %r338;
1265
+ mov.b32 %r339, %f413;
1266
+ cvt.rn.bf16.f32 %rs7, %r339;
1267
+ mov.b32 %r340, %f414;
1268
+ cvt.rn.bf16.f32 %rs8, %r340;
1269
+ mov.b32 %r341, %f415;
1270
+ cvt.rn.bf16.f32 %rs9, %r341;
1271
+ mov.b32 %r342, %f416;
1272
+ cvt.rn.bf16.f32 %rs10, %r342;
1273
+ mov.b32 %r343, %f417;
1274
+ cvt.rn.bf16.f32 %rs11, %r343;
1275
+ mov.b32 %r344, %f418;
1276
+ cvt.rn.bf16.f32 %rs12, %r344;
1277
+ mov.b32 %r345, %f419;
1278
+ cvt.rn.bf16.f32 %rs13, %r345;
1279
+ mov.b32 %r346, %f420;
1280
+ cvt.rn.bf16.f32 %rs14, %r346;
1281
+ mov.b32 %r347, %f421;
1282
+ cvt.rn.bf16.f32 %rs15, %r347;
1283
+ mov.b32 %r348, %f422;
1284
+ cvt.rn.bf16.f32 %rs16, %r348;
1285
+ mov.b32 %r366, {%rs1, %rs2};
1286
+ mov.b32 %r367, {%rs3, %rs4};
1287
+ mov.b32 %r368, {%rs5, %rs6};
1288
+ mov.b32 %r369, {%rs7, %rs8};
1289
+ @%p93 st.global.v4.b32 [ %rd108 + 0 ], { %r366, %r367, %r368, %r369 };
1290
+ mov.b32 %r370, {%rs9, %rs10};
1291
+ mov.b32 %r371, {%rs11, %rs12};
1292
+ mov.b32 %r372, {%rs13, %rs14};
1293
+ mov.b32 %r373, {%rs15, %rs16};
1294
+ @%p93 st.global.v4.b32 [ %rd109 + 0 ], { %r370, %r371, %r372, %r373 };
1295
+ .loc 1 55 4
1296
+ ret;
1297
+ $L__tmp37:
1298
+ $L__func_end0:
1299
+
1300
+ }
1301
+ // .globl __nv_rsqrtf
1302
+ .visible .func (.param .b32 func_retval0) __nv_rsqrtf(
1303
+ .param .b32 __nv_rsqrtf_param_0
1304
+ )
1305
+ {
1306
+ .reg .f32 %f<3>;
1307
+ $L__func_begin1:
1308
+
1309
+ ld.param.f32 %f1, [__nv_rsqrtf_param_0];
1310
+ rsqrt.approx.ftz.f32 %f2, %f1;
1311
+ st.param.f32 [func_retval0+0], %f2;
1312
+ ret;
1313
+ $L__func_end1:
1314
+
1315
+ }
1316
+ .file 1 "/tmp/torchinductor_root/lh/clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py"
1317
+ .file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
1318
+ .section .debug_abbrev
1319
+ {
1320
+ .b8 1
1321
+ .b8 17
1322
+ .b8 1
1323
+ .b8 37
1324
+ .b8 8
1325
+ .b8 19
1326
+ .b8 5
1327
+ .b8 3
1328
+ .b8 8
1329
+ .b8 16
1330
+ .b8 6
1331
+ .b8 27
1332
+ .b8 8
1333
+ .b8 180
1334
+ .b8 66
1335
+ .b8 12
1336
+ .b8 17
1337
+ .b8 1
1338
+ .b8 18
1339
+ .b8 1
1340
+ .b8 0
1341
+ .b8 0
1342
+ .b8 2
1343
+ .b8 46
1344
+ .b8 0
1345
+ .b8 135
1346
+ .b8 64
1347
+ .b8 8
1348
+ .b8 3
1349
+ .b8 8
1350
+ .b8 58
1351
+ .b8 11
1352
+ .b8 59
1353
+ .b8 11
1354
+ .b8 63
1355
+ .b8 12
1356
+ .b8 32
1357
+ .b8 11
1358
+ .b8 0
1359
+ .b8 0
1360
+ .b8 3
1361
+ .b8 46
1362
+ .b8 1
1363
+ .b8 17
1364
+ .b8 1
1365
+ .b8 18
1366
+ .b8 1
1367
+ .b8 64
1368
+ .b8 10
1369
+ .b8 49
1370
+ .b8 19
1371
+ .b8 0
1372
+ .b8 0
1373
+ .b8 4
1374
+ .b8 29
1375
+ .b8 0
1376
+ .b8 49
1377
+ .b8 19
1378
+ .b8 17
1379
+ .b8 1
1380
+ .b8 18
1381
+ .b8 1
1382
+ .b8 88
1383
+ .b8 11
1384
+ .b8 89
1385
+ .b8 11
1386
+ .b8 87
1387
+ .b8 11
1388
+ .b8 0
1389
+ .b8 0
1390
+ .b8 5
1391
+ .b8 29
1392
+ .b8 1
1393
+ .b8 49
1394
+ .b8 19
1395
+ .b8 17
1396
+ .b8 1
1397
+ .b8 18
1398
+ .b8 1
1399
+ .b8 88
1400
+ .b8 11
1401
+ .b8 89
1402
+ .b8 11
1403
+ .b8 87
1404
+ .b8 11
1405
+ .b8 0
1406
+ .b8 0
1407
+ .b8 0
1408
+ }
1409
+ .section .debug_info
1410
+ {
1411
+ .b32 298
1412
+ .b8 2
1413
+ .b8 0
1414
+ .b32 .debug_abbrev
1415
+ .b8 8
1416
+ .b8 1
1417
+ .b8 116
1418
+ .b8 114
1419
+ .b8 105
1420
+ .b8 116
1421
+ .b8 111
1422
+ .b8 110
1423
+ .b8 0
1424
+ .b8 2
1425
+ .b8 0
1426
+ .b8 99
1427
+ .b8 108
1428
+ .b8 104
1429
+ .b8 101
1430
+ .b8 52
1431
+ .b8 97
1432
+ .b8 51
1433
+ .b8 115
1434
+ .b8 116
1435
+ .b8 118
1436
+ .b8 117
1437
+ .b8 102
1438
+ .b8 120
1439
+ .b8 97
1440
+ .b8 102
1441
+ .b8 109
1442
+ .b8 113
1443
+ .b8 51
1444
+ .b8 107
1445
+ .b8 107
1446
+ .b8 53
1447
+ .b8 104
1448
+ .b8 111
1449
+ .b8 100
1450
+ .b8 97
1451
+ .b8 122
1452
+ .b8 122
1453
+ .b8 50
1454
+ .b8 101
1455
+ .b8 102
1456
+ .b8 99
1457
+ .b8 116
1458
+ .b8 102
1459
+ .b8 102
1460
+ .b8 116
1461
+ .b8 101
1462
+ .b8 54
1463
+ .b8 52
1464
+ .b8 54
1465
+ .b8 122
1466
+ .b8 110
1467
+ .b8 106
1468
+ .b8 100
1469
+ .b8 110
1470
+ .b8 118
1471
+ .b8 51
1472
+ .b8 108
1473
+ .b8 113
1474
+ .b8 105
1475
+ .b8 53
1476
+ .b8 111
1477
+ .b8 97
1478
+ .b8 46
1479
+ .b8 112
1480
+ .b8 121
1481
+ .b8 0
1482
+ .b32 .debug_line
1483
+ .b8 47
1484
+ .b8 116
1485
+ .b8 109
1486
+ .b8 112
1487
+ .b8 47
1488
+ .b8 116
1489
+ .b8 111
1490
+ .b8 114
1491
+ .b8 99
1492
+ .b8 104
1493
+ .b8 105
1494
+ .b8 110
1495
+ .b8 100
1496
+ .b8 117
1497
+ .b8 99
1498
+ .b8 116
1499
+ .b8 111
1500
+ .b8 114
1501
+ .b8 95
1502
+ .b8 114
1503
+ .b8 111
1504
+ .b8 111
1505
+ .b8 116
1506
+ .b8 47
1507
+ .b8 108
1508
+ .b8 104
1509
+ .b8 0
1510
+ .b8 1
1511
+ .b64 $L__func_begin0
1512
+ .b64 $L__func_end0
1513
+ .b8 2
1514
+ .b8 116
1515
+ .b8 114
1516
+ .b8 105
1517
+ .b8 116
1518
+ .b8 111
1519
+ .b8 110
1520
+ .b8 95
1521
+ .b8 95
1522
+ .b8 48
1523
+ .b8 100
1524
+ .b8 49
1525
+ .b8 100
1526
+ .b8 50
1527
+ .b8 100
1528
+ .b8 51
1529
+ .b8 100
1530
+ .b8 52
1531
+ .b8 100
1532
+ .b8 53
1533
+ .b8 100
1534
+ .b8 101
1535
+ .b8 54
1536
+ .b8 100
1537
+ .b8 101
1538
+ .b8 0
1539
+ .b8 116
1540
+ .b8 114
1541
+ .b8 105
1542
+ .b8 116
1543
+ .b8 111
1544
+ .b8 110
1545
+ .b8 95
1546
+ .b8 95
1547
+ .b8 48
1548
+ .b8 100
1549
+ .b8 49
1550
+ .b8 100
1551
+ .b8 50
1552
+ .b8 100
1553
+ .b8 51
1554
+ .b8 100
1555
+ .b8 52
1556
+ .b8 100
1557
+ .b8 53
1558
+ .b8 100
1559
+ .b8 101
1560
+ .b8 54
1561
+ .b8 100
1562
+ .b8 101
1563
+ .b8 0
1564
+ .b8 1
1565
+ .b8 18
1566
+ .b8 1
1567
+ .b8 1
1568
+ .b8 3
1569
+ .b64 $L__func_begin0
1570
+ .b64 $L__func_end0
1571
+ .b8 1
1572
+ .b8 156
1573
+ .b32 125
1574
+ .b8 4
1575
+ .b32 125
1576
+ .b64 $L__tmp1
1577
+ .b64 $L__tmp2
1578
+ .b8 2
1579
+ .b8 44
1580
+ .b8 38
1581
+ .b8 5
1582
+ .b32 125
1583
+ .b64 $L__tmp2
1584
+ .b64 $L__tmp36
1585
+ .b8 2
1586
+ .b8 50
1587
+ .b8 41
1588
+ .b8 4
1589
+ .b32 125
1590
+ .b64 $L__tmp2
1591
+ .b64 $L__tmp36
1592
+ .b8 2
1593
+ .b8 120
1594
+ .b8 46
1595
+ .b8 0
1596
+ .b8 4
1597
+ .b32 125
1598
+ .b64 $L__tmp3
1599
+ .b64 $L__tmp31
1600
+ .b8 2
1601
+ .b8 50
1602
+ .b8 41
1603
+ .b8 0
1604
+ .b8 0
1605
+ }
1606
+ .section .debug_pubnames
1607
+ {
1608
+ .b32 $L__pubNames_end0-$L__pubNames_start0
1609
+ $L__pubNames_start0:
1610
+ .b8 2
1611
+ .b8 0
1612
+ .b32 .debug_info
1613
+ .b32 302
1614
+ .b32 125
1615
+ .b8 116
1616
+ .b8 114
1617
+ .b8 105
1618
+ .b8 116
1619
+ .b8 111
1620
+ .b8 110
1621
+ .b8 95
1622
+ .b8 95
1623
+ .b8 48
1624
+ .b8 100
1625
+ .b8 49
1626
+ .b8 100
1627
+ .b8 50
1628
+ .b8 100
1629
+ .b8 51
1630
+ .b8 100
1631
+ .b8 52
1632
+ .b8 100
1633
+ .b8 53
1634
+ .b8 100
1635
+ .b8 101
1636
+ .b8 54
1637
+ .b8 100
1638
+ .b8 101
1639
+ .b8 0
1640
+ .b32 0
1641
+ $L__pubNames_end0:
1642
+ }
1643
+ .section .debug_pubtypes
1644
+ {
1645
+ .b32 $L__pubTypes_end0-$L__pubTypes_start0
1646
+ $L__pubTypes_start0:
1647
+ .b8 2
1648
+ .b8 0
1649
+ .b32 .debug_info
1650
+ .b32 302
1651
+ .b32 0
1652
+ $L__pubTypes_end0:
1653
+ }
1654
+ .section .debug_loc { }
.triton/dump/345a87a492fd703c73ab83265a21fcb6/triton_.ttir ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ module {
2
+ tt.func public @triton__0d1d2d3d4d5de6de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
3
+ %cst = arith.constant dense<1.000000e+00> : tensor<1x256xf32>
4
+ %cst_0 = arith.constant dense<0.000000e+00> : tensor<1x256xf32>
5
+ %cst_1 = arith.constant 0.000000e+00 : f32
6
+ %cst_2 = arith.constant dense<256> : tensor<16x1xi64>
7
+ %cst_3 = arith.constant dense<50257> : tensor<16x1xi64>
8
+ %cst_4 = arith.constant dense<0> : tensor<16x1xi64>
9
+ %cst_5 = arith.constant dense<9.99999974E-6> : tensor<16x1xf32>
10
+ %cst_6 = arith.constant dense<2.560000e+02> : tensor<16x1xf32>
11
+ %cst_7 = arith.constant dense<0.000000e+00> : tensor<16x256xf32>
12
+ %cst_8 = arith.constant dense<256> : tensor<16x1xi32>
13
+ %cst_9 = arith.constant dense<256> : tensor<1x256xi32>
14
+ %cst_10 = arith.constant dense<512> : tensor<16x1xi32>
15
+ %c16_i32 = arith.constant 16 : i32
16
+ %0 = tt.get_program_id x : i32
17
+ %1 = arith.muli %0, %c16_i32 : i32
18
+ %2 = tt.make_range {end = 16 : i32, start = 0 : i32} : tensor<16xi32>
19
+ %3 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<16xi32>) -> tensor<16x1xi32>
20
+ %4 = tt.splat %1 : (i32) -> tensor<16x1xi32>
21
+ %5 = arith.addi %4, %3 : tensor<16x1xi32>
22
+ %6 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32>
23
+ %7 = tt.expand_dims %6 {axis = 0 : i32} : (tensor<256xi32>) -> tensor<1x256xi32>
24
+ %8 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<16x1x!tt.ptr<i64, 1>>
25
+ %9 = tt.addptr %8, %5 : tensor<16x1x!tt.ptr<i64, 1>>, tensor<16x1xi32>
26
+ %10 = tt.load %9 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x1xi64>
27
+ %11 = arith.remsi %5, %cst_10 : tensor<16x1xi32>
28
+ %12 = arith.cmpi slt, %7, %cst_9 : tensor<1x256xi32>
29
+ %13 = arith.muli %11, %cst_8 : tensor<16x1xi32>
30
+ %14 = tt.broadcast %7 : (tensor<1x256xi32>) -> tensor<16x256xi32>
31
+ %15 = tt.broadcast %13 : (tensor<16x1xi32>) -> tensor<16x256xi32>
32
+ %16 = arith.addi %14, %15 : tensor<16x256xi32>
33
+ %17 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<16x256x!tt.ptr<f32, 1>>
34
+ %18 = tt.addptr %17, %16 : tensor<16x256x!tt.ptr<f32, 1>>, tensor<16x256xi32>
35
+ %19 = tt.broadcast %12 : (tensor<1x256xi1>) -> tensor<16x256xi1>
36
+ %20 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x256xf32>
37
+ %21 = arith.addi %10, %cst_3 : tensor<16x1xi64>
38
+ %22 = arith.cmpi slt, %10, %cst_4 : tensor<16x1xi64>
39
+ %23 = arith.select %22, %21, %10 : tensor<16x1xi1>, tensor<16x1xi64>
40
+ %24 = arith.cmpi sge, %23, %cst_4 : tensor<16x1xi64>
41
+ %25 = arith.cmpi slt, %23, %cst_3 : tensor<16x1xi64>
42
+ %26 = arith.andi %24, %25 : tensor<16x1xi1>
43
+ tt.assert %26, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<16x1xi1>
44
+ %27 = arith.muli %23, %cst_2 : tensor<16x1xi64>
45
+ %28 = tt.broadcast %27 : (tensor<16x1xi64>) -> tensor<16x256xi64>
46
+ %29 = arith.extsi %7 : tensor<1x256xi32> to tensor<1x256xi64>
47
+ %30 = tt.broadcast %29 : (tensor<1x256xi64>) -> tensor<16x256xi64>
48
+ %31 = arith.addi %30, %28 : tensor<16x256xi64>
49
+ %32 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<16x256x!tt.ptr<f32, 1>>
50
+ %33 = tt.addptr %32, %31 : tensor<16x256x!tt.ptr<f32, 1>>, tensor<16x256xi64>
51
+ %34 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x256xf32>
52
+ %35 = arith.addf %34, %20 : tensor<16x256xf32>
53
+ %36 = arith.addf %35, %cst_7 : tensor<16x256xf32>
54
+ %37 = arith.subf %35, %36 : tensor<16x256xf32>
55
+ %38 = arith.mulf %35, %37 : tensor<16x256xf32>
56
+ %39 = arith.addf %38, %cst_7 : tensor<16x256xf32>
57
+ %40 = arith.select %19, %36, %cst_7 : tensor<16x256xi1>, tensor<16x256xf32>
58
+ %41 = arith.select %19, %39, %cst_7 : tensor<16x256xi1>, tensor<16x256xf32>
59
+ %42 = arith.select %12, %cst, %cst_0 : tensor<1x256xi1>, tensor<1x256xf32>
60
+ %43 = tt.broadcast %42 : (tensor<1x256xf32>) -> tensor<16x256xf32>
61
+ %44:3 = "tt.reduce"(%40, %41, %43) <{axis = 1 : i32}> ({
62
+ ^bb0(%arg7: f32, %arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32):
63
+ %68 = arith.subf %arg10, %arg7 : f32
64
+ %69 = arith.addf %arg9, %arg12 : f32
65
+ %70 = arith.cmpf oeq, %69, %cst_1 : f32
66
+ %71 = arith.divf %arg12, %69 : f32
67
+ %72 = arith.select %70, %cst_1, %71 : f32
68
+ %73 = arith.mulf %68, %72 : f32
69
+ %74 = arith.addf %arg7, %73 : f32
70
+ %75 = arith.addf %arg8, %arg11 : f32
71
+ %76 = arith.mulf %68, %68 : f32
72
+ %77 = arith.mulf %76, %arg9 : f32
73
+ %78 = arith.mulf %77, %72 : f32
74
+ %79 = arith.addf %75, %78 : f32
75
+ tt.reduce.return %74, %79, %69 : f32, f32, f32
76
+ }) : (tensor<16x256xf32>, tensor<16x256xf32>, tensor<16x256xf32>) -> (tensor<16xf32>, tensor<16xf32>, tensor<16xf32>)
77
+ %45 = tt.expand_dims %44#0 {axis = 1 : i32} : (tensor<16xf32>) -> tensor<16x1xf32>
78
+ %46 = tt.expand_dims %44#1 {axis = 1 : i32} : (tensor<16xf32>) -> tensor<16x1xf32>
79
+ %47 = tt.load %18, %19, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<16x256xf32>
80
+ %48 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x256x!tt.ptr<f32, 1>>
81
+ %49 = tt.addptr %48, %7 : tensor<1x256x!tt.ptr<f32, 1>>, tensor<1x256xi32>
82
+ %50 = tt.load %49, %12, %cst_0 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x256xf32>
83
+ tt.assert %26, "index out of bounds: 0 <= tmp13 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<16x1xi1>
84
+ %51 = tt.load %33, %19, %cst_7 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<16x256xf32>
85
+ %52 = arith.addf %51, %47 : tensor<16x256xf32>
86
+ %53 = tt.broadcast %45 : (tensor<16x1xf32>) -> tensor<16x256xf32>
87
+ %54 = arith.subf %52, %53 : tensor<16x256xf32>
88
+ %55 = arith.divf %46, %cst_6 : tensor<16x1xf32>
89
+ %56 = arith.addf %55, %cst_5 : tensor<16x1xf32>
90
+ %57 = tt.extern_elementwise %56 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<16x1xf32>) -> tensor<16x1xf32>
91
+ %58 = tt.broadcast %57 : (tensor<16x1xf32>) -> tensor<16x256xf32>
92
+ %59 = arith.mulf %54, %58 : tensor<16x256xf32>
93
+ %60 = tt.broadcast %50 : (tensor<1x256xf32>) -> tensor<16x256xf32>
94
+ %61 = arith.mulf %59, %60 : tensor<16x256xf32>
95
+ %62 = arith.muli %5, %cst_8 : tensor<16x1xi32>
96
+ %63 = tt.broadcast %62 : (tensor<16x1xi32>) -> tensor<16x256xi32>
97
+ %64 = arith.addi %14, %63 : tensor<16x256xi32>
98
+ %65 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<16x256x!tt.ptr<bf16, 1>>
99
+ %66 = tt.addptr %65, %64 : tensor<16x256x!tt.ptr<bf16, 1>>, tensor<16x256xi32>
100
+ %67 = arith.truncf %61 : tensor<16x256xf32> to tensor<16x256xbf16>
101
+ tt.store %66, %67, %19 {cache = 1 : i32, evict = 1 : i32} : tensor<16x256xbf16>
102
+ tt.return
103
+ }
104
+ }
.triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ptx ADDED
@@ -0,0 +1,465 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.2
6
+ .target sm_89
7
+ .address_size 64
8
+
9
+ // .globl triton__0d1d2d34e
10
+
11
+ .visible .entry triton__0d1d2d34e(
12
+ .param .u64 triton__0d1d2d34e_param_0,
13
+ .param .u64 triton__0d1d2d34e_param_1,
14
+ .param .u64 triton__0d1d2d34e_param_2,
15
+ .param .u32 triton__0d1d2d34e_param_3,
16
+ .param .u32 triton__0d1d2d34e_param_4
17
+ )
18
+ .maxntid 64, 1, 1
19
+ {
20
+ .reg .pred %p<6>;
21
+ .reg .b32 %r<27>;
22
+ .reg .f32 %f<9>;
23
+ .reg .b64 %rd<24>;
24
+ .loc 1 18 0
25
+ $L__func_begin0:
26
+ .loc 1 18 0
27
+
28
+ ld.param.u64 %rd4, [triton__0d1d2d34e_param_0];
29
+ ld.param.u64 %rd5, [triton__0d1d2d34e_param_1];
30
+ $L__tmp0:
31
+ .loc 1 25 34
32
+ mov.u32 %r7, %tid.x;
33
+ and.b32 %r8, %r7, 7;
34
+ ld.param.u64 %rd6, [triton__0d1d2d34e_param_2];
35
+ .loc 1 28 30
36
+ mul.wide.u32 %rd7, %r8, 4;
37
+ add.s64 %rd1, %rd5, %rd7;
38
+ mov.b32 %r2, 0;
39
+ mov.pred %p1, -1;
40
+ .loc 1 28 35
41
+ mov.u32 %r1, 0x0;
42
+ @%p1 ld.global.b32 { %r1 }, [ %rd1 + 0 ];
43
+ @!%p1 mov.u32 %r1, %r2;
44
+ mov.b32 %f1, %r1;
45
+ .loc 1 29 30
46
+ mul.wide.u32 %rd8, %r8, 8;
47
+ add.s64 %rd3, %rd6, %rd8;
48
+ .loc 1 29 35
49
+ mov.u64 %rd2, 0x0;
50
+ @%p1 ld.global.b64 { %rd2 }, [ %rd3 + 0 ];
51
+ @!%p1 mov.u64 %rd2, 0x0;
52
+ $L__tmp1:
53
+ .loc 2 243 36
54
+ shfl.sync.bfly.b32 %r9, %r1, 4, 31, -1;
55
+ mov.b32 %f2, %r9;
56
+ $L__tmp2:
57
+ .loc 2 233 15
58
+ add.f32 %f3, %f1, %f2;
59
+ $L__tmp3:
60
+ .loc 2 243 36
61
+ mov.b32 %r10, %f3;
62
+ shfl.sync.bfly.b32 %r11, %r10, 2, 31, -1;
63
+ mov.b32 %f4, %r11;
64
+ $L__tmp4:
65
+ .loc 2 233 15
66
+ add.f32 %f5, %f3, %f4;
67
+ $L__tmp5:
68
+ .loc 2 243 36
69
+ mov.b32 %r12, %f5;
70
+ shfl.sync.bfly.b32 %r13, %r12, 1, 31, -1;
71
+ mov.b32 %f6, %r13;
72
+ $L__tmp6:
73
+ .loc 2 233 15
74
+ add.f32 %f7, %f5, %f6;
75
+ $L__tmp7:
76
+ .loc 2 243 36
77
+ cvt.u32.u64 %r14, %rd2;
78
+ shfl.sync.bfly.b32 %r15, %r14, 4, 31, -1;
79
+ { .reg .b32 tmp; mov.b64 {tmp, %r16}, %rd2; }
80
+ shfl.sync.bfly.b32 %r17, %r16, 4, 31, -1;
81
+ cvt.u64.u32 %rd9, %r15;
82
+ cvt.u64.u32 %rd10, %r17;
83
+ shl.b64 %rd11, %rd10, 32;
84
+ or.b64 %rd12, %rd9, %rd11;
85
+ $L__tmp8:
86
+ .loc 2 233 15
87
+ add.s64 %rd13, %rd2, %rd12;
88
+ $L__tmp9:
89
+ .loc 2 243 36
90
+ cvt.u32.u64 %r18, %rd13;
91
+ shfl.sync.bfly.b32 %r19, %r18, 2, 31, -1;
92
+ { .reg .b32 tmp; mov.b64 {tmp, %r20}, %rd13; }
93
+ shfl.sync.bfly.b32 %r21, %r20, 2, 31, -1;
94
+ cvt.u64.u32 %rd14, %r19;
95
+ cvt.u64.u32 %rd15, %r21;
96
+ shl.b64 %rd16, %rd15, 32;
97
+ or.b64 %rd17, %rd14, %rd16;
98
+ $L__tmp10:
99
+ .loc 2 233 15
100
+ add.s64 %rd18, %rd13, %rd17;
101
+ $L__tmp11:
102
+ .loc 2 243 36
103
+ cvt.u32.u64 %r22, %rd18;
104
+ shfl.sync.bfly.b32 %r23, %r22, 1, 31, -1;
105
+ { .reg .b32 tmp; mov.b64 {tmp, %r24}, %rd18; }
106
+ shfl.sync.bfly.b32 %r25, %r24, 1, 31, -1;
107
+ cvt.u64.u32 %rd19, %r23;
108
+ cvt.u64.u32 %rd20, %r25;
109
+ shl.b64 %rd21, %rd20, 32;
110
+ or.b64 %rd22, %rd19, %rd21;
111
+ $L__tmp12:
112
+ .loc 2 233 15
113
+ add.s64 %rd23, %rd18, %rd22;
114
+ $L__tmp13:
115
+ .loc 1 36 20
116
+ cvt.rn.f32.s64 %f8, %rd23;
117
+ .loc 1 37 19
118
+ mov.b32 %r4, %f7;
119
+ mov.b32 %r5, %f8;
120
+ div.full.f32 %r6, %r4, %r5;
121
+ .loc 1 38 4
122
+ bar.sync 0;
123
+ .loc 1 39 71
124
+ and.b32 %r26, %r7, 63;
125
+ setp.eq.s32 %p5, %r26, 0;
126
+ @%p5 st.global.b32 [ %rd4 + 0 ], { %r6 };
127
+ .loc 1 39 4
128
+ ret;
129
+ $L__tmp14:
130
+ $L__func_end0:
131
+
132
+ }
133
+ .file 1 "/tmp/torchinductor_root/2q/c2qomesxoic3sfzpdzftrhej7z6hhd6pritis2f4ye2ckqoetmyt.py"
134
+ .file 2 "/usr/local/lib/python3.10/dist-packages/triton/language/standard.py"
135
+ .section .debug_abbrev
136
+ {
137
+ .b8 1
138
+ .b8 17
139
+ .b8 1
140
+ .b8 37
141
+ .b8 8
142
+ .b8 19
143
+ .b8 5
144
+ .b8 3
145
+ .b8 8
146
+ .b8 16
147
+ .b8 6
148
+ .b8 27
149
+ .b8 8
150
+ .b8 180
151
+ .b8 66
152
+ .b8 12
153
+ .b8 17
154
+ .b8 1
155
+ .b8 18
156
+ .b8 1
157
+ .b8 0
158
+ .b8 0
159
+ .b8 2
160
+ .b8 46
161
+ .b8 0
162
+ .b8 135
163
+ .b8 64
164
+ .b8 8
165
+ .b8 3
166
+ .b8 8
167
+ .b8 58
168
+ .b8 11
169
+ .b8 59
170
+ .b8 11
171
+ .b8 63
172
+ .b8 12
173
+ .b8 32
174
+ .b8 11
175
+ .b8 0
176
+ .b8 0
177
+ .b8 3
178
+ .b8 46
179
+ .b8 1
180
+ .b8 17
181
+ .b8 1
182
+ .b8 18
183
+ .b8 1
184
+ .b8 64
185
+ .b8 10
186
+ .b8 49
187
+ .b8 19
188
+ .b8 0
189
+ .b8 0
190
+ .b8 4
191
+ .b8 29
192
+ .b8 0
193
+ .b8 49
194
+ .b8 19
195
+ .b8 17
196
+ .b8 1
197
+ .b8 18
198
+ .b8 1
199
+ .b8 88
200
+ .b8 11
201
+ .b8 89
202
+ .b8 11
203
+ .b8 87
204
+ .b8 11
205
+ .b8 0
206
+ .b8 0
207
+ .b8 5
208
+ .b8 29
209
+ .b8 1
210
+ .b8 49
211
+ .b8 19
212
+ .b8 17
213
+ .b8 1
214
+ .b8 18
215
+ .b8 1
216
+ .b8 88
217
+ .b8 11
218
+ .b8 89
219
+ .b8 11
220
+ .b8 87
221
+ .b8 11
222
+ .b8 0
223
+ .b8 0
224
+ .b8 0
225
+ }
226
+ .section .debug_info
227
+ {
228
+ .b32 333
229
+ .b8 2
230
+ .b8 0
231
+ .b32 .debug_abbrev
232
+ .b8 8
233
+ .b8 1
234
+ .b8 116
235
+ .b8 114
236
+ .b8 105
237
+ .b8 116
238
+ .b8 111
239
+ .b8 110
240
+ .b8 0
241
+ .b8 2
242
+ .b8 0
243
+ .b8 99
244
+ .b8 50
245
+ .b8 113
246
+ .b8 111
247
+ .b8 109
248
+ .b8 101
249
+ .b8 115
250
+ .b8 120
251
+ .b8 111
252
+ .b8 105
253
+ .b8 99
254
+ .b8 51
255
+ .b8 115
256
+ .b8 102
257
+ .b8 122
258
+ .b8 112
259
+ .b8 100
260
+ .b8 122
261
+ .b8 102
262
+ .b8 116
263
+ .b8 114
264
+ .b8 104
265
+ .b8 101
266
+ .b8 106
267
+ .b8 55
268
+ .b8 122
269
+ .b8 54
270
+ .b8 104
271
+ .b8 104
272
+ .b8 100
273
+ .b8 54
274
+ .b8 112
275
+ .b8 114
276
+ .b8 105
277
+ .b8 116
278
+ .b8 105
279
+ .b8 115
280
+ .b8 50
281
+ .b8 102
282
+ .b8 52
283
+ .b8 121
284
+ .b8 101
285
+ .b8 50
286
+ .b8 99
287
+ .b8 107
288
+ .b8 113
289
+ .b8 111
290
+ .b8 101
291
+ .b8 116
292
+ .b8 109
293
+ .b8 121
294
+ .b8 116
295
+ .b8 46
296
+ .b8 112
297
+ .b8 121
298
+ .b8 0
299
+ .b32 .debug_line
300
+ .b8 47
301
+ .b8 116
302
+ .b8 109
303
+ .b8 112
304
+ .b8 47
305
+ .b8 116
306
+ .b8 111
307
+ .b8 114
308
+ .b8 99
309
+ .b8 104
310
+ .b8 105
311
+ .b8 110
312
+ .b8 100
313
+ .b8 117
314
+ .b8 99
315
+ .b8 116
316
+ .b8 111
317
+ .b8 114
318
+ .b8 95
319
+ .b8 114
320
+ .b8 111
321
+ .b8 111
322
+ .b8 116
323
+ .b8 47
324
+ .b8 50
325
+ .b8 113
326
+ .b8 0
327
+ .b8 1
328
+ .b64 $L__func_begin0
329
+ .b64 $L__func_end0
330
+ .b8 2
331
+ .b8 116
332
+ .b8 114
333
+ .b8 105
334
+ .b8 116
335
+ .b8 111
336
+ .b8 110
337
+ .b8 95
338
+ .b8 95
339
+ .b8 48
340
+ .b8 100
341
+ .b8 49
342
+ .b8 100
343
+ .b8 50
344
+ .b8 100
345
+ .b8 51
346
+ .b8 52
347
+ .b8 101
348
+ .b8 0
349
+ .b8 116
350
+ .b8 114
351
+ .b8 105
352
+ .b8 116
353
+ .b8 111
354
+ .b8 110
355
+ .b8 95
356
+ .b8 95
357
+ .b8 48
358
+ .b8 100
359
+ .b8 49
360
+ .b8 100
361
+ .b8 50
362
+ .b8 100
363
+ .b8 51
364
+ .b8 52
365
+ .b8 101
366
+ .b8 0
367
+ .b8 1
368
+ .b8 18
369
+ .b8 1
370
+ .b8 1
371
+ .b8 3
372
+ .b64 $L__func_begin0
373
+ .b64 $L__func_end0
374
+ .b8 1
375
+ .b8 156
376
+ .b32 125
377
+ .b8 4
378
+ .b32 125
379
+ .b64 $L__tmp1
380
+ .b64 $L__tmp6
381
+ .b8 2
382
+ .b8 32
383
+ .b8 24
384
+ .b8 5
385
+ .b32 125
386
+ .b64 $L__tmp2
387
+ .b64 $L__tmp7
388
+ .b8 2
389
+ .b8 32
390
+ .b8 24
391
+ .b8 4
392
+ .b32 125
393
+ .b64 $L__tmp2
394
+ .b64 $L__tmp7
395
+ .b8 2
396
+ .b8 243
397
+ .b8 36
398
+ .b8 0
399
+ .b8 4
400
+ .b32 125
401
+ .b64 $L__tmp7
402
+ .b64 $L__tmp12
403
+ .b8 2
404
+ .b8 35
405
+ .b8 24
406
+ .b8 5
407
+ .b32 125
408
+ .b64 $L__tmp8
409
+ .b64 $L__tmp13
410
+ .b8 2
411
+ .b8 35
412
+ .b8 24
413
+ .b8 4
414
+ .b32 125
415
+ .b64 $L__tmp8
416
+ .b64 $L__tmp13
417
+ .b8 2
418
+ .b8 243
419
+ .b8 36
420
+ .b8 0
421
+ .b8 0
422
+ .b8 0
423
+ }
424
+ .section .debug_pubnames
425
+ {
426
+ .b32 $L__pubNames_end0-$L__pubNames_start0
427
+ $L__pubNames_start0:
428
+ .b8 2
429
+ .b8 0
430
+ .b32 .debug_info
431
+ .b32 337
432
+ .b32 125
433
+ .b8 116
434
+ .b8 114
435
+ .b8 105
436
+ .b8 116
437
+ .b8 111
438
+ .b8 110
439
+ .b8 95
440
+ .b8 95
441
+ .b8 48
442
+ .b8 100
443
+ .b8 49
444
+ .b8 100
445
+ .b8 50
446
+ .b8 100
447
+ .b8 51
448
+ .b8 52
449
+ .b8 101
450
+ .b8 0
451
+ .b32 0
452
+ $L__pubNames_end0:
453
+ }
454
+ .section .debug_pubtypes
455
+ {
456
+ .b32 $L__pubTypes_end0-$L__pubTypes_start0
457
+ $L__pubTypes_start0:
458
+ .b8 2
459
+ .b8 0
460
+ .b32 .debug_info
461
+ .b32 337
462
+ .b32 0
463
+ $L__pubTypes_end0:
464
+ }
465
+ .section .debug_loc { }
.triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttgir ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 2], order = [0, 1], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
2
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
3
+ tt.func public @triton__0d1d2d34e(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg3: i32, %arg4: i32 {tt.max_divisibility = 8 : i32}) attributes {noinline = false} {
4
+ %cst = arith.constant dense<0> : tensor<1x8xi64, #blocked>
5
+ %cst_0 = arith.constant dense<0.000000e+00> : tensor<1x8xf32, #blocked>
6
+ %cst_1 = arith.constant dense<8> : tensor<1x8xi32, #blocked>
7
+ %c0_i32 = arith.constant 0 : i32
8
+ %0 = tt.make_range {end = 8 : i32, start = 0 : i32} : tensor<8xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
9
+ %1 = tt.expand_dims %0 {axis = 0 : i32} : (tensor<8xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x8xi32, #blocked>
10
+ %2 = arith.cmpi slt, %1, %cst_1 : tensor<1x8xi32, #blocked>
11
+ %3 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<1x8x!tt.ptr<f32, 1>, #blocked>
12
+ %4 = tt.addptr %3, %1 : tensor<1x8x!tt.ptr<f32, 1>, #blocked>, tensor<1x8xi32, #blocked>
13
+ %5 = tt.load %4, %2, %cst_0 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xf32, #blocked>
14
+ %6 = tt.splat %arg2 : (!tt.ptr<i64, 1>) -> tensor<1x8x!tt.ptr<i64, 1>, #blocked>
15
+ %7 = tt.addptr %6, %1 : tensor<1x8x!tt.ptr<i64, 1>, #blocked>, tensor<1x8xi32, #blocked>
16
+ %8 = tt.load %7, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xi64, #blocked>
17
+ %9 = arith.select %2, %5, %cst_0 : tensor<1x8xi1, #blocked>, tensor<1x8xf32, #blocked>
18
+ %10 = "tt.reduce"(%9) <{axis = 1 : i32}> ({
19
+ ^bb0(%arg5: f32, %arg6: f32):
20
+ %19 = arith.addf %arg5, %arg6 : f32
21
+ tt.reduce.return %19 : f32
22
+ }) : (tensor<1x8xf32, #blocked>) -> tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
23
+ %11 = tt.expand_dims %10 {axis = 1 : i32} : (tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<1x1xf32, #blocked>
24
+ %12 = arith.select %2, %8, %cst : tensor<1x8xi1, #blocked>, tensor<1x8xi64, #blocked>
25
+ %13 = "tt.reduce"(%12) <{axis = 1 : i32}> ({
26
+ ^bb0(%arg5: i64, %arg6: i64):
27
+ %19 = arith.addi %arg5, %arg6 : i64
28
+ tt.reduce.return %19 : i64
29
+ }) : (tensor<1x8xi64, #blocked>) -> tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
30
+ %14 = tt.expand_dims %13 {axis = 1 : i32} : (tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<1x1xi64, #blocked>
31
+ %15 = arith.sitofp %14 : tensor<1x1xi64, #blocked> to tensor<1x1xf32, #blocked>
32
+ %16 = arith.divf %11, %15 : tensor<1x1xf32, #blocked>
33
+ gpu.barrier
34
+ %17 = tt.addptr %arg0, %c0_i32 : !tt.ptr<f32, 1>, i32
35
+ %18 = tt.splat %17 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>, #blocked>
36
+ tt.store %18, %16 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32, #blocked>
37
+ tt.return
38
+ }
39
+ }
.triton/dump/4710f23a3addbad00b260d7a02366fe0/triton_.ttir ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ module {
2
+ tt.func public @triton__0d1d2d34e(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg3: i32, %arg4: i32 {tt.max_divisibility = 8 : i32}) attributes {noinline = false} {
3
+ %c0_i32 = arith.constant 0 : i32
4
+ %cst = arith.constant dense<0> : tensor<1x8xi64>
5
+ %cst_0 = arith.constant dense<0.000000e+00> : tensor<1x8xf32>
6
+ %cst_1 = arith.constant dense<8> : tensor<1x8xi32>
7
+ %0 = tt.make_range {end = 8 : i32, start = 0 : i32} : tensor<8xi32>
8
+ %1 = tt.expand_dims %0 {axis = 0 : i32} : (tensor<8xi32>) -> tensor<1x8xi32>
9
+ %2 = arith.cmpi slt, %1, %cst_1 : tensor<1x8xi32>
10
+ %3 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<1x8x!tt.ptr<f32, 1>>
11
+ %4 = tt.addptr %3, %1 : tensor<1x8x!tt.ptr<f32, 1>>, tensor<1x8xi32>
12
+ %5 = tt.load %4, %2, %cst_0 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xf32>
13
+ %6 = tt.splat %arg2 : (!tt.ptr<i64, 1>) -> tensor<1x8x!tt.ptr<i64, 1>>
14
+ %7 = tt.addptr %6, %1 : tensor<1x8x!tt.ptr<i64, 1>>, tensor<1x8xi32>
15
+ %8 = tt.load %7, %2, %cst {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<1x8xi64>
16
+ %9 = arith.select %2, %5, %cst_0 : tensor<1x8xi1>, tensor<1x8xf32>
17
+ %10 = "tt.reduce"(%9) <{axis = 1 : i32}> ({
18
+ ^bb0(%arg5: f32, %arg6: f32):
19
+ %19 = arith.addf %arg5, %arg6 : f32
20
+ tt.reduce.return %19 : f32
21
+ }) : (tensor<1x8xf32>) -> tensor<1xf32>
22
+ %11 = tt.expand_dims %10 {axis = 1 : i32} : (tensor<1xf32>) -> tensor<1x1xf32>
23
+ %12 = arith.select %2, %8, %cst : tensor<1x8xi1>, tensor<1x8xi64>
24
+ %13 = "tt.reduce"(%12) <{axis = 1 : i32}> ({
25
+ ^bb0(%arg5: i64, %arg6: i64):
26
+ %19 = arith.addi %arg5, %arg6 : i64
27
+ tt.reduce.return %19 : i64
28
+ }) : (tensor<1x8xi64>) -> tensor<1xi64>
29
+ %14 = tt.expand_dims %13 {axis = 1 : i32} : (tensor<1xi64>) -> tensor<1x1xi64>
30
+ %15 = arith.sitofp %14 : tensor<1x1xi64> to tensor<1x1xf32>
31
+ %16 = arith.divf %11, %15 : tensor<1x1xf32>
32
+ gpu.barrier
33
+ %17 = tt.addptr %arg0, %c0_i32 : !tt.ptr<f32, 1>, i32
34
+ %18 = tt.splat %17 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>>
35
+ tt.store %18, %16 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32>
36
+ tt.return
37
+ }
38
+ }
.triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.cubin ADDED
Binary file (31.3 kB). View file
 
.triton/dump/473cf6e25c3e63117cd59fc0ed04b89f/triton_.llir ADDED
@@ -0,0 +1,550 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+
4
+ @assertFunc_1 = internal constant [25 x i8] c"_call_with_frames_removed"
5
+ @assertFile_1 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
6
+ @assertMessage_1 = internal constant [39 x i8] c"index out of bounds: 0 <= tmp16 < 50257"
7
+ @assertFunc_0 = internal constant [25 x i8] c"_call_with_frames_removed"
8
+ @assertFile_0 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
9
+ @assertMessage_0 = internal constant [38 x i8] c"index out of bounds: 0 <= tmp3 < 50257"
10
+ @global_smem = external addrspace(3) global [0 x i8]
11
+ @.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
12
+
13
+ declare void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr
14
+
15
+ define void @triton__0d1d2d3d4d5d6de7de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, i32 %6, i32 %7) local_unnamed_addr !dbg !7 {
16
+ %9 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
17
+ %10 = and i32 %9, 31, !dbg !10
18
+ %11 = lshr i32 %9, 5, !dbg !10
19
+ %12 = lshr i32 %9, 6, !dbg !10
20
+ %13 = and i32 %12, 1, !dbg !10
21
+ %14 = and i32 %9, 1, !dbg !10
22
+ %15 = and i32 %11, 1, !dbg !11
23
+ %urem = shl i32 %9, 2, !dbg !11
24
+ %16 = and i32 %urem, 252, !dbg !11
25
+ %17 = shl i32 %9, 1, !dbg !11
26
+ %18 = and i32 %17, 254, !dbg !11
27
+ %19 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !12
28
+ %20 = shl i32 %19, 1, !dbg !13
29
+ %21 = or i32 %20, %13, !dbg !14
30
+ %22 = or i32 %20, %14, !dbg !14
31
+ %23 = sext i32 %21 to i64, !dbg !15
32
+ %24 = getelementptr i64, ptr addrspace(1) %0, i64 %23, !dbg !15
33
+ %25 = sext i32 %22 to i64, !dbg !15
34
+ %26 = getelementptr i64, ptr addrspace(1) %0, i64 %25, !dbg !15
35
+ %27 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
36
+ %28 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
37
+ %29 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
38
+ %30 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %24, i1 true) #6, !dbg !16
39
+ %31 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %26, i1 true) #6, !dbg !16
40
+ %32 = srem i32 %21, 512, !dbg !17
41
+ %33 = shl nsw i32 %32, 8, !dbg !18
42
+ %34 = or i32 %33, %16, !dbg !19
43
+ %35 = sext i32 %34 to i64, !dbg !20
44
+ %36 = getelementptr float, ptr addrspace(1) %2, i64 %35, !dbg !20
45
+ %37 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %36, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !21
46
+ %38 = extractvalue { i32, i32, i32, i32 } %37, 0, !dbg !21
47
+ %39 = extractvalue { i32, i32, i32, i32 } %37, 1, !dbg !21
48
+ %40 = extractvalue { i32, i32, i32, i32 } %37, 2, !dbg !21
49
+ %41 = extractvalue { i32, i32, i32, i32 } %37, 3, !dbg !21
50
+ %42 = insertelement <2 x i32> poison, i32 %39, i64 0, !dbg !21
51
+ %43 = insertelement <2 x i32> %42, i32 %38, i64 1, !dbg !21
52
+ %44 = bitcast <2 x i32> %43 to <2 x float>, !dbg !21
53
+ %45 = bitcast i32 %40 to float, !dbg !21
54
+ %46 = bitcast i32 %41 to float, !dbg !21
55
+ %47 = shl i32 %21, 8, !dbg !22
56
+ %48 = or i32 %47, %16, !dbg !23
57
+ %49 = sext i32 %48 to i64, !dbg !24
58
+ %50 = getelementptr i16, ptr addrspace(1) %3, i64 %49, !dbg !24
59
+ %51 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.L1::evict_last.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %50, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !25
60
+ %52 = extractvalue { i32, i32 } %51, 0, !dbg !25
61
+ %53 = extractvalue { i32, i32 } %51, 1, !dbg !25
62
+ %54 = trunc i32 %52 to i16, !dbg !25
63
+ %extelt.offset = lshr i32 %52, 16, !dbg !25
64
+ %55 = trunc i32 %extelt.offset to i16, !dbg !25
65
+ %56 = trunc i32 %53 to i16, !dbg !25
66
+ %extelt.offset1 = lshr i32 %53, 16, !dbg !25
67
+ %57 = trunc i32 %extelt.offset1 to i16, !dbg !25
68
+ %58 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %54) #6, !dbg !26
69
+ %59 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %55) #6, !dbg !26
70
+ %60 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %56) #6, !dbg !26
71
+ %61 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %57) #6, !dbg !26
72
+ %62 = add i64 %31, 50257, !dbg !27
73
+ %63 = icmp slt i64 %27, 0, !dbg !28
74
+ %64 = icmp slt i64 %31, 0, !dbg !28
75
+ %65 = select i1 %64, i64 %62, i64 %31, !dbg !29
76
+ %66 = icmp ugt i64 %65, 50256, !dbg !30
77
+ br i1 %66, label %67, label %68, !dbg !31
78
+
79
+ 67: ; preds = %8
80
+ tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 883, ptr nonnull @assertFunc_0, i64 1), !dbg !31
81
+ br label %68, !dbg !31
82
+
83
+ 68: ; preds = %67, %8
84
+ %69 = shl i64 %27, 8, !dbg !32
85
+ %70 = add i64 %69, 12865792, !dbg !32
86
+ %71 = select i1 %63, i64 %70, i64 %69, !dbg !32
87
+ %72 = zext nneg i32 %16 to i64
88
+ %73 = or i64 %71, %72, !dbg !33
89
+ %74 = getelementptr float, ptr addrspace(1) %1, i64 %73, !dbg !34
90
+ %75 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %74, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !35
91
+ %76 = extractvalue { i32, i32, i32, i32 } %75, 0, !dbg !35
92
+ %77 = extractvalue { i32, i32, i32, i32 } %75, 1, !dbg !35
93
+ %78 = extractvalue { i32, i32, i32, i32 } %75, 2, !dbg !35
94
+ %79 = extractvalue { i32, i32, i32, i32 } %75, 3, !dbg !35
95
+ %80 = bitcast i32 %78 to float, !dbg !35
96
+ %81 = bitcast i32 %79 to float, !dbg !35
97
+ %82 = fadd float %45, %80, !dbg !36
98
+ %83 = fadd float %46, %81, !dbg !36
99
+ %84 = fadd float %60, %82, !dbg !37
100
+ %85 = fadd float %61, %83, !dbg !37
101
+ %86 = insertelement <2 x i32> poison, i32 %77, i64 0, !dbg !35
102
+ %87 = insertelement <2 x i32> %86, i32 %76, i64 1, !dbg !35
103
+ %88 = bitcast <2 x i32> %87 to <2 x float>, !dbg !35
104
+ %89 = fadd <2 x float> %44, %88, !dbg !36
105
+ %90 = insertelement <2 x float> poison, float %59, i64 0, !dbg !37
106
+ %91 = insertelement <2 x float> %90, float %58, i64 1, !dbg !37
107
+ %92 = fadd <2 x float> %91, %89, !dbg !37
108
+ %93 = fadd <2 x float> %92, zeroinitializer, !dbg !38
109
+ %94 = fadd float %84, 0.000000e+00, !dbg !38
110
+ %95 = fadd float %85, 0.000000e+00, !dbg !38
111
+ %96 = extractelement <2 x float> %93, i64 1, !dbg !42
112
+ %97 = extractelement <2 x float> %92, i64 1, !dbg !46
113
+ %98 = fsub float %97, %96, !dbg !47
114
+ %99 = extractelement <2 x float> %93, i64 0, !dbg !42
115
+ %100 = extractelement <2 x float> %92, i64 0, !dbg !46
116
+ %101 = fsub float %100, %99, !dbg !47
117
+ %102 = fsub float %84, %94, !dbg !47
118
+ %103 = fsub float %85, %95, !dbg !47
119
+ %104 = fmul float %97, %98, !dbg !46
120
+ %105 = fmul float %100, %101, !dbg !46
121
+ %106 = fmul float %84, %102, !dbg !46
122
+ %107 = fmul float %85, %103, !dbg !46
123
+ %108 = fadd float %104, 0.000000e+00, !dbg !48
124
+ %109 = fadd float %105, 0.000000e+00, !dbg !48
125
+ %110 = fadd float %106, 0.000000e+00, !dbg !48
126
+ %111 = fadd float %107, 0.000000e+00, !dbg !48
127
+ %112 = fsub float %99, %96, !dbg !42
128
+ %113 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 2.000000e+00) #6, !dbg !49
129
+ %114 = fmul float %113, %112, !dbg !50
130
+ %115 = fadd float %96, %114, !dbg !51
131
+ %116 = fadd float %108, %109, !dbg !52
132
+ %117 = fmul float %112, %112, !dbg !53
133
+ %118 = fmul float %113, %117, !dbg !54
134
+ %119 = fadd float %118, %116, !dbg !55
135
+ %120 = fsub float %94, %115, !dbg !42
136
+ %121 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 3.000000e+00) #6, !dbg !49
137
+ %122 = fmul float %121, %120, !dbg !50
138
+ %123 = fadd float %115, %122, !dbg !51
139
+ %124 = fadd float %110, %119, !dbg !52
140
+ %125 = fmul float %120, %120, !dbg !53
141
+ %126 = fmul float %125, 2.000000e+00, !dbg !56
142
+ %127 = fmul float %121, %126, !dbg !54
143
+ %128 = fadd float %124, %127, !dbg !55
144
+ %129 = fsub float %95, %123, !dbg !42
145
+ %130 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float 1.000000e+00, float 4.000000e+00) #6, !dbg !49
146
+ %131 = fmul float %130, %129, !dbg !50
147
+ %132 = fadd float %123, %131, !dbg !51
148
+ %133 = fadd float %111, %128, !dbg !52
149
+ %134 = fmul float %129, %129, !dbg !53
150
+ %135 = fmul float %134, 3.000000e+00, !dbg !56
151
+ %136 = fmul float %130, %135, !dbg !54
152
+ %137 = fadd float %133, %136, !dbg !55
153
+ %138 = bitcast float %132 to i32, !dbg !57
154
+ %139 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %138, i32 16, i32 31), !dbg !57
155
+ %140 = bitcast i32 %139 to float, !dbg !57
156
+ %141 = bitcast float %137 to i32, !dbg !57
157
+ %142 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %141, i32 16, i32 31), !dbg !57
158
+ %143 = bitcast i32 %142 to float, !dbg !57
159
+ %144 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 1082130432, i32 16, i32 31), !dbg !57
160
+ %145 = bitcast i32 %144 to float, !dbg !57
161
+ %146 = fsub float %140, %132, !dbg !42
162
+ %147 = fadd float %145, 4.000000e+00, !dbg !59
163
+ %148 = fcmp oeq float %147, 0.000000e+00, !dbg !60
164
+ %149 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %145, float %147) #6, !dbg !49
165
+ %150 = select i1 %148, float 0.000000e+00, float %149, !dbg !61
166
+ %151 = fmul float %150, %146, !dbg !50
167
+ %152 = fadd float %132, %151, !dbg !51
168
+ %153 = fadd float %137, %143, !dbg !52
169
+ %154 = fmul float %146, %146, !dbg !53
170
+ %155 = fmul float %154, 4.000000e+00, !dbg !56
171
+ %156 = fmul float %150, %155, !dbg !54
172
+ %157 = fadd float %153, %156, !dbg !55
173
+ %158 = bitcast float %152 to i32, !dbg !57
174
+ %159 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %158, i32 8, i32 31), !dbg !57
175
+ %160 = bitcast i32 %159 to float, !dbg !57
176
+ %161 = bitcast float %157 to i32, !dbg !57
177
+ %162 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %161, i32 8, i32 31), !dbg !57
178
+ %163 = bitcast i32 %162 to float, !dbg !57
179
+ %164 = bitcast float %147 to i32, !dbg !57
180
+ %165 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %164, i32 8, i32 31), !dbg !57
181
+ %166 = bitcast i32 %165 to float, !dbg !57
182
+ %167 = fsub float %160, %152, !dbg !42
183
+ %168 = fadd float %147, %166, !dbg !59
184
+ %169 = fcmp oeq float %168, 0.000000e+00, !dbg !60
185
+ %170 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %166, float %168) #6, !dbg !49
186
+ %171 = select i1 %169, float 0.000000e+00, float %170, !dbg !61
187
+ %172 = fmul float %171, %167, !dbg !50
188
+ %173 = fadd float %152, %172, !dbg !51
189
+ %174 = fadd float %157, %163, !dbg !52
190
+ %175 = fmul float %167, %167, !dbg !53
191
+ %176 = fmul float %147, %175, !dbg !56
192
+ %177 = fmul float %171, %176, !dbg !54
193
+ %178 = fadd float %174, %177, !dbg !55
194
+ %179 = bitcast float %173 to i32, !dbg !57
195
+ %180 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %179, i32 4, i32 31), !dbg !57
196
+ %181 = bitcast i32 %180 to float, !dbg !57
197
+ %182 = bitcast float %178 to i32, !dbg !57
198
+ %183 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %182, i32 4, i32 31), !dbg !57
199
+ %184 = bitcast i32 %183 to float, !dbg !57
200
+ %185 = bitcast float %168 to i32, !dbg !57
201
+ %186 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %185, i32 4, i32 31), !dbg !57
202
+ %187 = bitcast i32 %186 to float, !dbg !57
203
+ %188 = fsub float %181, %173, !dbg !42
204
+ %189 = fadd float %168, %187, !dbg !59
205
+ %190 = fcmp oeq float %189, 0.000000e+00, !dbg !60
206
+ %191 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %187, float %189) #6, !dbg !49
207
+ %192 = select i1 %190, float 0.000000e+00, float %191, !dbg !61
208
+ %193 = fmul float %192, %188, !dbg !50
209
+ %194 = fadd float %173, %193, !dbg !51
210
+ %195 = fadd float %178, %184, !dbg !52
211
+ %196 = fmul float %188, %188, !dbg !53
212
+ %197 = fmul float %168, %196, !dbg !56
213
+ %198 = fmul float %192, %197, !dbg !54
214
+ %199 = fadd float %195, %198, !dbg !55
215
+ %200 = bitcast float %194 to i32, !dbg !57
216
+ %201 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %200, i32 2, i32 31), !dbg !57
217
+ %202 = bitcast i32 %201 to float, !dbg !57
218
+ %203 = bitcast float %199 to i32, !dbg !57
219
+ %204 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %203, i32 2, i32 31), !dbg !57
220
+ %205 = bitcast i32 %204 to float, !dbg !57
221
+ %206 = bitcast float %189 to i32, !dbg !57
222
+ %207 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %206, i32 2, i32 31), !dbg !57
223
+ %208 = bitcast i32 %207 to float, !dbg !57
224
+ %209 = fsub float %202, %194, !dbg !42
225
+ %210 = fadd float %189, %208, !dbg !59
226
+ %211 = fcmp oeq float %210, 0.000000e+00, !dbg !60
227
+ %212 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %208, float %210) #6, !dbg !49
228
+ %213 = select i1 %211, float 0.000000e+00, float %212, !dbg !61
229
+ %214 = fmul float %213, %209, !dbg !50
230
+ %215 = fadd float %194, %214, !dbg !51
231
+ %216 = fadd float %199, %205, !dbg !52
232
+ %217 = fmul float %209, %209, !dbg !53
233
+ %218 = fmul float %189, %217, !dbg !56
234
+ %219 = fmul float %213, %218, !dbg !54
235
+ %220 = fadd float %216, %219, !dbg !55
236
+ %221 = bitcast float %215 to i32, !dbg !57
237
+ %222 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %221, i32 1, i32 31), !dbg !57
238
+ %223 = bitcast i32 %222 to float, !dbg !57
239
+ %224 = bitcast float %220 to i32, !dbg !57
240
+ %225 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %224, i32 1, i32 31), !dbg !57
241
+ %226 = bitcast i32 %225 to float, !dbg !57
242
+ %227 = bitcast float %210 to i32, !dbg !57
243
+ %228 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %227, i32 1, i32 31), !dbg !57
244
+ %229 = bitcast i32 %228 to float, !dbg !57
245
+ %230 = fsub float %223, %215, !dbg !42
246
+ %231 = fadd float %210, %229, !dbg !59
247
+ %232 = fcmp oeq float %231, 0.000000e+00, !dbg !60
248
+ %233 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %229, float %231) #6, !dbg !49
249
+ %234 = select i1 %232, float 0.000000e+00, float %233, !dbg !61
250
+ %235 = fmul float %234, %230, !dbg !50
251
+ %236 = fadd float %215, %235, !dbg !51
252
+ %237 = fadd float %220, %226, !dbg !52
253
+ %238 = fmul float %230, %230, !dbg !53
254
+ %239 = fmul float %210, %238, !dbg !56
255
+ %240 = fmul float %234, %239, !dbg !54
256
+ %241 = fadd float %237, %240, !dbg !55
257
+ %242 = icmp eq i32 %10, 0, !dbg !57
258
+ %243 = shl nuw nsw i32 %13, 1, !dbg !57
259
+ %244 = or i32 %243, %15, !dbg !57
260
+ %245 = zext nneg i32 %244 to i64, !dbg !57
261
+ %246 = getelementptr float, ptr addrspace(3) @global_smem, i64 %245, !dbg !57
262
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %246, float %236, i1 %242) #6, !dbg !57
263
+ %247 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 16), i64 %245, !dbg !57
264
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %247, float %241, i1 %242) #6, !dbg !57
265
+ %248 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 32), i64 %245, !dbg !57
266
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %248, float %231, i1 %242) #6, !dbg !57
267
+ tail call void @llvm.nvvm.barrier0(), !dbg !57
268
+ %249 = icmp slt i32 %9, 4, !dbg !57
269
+ %250 = sext i32 %9 to i64, !dbg !57
270
+ %251 = getelementptr float, ptr addrspace(3) @global_smem, i64 %250, !dbg !57
271
+ %252 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %251, i1 %249) #6, !dbg !57
272
+ %253 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 16), i64 %250, !dbg !57
273
+ %254 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %253, i1 %249) #6, !dbg !57
274
+ %255 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 32), i64 %250, !dbg !57
275
+ %256 = tail call float asm sideeffect "@$2 ld.shared.b32 $0, [ $1 + 0 ];", "=r,r,b"(ptr addrspace(3) %255, i1 %249) #6, !dbg !57
276
+ %257 = bitcast float %252 to i32, !dbg !57
277
+ %258 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %257, i32 1, i32 31), !dbg !57
278
+ %259 = bitcast i32 %258 to float, !dbg !57
279
+ %260 = bitcast float %254 to i32, !dbg !57
280
+ %261 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %260, i32 1, i32 31), !dbg !57
281
+ %262 = bitcast i32 %261 to float, !dbg !57
282
+ %263 = bitcast float %256 to i32, !dbg !57
283
+ %264 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %263, i32 1, i32 31), !dbg !57
284
+ %265 = bitcast i32 %264 to float, !dbg !57
285
+ %266 = fsub float %259, %252, !dbg !42
286
+ %267 = fadd float %256, %265, !dbg !59
287
+ %268 = fcmp oeq float %267, 0.000000e+00, !dbg !60
288
+ %269 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %265, float %267) #6, !dbg !49
289
+ %270 = select i1 %268, float 0.000000e+00, float %269, !dbg !61
290
+ %271 = fmul float %266, %270, !dbg !50
291
+ %272 = fadd float %252, %271, !dbg !51
292
+ %273 = fadd float %254, %262, !dbg !52
293
+ %274 = fmul float %266, %266, !dbg !53
294
+ %275 = fmul float %256, %274, !dbg !56
295
+ %276 = fmul float %275, %270, !dbg !54
296
+ %277 = fadd float %273, %276, !dbg !55
297
+ %278 = icmp eq i32 %14, 0, !dbg !57
298
+ %279 = and i1 %249, %278, !dbg !57
299
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %251, float %272, i1 %279) #6, !dbg !57
300
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %253, float %277, i1 %279) #6, !dbg !57
301
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %255, float %267, i1 %279) #6, !dbg !57
302
+ tail call void @llvm.nvvm.barrier0(), !dbg !57
303
+ %280 = zext nneg i32 %243 to i64, !dbg !57
304
+ %281 = getelementptr float, ptr addrspace(3) @global_smem, i64 %280, !dbg !57
305
+ %282 = load float, ptr addrspace(3) %281, align 4, !dbg !57
306
+ %283 = getelementptr float, ptr addrspace(3) getelementptr ([0 x i8], ptr addrspace(3) @global_smem, i64 0, i64 16), i64 %280, !dbg !57
307
+ %284 = load float, ptr addrspace(3) %283, align 4, !dbg !57
308
+ %285 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %36, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !62
309
+ %286 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.L1::evict_first.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %50, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !63
310
+ %287 = extractvalue { i32, i32 } %286, 0, !dbg !63
311
+ %288 = extractvalue { i32, i32 } %286, 1, !dbg !63
312
+ %289 = trunc i32 %287 to i16, !dbg !63
313
+ %extelt.offset2 = lshr i32 %287, 16, !dbg !63
314
+ %290 = trunc i32 %extelt.offset2 to i16, !dbg !63
315
+ %291 = trunc i32 %288 to i16, !dbg !63
316
+ %extelt.offset3 = lshr i32 %288, 16, !dbg !63
317
+ %292 = trunc i32 %extelt.offset3 to i16, !dbg !63
318
+ %293 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %289) #6, !dbg !64
319
+ %294 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %290) #6, !dbg !64
320
+ %295 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %291) #6, !dbg !64
321
+ %296 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %292) #6, !dbg !64
322
+ %297 = zext nneg i32 %18 to i64, !dbg !65
323
+ %298 = getelementptr float, ptr addrspace(1) %4, i64 %297, !dbg !65
324
+ %299 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09@$3 ld.global.L1::evict_last.v2.b32 { $0, $1 }, [ $2 + 0 ];\0A\09@!$5 mov.u32 $0, $4;\0A\09@!$7 mov.u32 $1, $6;", "=r,=r,l,b,r,b,r,b"(ptr addrspace(1) %298, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !66
325
+ br i1 %66, label %300, label %301, !dbg !67
326
+
327
+ 300: ; preds = %68
328
+ tail call void @__assertfail(ptr nonnull @assertMessage_1, ptr nonnull @assertFile_1, i32 883, ptr nonnull @assertFunc_1, i64 1), !dbg !67
329
+ br label %301, !dbg !67
330
+
331
+ 301: ; preds = %300, %68
332
+ %302 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %74, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !68
333
+ %303 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
334
+ %304 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
335
+ %305 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
336
+ %306 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %284, float 2.560000e+02) #6, !dbg !69
337
+ %307 = fadd float %303, 0x3EE4F8B580000000, !dbg !70
338
+ %308 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
339
+ %.not.i = icmp eq i32 %308, 0, !dbg !71
340
+ br i1 %.not.i, label %311, label %309, !dbg !71
341
+
342
+ 309: ; preds = %301
343
+ %310 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %307), !dbg !71
344
+ br label %__nv_rsqrtf.exit, !dbg !71
345
+
346
+ 311: ; preds = %301
347
+ %312 = tail call float @llvm.nvvm.rsqrt.approx.f(float %307), !dbg !71
348
+ br label %__nv_rsqrtf.exit, !dbg !71
349
+
350
+ __nv_rsqrtf.exit: ; preds = %309, %311
351
+ %.0.i = phi float [ %310, %309 ], [ %312, %311 ], !dbg !71
352
+ %313 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
353
+ %314 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
354
+ %315 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !71
355
+ %316 = extractvalue { i32, i32, i32, i32 } %302, 3, !dbg !68
356
+ %317 = bitcast i32 %316 to float, !dbg !68
357
+ %318 = extractvalue { i32, i32, i32, i32 } %285, 3, !dbg !62
358
+ %319 = bitcast i32 %318 to float, !dbg !62
359
+ %320 = fadd float %319, %317, !dbg !72
360
+ %321 = fadd float %296, %320, !dbg !73
361
+ %322 = fsub float %321, %282, !dbg !74
362
+ %323 = extractvalue { i32, i32, i32, i32 } %302, 2, !dbg !68
363
+ %324 = bitcast i32 %323 to float, !dbg !68
364
+ %325 = extractvalue { i32, i32, i32, i32 } %285, 2, !dbg !62
365
+ %326 = bitcast i32 %325 to float, !dbg !62
366
+ %327 = fadd float %326, %324, !dbg !72
367
+ %328 = fadd float %295, %327, !dbg !73
368
+ %329 = fsub float %328, %282, !dbg !74
369
+ %330 = extractvalue { i32, i32, i32, i32 } %302, 1, !dbg !68
370
+ %331 = bitcast i32 %330 to float, !dbg !68
371
+ %332 = extractvalue { i32, i32, i32, i32 } %285, 1, !dbg !62
372
+ %333 = bitcast i32 %332 to float, !dbg !62
373
+ %334 = fadd float %333, %331, !dbg !72
374
+ %335 = fadd float %294, %334, !dbg !73
375
+ %336 = fsub float %335, %282, !dbg !74
376
+ %337 = extractvalue { i32, i32, i32, i32 } %302, 0, !dbg !68
377
+ %338 = bitcast i32 %337 to float, !dbg !68
378
+ %339 = extractvalue { i32, i32, i32, i32 } %285, 0, !dbg !62
379
+ %340 = bitcast i32 %339 to float, !dbg !62
380
+ %341 = fadd float %340, %338, !dbg !72
381
+ %342 = fadd float %293, %341, !dbg !73
382
+ %343 = fsub float %342, %282, !dbg !74
383
+ %344 = extractvalue { i32, i32 } %299, 0, !dbg !66
384
+ %345 = extractvalue { i32, i32 } %299, 1, !dbg !66
385
+ %346 = fmul float %343, %.0.i, !dbg !75
386
+ %347 = fmul float %336, %.0.i, !dbg !75
387
+ %348 = fmul float %329, %.0.i, !dbg !75
388
+ %349 = fmul float %322, %.0.i, !dbg !75
389
+ tail call void @llvm.nvvm.barrier0(), !dbg !76
390
+ %350 = getelementptr float, ptr addrspace(3) @global_smem, i64 %297, !dbg !76
391
+ %351 = insertelement <2 x i32> undef, i32 %344, i64 0, !dbg !76
392
+ %352 = insertelement <2 x i32> %351, i32 %345, i64 1, !dbg !76
393
+ store <2 x i32> %352, ptr addrspace(3) %350, align 8, !dbg !76
394
+ tail call void @llvm.nvvm.barrier0(), !dbg !76
395
+ %353 = getelementptr float, ptr addrspace(3) @global_smem, i64 %72, !dbg !76
396
+ %354 = load float, ptr addrspace(3) %353, align 16, !dbg !76
397
+ %355 = getelementptr inbounds <4 x float>, ptr addrspace(3) %353, i64 0, i64 1, !dbg !76
398
+ %356 = load float, ptr addrspace(3) %355, align 4, !dbg !76
399
+ %357 = getelementptr inbounds <4 x float>, ptr addrspace(3) %353, i64 0, i64 2, !dbg !76
400
+ %358 = load float, ptr addrspace(3) %357, align 8, !dbg !76
401
+ %359 = getelementptr inbounds <4 x float>, ptr addrspace(3) %353, i64 0, i64 3, !dbg !76
402
+ %360 = load float, ptr addrspace(3) %359, align 4, !dbg !76
403
+ %361 = fmul float %346, %354, !dbg !76
404
+ %362 = fmul float %347, %356, !dbg !76
405
+ %363 = fmul float %348, %358, !dbg !76
406
+ %364 = fmul float %349, %360, !dbg !76
407
+ %365 = getelementptr i16, ptr addrspace(1) %5, i64 %49, !dbg !77
408
+ %366 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %361) #6, !dbg !78
409
+ %367 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %362) #6, !dbg !78
410
+ %368 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %363) #6, !dbg !78
411
+ %369 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %364) #6, !dbg !78
412
+ %370 = insertelement <2 x i16> undef, i16 %366, i64 0, !dbg !78
413
+ %371 = insertelement <2 x i16> %370, i16 %367, i64 1, !dbg !78
414
+ %372 = bitcast <2 x i16> %371 to i32, !dbg !78
415
+ %373 = insertelement <2 x i16> undef, i16 %368, i64 0, !dbg !78
416
+ %374 = insertelement <2 x i16> %373, i16 %369, i64 1, !dbg !78
417
+ %375 = bitcast <2 x i16> %374 to i32, !dbg !78
418
+ tail call void asm sideeffect "@$3 st.global.v2.b32 [ $2 + 0 ], { $0, $1 };", "r,r,l,b"(i32 %372, i32 %375, ptr addrspace(1) %365, i1 true) #6, !dbg !78
419
+ ret void, !dbg !79
420
+ }
421
+
422
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
423
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
424
+
425
+ ; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
426
+ declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
427
+
428
+ ; Function Attrs: convergent nocallback nounwind
429
+ declare void @llvm.nvvm.barrier0() #2
430
+
431
+ ; Function Attrs: alwaysinline nounwind
432
+ define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
433
+ %1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
434
+ %.not = icmp eq i32 %1, 0
435
+ br i1 %.not, label %4, label %2
436
+
437
+ 2: ; preds = %0
438
+ %3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
439
+ br label %6
440
+
441
+ 4: ; preds = %0
442
+ %5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
443
+ br label %6
444
+
445
+ 6: ; preds = %4, %2
446
+ %.0 = phi float [ %3, %2 ], [ %5, %4 ]
447
+ ret float %.0
448
+ }
449
+
450
+ declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
451
+
452
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
453
+ declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
454
+
455
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
456
+ declare float @llvm.nvvm.rsqrt.approx.f(float) #5
457
+
458
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
459
+ attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
460
+ attributes #2 = { convergent nocallback nounwind }
461
+ attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
462
+ attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
463
+ attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
464
+ attributes #6 = { nounwind }
465
+
466
+ !llvm.module.flags = !{!0, !1}
467
+ !llvm.dbg.cu = !{!2}
468
+ !nvvm.annotations = !{!4, !5, !5, !4}
469
+ !llvm.ident = !{!6}
470
+
471
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
472
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
473
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
474
+ !3 = !DIFile(filename: "ccig6fki6p4lxrdmgg6eudahiexcvueeol2p4qp532pvve2y463y.py", directory: "/tmp/torchinductor_root/ci")
475
+ !4 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"kernel", i32 1}
476
+ !5 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"maxntidx", i32 128}
477
+ !6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
478
+ !7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5d6de7de", linkageName: "triton__0d1d2d3d4d5d6de7de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
479
+ !8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
480
+ !9 = !{}
481
+ !10 = !DILocation(line: 22, column: 44, scope: !7)
482
+ !11 = !DILocation(line: 24, column: 33, scope: !7)
483
+ !12 = !DILocation(line: 21, column: 28, scope: !7)
484
+ !13 = !DILocation(line: 21, column: 33, scope: !7)
485
+ !14 = !DILocation(line: 22, column: 23, scope: !7)
486
+ !15 = !DILocation(line: 26, column: 30, scope: !7)
487
+ !16 = !DILocation(line: 26, column: 35, scope: !7)
488
+ !17 = !DILocation(line: 27, column: 18, scope: !7)
489
+ !18 = !DILocation(line: 35, column: 44, scope: !7)
490
+ !19 = !DILocation(line: 35, column: 40, scope: !7)
491
+ !20 = !DILocation(line: 35, column: 34, scope: !7)
492
+ !21 = !DILocation(line: 35, column: 50, scope: !7)
493
+ !22 = !DILocation(line: 36, column: 44, scope: !7)
494
+ !23 = !DILocation(line: 36, column: 40, scope: !7)
495
+ !24 = !DILocation(line: 36, column: 34, scope: !7)
496
+ !25 = !DILocation(line: 36, column: 50, scope: !7)
497
+ !26 = !DILocation(line: 36, column: 101, scope: !7)
498
+ !27 = !DILocation(line: 37, column: 22, scope: !7)
499
+ !28 = !DILocation(line: 38, column: 22, scope: !7)
500
+ !29 = !DILocation(line: 39, column: 36, scope: !7)
501
+ !30 = !DILocation(line: 40, column: 40, scope: !7)
502
+ !31 = !DILocation(line: 40, column: 55, scope: !7)
503
+ !32 = !DILocation(line: 41, column: 44, scope: !7)
504
+ !33 = !DILocation(line: 41, column: 40, scope: !7)
505
+ !34 = !DILocation(line: 41, column: 34, scope: !7)
506
+ !35 = !DILocation(line: 41, column: 52, scope: !7)
507
+ !36 = !DILocation(line: 42, column: 22, scope: !7)
508
+ !37 = !DILocation(line: 44, column: 22, scope: !7)
509
+ !38 = !DILocation(line: 98, column: 22, scope: !39, inlinedAt: !41)
510
+ !39 = distinct !DILexicalBlockFile(scope: !7, file: !40, discriminator: 0)
511
+ !40 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
512
+ !41 = !DILocation(line: 47, column: 41, scope: !39)
513
+ !42 = !DILocation(line: 108, column: 21, scope: !43, inlinedAt: !44)
514
+ !43 = distinct !DILexicalBlockFile(scope: !39, file: !40, discriminator: 0)
515
+ !44 = !DILocation(line: 120, column: 46, scope: !43, inlinedAt: !45)
516
+ !45 = !DILocation(line: 53, column: 44, scope: !43)
517
+ !46 = !DILocation(line: 101, column: 22, scope: !39, inlinedAt: !41)
518
+ !47 = !DILocation(line: 101, column: 30, scope: !39, inlinedAt: !41)
519
+ !48 = !DILocation(line: 101, column: 13, scope: !39, inlinedAt: !41)
520
+ !49 = !DILocation(line: 110, column: 60, scope: !43, inlinedAt: !44)
521
+ !50 = !DILocation(line: 112, column: 25, scope: !43, inlinedAt: !44)
522
+ !51 = !DILocation(line: 112, column: 17, scope: !43, inlinedAt: !44)
523
+ !52 = !DILocation(line: 113, column: 15, scope: !43, inlinedAt: !44)
524
+ !53 = !DILocation(line: 113, column: 30, scope: !43, inlinedAt: !44)
525
+ !54 = !DILocation(line: 113, column: 49, scope: !43, inlinedAt: !44)
526
+ !55 = !DILocation(line: 113, column: 22, scope: !43, inlinedAt: !44)
527
+ !56 = !DILocation(line: 113, column: 38, scope: !43, inlinedAt: !44)
528
+ !57 = !DILocation(line: 120, column: 46, scope: !39, inlinedAt: !58)
529
+ !58 = !DILocation(line: 53, column: 44, scope: !39)
530
+ !59 = !DILocation(line: 109, column: 28, scope: !43, inlinedAt: !44)
531
+ !60 = !DILocation(line: 110, column: 39, scope: !43, inlinedAt: !44)
532
+ !61 = !DILocation(line: 110, column: 49, scope: !43, inlinedAt: !44)
533
+ !62 = !DILocation(line: 62, column: 51, scope: !7)
534
+ !63 = !DILocation(line: 63, column: 51, scope: !7)
535
+ !64 = !DILocation(line: 63, column: 103, scope: !7)
536
+ !65 = !DILocation(line: 64, column: 35, scope: !7)
537
+ !66 = !DILocation(line: 64, column: 40, scope: !7)
538
+ !67 = !DILocation(line: 68, column: 57, scope: !7)
539
+ !68 = !DILocation(line: 69, column: 54, scope: !7)
540
+ !69 = !DILocation(line: 75, column: 24, scope: !7)
541
+ !70 = !DILocation(line: 77, column: 24, scope: !7)
542
+ !71 = !DILocation(line: 78, column: 30, scope: !7)
543
+ !72 = !DILocation(line: 70, column: 24, scope: !7)
544
+ !73 = !DILocation(line: 72, column: 24, scope: !7)
545
+ !74 = !DILocation(line: 73, column: 24, scope: !7)
546
+ !75 = !DILocation(line: 79, column: 24, scope: !7)
547
+ !76 = !DILocation(line: 80, column: 24, scope: !7)
548
+ !77 = !DILocation(line: 82, column: 29, scope: !7)
549
+ !78 = !DILocation(line: 82, column: 52, scope: !7)
550
+ !79 = !DILocation(line: 58, column: 4, scope: !7)
.triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.cubin ADDED
Binary file (5.54 kB). View file
 
.triton/dump/4c6ad48573c74d55ed79384f6b432d50/triton_.llir ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+
4
+ define void @triton__0d1d2de(ptr addrspace(1) %0, ptr addrspace(1) %1, i32 %2) local_unnamed_addr !dbg !5 {
5
+ %4 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !8
6
+ %5 = shl i32 %4, 3, !dbg !8
7
+ %6 = and i32 %5, 1016, !dbg !8
8
+ %7 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #1, !dbg !9
9
+ %8 = shl i32 %7, 10, !dbg !10
10
+ %9 = or i32 %8, %6, !dbg !11
11
+ %10 = or i32 %9, 4, !dbg !11
12
+ %11 = sext i32 %9 to i64, !dbg !12
13
+ %12 = getelementptr float, ptr addrspace(1) %0, i64 %11, !dbg !12
14
+ %13 = sext i32 %10 to i64, !dbg !12
15
+ %14 = getelementptr float, ptr addrspace(1) %0, i64 %13, !dbg !12
16
+ %15 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];", "=r,=r,=r,=r,l,b"(ptr addrspace(1) %12, i1 true) #1, !dbg !13
17
+ %16 = extractvalue { i32, i32, i32, i32 } %15, 0, !dbg !13
18
+ %17 = extractvalue { i32, i32, i32, i32 } %15, 1, !dbg !13
19
+ %18 = extractvalue { i32, i32, i32, i32 } %15, 2, !dbg !13
20
+ %19 = extractvalue { i32, i32, i32, i32 } %15, 3, !dbg !13
21
+ %20 = bitcast i32 %16 to float, !dbg !13
22
+ %21 = bitcast i32 %17 to float, !dbg !13
23
+ %22 = bitcast i32 %18 to float, !dbg !13
24
+ %23 = bitcast i32 %19 to float, !dbg !13
25
+ %24 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];", "=r,=r,=r,=r,l,b"(ptr addrspace(1) %14, i1 true) #1, !dbg !13
26
+ %25 = extractvalue { i32, i32, i32, i32 } %24, 0, !dbg !13
27
+ %26 = extractvalue { i32, i32, i32, i32 } %24, 1, !dbg !13
28
+ %27 = extractvalue { i32, i32, i32, i32 } %24, 2, !dbg !13
29
+ %28 = extractvalue { i32, i32, i32, i32 } %24, 3, !dbg !13
30
+ %29 = bitcast i32 %25 to float, !dbg !13
31
+ %30 = bitcast i32 %26 to float, !dbg !13
32
+ %31 = bitcast i32 %27 to float, !dbg !13
33
+ %32 = bitcast i32 %28 to float, !dbg !13
34
+ %33 = getelementptr i16, ptr addrspace(1) %1, i64 %11, !dbg !14
35
+ %34 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %20) #1, !dbg !15
36
+ %35 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %21) #1, !dbg !15
37
+ %36 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %22) #1, !dbg !15
38
+ %37 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %23) #1, !dbg !15
39
+ %38 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %29) #1, !dbg !15
40
+ %39 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %30) #1, !dbg !15
41
+ %40 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %31) #1, !dbg !15
42
+ %41 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %32) #1, !dbg !15
43
+ %42 = insertelement <2 x i16> undef, i16 %34, i64 0, !dbg !15
44
+ %43 = insertelement <2 x i16> %42, i16 %35, i64 1, !dbg !15
45
+ %44 = bitcast <2 x i16> %43 to i32, !dbg !15
46
+ %45 = insertelement <2 x i16> undef, i16 %36, i64 0, !dbg !15
47
+ %46 = insertelement <2 x i16> %45, i16 %37, i64 1, !dbg !15
48
+ %47 = bitcast <2 x i16> %46 to i32, !dbg !15
49
+ %48 = insertelement <2 x i16> undef, i16 %38, i64 0, !dbg !15
50
+ %49 = insertelement <2 x i16> %48, i16 %39, i64 1, !dbg !15
51
+ %50 = bitcast <2 x i16> %49 to i32, !dbg !15
52
+ %51 = insertelement <2 x i16> undef, i16 %40, i64 0, !dbg !15
53
+ %52 = insertelement <2 x i16> %51, i16 %41, i64 1, !dbg !15
54
+ %53 = bitcast <2 x i16> %52 to i32, !dbg !15
55
+ tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %44, i32 %47, i32 %50, i32 %53, ptr addrspace(1) %33, i1 true) #1, !dbg !15
56
+ ret void, !dbg !16
57
+ }
58
+
59
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
60
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
61
+
62
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
63
+ attributes #1 = { nounwind }
64
+
65
+ !llvm.module.flags = !{!0}
66
+ !llvm.dbg.cu = !{!1}
67
+ !nvvm.annotations = !{!3, !4, !4, !3}
68
+
69
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
70
+ !1 = distinct !DICompileUnit(language: DW_LANG_C, file: !2, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
71
+ !2 = !DIFile(filename: "czjxjqxojsyyr4zmce6q6twysnucw6p4l5ujgp6ts2ecrm3ue3ex.py", directory: "/tmp/torchinductor_root/zj")
72
+ !3 = !{ptr @triton__0d1d2de, !"kernel", i32 1}
73
+ !4 = !{ptr @triton__0d1d2de, !"maxntidx", i32 128}
74
+ !5 = distinct !DISubprogram(name: "triton__0d1d2de", linkageName: "triton__0d1d2de", scope: !2, file: !2, line: 18, type: !6, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !1)
75
+ !6 = !DISubroutineType(cc: DW_CC_normal, types: !7)
76
+ !7 = !{}
77
+ !8 = !DILocation(line: 21, column: 36, scope: !5)
78
+ !9 = !DILocation(line: 20, column: 28, scope: !5)
79
+ !10 = !DILocation(line: 20, column: 33, scope: !5)
80
+ !11 = !DILocation(line: 21, column: 23, scope: !5)
81
+ !12 = !DILocation(line: 24, column: 30, scope: !5)
82
+ !13 = !DILocation(line: 24, column: 35, scope: !5)
83
+ !14 = !DILocation(line: 26, column: 25, scope: !5)
84
+ !15 = !DILocation(line: 26, column: 36, scope: !5)
85
+ !16 = !DILocation(line: 26, column: 4, scope: !5)
.triton/dump/510522bb05917b836ed253751364fcad/triton_.cubin ADDED
Binary file (66.2 kB). View file
 
.triton/dump/510522bb05917b836ed253751364fcad/triton_.ptx ADDED
@@ -0,0 +1,1810 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.2
6
+ .target sm_89
7
+ .address_size 64
8
+
9
+ // .globl triton__0d1d2d3d4d5de6de
10
+ .extern .func __assertfail
11
+ (
12
+ .param .b64 __assertfail_param_0,
13
+ .param .b64 __assertfail_param_1,
14
+ .param .b32 __assertfail_param_2,
15
+ .param .b64 __assertfail_param_3,
16
+ .param .b64 __assertfail_param_4
17
+ )
18
+ ;
19
+ .global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
20
+ .global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
21
+ .global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 51, 32, 60, 32, 53, 48, 50, 53, 55};
22
+ .global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
23
+ .global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
24
+ .global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
25
+ .extern .shared .align 1 .b8 global_smem[];
26
+ .global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
27
+
28
+ .visible .entry triton__0d1d2d3d4d5de6de(
29
+ .param .u64 triton__0d1d2d3d4d5de6de_param_0,
30
+ .param .u64 triton__0d1d2d3d4d5de6de_param_1,
31
+ .param .u64 triton__0d1d2d3d4d5de6de_param_2,
32
+ .param .u64 triton__0d1d2d3d4d5de6de_param_3,
33
+ .param .u64 triton__0d1d2d3d4d5de6de_param_4,
34
+ .param .u32 triton__0d1d2d3d4d5de6de_param_5,
35
+ .param .u32 triton__0d1d2d3d4d5de6de_param_6
36
+ )
37
+ .maxntid 256, 1, 1
38
+ {
39
+ .reg .pred %p<137>;
40
+ .reg .b16 %rs<17>;
41
+ .reg .b32 %r<408>;
42
+ .reg .f32 %f<614>;
43
+ .reg .b64 %rd<107>;
44
+ .loc 1 18 0
45
+ $L__func_begin0:
46
+ .loc 1 18 0
47
+
48
+ ld.param.u64 %rd13, [triton__0d1d2d3d4d5de6de_param_4];
49
+ ld.param.u64 %rd12, [triton__0d1d2d3d4d5de6de_param_3];
50
+ ld.param.u64 %rd49, [triton__0d1d2d3d4d5de6de_param_0];
51
+ ld.param.u64 %rd50, [triton__0d1d2d3d4d5de6de_param_1];
52
+ $L__tmp0:
53
+ .loc 1 22 44
54
+ mov.u32 %r13, %tid.x;
55
+ ld.param.u64 %rd51, [triton__0d1d2d3d4d5de6de_param_2];
56
+ bfe.u32 %r1, %r13, 3, 5;
57
+ and.b32 %r2, %r13, 63;
58
+ .loc 1 24 33
59
+ shl.b32 %r14, %r13, 3;
60
+ and.b32 %r3, %r14, 56;
61
+ .loc 1 31 36
62
+ shr.u32 %r4, %r13, 6;
63
+ .loc 1 21 28
64
+ mov.u32 %r11, %ctaid.x;
65
+ .loc 1 21 33
66
+ shl.b32 %r15, %r11, 6;
67
+ .loc 1 22 23
68
+ or.b32 %r16, %r15, %r1;
69
+ or.b32 %r17, %r16, 32;
70
+ or.b32 %r18, %r15, %r2;
71
+ .loc 1 26 30
72
+ mul.wide.s32 %rd52, %r16, 8;
73
+ add.s64 %rd15, %rd49, %rd52;
74
+ add.s64 %rd31, %rd15, 256;
75
+ mul.wide.s32 %rd53, %r18, 8;
76
+ add.s64 %rd47, %rd49, %rd53;
77
+ mov.pred %p1, -1;
78
+ .loc 1 26 35
79
+ mov.u64 %rd14, 0x0;
80
+ @%p1 ld.global.L1::evict_last.b64 { %rd14 }, [ %rd15 + 0 ];
81
+ mov.u64 %rd16, 0x0;
82
+ @%p1 ld.global.L1::evict_last.b64 { %rd16 }, [ %rd15 + 0 ];
83
+ mov.u64 %rd18, 0x0;
84
+ @%p1 ld.global.L1::evict_last.b64 { %rd18 }, [ %rd15 + 0 ];
85
+ mov.u64 %rd20, 0x0;
86
+ @%p1 ld.global.L1::evict_last.b64 { %rd20 }, [ %rd15 + 0 ];
87
+ mov.u64 %rd22, 0x0;
88
+ @%p1 ld.global.L1::evict_last.b64 { %rd22 }, [ %rd15 + 0 ];
89
+ mov.u64 %rd24, 0x0;
90
+ @%p1 ld.global.L1::evict_last.b64 { %rd24 }, [ %rd15 + 0 ];
91
+ mov.u64 %rd26, 0x0;
92
+ @%p1 ld.global.L1::evict_last.b64 { %rd26 }, [ %rd15 + 0 ];
93
+ mov.u64 %rd28, 0x0;
94
+ @%p1 ld.global.L1::evict_last.b64 { %rd28 }, [ %rd15 + 0 ];
95
+ mov.u64 %rd30, 0x0;
96
+ @%p1 ld.global.L1::evict_last.b64 { %rd30 }, [ %rd31 + 0 ];
97
+ mov.u64 %rd32, 0x0;
98
+ @%p1 ld.global.L1::evict_last.b64 { %rd32 }, [ %rd31 + 0 ];
99
+ mov.u64 %rd34, 0x0;
100
+ @%p1 ld.global.L1::evict_last.b64 { %rd34 }, [ %rd31 + 0 ];
101
+ mov.u64 %rd36, 0x0;
102
+ @%p1 ld.global.L1::evict_last.b64 { %rd36 }, [ %rd31 + 0 ];
103
+ mov.u64 %rd38, 0x0;
104
+ @%p1 ld.global.L1::evict_last.b64 { %rd38 }, [ %rd31 + 0 ];
105
+ mov.u64 %rd40, 0x0;
106
+ @%p1 ld.global.L1::evict_last.b64 { %rd40 }, [ %rd31 + 0 ];
107
+ mov.u64 %rd42, 0x0;
108
+ @%p1 ld.global.L1::evict_last.b64 { %rd42 }, [ %rd31 + 0 ];
109
+ mov.u64 %rd44, 0x0;
110
+ @%p1 ld.global.L1::evict_last.b64 { %rd44 }, [ %rd31 + 0 ];
111
+ mov.u64 %rd46, 0x0;
112
+ @%p1 ld.global.L1::evict_last.b64 { %rd46 }, [ %rd47 + 0 ];
113
+ .loc 1 27 18
114
+ bfe.s32 %r19, %r11, 25, 1;
115
+ shr.u32 %r20, %r19, 23;
116
+ add.s32 %r21, %r16, %r20;
117
+ and.b32 %r22, %r21, 16776704;
118
+ sub.s32 %r23, %r16, %r22;
119
+ add.s32 %r24, %r17, %r20;
120
+ and.b32 %r25, %r24, 16776704;
121
+ sub.s32 %r26, %r17, %r25;
122
+ .loc 1 35 44
123
+ shl.b32 %r27, %r23, 8;
124
+ shl.b32 %r28, %r26, 8;
125
+ .loc 1 36 22
126
+ add.s64 %rd54, %rd46, 50257;
127
+ .loc 1 37 22
128
+ setp.lt.s64 %p18, %rd14, 0;
129
+ setp.lt.s64 %p19, %rd30, 0;
130
+ setp.lt.s64 %p20, %rd46, 0;
131
+ .loc 1 38 36
132
+ selp.b64 %rd1, %rd54, %rd46, %p20;
133
+ .loc 1 40 44
134
+ shl.b64 %rd55, %rd14, 8;
135
+ add.s64 %rd56, %rd55, 12865792;
136
+ selp.b64 %rd57, %rd56, %rd55, %p18;
137
+ shl.b64 %rd58, %rd30, 8;
138
+ add.s64 %rd59, %rd58, 12865792;
139
+ selp.b64 %rd60, %rd59, %rd58, %p19;
140
+ .loc 1 31 36
141
+ and.b32 %r29, %r13, 7;
142
+ mul.wide.u32 %rd2, %r29, 32;
143
+ shl.b64 %rd61, %rd60, 2;
144
+ or.b64 %rd62, %rd2, %rd61;
145
+ add.s64 %rd3, %rd50, %rd62;
146
+ shl.b64 %rd63, %rd57, 2;
147
+ or.b64 %rd64, %rd2, %rd63;
148
+ add.s64 %rd4, %rd50, %rd64;
149
+ or.b32 %r30, %r28, %r3;
150
+ mul.wide.s32 %rd65, %r30, 4;
151
+ add.s64 %rd5, %rd51, %rd65;
152
+ or.b32 %r31, %r27, %r3;
153
+ mul.wide.s32 %rd66, %r31, 4;
154
+ add.s64 %rd6, %rd51, %rd66;
155
+ mov.f32 %f550, 0f00000000;
156
+ mov.u64 %rd105, 0;
157
+ mov.b32 %r406, -64;
158
+ mov.f32 %f551, %f550;
159
+ mov.f32 %f552, %f550;
160
+ mov.f32 %f553, %f550;
161
+ mov.f32 %f554, %f550;
162
+ mov.f32 %f555, %f550;
163
+ mov.f32 %f556, %f550;
164
+ mov.f32 %f557, %f550;
165
+ mov.f32 %f558, %f550;
166
+ mov.f32 %f559, %f550;
167
+ mov.f32 %f560, %f550;
168
+ mov.f32 %f561, %f550;
169
+ mov.f32 %f562, %f550;
170
+ mov.f32 %f563, %f550;
171
+ mov.f32 %f564, %f550;
172
+ mov.f32 %f565, %f550;
173
+ mov.f32 %f566, %f550;
174
+ mov.f32 %f567, %f550;
175
+ mov.f32 %f568, %f550;
176
+ mov.f32 %f569, %f550;
177
+ mov.f32 %f570, %f550;
178
+ mov.f32 %f571, %f550;
179
+ mov.f32 %f572, %f550;
180
+ mov.f32 %f573, %f550;
181
+ mov.f32 %f574, %f550;
182
+ mov.f32 %f575, %f550;
183
+ mov.f32 %f576, %f550;
184
+ mov.f32 %f577, %f550;
185
+ mov.f32 %f578, %f550;
186
+ mov.f32 %f579, %f550;
187
+ mov.f32 %f580, %f550;
188
+ mov.f32 %f581, %f550;
189
+ mov.f32 %f582, %f550;
190
+ mov.f32 %f583, %f550;
191
+ mov.f32 %f584, %f550;
192
+ mov.f32 %f585, %f550;
193
+ mov.f32 %f586, %f550;
194
+ mov.f32 %f587, %f550;
195
+ mov.f32 %f588, %f550;
196
+ mov.f32 %f589, %f550;
197
+ mov.f32 %f590, %f550;
198
+ mov.f32 %f591, %f550;
199
+ mov.f32 %f592, %f550;
200
+ mov.f32 %f593, %f550;
201
+ mov.f32 %f594, %f550;
202
+ mov.f32 %f595, %f550;
203
+ mov.f32 %f596, %f550;
204
+ mov.f32 %f597, %f550;
205
+ mov.f32 %f598, %f550;
206
+ mov.f32 %f599, %f550;
207
+ mov.f32 %f600, %f550;
208
+ mov.f32 %f601, %f550;
209
+ mov.f32 %f602, %f550;
210
+ mov.f32 %f603, %f550;
211
+ mov.f32 %f604, %f550;
212
+ mov.f32 %f605, %f550;
213
+ mov.f32 %f606, %f550;
214
+ mov.f32 %f607, %f550;
215
+ mov.f32 %f608, %f550;
216
+ mov.f32 %f609, %f550;
217
+ mov.f32 %f610, %f550;
218
+ mov.f32 %f611, %f550;
219
+ mov.f32 %f612, %f550;
220
+ mov.f32 %f613, %f550;
221
+ bra.uni $L__BB0_1;
222
+ $L__BB0_3:
223
+ .loc 1 40 40
224
+ add.s64 %rd78, %rd4, %rd105;
225
+ .loc 1 40 34
226
+ add.s64 %rd79, %rd78, 16;
227
+ add.s64 %rd80, %rd3, %rd105;
228
+ .loc 1 40 52
229
+ add.s64 %rd81, %rd80, 16;
230
+ mov.u32 %r65, 0x0;
231
+ mov.u32 %r66, 0x0;
232
+ mov.u32 %r67, 0x0;
233
+ mov.u32 %r68, 0x0;
234
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r65, %r66, %r67, %r68 }, [ %rd78 + 0 ];
235
+ @!%p1 mov.u32 %r65, %r342;
236
+ @!%p1 mov.u32 %r66, %r342;
237
+ @!%p1 mov.u32 %r67, %r342;
238
+ @!%p1 mov.u32 %r68, %r342;
239
+ mov.b32 %f174, %r65;
240
+ mov.b32 %f175, %r66;
241
+ mov.b32 %f176, %r67;
242
+ mov.b32 %f177, %r68;
243
+ mov.u32 %r73, 0x0;
244
+ mov.u32 %r74, 0x0;
245
+ mov.u32 %r75, 0x0;
246
+ mov.u32 %r76, 0x0;
247
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r73, %r74, %r75, %r76 }, [ %rd79 + 0 ];
248
+ @!%p1 mov.u32 %r73, %r342;
249
+ @!%p1 mov.u32 %r74, %r342;
250
+ @!%p1 mov.u32 %r75, %r342;
251
+ @!%p1 mov.u32 %r76, %r342;
252
+ mov.b32 %f178, %r73;
253
+ mov.b32 %f179, %r74;
254
+ mov.b32 %f180, %r75;
255
+ mov.b32 %f181, %r76;
256
+ mov.u32 %r81, 0x0;
257
+ mov.u32 %r82, 0x0;
258
+ mov.u32 %r83, 0x0;
259
+ mov.u32 %r84, 0x0;
260
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r81, %r82, %r83, %r84 }, [ %rd80 + 0 ];
261
+ @!%p1 mov.u32 %r81, %r342;
262
+ @!%p1 mov.u32 %r82, %r342;
263
+ @!%p1 mov.u32 %r83, %r342;
264
+ @!%p1 mov.u32 %r84, %r342;
265
+ mov.b32 %f182, %r81;
266
+ mov.b32 %f183, %r82;
267
+ mov.b32 %f184, %r83;
268
+ mov.b32 %f185, %r84;
269
+ mov.u32 %r89, 0x0;
270
+ mov.u32 %r90, 0x0;
271
+ mov.u32 %r91, 0x0;
272
+ mov.u32 %r92, 0x0;
273
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r89, %r90, %r91, %r92 }, [ %rd81 + 0 ];
274
+ @!%p1 mov.u32 %r89, %r342;
275
+ @!%p1 mov.u32 %r90, %r342;
276
+ @!%p1 mov.u32 %r91, %r342;
277
+ @!%p1 mov.u32 %r92, %r342;
278
+ mov.b32 %f186, %r89;
279
+ mov.b32 %f187, %r90;
280
+ mov.b32 %f188, %r91;
281
+ mov.b32 %f189, %r92;
282
+ .loc 1 41 22
283
+ add.f32 %f190, %f65, %f174;
284
+ add.f32 %f191, %f66, %f175;
285
+ add.f32 %f192, %f67, %f176;
286
+ add.f32 %f193, %f68, %f177;
287
+ add.f32 %f194, %f69, %f178;
288
+ add.f32 %f195, %f70, %f179;
289
+ add.f32 %f196, %f71, %f180;
290
+ add.f32 %f197, %f72, %f181;
291
+ add.f32 %f198, %f73, %f182;
292
+ add.f32 %f199, %f74, %f183;
293
+ add.f32 %f200, %f75, %f184;
294
+ add.f32 %f201, %f76, %f185;
295
+ add.f32 %f202, %f77, %f186;
296
+ add.f32 %f203, %f78, %f187;
297
+ add.f32 %f204, %f79, %f188;
298
+ add.f32 %f205, %f80, %f189;
299
+ $L__tmp1:
300
+ .loc 2 96 20
301
+ sub.f32 %f206, %f190, %f598;
302
+ sub.f32 %f207, %f191, %f599;
303
+ sub.f32 %f208, %f192, %f600;
304
+ sub.f32 %f209, %f193, %f601;
305
+ sub.f32 %f210, %f194, %f602;
306
+ sub.f32 %f211, %f195, %f603;
307
+ sub.f32 %f212, %f196, %f604;
308
+ sub.f32 %f213, %f197, %f605;
309
+ sub.f32 %f214, %f198, %f606;
310
+ sub.f32 %f215, %f199, %f607;
311
+ sub.f32 %f216, %f200, %f608;
312
+ sub.f32 %f217, %f201, %f609;
313
+ sub.f32 %f218, %f202, %f610;
314
+ sub.f32 %f219, %f203, %f611;
315
+ sub.f32 %f220, %f204, %f612;
316
+ sub.f32 %f221, %f205, %f613;
317
+ .loc 2 97 26
318
+ add.f32 %f550, %f550, 0f3F800000;
319
+ add.f32 %f551, %f551, 0f3F800000;
320
+ add.f32 %f552, %f552, 0f3F800000;
321
+ add.f32 %f553, %f553, 0f3F800000;
322
+ add.f32 %f554, %f554, 0f3F800000;
323
+ add.f32 %f555, %f555, 0f3F800000;
324
+ add.f32 %f556, %f556, 0f3F800000;
325
+ add.f32 %f557, %f557, 0f3F800000;
326
+ add.f32 %f558, %f558, 0f3F800000;
327
+ add.f32 %f559, %f559, 0f3F800000;
328
+ add.f32 %f560, %f560, 0f3F800000;
329
+ add.f32 %f561, %f561, 0f3F800000;
330
+ add.f32 %f562, %f562, 0f3F800000;
331
+ add.f32 %f563, %f563, 0f3F800000;
332
+ add.f32 %f564, %f564, 0f3F800000;
333
+ add.f32 %f565, %f565, 0f3F800000;
334
+ add.f32 %f566, %f566, 0f3F800000;
335
+ add.f32 %f567, %f567, 0f3F800000;
336
+ add.f32 %f568, %f568, 0f3F800000;
337
+ add.f32 %f569, %f569, 0f3F800000;
338
+ add.f32 %f570, %f570, 0f3F800000;
339
+ add.f32 %f571, %f571, 0f3F800000;
340
+ add.f32 %f572, %f572, 0f3F800000;
341
+ add.f32 %f573, %f573, 0f3F800000;
342
+ add.f32 %f574, %f574, 0f3F800000;
343
+ add.f32 %f575, %f575, 0f3F800000;
344
+ add.f32 %f576, %f576, 0f3F800000;
345
+ add.f32 %f577, %f577, 0f3F800000;
346
+ add.f32 %f578, %f578, 0f3F800000;
347
+ add.f32 %f579, %f579, 0f3F800000;
348
+ add.f32 %f580, %f580, 0f3F800000;
349
+ add.f32 %f581, %f581, 0f3F800000;
350
+ .loc 2 98 30
351
+ mov.b32 %r98, %f206;
352
+ mov.b32 %r99, %f550;
353
+ div.full.f32 %r97, %r98, %r99;
354
+ mov.b32 %f222, %r97;
355
+ mov.b32 %r101, %f207;
356
+ mov.b32 %r102, %f551;
357
+ div.full.f32 %r100, %r101, %r102;
358
+ mov.b32 %f223, %r100;
359
+ mov.b32 %r104, %f208;
360
+ mov.b32 %r105, %f552;
361
+ div.full.f32 %r103, %r104, %r105;
362
+ mov.b32 %f224, %r103;
363
+ mov.b32 %r107, %f209;
364
+ mov.b32 %r108, %f553;
365
+ div.full.f32 %r106, %r107, %r108;
366
+ mov.b32 %f225, %r106;
367
+ mov.b32 %r110, %f210;
368
+ mov.b32 %r111, %f554;
369
+ div.full.f32 %r109, %r110, %r111;
370
+ mov.b32 %f226, %r109;
371
+ mov.b32 %r113, %f211;
372
+ mov.b32 %r114, %f555;
373
+ div.full.f32 %r112, %r113, %r114;
374
+ mov.b32 %f227, %r112;
375
+ mov.b32 %r116, %f212;
376
+ mov.b32 %r117, %f556;
377
+ div.full.f32 %r115, %r116, %r117;
378
+ mov.b32 %f228, %r115;
379
+ mov.b32 %r119, %f213;
380
+ mov.b32 %r120, %f557;
381
+ div.full.f32 %r118, %r119, %r120;
382
+ mov.b32 %f229, %r118;
383
+ mov.b32 %r122, %f214;
384
+ mov.b32 %r123, %f558;
385
+ div.full.f32 %r121, %r122, %r123;
386
+ mov.b32 %f230, %r121;
387
+ mov.b32 %r125, %f215;
388
+ mov.b32 %r126, %f559;
389
+ div.full.f32 %r124, %r125, %r126;
390
+ mov.b32 %f231, %r124;
391
+ mov.b32 %r128, %f216;
392
+ mov.b32 %r129, %f560;
393
+ div.full.f32 %r127, %r128, %r129;
394
+ mov.b32 %f232, %r127;
395
+ mov.b32 %r131, %f217;
396
+ mov.b32 %r132, %f561;
397
+ div.full.f32 %r130, %r131, %r132;
398
+ mov.b32 %f233, %r130;
399
+ mov.b32 %r134, %f218;
400
+ mov.b32 %r135, %f562;
401
+ div.full.f32 %r133, %r134, %r135;
402
+ mov.b32 %f234, %r133;
403
+ mov.b32 %r137, %f219;
404
+ mov.b32 %r138, %f563;
405
+ div.full.f32 %r136, %r137, %r138;
406
+ mov.b32 %f235, %r136;
407
+ mov.b32 %r140, %f220;
408
+ mov.b32 %r141, %f564;
409
+ div.full.f32 %r139, %r140, %r141;
410
+ mov.b32 %f236, %r139;
411
+ mov.b32 %r143, %f221;
412
+ mov.b32 %r144, %f565;
413
+ div.full.f32 %r142, %r143, %r144;
414
+ mov.b32 %f237, %r142;
415
+ .loc 2 98 22
416
+ add.f32 %f598, %f598, %f222;
417
+ add.f32 %f599, %f599, %f223;
418
+ add.f32 %f600, %f600, %f224;
419
+ add.f32 %f601, %f601, %f225;
420
+ add.f32 %f602, %f602, %f226;
421
+ add.f32 %f603, %f603, %f227;
422
+ add.f32 %f604, %f604, %f228;
423
+ add.f32 %f605, %f605, %f229;
424
+ add.f32 %f606, %f606, %f230;
425
+ add.f32 %f607, %f607, %f231;
426
+ add.f32 %f608, %f608, %f232;
427
+ add.f32 %f609, %f609, %f233;
428
+ add.f32 %f610, %f610, %f234;
429
+ add.f32 %f611, %f611, %f235;
430
+ add.f32 %f612, %f612, %f236;
431
+ add.f32 %f613, %f613, %f237;
432
+ .loc 2 101 30
433
+ sub.f32 %f238, %f190, %f598;
434
+ sub.f32 %f239, %f191, %f599;
435
+ sub.f32 %f240, %f192, %f600;
436
+ sub.f32 %f241, %f193, %f601;
437
+ sub.f32 %f242, %f194, %f602;
438
+ sub.f32 %f243, %f195, %f603;
439
+ sub.f32 %f244, %f196, %f604;
440
+ sub.f32 %f245, %f197, %f605;
441
+ sub.f32 %f246, %f198, %f606;
442
+ sub.f32 %f247, %f199, %f607;
443
+ sub.f32 %f248, %f200, %f608;
444
+ sub.f32 %f249, %f201, %f609;
445
+ sub.f32 %f250, %f202, %f610;
446
+ sub.f32 %f251, %f203, %f611;
447
+ sub.f32 %f252, %f204, %f612;
448
+ sub.f32 %f253, %f205, %f613;
449
+ $L__tmp2:
450
+ .loc 1 47 48
451
+ fma.rn.f32 %f582, %f206, %f238, %f582;
452
+ fma.rn.f32 %f583, %f207, %f239, %f583;
453
+ fma.rn.f32 %f584, %f208, %f240, %f584;
454
+ fma.rn.f32 %f585, %f209, %f241, %f585;
455
+ fma.rn.f32 %f586, %f210, %f242, %f586;
456
+ fma.rn.f32 %f587, %f211, %f243, %f587;
457
+ fma.rn.f32 %f588, %f212, %f244, %f588;
458
+ fma.rn.f32 %f589, %f213, %f245, %f589;
459
+ fma.rn.f32 %f590, %f214, %f246, %f590;
460
+ fma.rn.f32 %f591, %f215, %f247, %f591;
461
+ fma.rn.f32 %f592, %f216, %f248, %f592;
462
+ fma.rn.f32 %f593, %f217, %f249, %f593;
463
+ fma.rn.f32 %f594, %f218, %f250, %f594;
464
+ fma.rn.f32 %f595, %f219, %f251, %f595;
465
+ fma.rn.f32 %f596, %f220, %f252, %f596;
466
+ fma.rn.f32 %f597, %f221, %f253, %f597;
467
+ .loc 1 31 36
468
+ add.s64 %rd105, %rd105, 256;
469
+ add.s32 %r406, %r406, 64;
470
+ setp.lt.u32 %p62, %r406, 192;
471
+ @%p62 bra $L__BB0_1;
472
+ bra.uni $L__BB0_4;
473
+ $L__BB0_1:
474
+ .loc 1 39 40
475
+ setp.lt.u64 %p41, %rd1, 50257;
476
+ .loc 1 35 34
477
+ add.s64 %rd67, %rd6, %rd105;
478
+ add.s64 %rd68, %rd67, 16;
479
+ add.s64 %rd69, %rd5, %rd105;
480
+ .loc 1 35 50
481
+ add.s64 %rd70, %rd69, 16;
482
+ mov.b32 %r342, 0;
483
+ mov.u32 %r32, 0x0;
484
+ mov.u32 %r33, 0x0;
485
+ mov.u32 %r34, 0x0;
486
+ mov.u32 %r35, 0x0;
487
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r32, %r33, %r34, %r35 }, [ %rd67 + 0 ];
488
+ @!%p1 mov.u32 %r32, %r342;
489
+ @!%p1 mov.u32 %r33, %r342;
490
+ @!%p1 mov.u32 %r34, %r342;
491
+ @!%p1 mov.u32 %r35, %r342;
492
+ mov.b32 %f65, %r32;
493
+ mov.b32 %f66, %r33;
494
+ mov.b32 %f67, %r34;
495
+ mov.b32 %f68, %r35;
496
+ mov.u32 %r40, 0x0;
497
+ mov.u32 %r41, 0x0;
498
+ mov.u32 %r42, 0x0;
499
+ mov.u32 %r43, 0x0;
500
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r40, %r41, %r42, %r43 }, [ %rd68 + 0 ];
501
+ @!%p1 mov.u32 %r40, %r342;
502
+ @!%p1 mov.u32 %r41, %r342;
503
+ @!%p1 mov.u32 %r42, %r342;
504
+ @!%p1 mov.u32 %r43, %r342;
505
+ mov.b32 %f69, %r40;
506
+ mov.b32 %f70, %r41;
507
+ mov.b32 %f71, %r42;
508
+ mov.b32 %f72, %r43;
509
+ mov.u32 %r48, 0x0;
510
+ mov.u32 %r49, 0x0;
511
+ mov.u32 %r50, 0x0;
512
+ mov.u32 %r51, 0x0;
513
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r48, %r49, %r50, %r51 }, [ %rd69 + 0 ];
514
+ @!%p1 mov.u32 %r48, %r342;
515
+ @!%p1 mov.u32 %r49, %r342;
516
+ @!%p1 mov.u32 %r50, %r342;
517
+ @!%p1 mov.u32 %r51, %r342;
518
+ mov.b32 %f73, %r48;
519
+ mov.b32 %f74, %r49;
520
+ mov.b32 %f75, %r50;
521
+ mov.b32 %f76, %r51;
522
+ mov.u32 %r56, 0x0;
523
+ mov.u32 %r57, 0x0;
524
+ mov.u32 %r58, 0x0;
525
+ mov.u32 %r59, 0x0;
526
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r56, %r57, %r58, %r59 }, [ %rd70 + 0 ];
527
+ @!%p1 mov.u32 %r56, %r342;
528
+ @!%p1 mov.u32 %r57, %r342;
529
+ @!%p1 mov.u32 %r58, %r342;
530
+ @!%p1 mov.u32 %r59, %r342;
531
+ mov.b32 %f77, %r56;
532
+ mov.b32 %f78, %r57;
533
+ mov.b32 %f79, %r58;
534
+ mov.b32 %f80, %r59;
535
+ mov.b32 %r405, 883;
536
+ mov.u64 %rd104, 1;
537
+ .loc 1 39 55
538
+ @%p41 bra $L__BB0_3;
539
+ mov.u64 %rd71, assertMessage_0;
540
+ cvta.global.u64 %rd72, %rd71;
541
+ mov.u64 %rd73, assertFile_0;
542
+ cvta.global.u64 %rd74, %rd73;
543
+ mov.u64 %rd75, assertFunc_0;
544
+ cvta.global.u64 %rd76, %rd75;
545
+ { // callseq 6, 0
546
+ .reg .b32 temp_param_reg;
547
+ .param .b64 param0;
548
+ st.param.b64 [param0+0], %rd72;
549
+ .param .b64 param1;
550
+ st.param.b64 [param1+0], %rd74;
551
+ .param .b32 param2;
552
+ st.param.b32 [param2+0], %r405;
553
+ .param .b64 param3;
554
+ st.param.b64 [param3+0], %rd76;
555
+ .param .b64 param4;
556
+ st.param.b64 [param4+0], %rd104;
557
+ call.uni
558
+ __assertfail,
559
+ (
560
+ param0,
561
+ param1,
562
+ param2,
563
+ param3,
564
+ param4
565
+ );
566
+ } // callseq 6
567
+ bra.uni $L__BB0_3;
568
+ $L__BB0_4:
569
+ .loc 1 31 36
570
+ and.b32 %r254, %r4, 3;
571
+ mad.lo.s32 %r255, %r254, 72, %r2;
572
+ shl.b32 %r256, %r255, 2;
573
+ mov.u32 %r257, global_smem;
574
+ add.s32 %r258, %r257, %r256;
575
+ st.shared.f32 [%r258], %f566;
576
+ st.shared.f32 [%r258+1152], %f567;
577
+ st.shared.f32 [%r258+2304], %f568;
578
+ st.shared.f32 [%r258+3456], %f569;
579
+ st.shared.f32 [%r258+4608], %f570;
580
+ st.shared.f32 [%r258+5760], %f571;
581
+ st.shared.f32 [%r258+6912], %f572;
582
+ st.shared.f32 [%r258+8064], %f573;
583
+ bar.sync 0;
584
+ mad.lo.s32 %r259, %r1, 72, %r3;
585
+ shl.b32 %r260, %r259, 2;
586
+ add.s32 %r261, %r257, %r260;
587
+ ld.shared.v4.f32 {%f254, %f255, %f256, %f257}, [%r261];
588
+ ld.shared.v4.f32 {%f258, %f259, %f260, %f261}, [%r261+16];
589
+ bar.sync 0;
590
+ st.shared.f32 [%r258], %f574;
591
+ st.shared.f32 [%r258+1152], %f575;
592
+ st.shared.f32 [%r258+2304], %f576;
593
+ st.shared.f32 [%r258+3456], %f577;
594
+ st.shared.f32 [%r258+4608], %f578;
595
+ st.shared.f32 [%r258+5760], %f579;
596
+ st.shared.f32 [%r258+6912], %f580;
597
+ st.shared.f32 [%r258+8064], %f581;
598
+ bar.sync 0;
599
+ ld.shared.v4.f32 {%f262, %f263, %f264, %f265}, [%r261];
600
+ ld.shared.v4.f32 {%f266, %f267, %f268, %f269}, [%r261+16];
601
+ $L__tmp3:
602
+ .loc 2 108 21
603
+ sub.f32 %f270, %f599, %f598;
604
+ .loc 2 109 28
605
+ add.f32 %f271, %f254, %f255;
606
+ .loc 2 110 39
607
+ setp.eq.f32 %p63, %f271, 0f00000000;
608
+ .loc 2 110 60
609
+ mov.b32 %r146, %f255;
610
+ mov.b32 %r147, %f271;
611
+ div.full.f32 %r145, %r146, %r147;
612
+ mov.b32 %f272, %r145;
613
+ .loc 2 110 49
614
+ selp.f32 %f273, 0f00000000, %f272, %p63;
615
+ .loc 2 112 17
616
+ fma.rn.f32 %f274, %f270, %f273, %f598;
617
+ .loc 2 113 15
618
+ add.f32 %f275, %f582, %f583;
619
+ .loc 2 113 30
620
+ mul.f32 %f276, %f270, %f270;
621
+ .loc 2 113 38
622
+ mul.f32 %f277, %f276, %f254;
623
+ .loc 2 113 22
624
+ fma.rn.f32 %f278, %f277, %f273, %f275;
625
+ .loc 2 108 21
626
+ sub.f32 %f279, %f600, %f274;
627
+ .loc 2 109 28
628
+ add.f32 %f280, %f256, %f271;
629
+ .loc 2 110 39
630
+ setp.eq.f32 %p64, %f280, 0f00000000;
631
+ .loc 2 110 60
632
+ mov.b32 %r150, %f280;
633
+ mov.b32 %r149, %f256;
634
+ div.full.f32 %r148, %r149, %r150;
635
+ mov.b32 %f281, %r148;
636
+ .loc 2 110 49
637
+ selp.f32 %f282, 0f00000000, %f281, %p64;
638
+ .loc 2 112 17
639
+ fma.rn.f32 %f283, %f282, %f279, %f274;
640
+ .loc 2 113 15
641
+ add.f32 %f284, %f584, %f278;
642
+ .loc 2 113 30
643
+ mul.f32 %f285, %f279, %f279;
644
+ .loc 2 113 38
645
+ mul.f32 %f286, %f271, %f285;
646
+ .loc 2 113 22
647
+ fma.rn.f32 %f287, %f282, %f286, %f284;
648
+ .loc 2 108 21
649
+ sub.f32 %f288, %f601, %f283;
650
+ .loc 2 109 28
651
+ add.f32 %f289, %f257, %f280;
652
+ .loc 2 110 39
653
+ setp.eq.f32 %p65, %f289, 0f00000000;
654
+ .loc 2 110 60
655
+ mov.b32 %r153, %f289;
656
+ mov.b32 %r152, %f257;
657
+ div.full.f32 %r151, %r152, %r153;
658
+ mov.b32 %f290, %r151;
659
+ .loc 2 110 49
660
+ selp.f32 %f291, 0f00000000, %f290, %p65;
661
+ .loc 2 112 17
662
+ fma.rn.f32 %f292, %f291, %f288, %f283;
663
+ .loc 2 113 15
664
+ add.f32 %f293, %f585, %f287;
665
+ .loc 2 113 30
666
+ mul.f32 %f294, %f288, %f288;
667
+ .loc 2 113 38
668
+ mul.f32 %f295, %f280, %f294;
669
+ .loc 2 113 22
670
+ fma.rn.f32 %f296, %f291, %f295, %f293;
671
+ .loc 2 108 21
672
+ sub.f32 %f297, %f602, %f292;
673
+ .loc 2 109 28
674
+ add.f32 %f298, %f258, %f289;
675
+ .loc 2 110 39
676
+ setp.eq.f32 %p66, %f298, 0f00000000;
677
+ .loc 2 110 60
678
+ mov.b32 %r156, %f298;
679
+ mov.b32 %r155, %f258;
680
+ div.full.f32 %r154, %r155, %r156;
681
+ mov.b32 %f299, %r154;
682
+ .loc 2 110 49
683
+ selp.f32 %f300, 0f00000000, %f299, %p66;
684
+ .loc 2 112 17
685
+ fma.rn.f32 %f301, %f300, %f297, %f292;
686
+ .loc 2 113 15
687
+ add.f32 %f302, %f586, %f296;
688
+ .loc 2 113 30
689
+ mul.f32 %f303, %f297, %f297;
690
+ .loc 2 113 38
691
+ mul.f32 %f304, %f289, %f303;
692
+ .loc 2 113 22
693
+ fma.rn.f32 %f305, %f300, %f304, %f302;
694
+ .loc 2 108 21
695
+ sub.f32 %f306, %f603, %f301;
696
+ .loc 2 109 28
697
+ add.f32 %f307, %f259, %f298;
698
+ .loc 2 110 39
699
+ setp.eq.f32 %p67, %f307, 0f00000000;
700
+ .loc 2 110 60
701
+ mov.b32 %r159, %f307;
702
+ mov.b32 %r158, %f259;
703
+ div.full.f32 %r157, %r158, %r159;
704
+ mov.b32 %f308, %r157;
705
+ .loc 2 110 49
706
+ selp.f32 %f309, 0f00000000, %f308, %p67;
707
+ .loc 2 112 17
708
+ fma.rn.f32 %f310, %f309, %f306, %f301;
709
+ .loc 2 113 15
710
+ add.f32 %f311, %f587, %f305;
711
+ .loc 2 113 30
712
+ mul.f32 %f312, %f306, %f306;
713
+ .loc 2 113 38
714
+ mul.f32 %f313, %f298, %f312;
715
+ .loc 2 113 22
716
+ fma.rn.f32 %f314, %f309, %f313, %f311;
717
+ .loc 2 108 21
718
+ sub.f32 %f315, %f604, %f310;
719
+ .loc 2 109 28
720
+ add.f32 %f316, %f260, %f307;
721
+ .loc 2 110 39
722
+ setp.eq.f32 %p68, %f316, 0f00000000;
723
+ .loc 2 110 60
724
+ mov.b32 %r162, %f316;
725
+ mov.b32 %r161, %f260;
726
+ div.full.f32 %r160, %r161, %r162;
727
+ mov.b32 %f317, %r160;
728
+ .loc 2 110 49
729
+ selp.f32 %f318, 0f00000000, %f317, %p68;
730
+ .loc 2 112 17
731
+ fma.rn.f32 %f319, %f318, %f315, %f310;
732
+ .loc 2 113 15
733
+ add.f32 %f320, %f588, %f314;
734
+ .loc 2 113 30
735
+ mul.f32 %f321, %f315, %f315;
736
+ .loc 2 113 38
737
+ mul.f32 %f322, %f307, %f321;
738
+ .loc 2 113 22
739
+ fma.rn.f32 %f323, %f318, %f322, %f320;
740
+ .loc 2 108 21
741
+ sub.f32 %f324, %f605, %f319;
742
+ .loc 2 109 28
743
+ add.f32 %f325, %f261, %f316;
744
+ .loc 2 110 39
745
+ setp.eq.f32 %p69, %f325, 0f00000000;
746
+ .loc 2 110 60
747
+ mov.b32 %r165, %f325;
748
+ mov.b32 %r164, %f261;
749
+ div.full.f32 %r163, %r164, %r165;
750
+ mov.b32 %f326, %r163;
751
+ .loc 2 110 49
752
+ selp.f32 %f327, 0f00000000, %f326, %p69;
753
+ .loc 2 112 17
754
+ fma.rn.f32 %f328, %f327, %f324, %f319;
755
+ .loc 2 113 15
756
+ add.f32 %f329, %f589, %f323;
757
+ .loc 2 113 30
758
+ mul.f32 %f330, %f324, %f324;
759
+ .loc 2 113 38
760
+ mul.f32 %f331, %f316, %f330;
761
+ .loc 2 113 22
762
+ fma.rn.f32 %f332, %f327, %f331, %f329;
763
+ .loc 2 108 21
764
+ sub.f32 %f333, %f607, %f606;
765
+ .loc 2 109 28
766
+ add.f32 %f334, %f262, %f263;
767
+ .loc 2 110 39
768
+ setp.eq.f32 %p70, %f334, 0f00000000;
769
+ .loc 2 110 60
770
+ mov.b32 %r167, %f263;
771
+ mov.b32 %r168, %f334;
772
+ div.full.f32 %r166, %r167, %r168;
773
+ mov.b32 %f335, %r166;
774
+ .loc 2 110 49
775
+ selp.f32 %f336, 0f00000000, %f335, %p70;
776
+ .loc 2 112 17
777
+ fma.rn.f32 %f337, %f333, %f336, %f606;
778
+ .loc 2 113 15
779
+ add.f32 %f338, %f590, %f591;
780
+ .loc 2 113 30
781
+ mul.f32 %f339, %f333, %f333;
782
+ .loc 2 113 38
783
+ mul.f32 %f340, %f339, %f262;
784
+ .loc 2 113 22
785
+ fma.rn.f32 %f341, %f340, %f336, %f338;
786
+ .loc 2 108 21
787
+ sub.f32 %f342, %f608, %f337;
788
+ .loc 2 109 28
789
+ add.f32 %f343, %f264, %f334;
790
+ .loc 2 110 39
791
+ setp.eq.f32 %p71, %f343, 0f00000000;
792
+ .loc 2 110 60
793
+ mov.b32 %r171, %f343;
794
+ mov.b32 %r170, %f264;
795
+ div.full.f32 %r169, %r170, %r171;
796
+ mov.b32 %f344, %r169;
797
+ .loc 2 110 49
798
+ selp.f32 %f345, 0f00000000, %f344, %p71;
799
+ .loc 2 112 17
800
+ fma.rn.f32 %f346, %f345, %f342, %f337;
801
+ .loc 2 113 15
802
+ add.f32 %f347, %f592, %f341;
803
+ .loc 2 113 30
804
+ mul.f32 %f348, %f342, %f342;
805
+ .loc 2 113 38
806
+ mul.f32 %f349, %f334, %f348;
807
+ .loc 2 113 22
808
+ fma.rn.f32 %f350, %f345, %f349, %f347;
809
+ .loc 2 108 21
810
+ sub.f32 %f351, %f609, %f346;
811
+ .loc 2 109 28
812
+ add.f32 %f352, %f265, %f343;
813
+ .loc 2 110 39
814
+ setp.eq.f32 %p72, %f352, 0f00000000;
815
+ .loc 2 110 60
816
+ mov.b32 %r174, %f352;
817
+ mov.b32 %r173, %f265;
818
+ div.full.f32 %r172, %r173, %r174;
819
+ mov.b32 %f353, %r172;
820
+ .loc 2 110 49
821
+ selp.f32 %f354, 0f00000000, %f353, %p72;
822
+ .loc 2 112 17
823
+ fma.rn.f32 %f355, %f354, %f351, %f346;
824
+ .loc 2 113 15
825
+ add.f32 %f356, %f593, %f350;
826
+ .loc 2 113 30
827
+ mul.f32 %f357, %f351, %f351;
828
+ .loc 2 113 38
829
+ mul.f32 %f358, %f343, %f357;
830
+ .loc 2 113 22
831
+ fma.rn.f32 %f359, %f354, %f358, %f356;
832
+ .loc 2 108 21
833
+ sub.f32 %f360, %f610, %f355;
834
+ .loc 2 109 28
835
+ add.f32 %f361, %f266, %f352;
836
+ .loc 2 110 39
837
+ setp.eq.f32 %p73, %f361, 0f00000000;
838
+ .loc 2 110 60
839
+ mov.b32 %r177, %f361;
840
+ mov.b32 %r176, %f266;
841
+ div.full.f32 %r175, %r176, %r177;
842
+ mov.b32 %f362, %r175;
843
+ .loc 2 110 49
844
+ selp.f32 %f363, 0f00000000, %f362, %p73;
845
+ .loc 2 112 17
846
+ fma.rn.f32 %f364, %f363, %f360, %f355;
847
+ .loc 2 113 15
848
+ add.f32 %f365, %f594, %f359;
849
+ .loc 2 113 30
850
+ mul.f32 %f366, %f360, %f360;
851
+ .loc 2 113 38
852
+ mul.f32 %f367, %f352, %f366;
853
+ .loc 2 113 22
854
+ fma.rn.f32 %f368, %f363, %f367, %f365;
855
+ .loc 2 108 21
856
+ sub.f32 %f369, %f611, %f364;
857
+ .loc 2 109 28
858
+ add.f32 %f370, %f267, %f361;
859
+ .loc 2 110 39
860
+ setp.eq.f32 %p74, %f370, 0f00000000;
861
+ .loc 2 110 60
862
+ mov.b32 %r180, %f370;
863
+ mov.b32 %r179, %f267;
864
+ div.full.f32 %r178, %r179, %r180;
865
+ mov.b32 %f371, %r178;
866
+ .loc 2 110 49
867
+ selp.f32 %f372, 0f00000000, %f371, %p74;
868
+ .loc 2 112 17
869
+ fma.rn.f32 %f373, %f372, %f369, %f364;
870
+ .loc 2 113 15
871
+ add.f32 %f374, %f595, %f368;
872
+ .loc 2 113 30
873
+ mul.f32 %f375, %f369, %f369;
874
+ .loc 2 113 38
875
+ mul.f32 %f376, %f361, %f375;
876
+ .loc 2 113 22
877
+ fma.rn.f32 %f377, %f372, %f376, %f374;
878
+ .loc 2 108 21
879
+ sub.f32 %f378, %f612, %f373;
880
+ .loc 2 109 28
881
+ add.f32 %f379, %f268, %f370;
882
+ .loc 2 110 39
883
+ setp.eq.f32 %p75, %f379, 0f00000000;
884
+ .loc 2 110 60
885
+ mov.b32 %r183, %f379;
886
+ mov.b32 %r182, %f268;
887
+ div.full.f32 %r181, %r182, %r183;
888
+ mov.b32 %f380, %r181;
889
+ .loc 2 110 49
890
+ selp.f32 %f381, 0f00000000, %f380, %p75;
891
+ .loc 2 112 17
892
+ fma.rn.f32 %f382, %f381, %f378, %f373;
893
+ .loc 2 113 15
894
+ add.f32 %f383, %f596, %f377;
895
+ .loc 2 113 30
896
+ mul.f32 %f384, %f378, %f378;
897
+ .loc 2 113 38
898
+ mul.f32 %f385, %f370, %f384;
899
+ .loc 2 113 22
900
+ fma.rn.f32 %f386, %f381, %f385, %f383;
901
+ .loc 2 108 21
902
+ sub.f32 %f387, %f613, %f382;
903
+ .loc 2 109 28
904
+ add.f32 %f388, %f269, %f379;
905
+ .loc 2 110 39
906
+ setp.eq.f32 %p76, %f388, 0f00000000;
907
+ .loc 2 110 60
908
+ mov.b32 %r186, %f388;
909
+ mov.b32 %r185, %f269;
910
+ div.full.f32 %r184, %r185, %r186;
911
+ mov.b32 %f389, %r184;
912
+ .loc 2 110 49
913
+ selp.f32 %f390, 0f00000000, %f389, %p76;
914
+ .loc 2 112 17
915
+ fma.rn.f32 %f391, %f390, %f387, %f382;
916
+ .loc 2 113 15
917
+ add.f32 %f392, %f597, %f386;
918
+ .loc 2 113 30
919
+ mul.f32 %f393, %f387, %f387;
920
+ .loc 2 113 38
921
+ mul.f32 %f394, %f379, %f393;
922
+ .loc 2 113 22
923
+ fma.rn.f32 %f395, %f390, %f394, %f392;
924
+ $L__tmp4:
925
+ .loc 2 120 46
926
+ mov.b32 %r262, %f328;
927
+ shfl.sync.bfly.b32 %r263, %r262, 4, 31, -1;
928
+ mov.b32 %f396, %r263;
929
+ mov.b32 %r264, %f332;
930
+ shfl.sync.bfly.b32 %r265, %r264, 4, 31, -1;
931
+ mov.b32 %f397, %r265;
932
+ shfl.sync.bfly.b32 %r188, %r165, 4, 31, -1;
933
+ mov.b32 %f398, %r188;
934
+ $L__tmp5:
935
+ .loc 2 108 21
936
+ sub.f32 %f399, %f396, %f328;
937
+ .loc 2 109 28
938
+ add.f32 %f400, %f325, %f398;
939
+ .loc 2 110 39
940
+ setp.eq.f32 %p77, %f400, 0f00000000;
941
+ .loc 2 110 60
942
+ mov.b32 %r189, %f400;
943
+ div.full.f32 %r187, %r188, %r189;
944
+ mov.b32 %f401, %r187;
945
+ .loc 2 110 49
946
+ selp.f32 %f402, 0f00000000, %f401, %p77;
947
+ .loc 2 112 17
948
+ fma.rn.f32 %f403, %f402, %f399, %f328;
949
+ .loc 2 113 15
950
+ add.f32 %f404, %f332, %f397;
951
+ .loc 2 113 30
952
+ mul.f32 %f405, %f399, %f399;
953
+ .loc 2 113 38
954
+ mul.f32 %f406, %f325, %f405;
955
+ .loc 2 113 22
956
+ fma.rn.f32 %f407, %f402, %f406, %f404;
957
+ $L__tmp6:
958
+ .loc 2 120 46
959
+ mov.b32 %r266, %f403;
960
+ shfl.sync.bfly.b32 %r267, %r266, 2, 31, -1;
961
+ mov.b32 %f408, %r267;
962
+ mov.b32 %r268, %f407;
963
+ shfl.sync.bfly.b32 %r269, %r268, 2, 31, -1;
964
+ mov.b32 %f409, %r269;
965
+ shfl.sync.bfly.b32 %r191, %r189, 2, 31, -1;
966
+ mov.b32 %f410, %r191;
967
+ $L__tmp7:
968
+ .loc 2 108 21
969
+ sub.f32 %f411, %f408, %f403;
970
+ .loc 2 109 28
971
+ add.f32 %f412, %f400, %f410;
972
+ .loc 2 110 39
973
+ setp.eq.f32 %p78, %f412, 0f00000000;
974
+ .loc 2 110 60
975
+ mov.b32 %r192, %f412;
976
+ div.full.f32 %r190, %r191, %r192;
977
+ mov.b32 %f413, %r190;
978
+ .loc 2 110 49
979
+ selp.f32 %f414, 0f00000000, %f413, %p78;
980
+ .loc 2 112 17
981
+ fma.rn.f32 %f415, %f414, %f411, %f403;
982
+ .loc 2 113 15
983
+ add.f32 %f416, %f407, %f409;
984
+ .loc 2 113 30
985
+ mul.f32 %f417, %f411, %f411;
986
+ .loc 2 113 38
987
+ mul.f32 %f418, %f400, %f417;
988
+ .loc 2 113 22
989
+ fma.rn.f32 %f419, %f414, %f418, %f416;
990
+ $L__tmp8:
991
+ .loc 2 120 46
992
+ mov.b32 %r270, %f415;
993
+ shfl.sync.bfly.b32 %r271, %r270, 1, 31, -1;
994
+ mov.b32 %f420, %r271;
995
+ mov.b32 %r272, %f419;
996
+ shfl.sync.bfly.b32 %r273, %r272, 1, 31, -1;
997
+ mov.b32 %f421, %r273;
998
+ shfl.sync.bfly.b32 %r194, %r192, 1, 31, -1;
999
+ mov.b32 %f422, %r194;
1000
+ $L__tmp9:
1001
+ .loc 2 108 21
1002
+ sub.f32 %f423, %f420, %f415;
1003
+ .loc 2 109 28
1004
+ add.f32 %f424, %f412, %f422;
1005
+ .loc 2 110 39
1006
+ setp.eq.f32 %p79, %f424, 0f00000000;
1007
+ .loc 2 110 60
1008
+ mov.b32 %r195, %f424;
1009
+ div.full.f32 %r193, %r194, %r195;
1010
+ mov.b32 %f425, %r193;
1011
+ .loc 2 110 49
1012
+ selp.f32 %f426, 0f00000000, %f425, %p79;
1013
+ .loc 2 112 17
1014
+ fma.rn.f32 %f145, %f423, %f426, %f415;
1015
+ .loc 2 113 15
1016
+ add.f32 %f427, %f419, %f421;
1017
+ .loc 2 113 30
1018
+ mul.f32 %f428, %f423, %f423;
1019
+ .loc 2 113 38
1020
+ mul.f32 %f429, %f412, %f428;
1021
+ .loc 2 113 22
1022
+ fma.rn.f32 %f430, %f426, %f429, %f427;
1023
+ $L__tmp10:
1024
+ .loc 2 120 46
1025
+ mov.b32 %r274, %f391;
1026
+ shfl.sync.bfly.b32 %r275, %r274, 4, 31, -1;
1027
+ mov.b32 %f431, %r275;
1028
+ mov.b32 %r276, %f395;
1029
+ shfl.sync.bfly.b32 %r277, %r276, 4, 31, -1;
1030
+ mov.b32 %f432, %r277;
1031
+ shfl.sync.bfly.b32 %r197, %r186, 4, 31, -1;
1032
+ mov.b32 %f433, %r197;
1033
+ $L__tmp11:
1034
+ .loc 2 108 21
1035
+ sub.f32 %f434, %f431, %f391;
1036
+ .loc 2 109 28
1037
+ add.f32 %f435, %f388, %f433;
1038
+ .loc 2 110 39
1039
+ setp.eq.f32 %p80, %f435, 0f00000000;
1040
+ .loc 2 110 60
1041
+ mov.b32 %r198, %f435;
1042
+ div.full.f32 %r196, %r197, %r198;
1043
+ mov.b32 %f436, %r196;
1044
+ .loc 2 110 49
1045
+ selp.f32 %f437, 0f00000000, %f436, %p80;
1046
+ .loc 2 112 17
1047
+ fma.rn.f32 %f438, %f434, %f437, %f391;
1048
+ .loc 2 113 15
1049
+ add.f32 %f439, %f395, %f432;
1050
+ .loc 2 113 30
1051
+ mul.f32 %f440, %f434, %f434;
1052
+ .loc 2 113 38
1053
+ mul.f32 %f441, %f388, %f440;
1054
+ .loc 2 113 22
1055
+ fma.rn.f32 %f442, %f441, %f437, %f439;
1056
+ $L__tmp12:
1057
+ .loc 2 120 46
1058
+ mov.b32 %r278, %f438;
1059
+ shfl.sync.bfly.b32 %r279, %r278, 2, 31, -1;
1060
+ mov.b32 %f443, %r279;
1061
+ mov.b32 %r280, %f442;
1062
+ shfl.sync.bfly.b32 %r281, %r280, 2, 31, -1;
1063
+ mov.b32 %f444, %r281;
1064
+ shfl.sync.bfly.b32 %r200, %r198, 2, 31, -1;
1065
+ mov.b32 %f445, %r200;
1066
+ $L__tmp13:
1067
+ .loc 2 108 21
1068
+ sub.f32 %f446, %f443, %f438;
1069
+ .loc 2 109 28
1070
+ add.f32 %f447, %f435, %f445;
1071
+ .loc 2 110 39
1072
+ setp.eq.f32 %p81, %f447, 0f00000000;
1073
+ .loc 2 110 60
1074
+ mov.b32 %r201, %f447;
1075
+ div.full.f32 %r199, %r200, %r201;
1076
+ mov.b32 %f448, %r199;
1077
+ .loc 2 110 49
1078
+ selp.f32 %f449, 0f00000000, %f448, %p81;
1079
+ .loc 2 112 17
1080
+ fma.rn.f32 %f450, %f446, %f449, %f438;
1081
+ .loc 2 113 15
1082
+ add.f32 %f451, %f442, %f444;
1083
+ .loc 2 113 30
1084
+ mul.f32 %f452, %f446, %f446;
1085
+ .loc 2 113 38
1086
+ mul.f32 %f453, %f435, %f452;
1087
+ .loc 2 113 22
1088
+ fma.rn.f32 %f454, %f449, %f453, %f451;
1089
+ $L__tmp14:
1090
+ .loc 2 120 46
1091
+ mov.b32 %r282, %f450;
1092
+ shfl.sync.bfly.b32 %r283, %r282, 1, 31, -1;
1093
+ mov.b32 %f455, %r283;
1094
+ mov.b32 %r284, %f454;
1095
+ shfl.sync.bfly.b32 %r285, %r284, 1, 31, -1;
1096
+ mov.b32 %f456, %r285;
1097
+ shfl.sync.bfly.b32 %r203, %r201, 1, 31, -1;
1098
+ mov.b32 %f457, %r203;
1099
+ $L__tmp15:
1100
+ .loc 2 108 21
1101
+ sub.f32 %f458, %f455, %f450;
1102
+ .loc 2 109 28
1103
+ add.f32 %f459, %f447, %f457;
1104
+ .loc 2 110 39
1105
+ setp.eq.f32 %p82, %f459, 0f00000000;
1106
+ .loc 2 110 60
1107
+ mov.b32 %r204, %f459;
1108
+ div.full.f32 %r202, %r203, %r204;
1109
+ mov.b32 %f460, %r202;
1110
+ .loc 2 110 49
1111
+ selp.f32 %f461, 0f00000000, %f460, %p82;
1112
+ .loc 2 112 17
1113
+ fma.rn.f32 %f146, %f458, %f461, %f450;
1114
+ .loc 2 113 15
1115
+ add.f32 %f462, %f454, %f456;
1116
+ .loc 2 113 30
1117
+ mul.f32 %f463, %f458, %f458;
1118
+ .loc 2 113 38
1119
+ mul.f32 %f464, %f447, %f463;
1120
+ .loc 2 113 22
1121
+ fma.rn.f32 %f465, %f461, %f464, %f462;
1122
+ $L__tmp16:
1123
+ .loc 1 69 23
1124
+ mov.b32 %r206, %f430;
1125
+ mov.b32 %r207, 1132462080;
1126
+ div.full.f32 %r205, %r206, %r207;
1127
+ mov.b32 %f466, %r205;
1128
+ mov.b32 %r230, %f465;
1129
+ div.full.f32 %r229, %r230, %r207;
1130
+ mov.b32 %f467, %r229;
1131
+ .loc 1 71 24
1132
+ add.f32 %f147, %f466, 0f3727C5AC;
1133
+ add.f32 %f148, %f467, 0f3727C5AC;
1134
+ .loc 1 55 36
1135
+ add.s64 %rd9, %rd12, %rd2;
1136
+ shl.b32 %r286, %r11, 14;
1137
+ shl.b32 %r287, %r1, 8;
1138
+ or.b32 %r288, %r286, %r287;
1139
+ or.b32 %r8, %r288, %r3;
1140
+ mov.u64 %rd106, 0;
1141
+ mov.b32 %r407, -64;
1142
+ rsqrt.approx.ftz.f32 %f516, %f147;
1143
+ rsqrt.approx.ftz.f32 %f517, %f148;
1144
+ bra.uni $L__BB0_5;
1145
+ $L__BB0_7:
1146
+ .loc 1 65 35
1147
+ add.s64 %rd96, %rd4, %rd106;
1148
+ add.s64 %rd97, %rd96, 16;
1149
+ add.s64 %rd98, %rd3, %rd106;
1150
+ .loc 1 65 54
1151
+ add.s64 %rd99, %rd98, 16;
1152
+ mov.u32 %r338, 0x0;
1153
+ mov.u32 %r339, 0x0;
1154
+ mov.u32 %r340, 0x0;
1155
+ mov.u32 %r341, 0x0;
1156
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r338, %r339, %r340, %r341 }, [ %rd96 + 0 ];
1157
+ @!%p1 mov.u32 %r338, %r342;
1158
+ @!%p1 mov.u32 %r339, %r342;
1159
+ @!%p1 mov.u32 %r340, %r342;
1160
+ @!%p1 mov.u32 %r341, %r342;
1161
+ mov.b32 %f468, %r338;
1162
+ mov.b32 %f469, %r339;
1163
+ mov.b32 %f470, %r340;
1164
+ mov.b32 %f471, %r341;
1165
+ mov.u32 %r346, 0x0;
1166
+ mov.u32 %r347, 0x0;
1167
+ mov.u32 %r348, 0x0;
1168
+ mov.u32 %r349, 0x0;
1169
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r346, %r347, %r348, %r349 }, [ %rd97 + 0 ];
1170
+ @!%p1 mov.u32 %r346, %r342;
1171
+ @!%p1 mov.u32 %r347, %r342;
1172
+ @!%p1 mov.u32 %r348, %r342;
1173
+ @!%p1 mov.u32 %r349, %r342;
1174
+ mov.b32 %f472, %r346;
1175
+ mov.b32 %f473, %r347;
1176
+ mov.b32 %f474, %r348;
1177
+ mov.b32 %f475, %r349;
1178
+ mov.u32 %r354, 0x0;
1179
+ mov.u32 %r355, 0x0;
1180
+ mov.u32 %r356, 0x0;
1181
+ mov.u32 %r357, 0x0;
1182
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r354, %r355, %r356, %r357 }, [ %rd98 + 0 ];
1183
+ @!%p1 mov.u32 %r354, %r342;
1184
+ @!%p1 mov.u32 %r355, %r342;
1185
+ @!%p1 mov.u32 %r356, %r342;
1186
+ @!%p1 mov.u32 %r357, %r342;
1187
+ mov.b32 %f476, %r354;
1188
+ mov.b32 %f477, %r355;
1189
+ mov.b32 %f478, %r356;
1190
+ mov.b32 %f479, %r357;
1191
+ mov.u32 %r362, 0x0;
1192
+ mov.u32 %r363, 0x0;
1193
+ mov.u32 %r364, 0x0;
1194
+ mov.u32 %r365, 0x0;
1195
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r362, %r363, %r364, %r365 }, [ %rd99 + 0 ];
1196
+ @!%p1 mov.u32 %r362, %r342;
1197
+ @!%p1 mov.u32 %r363, %r342;
1198
+ @!%p1 mov.u32 %r364, %r342;
1199
+ @!%p1 mov.u32 %r365, %r342;
1200
+ mov.b32 %f480, %r362;
1201
+ mov.b32 %f481, %r363;
1202
+ mov.b32 %f482, %r364;
1203
+ mov.b32 %f483, %r365;
1204
+ .loc 1 66 24
1205
+ add.f32 %f484, %f149, %f468;
1206
+ add.f32 %f485, %f150, %f469;
1207
+ add.f32 %f486, %f151, %f470;
1208
+ add.f32 %f487, %f152, %f471;
1209
+ add.f32 %f488, %f153, %f472;
1210
+ add.f32 %f489, %f154, %f473;
1211
+ add.f32 %f490, %f155, %f474;
1212
+ add.f32 %f491, %f156, %f475;
1213
+ add.f32 %f492, %f157, %f476;
1214
+ add.f32 %f493, %f158, %f477;
1215
+ add.f32 %f494, %f159, %f478;
1216
+ add.f32 %f495, %f160, %f479;
1217
+ add.f32 %f496, %f161, %f480;
1218
+ add.f32 %f497, %f162, %f481;
1219
+ add.f32 %f498, %f163, %f482;
1220
+ add.f32 %f499, %f164, %f483;
1221
+ .loc 1 67 24
1222
+ sub.f32 %f500, %f484, %f145;
1223
+ sub.f32 %f501, %f485, %f145;
1224
+ sub.f32 %f502, %f486, %f145;
1225
+ sub.f32 %f503, %f487, %f145;
1226
+ sub.f32 %f504, %f488, %f145;
1227
+ sub.f32 %f505, %f489, %f145;
1228
+ sub.f32 %f506, %f490, %f145;
1229
+ sub.f32 %f507, %f491, %f145;
1230
+ sub.f32 %f508, %f492, %f146;
1231
+ sub.f32 %f509, %f493, %f146;
1232
+ sub.f32 %f510, %f494, %f146;
1233
+ sub.f32 %f511, %f495, %f146;
1234
+ sub.f32 %f512, %f496, %f146;
1235
+ sub.f32 %f513, %f497, %f146;
1236
+ sub.f32 %f514, %f498, %f146;
1237
+ sub.f32 %f515, %f499, %f146;
1238
+ .loc 1 73 24
1239
+ mul.f32 %f518, %f500, %f516;
1240
+ mul.f32 %f519, %f501, %f516;
1241
+ mul.f32 %f520, %f502, %f516;
1242
+ mul.f32 %f521, %f503, %f516;
1243
+ mul.f32 %f522, %f504, %f516;
1244
+ mul.f32 %f523, %f505, %f516;
1245
+ mul.f32 %f524, %f506, %f516;
1246
+ mul.f32 %f525, %f507, %f516;
1247
+ mul.f32 %f526, %f508, %f517;
1248
+ mul.f32 %f527, %f509, %f517;
1249
+ mul.f32 %f528, %f510, %f517;
1250
+ mul.f32 %f529, %f511, %f517;
1251
+ mul.f32 %f530, %f512, %f517;
1252
+ mul.f32 %f531, %f513, %f517;
1253
+ mul.f32 %f532, %f514, %f517;
1254
+ mul.f32 %f533, %f515, %f517;
1255
+ .loc 1 74 24
1256
+ mul.f32 %f534, %f518, %f165;
1257
+ mul.f32 %f535, %f519, %f166;
1258
+ mul.f32 %f536, %f520, %f167;
1259
+ mul.f32 %f537, %f521, %f168;
1260
+ mul.f32 %f538, %f522, %f169;
1261
+ mul.f32 %f539, %f523, %f170;
1262
+ mul.f32 %f540, %f524, %f171;
1263
+ mul.f32 %f541, %f525, %f172;
1264
+ mul.f32 %f542, %f526, %f165;
1265
+ mul.f32 %f543, %f527, %f166;
1266
+ mul.f32 %f544, %f528, %f167;
1267
+ mul.f32 %f545, %f529, %f168;
1268
+ mul.f32 %f546, %f530, %f169;
1269
+ mul.f32 %f547, %f531, %f170;
1270
+ mul.f32 %f548, %f532, %f171;
1271
+ mul.f32 %f549, %f533, %f172;
1272
+ .loc 1 76 35
1273
+ add.s32 %r394, %r8, %r407;
1274
+ add.s32 %r395, %r394, 64;
1275
+ .loc 1 76 29
1276
+ add.s32 %r396, %r394, 8256;
1277
+ mul.wide.s32 %rd102, %r395, 2;
1278
+ add.s64 %rd100, %rd13, %rd102;
1279
+ mul.wide.s32 %rd103, %r396, 2;
1280
+ add.s64 %rd101, %rd13, %rd103;
1281
+ .loc 1 76 52
1282
+ mov.b32 %r370, %f534;
1283
+ cvt.rn.bf16.f32 %rs1, %r370;
1284
+ mov.b32 %r371, %f535;
1285
+ cvt.rn.bf16.f32 %rs2, %r371;
1286
+ mov.b32 %r372, %f536;
1287
+ cvt.rn.bf16.f32 %rs3, %r372;
1288
+ mov.b32 %r373, %f537;
1289
+ cvt.rn.bf16.f32 %rs4, %r373;
1290
+ mov.b32 %r374, %f538;
1291
+ cvt.rn.bf16.f32 %rs5, %r374;
1292
+ mov.b32 %r375, %f539;
1293
+ cvt.rn.bf16.f32 %rs6, %r375;
1294
+ mov.b32 %r376, %f540;
1295
+ cvt.rn.bf16.f32 %rs7, %r376;
1296
+ mov.b32 %r377, %f541;
1297
+ cvt.rn.bf16.f32 %rs8, %r377;
1298
+ mov.b32 %r378, %f542;
1299
+ cvt.rn.bf16.f32 %rs9, %r378;
1300
+ mov.b32 %r379, %f543;
1301
+ cvt.rn.bf16.f32 %rs10, %r379;
1302
+ mov.b32 %r380, %f544;
1303
+ cvt.rn.bf16.f32 %rs11, %r380;
1304
+ mov.b32 %r381, %f545;
1305
+ cvt.rn.bf16.f32 %rs12, %r381;
1306
+ mov.b32 %r382, %f546;
1307
+ cvt.rn.bf16.f32 %rs13, %r382;
1308
+ mov.b32 %r383, %f547;
1309
+ cvt.rn.bf16.f32 %rs14, %r383;
1310
+ mov.b32 %r384, %f548;
1311
+ cvt.rn.bf16.f32 %rs15, %r384;
1312
+ mov.b32 %r385, %f549;
1313
+ cvt.rn.bf16.f32 %rs16, %r385;
1314
+ mov.b32 %r397, {%rs1, %rs2};
1315
+ mov.b32 %r398, {%rs3, %rs4};
1316
+ mov.b32 %r399, {%rs5, %rs6};
1317
+ mov.b32 %r400, {%rs7, %rs8};
1318
+ @%p1 st.global.v4.b32 [ %rd100 + 0 ], { %r397, %r398, %r399, %r400 };
1319
+ mov.b32 %r401, {%rs9, %rs10};
1320
+ mov.b32 %r402, {%rs11, %rs12};
1321
+ mov.b32 %r403, {%rs13, %rs14};
1322
+ mov.b32 %r404, {%rs15, %rs16};
1323
+ @%p1 st.global.v4.b32 [ %rd101 + 0 ], { %r401, %r402, %r403, %r404 };
1324
+ .loc 1 55 36
1325
+ add.s64 %rd106, %rd106, 256;
1326
+ add.s32 %r407, %r407, 64;
1327
+ setp.lt.u32 %p136, %r407, 192;
1328
+ @%p136 bra $L__BB0_5;
1329
+ bra.uni $L__BB0_8;
1330
+ $L__BB0_5:
1331
+ .loc 1 59 35
1332
+ add.s64 %rd83, %rd6, %rd106;
1333
+ add.s64 %rd84, %rd83, 16;
1334
+ add.s64 %rd85, %rd5, %rd106;
1335
+ .loc 1 59 51
1336
+ add.s64 %rd86, %rd85, 16;
1337
+ mov.u32 %r289, 0x0;
1338
+ mov.u32 %r290, 0x0;
1339
+ mov.u32 %r291, 0x0;
1340
+ mov.u32 %r292, 0x0;
1341
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r289, %r290, %r291, %r292 }, [ %rd83 + 0 ];
1342
+ @!%p1 mov.u32 %r289, %r342;
1343
+ @!%p1 mov.u32 %r290, %r342;
1344
+ @!%p1 mov.u32 %r291, %r342;
1345
+ @!%p1 mov.u32 %r292, %r342;
1346
+ mov.b32 %f149, %r289;
1347
+ mov.b32 %f150, %r290;
1348
+ mov.b32 %f151, %r291;
1349
+ mov.b32 %f152, %r292;
1350
+ mov.u32 %r297, 0x0;
1351
+ mov.u32 %r298, 0x0;
1352
+ mov.u32 %r299, 0x0;
1353
+ mov.u32 %r300, 0x0;
1354
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r297, %r298, %r299, %r300 }, [ %rd84 + 0 ];
1355
+ @!%p1 mov.u32 %r297, %r342;
1356
+ @!%p1 mov.u32 %r298, %r342;
1357
+ @!%p1 mov.u32 %r299, %r342;
1358
+ @!%p1 mov.u32 %r300, %r342;
1359
+ mov.b32 %f153, %r297;
1360
+ mov.b32 %f154, %r298;
1361
+ mov.b32 %f155, %r299;
1362
+ mov.b32 %f156, %r300;
1363
+ mov.u32 %r305, 0x0;
1364
+ mov.u32 %r306, 0x0;
1365
+ mov.u32 %r307, 0x0;
1366
+ mov.u32 %r308, 0x0;
1367
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r305, %r306, %r307, %r308 }, [ %rd85 + 0 ];
1368
+ @!%p1 mov.u32 %r305, %r342;
1369
+ @!%p1 mov.u32 %r306, %r342;
1370
+ @!%p1 mov.u32 %r307, %r342;
1371
+ @!%p1 mov.u32 %r308, %r342;
1372
+ mov.b32 %f157, %r305;
1373
+ mov.b32 %f158, %r306;
1374
+ mov.b32 %f159, %r307;
1375
+ mov.b32 %f160, %r308;
1376
+ mov.u32 %r313, 0x0;
1377
+ mov.u32 %r314, 0x0;
1378
+ mov.u32 %r315, 0x0;
1379
+ mov.u32 %r316, 0x0;
1380
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r313, %r314, %r315, %r316 }, [ %rd86 + 0 ];
1381
+ @!%p1 mov.u32 %r313, %r342;
1382
+ @!%p1 mov.u32 %r314, %r342;
1383
+ @!%p1 mov.u32 %r315, %r342;
1384
+ @!%p1 mov.u32 %r316, %r342;
1385
+ mov.b32 %f161, %r313;
1386
+ mov.b32 %f162, %r314;
1387
+ mov.b32 %f163, %r315;
1388
+ mov.b32 %f164, %r316;
1389
+ .loc 1 60 35
1390
+ add.s64 %rd87, %rd9, %rd106;
1391
+ .loc 1 60 40
1392
+ add.s64 %rd88, %rd87, 16;
1393
+ mov.u32 %r321, 0x0;
1394
+ mov.u32 %r322, 0x0;
1395
+ mov.u32 %r323, 0x0;
1396
+ mov.u32 %r324, 0x0;
1397
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r321, %r322, %r323, %r324 }, [ %rd87 + 0 ];
1398
+ @!%p1 mov.u32 %r321, %r342;
1399
+ @!%p1 mov.u32 %r322, %r342;
1400
+ @!%p1 mov.u32 %r323, %r342;
1401
+ @!%p1 mov.u32 %r324, %r342;
1402
+ mov.b32 %f165, %r321;
1403
+ mov.b32 %f166, %r322;
1404
+ mov.b32 %f167, %r323;
1405
+ mov.b32 %f168, %r324;
1406
+ mov.u32 %r329, 0x0;
1407
+ mov.u32 %r330, 0x0;
1408
+ mov.u32 %r331, 0x0;
1409
+ mov.u32 %r332, 0x0;
1410
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r329, %r330, %r331, %r332 }, [ %rd88 + 0 ];
1411
+ @!%p1 mov.u32 %r329, %r342;
1412
+ @!%p1 mov.u32 %r330, %r342;
1413
+ @!%p1 mov.u32 %r331, %r342;
1414
+ @!%p1 mov.u32 %r332, %r342;
1415
+ mov.b32 %f169, %r329;
1416
+ mov.b32 %f170, %r330;
1417
+ mov.b32 %f171, %r331;
1418
+ mov.b32 %f172, %r332;
1419
+ .loc 1 64 57
1420
+ @%p41 bra $L__BB0_7;
1421
+ mov.u64 %rd89, assertMessage_1;
1422
+ cvta.global.u64 %rd90, %rd89;
1423
+ mov.u64 %rd91, assertFile_1;
1424
+ cvta.global.u64 %rd92, %rd91;
1425
+ mov.u64 %rd93, assertFunc_1;
1426
+ cvta.global.u64 %rd94, %rd93;
1427
+ { // callseq 7, 0
1428
+ .reg .b32 temp_param_reg;
1429
+ .param .b64 param0;
1430
+ st.param.b64 [param0+0], %rd90;
1431
+ .param .b64 param1;
1432
+ st.param.b64 [param1+0], %rd92;
1433
+ .param .b32 param2;
1434
+ st.param.b32 [param2+0], %r405;
1435
+ .param .b64 param3;
1436
+ st.param.b64 [param3+0], %rd94;
1437
+ .param .b64 param4;
1438
+ st.param.b64 [param4+0], %rd104;
1439
+ call.uni
1440
+ __assertfail,
1441
+ (
1442
+ param0,
1443
+ param1,
1444
+ param2,
1445
+ param3,
1446
+ param4
1447
+ );
1448
+ } // callseq 7
1449
+ bra.uni $L__BB0_7;
1450
+ $L__BB0_8:
1451
+ .loc 1 55 4
1452
+ ret;
1453
+ $L__tmp17:
1454
+ $L__func_end0:
1455
+
1456
+ }
1457
+ // .globl __nv_rsqrtf
1458
+ .visible .func (.param .b32 func_retval0) __nv_rsqrtf(
1459
+ .param .b32 __nv_rsqrtf_param_0
1460
+ )
1461
+ {
1462
+ .reg .f32 %f<3>;
1463
+ $L__func_begin1:
1464
+
1465
+ ld.param.f32 %f1, [__nv_rsqrtf_param_0];
1466
+ rsqrt.approx.ftz.f32 %f2, %f1;
1467
+ st.param.f32 [func_retval0+0], %f2;
1468
+ ret;
1469
+ $L__func_end1:
1470
+
1471
+ }
1472
+ .file 1 "/tmp/torchinductor_root/lh/clhe4a3stvufxafmq3kk5hodazz2efctffte646znjdnv3lqi5oa.py"
1473
+ .file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
1474
+ .section .debug_abbrev
1475
+ {
1476
+ .b8 1
1477
+ .b8 17
1478
+ .b8 1
1479
+ .b8 37
1480
+ .b8 8
1481
+ .b8 19
1482
+ .b8 5
1483
+ .b8 3
1484
+ .b8 8
1485
+ .b8 16
1486
+ .b8 6
1487
+ .b8 27
1488
+ .b8 8
1489
+ .b8 180
1490
+ .b8 66
1491
+ .b8 12
1492
+ .b8 17
1493
+ .b8 1
1494
+ .b8 18
1495
+ .b8 1
1496
+ .b8 0
1497
+ .b8 0
1498
+ .b8 2
1499
+ .b8 46
1500
+ .b8 0
1501
+ .b8 135
1502
+ .b8 64
1503
+ .b8 8
1504
+ .b8 3
1505
+ .b8 8
1506
+ .b8 58
1507
+ .b8 11
1508
+ .b8 59
1509
+ .b8 11
1510
+ .b8 63
1511
+ .b8 12
1512
+ .b8 32
1513
+ .b8 11
1514
+ .b8 0
1515
+ .b8 0
1516
+ .b8 3
1517
+ .b8 46
1518
+ .b8 1
1519
+ .b8 17
1520
+ .b8 1
1521
+ .b8 18
1522
+ .b8 1
1523
+ .b8 64
1524
+ .b8 10
1525
+ .b8 49
1526
+ .b8 19
1527
+ .b8 0
1528
+ .b8 0
1529
+ .b8 4
1530
+ .b8 29
1531
+ .b8 0
1532
+ .b8 49
1533
+ .b8 19
1534
+ .b8 17
1535
+ .b8 1
1536
+ .b8 18
1537
+ .b8 1
1538
+ .b8 88
1539
+ .b8 11
1540
+ .b8 89
1541
+ .b8 11
1542
+ .b8 87
1543
+ .b8 11
1544
+ .b8 0
1545
+ .b8 0
1546
+ .b8 5
1547
+ .b8 29
1548
+ .b8 1
1549
+ .b8 49
1550
+ .b8 19
1551
+ .b8 17
1552
+ .b8 1
1553
+ .b8 18
1554
+ .b8 1
1555
+ .b8 88
1556
+ .b8 11
1557
+ .b8 89
1558
+ .b8 11
1559
+ .b8 87
1560
+ .b8 11
1561
+ .b8 0
1562
+ .b8 0
1563
+ .b8 0
1564
+ }
1565
+ .section .debug_info
1566
+ {
1567
+ .b32 298
1568
+ .b8 2
1569
+ .b8 0
1570
+ .b32 .debug_abbrev
1571
+ .b8 8
1572
+ .b8 1
1573
+ .b8 116
1574
+ .b8 114
1575
+ .b8 105
1576
+ .b8 116
1577
+ .b8 111
1578
+ .b8 110
1579
+ .b8 0
1580
+ .b8 2
1581
+ .b8 0
1582
+ .b8 99
1583
+ .b8 108
1584
+ .b8 104
1585
+ .b8 101
1586
+ .b8 52
1587
+ .b8 97
1588
+ .b8 51
1589
+ .b8 115
1590
+ .b8 116
1591
+ .b8 118
1592
+ .b8 117
1593
+ .b8 102
1594
+ .b8 120
1595
+ .b8 97
1596
+ .b8 102
1597
+ .b8 109
1598
+ .b8 113
1599
+ .b8 51
1600
+ .b8 107
1601
+ .b8 107
1602
+ .b8 53
1603
+ .b8 104
1604
+ .b8 111
1605
+ .b8 100
1606
+ .b8 97
1607
+ .b8 122
1608
+ .b8 122
1609
+ .b8 50
1610
+ .b8 101
1611
+ .b8 102
1612
+ .b8 99
1613
+ .b8 116
1614
+ .b8 102
1615
+ .b8 102
1616
+ .b8 116
1617
+ .b8 101
1618
+ .b8 54
1619
+ .b8 52
1620
+ .b8 54
1621
+ .b8 122
1622
+ .b8 110
1623
+ .b8 106
1624
+ .b8 100
1625
+ .b8 110
1626
+ .b8 118
1627
+ .b8 51
1628
+ .b8 108
1629
+ .b8 113
1630
+ .b8 105
1631
+ .b8 53
1632
+ .b8 111
1633
+ .b8 97
1634
+ .b8 46
1635
+ .b8 112
1636
+ .b8 121
1637
+ .b8 0
1638
+ .b32 .debug_line
1639
+ .b8 47
1640
+ .b8 116
1641
+ .b8 109
1642
+ .b8 112
1643
+ .b8 47
1644
+ .b8 116
1645
+ .b8 111
1646
+ .b8 114
1647
+ .b8 99
1648
+ .b8 104
1649
+ .b8 105
1650
+ .b8 110
1651
+ .b8 100
1652
+ .b8 117
1653
+ .b8 99
1654
+ .b8 116
1655
+ .b8 111
1656
+ .b8 114
1657
+ .b8 95
1658
+ .b8 114
1659
+ .b8 111
1660
+ .b8 111
1661
+ .b8 116
1662
+ .b8 47
1663
+ .b8 108
1664
+ .b8 104
1665
+ .b8 0
1666
+ .b8 1
1667
+ .b64 $L__func_begin0
1668
+ .b64 $L__func_end0
1669
+ .b8 2
1670
+ .b8 116
1671
+ .b8 114
1672
+ .b8 105
1673
+ .b8 116
1674
+ .b8 111
1675
+ .b8 110
1676
+ .b8 95
1677
+ .b8 95
1678
+ .b8 48
1679
+ .b8 100
1680
+ .b8 49
1681
+ .b8 100
1682
+ .b8 50
1683
+ .b8 100
1684
+ .b8 51
1685
+ .b8 100
1686
+ .b8 52
1687
+ .b8 100
1688
+ .b8 53
1689
+ .b8 100
1690
+ .b8 101
1691
+ .b8 54
1692
+ .b8 100
1693
+ .b8 101
1694
+ .b8 0
1695
+ .b8 116
1696
+ .b8 114
1697
+ .b8 105
1698
+ .b8 116
1699
+ .b8 111
1700
+ .b8 110
1701
+ .b8 95
1702
+ .b8 95
1703
+ .b8 48
1704
+ .b8 100
1705
+ .b8 49
1706
+ .b8 100
1707
+ .b8 50
1708
+ .b8 100
1709
+ .b8 51
1710
+ .b8 100
1711
+ .b8 52
1712
+ .b8 100
1713
+ .b8 53
1714
+ .b8 100
1715
+ .b8 101
1716
+ .b8 54
1717
+ .b8 100
1718
+ .b8 101
1719
+ .b8 0
1720
+ .b8 1
1721
+ .b8 18
1722
+ .b8 1
1723
+ .b8 1
1724
+ .b8 3
1725
+ .b64 $L__func_begin0
1726
+ .b64 $L__func_end0
1727
+ .b8 1
1728
+ .b8 156
1729
+ .b32 125
1730
+ .b8 4
1731
+ .b32 125
1732
+ .b64 $L__tmp1
1733
+ .b64 $L__tmp2
1734
+ .b8 2
1735
+ .b8 44
1736
+ .b8 38
1737
+ .b8 5
1738
+ .b32 125
1739
+ .b64 $L__tmp3
1740
+ .b64 $L__tmp16
1741
+ .b8 2
1742
+ .b8 50
1743
+ .b8 41
1744
+ .b8 4
1745
+ .b32 125
1746
+ .b64 $L__tmp3
1747
+ .b64 $L__tmp16
1748
+ .b8 2
1749
+ .b8 120
1750
+ .b8 46
1751
+ .b8 0
1752
+ .b8 4
1753
+ .b32 125
1754
+ .b64 $L__tmp4
1755
+ .b64 $L__tmp15
1756
+ .b8 2
1757
+ .b8 50
1758
+ .b8 41
1759
+ .b8 0
1760
+ .b8 0
1761
+ }
1762
+ .section .debug_pubnames
1763
+ {
1764
+ .b32 $L__pubNames_end0-$L__pubNames_start0
1765
+ $L__pubNames_start0:
1766
+ .b8 2
1767
+ .b8 0
1768
+ .b32 .debug_info
1769
+ .b32 302
1770
+ .b32 125
1771
+ .b8 116
1772
+ .b8 114
1773
+ .b8 105
1774
+ .b8 116
1775
+ .b8 111
1776
+ .b8 110
1777
+ .b8 95
1778
+ .b8 95
1779
+ .b8 48
1780
+ .b8 100
1781
+ .b8 49
1782
+ .b8 100
1783
+ .b8 50
1784
+ .b8 100
1785
+ .b8 51
1786
+ .b8 100
1787
+ .b8 52
1788
+ .b8 100
1789
+ .b8 53
1790
+ .b8 100
1791
+ .b8 101
1792
+ .b8 54
1793
+ .b8 100
1794
+ .b8 101
1795
+ .b8 0
1796
+ .b32 0
1797
+ $L__pubNames_end0:
1798
+ }
1799
+ .section .debug_pubtypes
1800
+ {
1801
+ .b32 $L__pubTypes_end0-$L__pubTypes_start0
1802
+ $L__pubTypes_start0:
1803
+ .b8 2
1804
+ .b8 0
1805
+ .b32 .debug_info
1806
+ .b32 302
1807
+ .b32 0
1808
+ $L__pubTypes_end0:
1809
+ }
1810
+ .section .debug_loc { }
.triton/dump/510522bb05917b836ed253751364fcad/triton_.ttgir ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [1, 8], threadsPerWarp = [4, 8], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
2
+ #blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
3
+ #blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [4, 2], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
4
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 8 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
5
+ tt.func public @triton__0d1d2d3d4d5de6de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg5: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
6
+ %cst = arith.constant dense<512> : tensor<64x1xi32, #blocked>
7
+ %cst_0 = arith.constant dense<256> : tensor<1x64xi32, #blocked>
8
+ %cst_1 = arith.constant dense<256> : tensor<64x1xi32, #blocked>
9
+ %cst_2 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked>
10
+ %cst_3 = arith.constant dense<0.000000e+00> : tensor<1x64xf32, #blocked>
11
+ %cst_4 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked>
12
+ %cst_5 = arith.constant dense<256> : tensor<64x1xi64, #blocked>
13
+ %cst_6 = arith.constant dense<0> : tensor<64x1xi64, #blocked>
14
+ %cst_7 = arith.constant dense<50257> : tensor<64x1xi64, #blocked>
15
+ %cst_8 = arith.constant dense<50257> : tensor<64x1xi64, #blocked1>
16
+ %cst_9 = arith.constant dense<0> : tensor<64x1xi64, #blocked1>
17
+ %c0_i32 = arith.constant 0 : i32
18
+ %c64_i32 = arith.constant 64 : i32
19
+ %c256_i32 = arith.constant 256 : i32
20
+ %cst_10 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked2>
21
+ %cst_11 = arith.constant 0.000000e+00 : f32
22
+ %cst_12 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked2>
23
+ %cst_13 = arith.constant dense<256> : tensor<1x64xi32, #blocked2>
24
+ %cst_14 = arith.constant dense<9.99999974E-6> : tensor<64x1xf32, #blocked>
25
+ %cst_15 = arith.constant dense<2.560000e+02> : tensor<64x1xf32, #blocked>
26
+ %0 = tt.get_program_id x : i32
27
+ %1 = arith.muli %0, %c64_i32 : i32
28
+ %2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
29
+ %3 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
30
+ %4 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xi32, #blocked>
31
+ %5 = tt.expand_dims %3 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<64x1xi32, #blocked1>
32
+ %6 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked>
33
+ %7 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked1>
34
+ %8 = arith.addi %6, %4 : tensor<64x1xi32, #blocked>
35
+ %9 = arith.addi %7, %5 : tensor<64x1xi32, #blocked1>
36
+ %10 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
37
+ %11 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>
38
+ %12 = tt.expand_dims %10 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x64xi32, #blocked>
39
+ %13 = tt.expand_dims %11 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>) -> tensor<1x64xi32, #blocked2>
40
+ %14 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked>
41
+ %15 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked1>
42
+ %16 = tt.addptr %14, %8 : tensor<64x1x!tt.ptr<i64, 1>, #blocked>, tensor<64x1xi32, #blocked>
43
+ %17 = tt.addptr %15, %9 : tensor<64x1x!tt.ptr<i64, 1>, #blocked1>, tensor<64x1xi32, #blocked1>
44
+ %18 = tt.load %16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked>
45
+ %19 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked1>
46
+ %20 = arith.remsi %8, %cst : tensor<64x1xi32, #blocked>
47
+ %21 = arith.muli %20, %cst_1 : tensor<64x1xi32, #blocked>
48
+ %22 = tt.broadcast %21 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
49
+ %23 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
50
+ %24 = arith.addi %18, %cst_7 : tensor<64x1xi64, #blocked>
51
+ %25 = arith.addi %19, %cst_8 : tensor<64x1xi64, #blocked1>
52
+ %26 = arith.cmpi slt, %18, %cst_6 : tensor<64x1xi64, #blocked>
53
+ %27 = arith.cmpi slt, %19, %cst_9 : tensor<64x1xi64, #blocked1>
54
+ %28 = arith.select %26, %24, %18 : tensor<64x1xi1, #blocked>, tensor<64x1xi64, #blocked>
55
+ %29 = arith.select %27, %25, %19 : tensor<64x1xi1, #blocked1>, tensor<64x1xi64, #blocked1>
56
+ %30 = arith.cmpi sge, %29, %cst_9 : tensor<64x1xi64, #blocked1>
57
+ %31 = arith.cmpi slt, %29, %cst_8 : tensor<64x1xi64, #blocked1>
58
+ %32 = arith.andi %30, %31 : tensor<64x1xi1, #blocked1>
59
+ %33 = arith.muli %28, %cst_5 : tensor<64x1xi64, #blocked>
60
+ %34 = tt.broadcast %33 : (tensor<64x1xi64, #blocked>) -> tensor<64x64xi64, #blocked>
61
+ %35 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
62
+ %36:4 = scf.for %arg7 = %c0_i32 to %c256_i32 step %c64_i32 iter_args(%arg8 = %cst_2, %arg9 = %cst_2, %arg10 = %cst_12, %arg11 = %cst_2) -> (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>) : i32 {
63
+ %48 = tt.splat %arg7 : (i32) -> tensor<1x64xi32, #blocked>
64
+ %49 = tt.splat %arg7 : (i32) -> tensor<1x64xi32, #blocked2>
65
+ %50 = arith.addi %48, %12 : tensor<1x64xi32, #blocked>
66
+ %51 = arith.addi %49, %13 : tensor<1x64xi32, #blocked2>
67
+ %52 = arith.cmpi slt, %50, %cst_0 : tensor<1x64xi32, #blocked>
68
+ %53 = arith.cmpi slt, %51, %cst_13 : tensor<1x64xi32, #blocked2>
69
+ %54 = tt.broadcast %50 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
70
+ %55 = arith.addi %54, %22 : tensor<64x64xi32, #blocked>
71
+ %56 = tt.addptr %23, %55 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
72
+ %57 = tt.broadcast %52 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
73
+ %58 = tt.broadcast %53 : (tensor<1x64xi1, #blocked2>) -> tensor<64x64xi1, #blocked2>
74
+ %59 = tt.load %56, %57, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
75
+ tt.assert %32, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
76
+ %60 = arith.extsi %50 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
77
+ %61 = tt.broadcast %60 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
78
+ %62 = arith.addi %61, %34 : tensor<64x64xi64, #blocked>
79
+ %63 = tt.addptr %35, %62 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
80
+ %64 = tt.load %63, %57, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
81
+ %65 = arith.addf %64, %59 : tensor<64x64xf32, #blocked>
82
+ %66 = arith.subf %65, %arg8 : tensor<64x64xf32, #blocked>
83
+ %67 = arith.addf %arg11, %cst_4 : tensor<64x64xf32, #blocked>
84
+ %68 = arith.addf %arg10, %cst_10 : tensor<64x64xf32, #blocked2>
85
+ %69 = arith.divf %66, %67 : tensor<64x64xf32, #blocked>
86
+ %70 = arith.addf %arg8, %69 : tensor<64x64xf32, #blocked>
87
+ %71 = arith.subf %65, %70 : tensor<64x64xf32, #blocked>
88
+ %72 = arith.mulf %66, %71 : tensor<64x64xf32, #blocked>
89
+ %73 = arith.addf %arg9, %72 : tensor<64x64xf32, #blocked>
90
+ %74 = arith.select %57, %70, %arg8 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
91
+ %75 = arith.select %57, %73, %arg9 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
92
+ %76 = arith.select %57, %67, %arg11 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
93
+ %77 = arith.select %58, %68, %arg10 : tensor<64x64xi1, #blocked2>, tensor<64x64xf32, #blocked2>
94
+ scf.yield %74, %75, %77, %76 : tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>
95
+ }
96
+ %37 = triton_gpu.convert_layout %36#2 : (tensor<64x64xf32, #blocked2>) -> tensor<64x64xf32, #blocked>
97
+ %38:3 = "tt.reduce"(%36#0, %36#1, %37) <{axis = 1 : i32}> ({
98
+ ^bb0(%arg7: f32, %arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32):
99
+ %48 = arith.subf %arg10, %arg7 : f32
100
+ %49 = arith.addf %arg9, %arg12 : f32
101
+ %50 = arith.cmpf oeq, %49, %cst_11 : f32
102
+ %51 = arith.divf %arg12, %49 : f32
103
+ %52 = arith.select %50, %cst_11, %51 : f32
104
+ %53 = arith.mulf %48, %52 : f32
105
+ %54 = arith.addf %arg7, %53 : f32
106
+ %55 = arith.addf %arg8, %arg11 : f32
107
+ %56 = arith.mulf %48, %48 : f32
108
+ %57 = arith.mulf %56, %arg9 : f32
109
+ %58 = arith.mulf %57, %52 : f32
110
+ %59 = arith.addf %55, %58 : f32
111
+ tt.reduce.return %54, %59, %49 : f32, f32, f32
112
+ }) : (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>) -> (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>)
113
+ %39 = tt.expand_dims %38#0 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
114
+ %40 = tt.expand_dims %38#1 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
115
+ %41 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x64x!tt.ptr<f32, 1>, #blocked>
116
+ %42 = tt.broadcast %39 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
117
+ %43 = arith.divf %40, %cst_15 : tensor<64x1xf32, #blocked>
118
+ %44 = arith.addf %43, %cst_14 : tensor<64x1xf32, #blocked>
119
+ %45 = arith.muli %8, %cst_1 : tensor<64x1xi32, #blocked>
120
+ %46 = tt.broadcast %45 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
121
+ %47 = tt.splat %arg4 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>, #blocked>
122
+ scf.for %arg7 = %c0_i32 to %c256_i32 step %c64_i32 : i32 {
123
+ %48 = tt.splat %arg7 : (i32) -> tensor<1x64xi32, #blocked>
124
+ %49 = arith.addi %48, %12 : tensor<1x64xi32, #blocked>
125
+ %50 = arith.cmpi slt, %49, %cst_0 : tensor<1x64xi32, #blocked>
126
+ %51 = tt.broadcast %49 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
127
+ %52 = arith.addi %51, %22 : tensor<64x64xi32, #blocked>
128
+ %53 = tt.addptr %23, %52 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
129
+ %54 = tt.broadcast %50 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
130
+ %55 = tt.load %53, %54, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
131
+ %56 = tt.addptr %41, %49 : tensor<1x64x!tt.ptr<f32, 1>, #blocked>, tensor<1x64xi32, #blocked>
132
+ %57 = tt.load %56, %50, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x64xf32, #blocked>
133
+ tt.assert %32, "index out of bounds: 0 <= tmp13 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
134
+ %58 = arith.extsi %49 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
135
+ %59 = tt.broadcast %58 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
136
+ %60 = arith.addi %59, %34 : tensor<64x64xi64, #blocked>
137
+ %61 = tt.addptr %35, %60 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
138
+ %62 = tt.load %61, %54, %cst_2 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
139
+ %63 = arith.addf %62, %55 : tensor<64x64xf32, #blocked>
140
+ %64 = arith.subf %63, %42 : tensor<64x64xf32, #blocked>
141
+ %65 = tt.extern_elementwise %44 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<64x1xf32, #blocked>) -> tensor<64x1xf32, #blocked>
142
+ %66 = tt.broadcast %65 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
143
+ %67 = arith.mulf %64, %66 : tensor<64x64xf32, #blocked>
144
+ %68 = tt.broadcast %57 : (tensor<1x64xf32, #blocked>) -> tensor<64x64xf32, #blocked>
145
+ %69 = arith.mulf %67, %68 : tensor<64x64xf32, #blocked>
146
+ %70 = arith.addi %51, %46 : tensor<64x64xi32, #blocked>
147
+ %71 = tt.addptr %47, %70 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
148
+ %72 = arith.truncf %69 : tensor<64x64xf32, #blocked> to tensor<64x64xbf16, #blocked>
149
+ tt.store %71, %72, %54 {cache = 1 : i32, evict = 1 : i32} : tensor<64x64xbf16, #blocked>
150
+ }
151
+ tt.return
152
+ }
153
+ }
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.cubin ADDED
Binary file (73.7 kB). View file
 
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.llir ADDED
@@ -0,0 +1,1360 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+
4
+ @assertFunc_1 = internal constant [25 x i8] c"_call_with_frames_removed"
5
+ @assertFile_1 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
6
+ @assertMessage_1 = internal constant [39 x i8] c"index out of bounds: 0 <= tmp16 < 50257"
7
+ @assertFunc_0 = internal constant [25 x i8] c"_call_with_frames_removed"
8
+ @assertFile_0 = internal constant [38 x i8] c"<frozen importlib._bootstrap_external>"
9
+ @assertMessage_0 = internal constant [38 x i8] c"index out of bounds: 0 <= tmp3 < 50257"
10
+ @global_smem = external local_unnamed_addr addrspace(3) global [0 x i8]
11
+ @.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
12
+
13
+ declare void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr
14
+
15
+ define void @triton__0d1d2d3d4d5d6de7de(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, i32 %6, i32 %7) local_unnamed_addr !dbg !7 {
16
+ %9 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !10
17
+ %10 = lshr i32 %9, 3, !dbg !10
18
+ %11 = and i32 %10, 31, !dbg !10
19
+ %12 = and i32 %9, 63, !dbg !10
20
+ %13 = shl i32 %9, 3, !dbg !11
21
+ %14 = and i32 %13, 56, !dbg !11
22
+ %15 = or i32 %14, 4, !dbg !11
23
+ %16 = lshr i32 %9, 6, !dbg !12
24
+ %17 = tail call i32 asm "mov.u32 $0, %ctaid.x;", "=r"() #6, !dbg !13
25
+ %18 = shl i32 %17, 6, !dbg !14
26
+ %19 = or i32 %18, %11, !dbg !15
27
+ %20 = or i32 %19, 32, !dbg !15
28
+ %21 = or i32 %18, %12, !dbg !15
29
+ %22 = sext i32 %19 to i64, !dbg !16
30
+ %23 = getelementptr i64, ptr addrspace(1) %0, i64 %22, !dbg !16
31
+ %24 = sext i32 %20 to i64, !dbg !16
32
+ %25 = getelementptr i64, ptr addrspace(1) %0, i64 %24, !dbg !16
33
+ %26 = sext i32 %21 to i64, !dbg !16
34
+ %27 = getelementptr i64, ptr addrspace(1) %0, i64 %26, !dbg !16
35
+ %28 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
36
+ %29 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
37
+ %30 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
38
+ %31 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
39
+ %32 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
40
+ %33 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
41
+ %34 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
42
+ %35 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %23, i1 true) #6, !dbg !17
43
+ %36 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
44
+ %37 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
45
+ %38 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
46
+ %39 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
47
+ %40 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
48
+ %41 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
49
+ %42 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
50
+ %43 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %25, i1 true) #6, !dbg !17
51
+ %44 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %27, i1 true) #6, !dbg !17
52
+ %45 = srem i32 %19, 512, !dbg !18
53
+ %46 = srem i32 %20, 512, !dbg !18
54
+ %47 = shl nsw i32 %45, 8, !dbg !19
55
+ %48 = shl nsw i32 %46, 8, !dbg !19
56
+ %49 = shl i32 %19, 8, !dbg !20
57
+ %50 = shl i32 %20, 8, !dbg !20
58
+ %51 = add i64 %44, 50257, !dbg !21
59
+ %52 = icmp slt i64 %28, 0, !dbg !22
60
+ %53 = icmp slt i64 %36, 0, !dbg !22
61
+ %54 = icmp slt i64 %44, 0, !dbg !22
62
+ %55 = select i1 %54, i64 %51, i64 %44, !dbg !23
63
+ %56 = icmp ugt i64 %55, 50256, !dbg !24
64
+ %57 = shl i64 %28, 8, !dbg !25
65
+ %58 = add i64 %57, 12865792, !dbg !25
66
+ %59 = select i1 %52, i64 %58, i64 %57, !dbg !25
67
+ %60 = shl i64 %36, 8, !dbg !25
68
+ %61 = add i64 %60, 12865792, !dbg !25
69
+ %62 = select i1 %53, i64 %61, i64 %60, !dbg !25
70
+ %63 = getelementptr float, ptr addrspace(1) %1, i64 %59
71
+ %64 = getelementptr float, ptr addrspace(1) %1, i64 %62
72
+ br label %65, !dbg !12
73
+
74
+ 65: ; preds = %8, %230
75
+ %66 = phi float [ 0.000000e+00, %8 ], [ %321, %230 ]
76
+ %67 = phi float [ 0.000000e+00, %8 ], [ %322, %230 ]
77
+ %68 = phi float [ 0.000000e+00, %8 ], [ %323, %230 ]
78
+ %69 = phi float [ 0.000000e+00, %8 ], [ %324, %230 ]
79
+ %70 = phi float [ 0.000000e+00, %8 ], [ %325, %230 ]
80
+ %71 = phi float [ 0.000000e+00, %8 ], [ %326, %230 ]
81
+ %72 = phi float [ 0.000000e+00, %8 ], [ %327, %230 ]
82
+ %73 = phi float [ 0.000000e+00, %8 ], [ %328, %230 ]
83
+ %74 = phi float [ 0.000000e+00, %8 ], [ %329, %230 ]
84
+ %75 = phi float [ 0.000000e+00, %8 ], [ %330, %230 ]
85
+ %76 = phi float [ 0.000000e+00, %8 ], [ %331, %230 ]
86
+ %77 = phi float [ 0.000000e+00, %8 ], [ %332, %230 ]
87
+ %78 = phi float [ 0.000000e+00, %8 ], [ %333, %230 ]
88
+ %79 = phi float [ 0.000000e+00, %8 ], [ %334, %230 ]
89
+ %80 = phi float [ 0.000000e+00, %8 ], [ %335, %230 ]
90
+ %81 = phi float [ 0.000000e+00, %8 ], [ %336, %230 ]
91
+ %82 = phi float [ 0.000000e+00, %8 ], [ %337, %230 ]
92
+ %83 = phi float [ 0.000000e+00, %8 ], [ %338, %230 ]
93
+ %84 = phi float [ 0.000000e+00, %8 ], [ %339, %230 ]
94
+ %85 = phi float [ 0.000000e+00, %8 ], [ %340, %230 ]
95
+ %86 = phi float [ 0.000000e+00, %8 ], [ %341, %230 ]
96
+ %87 = phi float [ 0.000000e+00, %8 ], [ %342, %230 ]
97
+ %88 = phi float [ 0.000000e+00, %8 ], [ %343, %230 ]
98
+ %89 = phi float [ 0.000000e+00, %8 ], [ %344, %230 ]
99
+ %90 = phi float [ 0.000000e+00, %8 ], [ %345, %230 ]
100
+ %91 = phi float [ 0.000000e+00, %8 ], [ %346, %230 ]
101
+ %92 = phi float [ 0.000000e+00, %8 ], [ %347, %230 ]
102
+ %93 = phi float [ 0.000000e+00, %8 ], [ %348, %230 ]
103
+ %94 = phi float [ 0.000000e+00, %8 ], [ %349, %230 ]
104
+ %95 = phi float [ 0.000000e+00, %8 ], [ %350, %230 ]
105
+ %96 = phi float [ 0.000000e+00, %8 ], [ %351, %230 ]
106
+ %97 = phi float [ 0.000000e+00, %8 ], [ %352, %230 ]
107
+ %98 = phi float [ 0.000000e+00, %8 ], [ %417, %230 ]
108
+ %99 = phi float [ 0.000000e+00, %8 ], [ %418, %230 ]
109
+ %100 = phi float [ 0.000000e+00, %8 ], [ %419, %230 ]
110
+ %101 = phi float [ 0.000000e+00, %8 ], [ %420, %230 ]
111
+ %102 = phi float [ 0.000000e+00, %8 ], [ %421, %230 ]
112
+ %103 = phi float [ 0.000000e+00, %8 ], [ %422, %230 ]
113
+ %104 = phi float [ 0.000000e+00, %8 ], [ %423, %230 ]
114
+ %105 = phi float [ 0.000000e+00, %8 ], [ %424, %230 ]
115
+ %106 = phi float [ 0.000000e+00, %8 ], [ %425, %230 ]
116
+ %107 = phi float [ 0.000000e+00, %8 ], [ %426, %230 ]
117
+ %108 = phi float [ 0.000000e+00, %8 ], [ %427, %230 ]
118
+ %109 = phi float [ 0.000000e+00, %8 ], [ %428, %230 ]
119
+ %110 = phi float [ 0.000000e+00, %8 ], [ %429, %230 ]
120
+ %111 = phi float [ 0.000000e+00, %8 ], [ %430, %230 ]
121
+ %112 = phi float [ 0.000000e+00, %8 ], [ %431, %230 ]
122
+ %113 = phi float [ 0.000000e+00, %8 ], [ %432, %230 ]
123
+ %114 = phi float [ 0.000000e+00, %8 ], [ %369, %230 ]
124
+ %115 = phi float [ 0.000000e+00, %8 ], [ %370, %230 ]
125
+ %116 = phi float [ 0.000000e+00, %8 ], [ %371, %230 ]
126
+ %117 = phi float [ 0.000000e+00, %8 ], [ %372, %230 ]
127
+ %118 = phi float [ 0.000000e+00, %8 ], [ %373, %230 ]
128
+ %119 = phi float [ 0.000000e+00, %8 ], [ %374, %230 ]
129
+ %120 = phi float [ 0.000000e+00, %8 ], [ %375, %230 ]
130
+ %121 = phi float [ 0.000000e+00, %8 ], [ %376, %230 ]
131
+ %122 = phi float [ 0.000000e+00, %8 ], [ %377, %230 ]
132
+ %123 = phi float [ 0.000000e+00, %8 ], [ %378, %230 ]
133
+ %124 = phi float [ 0.000000e+00, %8 ], [ %379, %230 ]
134
+ %125 = phi float [ 0.000000e+00, %8 ], [ %380, %230 ]
135
+ %126 = phi float [ 0.000000e+00, %8 ], [ %381, %230 ]
136
+ %127 = phi float [ 0.000000e+00, %8 ], [ %382, %230 ]
137
+ %128 = phi float [ 0.000000e+00, %8 ], [ %383, %230 ]
138
+ %129 = phi float [ 0.000000e+00, %8 ], [ %384, %230 ]
139
+ %130 = phi i32 [ 0, %8 ], [ %433, %230 ]
140
+ %131 = or i32 %130, %14, !dbg !26
141
+ %132 = or i32 %130, %15, !dbg !26
142
+ %133 = add i32 %131, %47, !dbg !27
143
+ %134 = add i32 %132, %47, !dbg !27
144
+ %135 = add i32 %131, %48, !dbg !27
145
+ %136 = add i32 %132, %48, !dbg !27
146
+ %137 = sext i32 %133 to i64, !dbg !28
147
+ %138 = getelementptr float, ptr addrspace(1) %2, i64 %137, !dbg !28
148
+ %139 = sext i32 %134 to i64, !dbg !28
149
+ %140 = getelementptr float, ptr addrspace(1) %2, i64 %139, !dbg !28
150
+ %141 = sext i32 %135 to i64, !dbg !28
151
+ %142 = getelementptr float, ptr addrspace(1) %2, i64 %141, !dbg !28
152
+ %143 = sext i32 %136 to i64, !dbg !28
153
+ %144 = getelementptr float, ptr addrspace(1) %2, i64 %143, !dbg !28
154
+ %145 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %138, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
155
+ %146 = extractvalue { i32, i32, i32, i32 } %145, 0, !dbg !29
156
+ %147 = extractvalue { i32, i32, i32, i32 } %145, 1, !dbg !29
157
+ %148 = extractvalue { i32, i32, i32, i32 } %145, 2, !dbg !29
158
+ %149 = extractvalue { i32, i32, i32, i32 } %145, 3, !dbg !29
159
+ %150 = bitcast i32 %146 to float, !dbg !29
160
+ %151 = bitcast i32 %147 to float, !dbg !29
161
+ %152 = bitcast i32 %148 to float, !dbg !29
162
+ %153 = bitcast i32 %149 to float, !dbg !29
163
+ %154 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %140, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
164
+ %155 = extractvalue { i32, i32, i32, i32 } %154, 0, !dbg !29
165
+ %156 = extractvalue { i32, i32, i32, i32 } %154, 1, !dbg !29
166
+ %157 = extractvalue { i32, i32, i32, i32 } %154, 2, !dbg !29
167
+ %158 = extractvalue { i32, i32, i32, i32 } %154, 3, !dbg !29
168
+ %159 = bitcast i32 %155 to float, !dbg !29
169
+ %160 = bitcast i32 %156 to float, !dbg !29
170
+ %161 = bitcast i32 %157 to float, !dbg !29
171
+ %162 = bitcast i32 %158 to float, !dbg !29
172
+ %163 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %142, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
173
+ %164 = extractvalue { i32, i32, i32, i32 } %163, 0, !dbg !29
174
+ %165 = extractvalue { i32, i32, i32, i32 } %163, 1, !dbg !29
175
+ %166 = extractvalue { i32, i32, i32, i32 } %163, 2, !dbg !29
176
+ %167 = extractvalue { i32, i32, i32, i32 } %163, 3, !dbg !29
177
+ %168 = bitcast i32 %164 to float, !dbg !29
178
+ %169 = bitcast i32 %165 to float, !dbg !29
179
+ %170 = bitcast i32 %166 to float, !dbg !29
180
+ %171 = bitcast i32 %167 to float, !dbg !29
181
+ %172 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %144, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !29
182
+ %173 = extractvalue { i32, i32, i32, i32 } %172, 0, !dbg !29
183
+ %174 = extractvalue { i32, i32, i32, i32 } %172, 1, !dbg !29
184
+ %175 = extractvalue { i32, i32, i32, i32 } %172, 2, !dbg !29
185
+ %176 = extractvalue { i32, i32, i32, i32 } %172, 3, !dbg !29
186
+ %177 = bitcast i32 %173 to float, !dbg !29
187
+ %178 = bitcast i32 %174 to float, !dbg !29
188
+ %179 = bitcast i32 %175 to float, !dbg !29
189
+ %180 = bitcast i32 %176 to float, !dbg !29
190
+ %181 = add i32 %131, %49, !dbg !30
191
+ %182 = add i32 %131, %50, !dbg !30
192
+ %183 = sext i32 %181 to i64, !dbg !31
193
+ %184 = getelementptr i16, ptr addrspace(1) %3, i64 %183, !dbg !31
194
+ %185 = sext i32 %182 to i64, !dbg !31
195
+ %186 = getelementptr i16, ptr addrspace(1) %3, i64 %185, !dbg !31
196
+ %187 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %184, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !32
197
+ %188 = extractvalue { i32, i32, i32, i32 } %187, 0, !dbg !32
198
+ %189 = extractvalue { i32, i32, i32, i32 } %187, 1, !dbg !32
199
+ %190 = extractvalue { i32, i32, i32, i32 } %187, 2, !dbg !32
200
+ %191 = extractvalue { i32, i32, i32, i32 } %187, 3, !dbg !32
201
+ %192 = trunc i32 %188 to i16, !dbg !32
202
+ %extelt.offset9 = lshr i32 %188, 16, !dbg !32
203
+ %193 = trunc i32 %extelt.offset9 to i16, !dbg !32
204
+ %194 = trunc i32 %189 to i16, !dbg !32
205
+ %extelt.offset10 = lshr i32 %189, 16, !dbg !32
206
+ %195 = trunc i32 %extelt.offset10 to i16, !dbg !32
207
+ %196 = trunc i32 %190 to i16, !dbg !32
208
+ %extelt.offset11 = lshr i32 %190, 16, !dbg !32
209
+ %197 = trunc i32 %extelt.offset11 to i16, !dbg !32
210
+ %198 = trunc i32 %191 to i16, !dbg !32
211
+ %extelt.offset12 = lshr i32 %191, 16, !dbg !32
212
+ %199 = trunc i32 %extelt.offset12 to i16, !dbg !32
213
+ %200 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %186, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !32
214
+ %201 = extractvalue { i32, i32, i32, i32 } %200, 0, !dbg !32
215
+ %202 = extractvalue { i32, i32, i32, i32 } %200, 1, !dbg !32
216
+ %203 = extractvalue { i32, i32, i32, i32 } %200, 2, !dbg !32
217
+ %204 = extractvalue { i32, i32, i32, i32 } %200, 3, !dbg !32
218
+ %205 = trunc i32 %201 to i16, !dbg !32
219
+ %extelt.offset13 = lshr i32 %201, 16, !dbg !32
220
+ %206 = trunc i32 %extelt.offset13 to i16, !dbg !32
221
+ %207 = trunc i32 %202 to i16, !dbg !32
222
+ %extelt.offset14 = lshr i32 %202, 16, !dbg !32
223
+ %208 = trunc i32 %extelt.offset14 to i16, !dbg !32
224
+ %209 = trunc i32 %203 to i16, !dbg !32
225
+ %extelt.offset15 = lshr i32 %203, 16, !dbg !32
226
+ %210 = trunc i32 %extelt.offset15 to i16, !dbg !32
227
+ %211 = trunc i32 %204 to i16, !dbg !32
228
+ %extelt.offset16 = lshr i32 %204, 16, !dbg !32
229
+ %212 = trunc i32 %extelt.offset16 to i16, !dbg !32
230
+ %213 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %192) #6, !dbg !33
231
+ %214 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %193) #6, !dbg !33
232
+ %215 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %194) #6, !dbg !33
233
+ %216 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %195) #6, !dbg !33
234
+ %217 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %196) #6, !dbg !33
235
+ %218 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %197) #6, !dbg !33
236
+ %219 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %198) #6, !dbg !33
237
+ %220 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %199) #6, !dbg !33
238
+ %221 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %205) #6, !dbg !33
239
+ %222 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %206) #6, !dbg !33
240
+ %223 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %207) #6, !dbg !33
241
+ %224 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %208) #6, !dbg !33
242
+ %225 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %209) #6, !dbg !33
243
+ %226 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %210) #6, !dbg !33
244
+ %227 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %211) #6, !dbg !33
245
+ %228 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %212) #6, !dbg !33
246
+ br i1 %56, label %229, label %230, !dbg !34
247
+
248
+ 229: ; preds = %65
249
+ tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 883, ptr nonnull @assertFunc_0, i64 1), !dbg !34
250
+ br label %230, !dbg !34
251
+
252
+ 230: ; preds = %229, %65
253
+ %231 = zext nneg i32 %131 to i64, !dbg !35
254
+ %232 = zext nneg i32 %132 to i64, !dbg !35
255
+ %233 = getelementptr float, ptr addrspace(1) %63, i64 %231, !dbg !36
256
+ %234 = getelementptr float, ptr addrspace(1) %63, i64 %232, !dbg !36
257
+ %235 = getelementptr float, ptr addrspace(1) %64, i64 %231, !dbg !36
258
+ %236 = getelementptr float, ptr addrspace(1) %64, i64 %232, !dbg !36
259
+ %237 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %233, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
260
+ %238 = extractvalue { i32, i32, i32, i32 } %237, 0, !dbg !37
261
+ %239 = extractvalue { i32, i32, i32, i32 } %237, 1, !dbg !37
262
+ %240 = extractvalue { i32, i32, i32, i32 } %237, 2, !dbg !37
263
+ %241 = extractvalue { i32, i32, i32, i32 } %237, 3, !dbg !37
264
+ %242 = bitcast i32 %238 to float, !dbg !37
265
+ %243 = bitcast i32 %239 to float, !dbg !37
266
+ %244 = bitcast i32 %240 to float, !dbg !37
267
+ %245 = bitcast i32 %241 to float, !dbg !37
268
+ %246 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %234, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
269
+ %247 = extractvalue { i32, i32, i32, i32 } %246, 0, !dbg !37
270
+ %248 = extractvalue { i32, i32, i32, i32 } %246, 1, !dbg !37
271
+ %249 = extractvalue { i32, i32, i32, i32 } %246, 2, !dbg !37
272
+ %250 = extractvalue { i32, i32, i32, i32 } %246, 3, !dbg !37
273
+ %251 = bitcast i32 %247 to float, !dbg !37
274
+ %252 = bitcast i32 %248 to float, !dbg !37
275
+ %253 = bitcast i32 %249 to float, !dbg !37
276
+ %254 = bitcast i32 %250 to float, !dbg !37
277
+ %255 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %235, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
278
+ %256 = extractvalue { i32, i32, i32, i32 } %255, 0, !dbg !37
279
+ %257 = extractvalue { i32, i32, i32, i32 } %255, 1, !dbg !37
280
+ %258 = extractvalue { i32, i32, i32, i32 } %255, 2, !dbg !37
281
+ %259 = extractvalue { i32, i32, i32, i32 } %255, 3, !dbg !37
282
+ %260 = bitcast i32 %256 to float, !dbg !37
283
+ %261 = bitcast i32 %257 to float, !dbg !37
284
+ %262 = bitcast i32 %258 to float, !dbg !37
285
+ %263 = bitcast i32 %259 to float, !dbg !37
286
+ %264 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %236, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !37
287
+ %265 = extractvalue { i32, i32, i32, i32 } %264, 0, !dbg !37
288
+ %266 = extractvalue { i32, i32, i32, i32 } %264, 1, !dbg !37
289
+ %267 = extractvalue { i32, i32, i32, i32 } %264, 2, !dbg !37
290
+ %268 = extractvalue { i32, i32, i32, i32 } %264, 3, !dbg !37
291
+ %269 = bitcast i32 %265 to float, !dbg !37
292
+ %270 = bitcast i32 %266 to float, !dbg !37
293
+ %271 = bitcast i32 %267 to float, !dbg !37
294
+ %272 = bitcast i32 %268 to float, !dbg !37
295
+ %273 = fadd float %150, %242, !dbg !38
296
+ %274 = fadd float %151, %243, !dbg !38
297
+ %275 = fadd float %152, %244, !dbg !38
298
+ %276 = fadd float %153, %245, !dbg !38
299
+ %277 = fadd float %159, %251, !dbg !38
300
+ %278 = fadd float %160, %252, !dbg !38
301
+ %279 = fadd float %161, %253, !dbg !38
302
+ %280 = fadd float %162, %254, !dbg !38
303
+ %281 = fadd float %168, %260, !dbg !38
304
+ %282 = fadd float %169, %261, !dbg !38
305
+ %283 = fadd float %170, %262, !dbg !38
306
+ %284 = fadd float %171, %263, !dbg !38
307
+ %285 = fadd float %177, %269, !dbg !38
308
+ %286 = fadd float %178, %270, !dbg !38
309
+ %287 = fadd float %179, %271, !dbg !38
310
+ %288 = fadd float %180, %272, !dbg !38
311
+ %289 = fadd float %213, %273, !dbg !39
312
+ %290 = fadd float %214, %274, !dbg !39
313
+ %291 = fadd float %215, %275, !dbg !39
314
+ %292 = fadd float %216, %276, !dbg !39
315
+ %293 = fadd float %217, %277, !dbg !39
316
+ %294 = fadd float %218, %278, !dbg !39
317
+ %295 = fadd float %219, %279, !dbg !39
318
+ %296 = fadd float %220, %280, !dbg !39
319
+ %297 = fadd float %221, %281, !dbg !39
320
+ %298 = fadd float %222, %282, !dbg !39
321
+ %299 = fadd float %223, %283, !dbg !39
322
+ %300 = fadd float %224, %284, !dbg !39
323
+ %301 = fadd float %225, %285, !dbg !39
324
+ %302 = fadd float %226, %286, !dbg !39
325
+ %303 = fadd float %227, %287, !dbg !39
326
+ %304 = fadd float %228, %288, !dbg !39
327
+ %305 = fsub float %289, %114, !dbg !40
328
+ %306 = fsub float %290, %115, !dbg !40
329
+ %307 = fsub float %291, %116, !dbg !40
330
+ %308 = fsub float %292, %117, !dbg !40
331
+ %309 = fsub float %293, %118, !dbg !40
332
+ %310 = fsub float %294, %119, !dbg !40
333
+ %311 = fsub float %295, %120, !dbg !40
334
+ %312 = fsub float %296, %121, !dbg !40
335
+ %313 = fsub float %297, %122, !dbg !40
336
+ %314 = fsub float %298, %123, !dbg !40
337
+ %315 = fsub float %299, %124, !dbg !40
338
+ %316 = fsub float %300, %125, !dbg !40
339
+ %317 = fsub float %301, %126, !dbg !40
340
+ %318 = fsub float %302, %127, !dbg !40
341
+ %319 = fsub float %303, %128, !dbg !40
342
+ %320 = fsub float %304, %129, !dbg !40
343
+ %321 = fadd float %66, 1.000000e+00, !dbg !44
344
+ %322 = fadd float %67, 1.000000e+00, !dbg !44
345
+ %323 = fadd float %68, 1.000000e+00, !dbg !44
346
+ %324 = fadd float %69, 1.000000e+00, !dbg !44
347
+ %325 = fadd float %70, 1.000000e+00, !dbg !44
348
+ %326 = fadd float %71, 1.000000e+00, !dbg !44
349
+ %327 = fadd float %72, 1.000000e+00, !dbg !44
350
+ %328 = fadd float %73, 1.000000e+00, !dbg !44
351
+ %329 = fadd float %74, 1.000000e+00, !dbg !44
352
+ %330 = fadd float %75, 1.000000e+00, !dbg !44
353
+ %331 = fadd float %76, 1.000000e+00, !dbg !44
354
+ %332 = fadd float %77, 1.000000e+00, !dbg !44
355
+ %333 = fadd float %78, 1.000000e+00, !dbg !44
356
+ %334 = fadd float %79, 1.000000e+00, !dbg !44
357
+ %335 = fadd float %80, 1.000000e+00, !dbg !44
358
+ %336 = fadd float %81, 1.000000e+00, !dbg !44
359
+ %337 = fadd float %82, 1.000000e+00, !dbg !44
360
+ %338 = fadd float %83, 1.000000e+00, !dbg !44
361
+ %339 = fadd float %84, 1.000000e+00, !dbg !44
362
+ %340 = fadd float %85, 1.000000e+00, !dbg !44
363
+ %341 = fadd float %86, 1.000000e+00, !dbg !44
364
+ %342 = fadd float %87, 1.000000e+00, !dbg !44
365
+ %343 = fadd float %88, 1.000000e+00, !dbg !44
366
+ %344 = fadd float %89, 1.000000e+00, !dbg !44
367
+ %345 = fadd float %90, 1.000000e+00, !dbg !44
368
+ %346 = fadd float %91, 1.000000e+00, !dbg !44
369
+ %347 = fadd float %92, 1.000000e+00, !dbg !44
370
+ %348 = fadd float %93, 1.000000e+00, !dbg !44
371
+ %349 = fadd float %94, 1.000000e+00, !dbg !44
372
+ %350 = fadd float %95, 1.000000e+00, !dbg !44
373
+ %351 = fadd float %96, 1.000000e+00, !dbg !44
374
+ %352 = fadd float %97, 1.000000e+00, !dbg !44
375
+ %353 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %305, float %321) #6, !dbg !45
376
+ %354 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %306, float %322) #6, !dbg !45
377
+ %355 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %307, float %323) #6, !dbg !45
378
+ %356 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %308, float %324) #6, !dbg !45
379
+ %357 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %309, float %325) #6, !dbg !45
380
+ %358 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %310, float %326) #6, !dbg !45
381
+ %359 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %311, float %327) #6, !dbg !45
382
+ %360 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %312, float %328) #6, !dbg !45
383
+ %361 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %313, float %329) #6, !dbg !45
384
+ %362 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %314, float %330) #6, !dbg !45
385
+ %363 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %315, float %331) #6, !dbg !45
386
+ %364 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %316, float %332) #6, !dbg !45
387
+ %365 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %317, float %333) #6, !dbg !45
388
+ %366 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %318, float %334) #6, !dbg !45
389
+ %367 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %319, float %335) #6, !dbg !45
390
+ %368 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %320, float %336) #6, !dbg !45
391
+ %369 = fadd float %114, %353, !dbg !46
392
+ %370 = fadd float %115, %354, !dbg !46
393
+ %371 = fadd float %116, %355, !dbg !46
394
+ %372 = fadd float %117, %356, !dbg !46
395
+ %373 = fadd float %118, %357, !dbg !46
396
+ %374 = fadd float %119, %358, !dbg !46
397
+ %375 = fadd float %120, %359, !dbg !46
398
+ %376 = fadd float %121, %360, !dbg !46
399
+ %377 = fadd float %122, %361, !dbg !46
400
+ %378 = fadd float %123, %362, !dbg !46
401
+ %379 = fadd float %124, %363, !dbg !46
402
+ %380 = fadd float %125, %364, !dbg !46
403
+ %381 = fadd float %126, %365, !dbg !46
404
+ %382 = fadd float %127, %366, !dbg !46
405
+ %383 = fadd float %128, %367, !dbg !46
406
+ %384 = fadd float %129, %368, !dbg !46
407
+ %385 = fsub float %289, %369, !dbg !47
408
+ %386 = fsub float %290, %370, !dbg !47
409
+ %387 = fsub float %291, %371, !dbg !47
410
+ %388 = fsub float %292, %372, !dbg !47
411
+ %389 = fsub float %293, %373, !dbg !47
412
+ %390 = fsub float %294, %374, !dbg !47
413
+ %391 = fsub float %295, %375, !dbg !47
414
+ %392 = fsub float %296, %376, !dbg !47
415
+ %393 = fsub float %297, %377, !dbg !47
416
+ %394 = fsub float %298, %378, !dbg !47
417
+ %395 = fsub float %299, %379, !dbg !47
418
+ %396 = fsub float %300, %380, !dbg !47
419
+ %397 = fsub float %301, %381, !dbg !47
420
+ %398 = fsub float %302, %382, !dbg !47
421
+ %399 = fsub float %303, %383, !dbg !47
422
+ %400 = fsub float %304, %384, !dbg !47
423
+ %401 = fmul float %305, %385, !dbg !48
424
+ %402 = fmul float %306, %386, !dbg !48
425
+ %403 = fmul float %307, %387, !dbg !48
426
+ %404 = fmul float %308, %388, !dbg !48
427
+ %405 = fmul float %309, %389, !dbg !48
428
+ %406 = fmul float %310, %390, !dbg !48
429
+ %407 = fmul float %311, %391, !dbg !48
430
+ %408 = fmul float %312, %392, !dbg !48
431
+ %409 = fmul float %313, %393, !dbg !48
432
+ %410 = fmul float %314, %394, !dbg !48
433
+ %411 = fmul float %315, %395, !dbg !48
434
+ %412 = fmul float %316, %396, !dbg !48
435
+ %413 = fmul float %317, %397, !dbg !48
436
+ %414 = fmul float %318, %398, !dbg !48
437
+ %415 = fmul float %319, %399, !dbg !48
438
+ %416 = fmul float %320, %400, !dbg !48
439
+ %417 = fadd float %98, %401, !dbg !49
440
+ %418 = fadd float %99, %402, !dbg !49
441
+ %419 = fadd float %100, %403, !dbg !49
442
+ %420 = fadd float %101, %404, !dbg !49
443
+ %421 = fadd float %102, %405, !dbg !49
444
+ %422 = fadd float %103, %406, !dbg !49
445
+ %423 = fadd float %104, %407, !dbg !49
446
+ %424 = fadd float %105, %408, !dbg !49
447
+ %425 = fadd float %106, %409, !dbg !49
448
+ %426 = fadd float %107, %410, !dbg !49
449
+ %427 = fadd float %108, %411, !dbg !49
450
+ %428 = fadd float %109, %412, !dbg !49
451
+ %429 = fadd float %110, %413, !dbg !49
452
+ %430 = fadd float %111, %414, !dbg !49
453
+ %431 = fadd float %112, %415, !dbg !49
454
+ %432 = fadd float %113, %416, !dbg !49
455
+ %433 = add nuw nsw i32 %130, 64, !dbg !12
456
+ %434 = icmp ult i32 %130, 192, !dbg !12
457
+ br i1 %434, label %65, label %435, !dbg !12
458
+
459
+ 435: ; preds = %230
460
+ %436 = and i32 %16, 3, !dbg !12
461
+ %437 = mul nuw nsw i32 %436, 72, !dbg !12
462
+ %438 = add nuw nsw i32 %437, %12, !dbg !12
463
+ %439 = zext nneg i32 %438 to i64, !dbg !12
464
+ %440 = getelementptr float, ptr addrspace(3) @global_smem, i64 %439, !dbg !12
465
+ %441 = insertelement <1 x float> undef, float %337, i64 0, !dbg !12
466
+ store <1 x float> %441, ptr addrspace(3) %440, align 4, !dbg !12
467
+ %442 = add nuw nsw i32 %12, 288, !dbg !12
468
+ %443 = add nuw nsw i32 %442, %437, !dbg !12
469
+ %444 = zext nneg i32 %443 to i64, !dbg !12
470
+ %445 = getelementptr float, ptr addrspace(3) @global_smem, i64 %444, !dbg !12
471
+ %446 = insertelement <1 x float> undef, float %338, i64 0, !dbg !12
472
+ store <1 x float> %446, ptr addrspace(3) %445, align 4, !dbg !12
473
+ %447 = or i32 %12, 576, !dbg !12
474
+ %448 = add nuw nsw i32 %447, %437, !dbg !12
475
+ %449 = zext nneg i32 %448 to i64, !dbg !12
476
+ %450 = getelementptr float, ptr addrspace(3) @global_smem, i64 %449, !dbg !12
477
+ %451 = insertelement <1 x float> undef, float %339, i64 0, !dbg !12
478
+ store <1 x float> %451, ptr addrspace(3) %450, align 4, !dbg !12
479
+ %452 = add nuw nsw i32 %12, 864, !dbg !12
480
+ %453 = add nuw nsw i32 %452, %437, !dbg !12
481
+ %454 = zext nneg i32 %453 to i64, !dbg !12
482
+ %455 = getelementptr float, ptr addrspace(3) @global_smem, i64 %454, !dbg !12
483
+ %456 = insertelement <1 x float> undef, float %340, i64 0, !dbg !12
484
+ store <1 x float> %456, ptr addrspace(3) %455, align 4, !dbg !12
485
+ %457 = or i32 %12, 1152, !dbg !12
486
+ %458 = add nuw nsw i32 %457, %437, !dbg !12
487
+ %459 = zext nneg i32 %458 to i64, !dbg !12
488
+ %460 = getelementptr float, ptr addrspace(3) @global_smem, i64 %459, !dbg !12
489
+ %461 = insertelement <1 x float> undef, float %341, i64 0, !dbg !12
490
+ store <1 x float> %461, ptr addrspace(3) %460, align 4, !dbg !12
491
+ %462 = add nuw nsw i32 %12, 1440, !dbg !12
492
+ %463 = add nuw nsw i32 %462, %437, !dbg !12
493
+ %464 = zext nneg i32 %463 to i64, !dbg !12
494
+ %465 = getelementptr float, ptr addrspace(3) @global_smem, i64 %464, !dbg !12
495
+ %466 = insertelement <1 x float> undef, float %342, i64 0, !dbg !12
496
+ store <1 x float> %466, ptr addrspace(3) %465, align 4, !dbg !12
497
+ %467 = or i32 %12, 1728, !dbg !12
498
+ %468 = add nuw nsw i32 %467, %437, !dbg !12
499
+ %469 = zext nneg i32 %468 to i64, !dbg !12
500
+ %470 = getelementptr float, ptr addrspace(3) @global_smem, i64 %469, !dbg !12
501
+ %471 = insertelement <1 x float> undef, float %343, i64 0, !dbg !12
502
+ store <1 x float> %471, ptr addrspace(3) %470, align 4, !dbg !12
503
+ %472 = add nuw nsw i32 %12, 2016, !dbg !12
504
+ %473 = add nuw nsw i32 %472, %437, !dbg !12
505
+ %474 = zext nneg i32 %473 to i64, !dbg !12
506
+ %475 = getelementptr float, ptr addrspace(3) @global_smem, i64 %474, !dbg !12
507
+ %476 = insertelement <1 x float> undef, float %344, i64 0, !dbg !12
508
+ store <1 x float> %476, ptr addrspace(3) %475, align 4, !dbg !12
509
+ tail call void @llvm.nvvm.barrier0(), !dbg !12
510
+ %477 = mul nuw nsw i32 %11, 72, !dbg !12
511
+ %478 = add nuw nsw i32 %477, %14, !dbg !12
512
+ %479 = zext nneg i32 %478 to i64, !dbg !12
513
+ %480 = getelementptr float, ptr addrspace(3) @global_smem, i64 %479, !dbg !12
514
+ %481 = load float, ptr addrspace(3) %480, align 32, !dbg !12
515
+ %482 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 1, !dbg !12
516
+ %483 = load float, ptr addrspace(3) %482, align 4, !dbg !12
517
+ %484 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 2, !dbg !12
518
+ %485 = load float, ptr addrspace(3) %484, align 8, !dbg !12
519
+ %486 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 3, !dbg !12
520
+ %487 = load float, ptr addrspace(3) %486, align 4, !dbg !12
521
+ %488 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 4, !dbg !12
522
+ %489 = load float, ptr addrspace(3) %488, align 16, !dbg !12
523
+ %490 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 5, !dbg !12
524
+ %491 = load float, ptr addrspace(3) %490, align 4, !dbg !12
525
+ %492 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 6, !dbg !12
526
+ %493 = load float, ptr addrspace(3) %492, align 8, !dbg !12
527
+ %494 = getelementptr inbounds <8 x float>, ptr addrspace(3) %480, i64 0, i64 7, !dbg !12
528
+ %495 = load float, ptr addrspace(3) %494, align 4, !dbg !12
529
+ tail call void @llvm.nvvm.barrier0(), !dbg !12
530
+ %496 = insertelement <1 x float> undef, float %345, i64 0, !dbg !12
531
+ store <1 x float> %496, ptr addrspace(3) %440, align 4, !dbg !12
532
+ %497 = insertelement <1 x float> undef, float %346, i64 0, !dbg !12
533
+ store <1 x float> %497, ptr addrspace(3) %445, align 4, !dbg !12
534
+ %498 = insertelement <1 x float> undef, float %347, i64 0, !dbg !12
535
+ store <1 x float> %498, ptr addrspace(3) %450, align 4, !dbg !12
536
+ %499 = insertelement <1 x float> undef, float %348, i64 0, !dbg !12
537
+ store <1 x float> %499, ptr addrspace(3) %455, align 4, !dbg !12
538
+ %500 = insertelement <1 x float> undef, float %349, i64 0, !dbg !12
539
+ store <1 x float> %500, ptr addrspace(3) %460, align 4, !dbg !12
540
+ %501 = insertelement <1 x float> undef, float %350, i64 0, !dbg !12
541
+ store <1 x float> %501, ptr addrspace(3) %465, align 4, !dbg !12
542
+ %502 = insertelement <1 x float> undef, float %351, i64 0, !dbg !12
543
+ store <1 x float> %502, ptr addrspace(3) %470, align 4, !dbg !12
544
+ %503 = insertelement <1 x float> undef, float %352, i64 0, !dbg !12
545
+ store <1 x float> %503, ptr addrspace(3) %475, align 4, !dbg !12
546
+ tail call void @llvm.nvvm.barrier0(), !dbg !12
547
+ %504 = load float, ptr addrspace(3) %480, align 32, !dbg !12
548
+ %505 = load float, ptr addrspace(3) %482, align 4, !dbg !12
549
+ %506 = load float, ptr addrspace(3) %484, align 8, !dbg !12
550
+ %507 = load float, ptr addrspace(3) %486, align 4, !dbg !12
551
+ %508 = load float, ptr addrspace(3) %488, align 16, !dbg !12
552
+ %509 = load float, ptr addrspace(3) %490, align 4, !dbg !12
553
+ %510 = load float, ptr addrspace(3) %492, align 8, !dbg !12
554
+ %511 = load float, ptr addrspace(3) %494, align 4, !dbg !12
555
+ %512 = fsub float %370, %369, !dbg !50
556
+ %513 = fadd float %481, %483, !dbg !54
557
+ %514 = fcmp oeq float %513, 0.000000e+00, !dbg !55
558
+ %515 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %483, float %513) #6, !dbg !56
559
+ %516 = select i1 %514, float 0.000000e+00, float %515, !dbg !57
560
+ %517 = fmul float %512, %516, !dbg !58
561
+ %518 = fadd float %369, %517, !dbg !59
562
+ %519 = fadd float %417, %418, !dbg !60
563
+ %520 = fmul float %512, %512, !dbg !61
564
+ %521 = fmul float %520, %481, !dbg !62
565
+ %522 = fmul float %521, %516, !dbg !63
566
+ %523 = fadd float %519, %522, !dbg !64
567
+ %524 = fsub float %371, %518, !dbg !50
568
+ %525 = fadd float %485, %513, !dbg !54
569
+ %526 = fcmp oeq float %525, 0.000000e+00, !dbg !55
570
+ %527 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %485, float %525) #6, !dbg !56
571
+ %528 = select i1 %526, float 0.000000e+00, float %527, !dbg !57
572
+ %529 = fmul float %528, %524, !dbg !58
573
+ %530 = fadd float %518, %529, !dbg !59
574
+ %531 = fadd float %419, %523, !dbg !60
575
+ %532 = fmul float %524, %524, !dbg !61
576
+ %533 = fmul float %513, %532, !dbg !62
577
+ %534 = fmul float %528, %533, !dbg !63
578
+ %535 = fadd float %531, %534, !dbg !64
579
+ %536 = fsub float %372, %530, !dbg !50
580
+ %537 = fadd float %487, %525, !dbg !54
581
+ %538 = fcmp oeq float %537, 0.000000e+00, !dbg !55
582
+ %539 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %487, float %537) #6, !dbg !56
583
+ %540 = select i1 %538, float 0.000000e+00, float %539, !dbg !57
584
+ %541 = fmul float %540, %536, !dbg !58
585
+ %542 = fadd float %530, %541, !dbg !59
586
+ %543 = fadd float %420, %535, !dbg !60
587
+ %544 = fmul float %536, %536, !dbg !61
588
+ %545 = fmul float %525, %544, !dbg !62
589
+ %546 = fmul float %540, %545, !dbg !63
590
+ %547 = fadd float %543, %546, !dbg !64
591
+ %548 = fsub float %373, %542, !dbg !50
592
+ %549 = fadd float %489, %537, !dbg !54
593
+ %550 = fcmp oeq float %549, 0.000000e+00, !dbg !55
594
+ %551 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %489, float %549) #6, !dbg !56
595
+ %552 = select i1 %550, float 0.000000e+00, float %551, !dbg !57
596
+ %553 = fmul float %552, %548, !dbg !58
597
+ %554 = fadd float %542, %553, !dbg !59
598
+ %555 = fadd float %421, %547, !dbg !60
599
+ %556 = fmul float %548, %548, !dbg !61
600
+ %557 = fmul float %537, %556, !dbg !62
601
+ %558 = fmul float %552, %557, !dbg !63
602
+ %559 = fadd float %555, %558, !dbg !64
603
+ %560 = fsub float %374, %554, !dbg !50
604
+ %561 = fadd float %491, %549, !dbg !54
605
+ %562 = fcmp oeq float %561, 0.000000e+00, !dbg !55
606
+ %563 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %491, float %561) #6, !dbg !56
607
+ %564 = select i1 %562, float 0.000000e+00, float %563, !dbg !57
608
+ %565 = fmul float %564, %560, !dbg !58
609
+ %566 = fadd float %554, %565, !dbg !59
610
+ %567 = fadd float %422, %559, !dbg !60
611
+ %568 = fmul float %560, %560, !dbg !61
612
+ %569 = fmul float %549, %568, !dbg !62
613
+ %570 = fmul float %564, %569, !dbg !63
614
+ %571 = fadd float %567, %570, !dbg !64
615
+ %572 = fsub float %375, %566, !dbg !50
616
+ %573 = fadd float %493, %561, !dbg !54
617
+ %574 = fcmp oeq float %573, 0.000000e+00, !dbg !55
618
+ %575 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %493, float %573) #6, !dbg !56
619
+ %576 = select i1 %574, float 0.000000e+00, float %575, !dbg !57
620
+ %577 = fmul float %576, %572, !dbg !58
621
+ %578 = fadd float %566, %577, !dbg !59
622
+ %579 = fadd float %423, %571, !dbg !60
623
+ %580 = fmul float %572, %572, !dbg !61
624
+ %581 = fmul float %561, %580, !dbg !62
625
+ %582 = fmul float %576, %581, !dbg !63
626
+ %583 = fadd float %579, %582, !dbg !64
627
+ %584 = fsub float %376, %578, !dbg !50
628
+ %585 = fadd float %495, %573, !dbg !54
629
+ %586 = fcmp oeq float %585, 0.000000e+00, !dbg !55
630
+ %587 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %495, float %585) #6, !dbg !56
631
+ %588 = select i1 %586, float 0.000000e+00, float %587, !dbg !57
632
+ %589 = fmul float %588, %584, !dbg !58
633
+ %590 = fadd float %578, %589, !dbg !59
634
+ %591 = fadd float %424, %583, !dbg !60
635
+ %592 = fmul float %584, %584, !dbg !61
636
+ %593 = fmul float %573, %592, !dbg !62
637
+ %594 = fmul float %588, %593, !dbg !63
638
+ %595 = fadd float %591, %594, !dbg !64
639
+ %596 = fsub float %378, %377, !dbg !50
640
+ %597 = fadd float %504, %505, !dbg !54
641
+ %598 = fcmp oeq float %597, 0.000000e+00, !dbg !55
642
+ %599 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %505, float %597) #6, !dbg !56
643
+ %600 = select i1 %598, float 0.000000e+00, float %599, !dbg !57
644
+ %601 = fmul float %596, %600, !dbg !58
645
+ %602 = fadd float %377, %601, !dbg !59
646
+ %603 = fadd float %425, %426, !dbg !60
647
+ %604 = fmul float %596, %596, !dbg !61
648
+ %605 = fmul float %604, %504, !dbg !62
649
+ %606 = fmul float %605, %600, !dbg !63
650
+ %607 = fadd float %603, %606, !dbg !64
651
+ %608 = fsub float %379, %602, !dbg !50
652
+ %609 = fadd float %506, %597, !dbg !54
653
+ %610 = fcmp oeq float %609, 0.000000e+00, !dbg !55
654
+ %611 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %506, float %609) #6, !dbg !56
655
+ %612 = select i1 %610, float 0.000000e+00, float %611, !dbg !57
656
+ %613 = fmul float %612, %608, !dbg !58
657
+ %614 = fadd float %602, %613, !dbg !59
658
+ %615 = fadd float %427, %607, !dbg !60
659
+ %616 = fmul float %608, %608, !dbg !61
660
+ %617 = fmul float %597, %616, !dbg !62
661
+ %618 = fmul float %612, %617, !dbg !63
662
+ %619 = fadd float %615, %618, !dbg !64
663
+ %620 = fsub float %380, %614, !dbg !50
664
+ %621 = fadd float %507, %609, !dbg !54
665
+ %622 = fcmp oeq float %621, 0.000000e+00, !dbg !55
666
+ %623 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %507, float %621) #6, !dbg !56
667
+ %624 = select i1 %622, float 0.000000e+00, float %623, !dbg !57
668
+ %625 = fmul float %624, %620, !dbg !58
669
+ %626 = fadd float %614, %625, !dbg !59
670
+ %627 = fadd float %428, %619, !dbg !60
671
+ %628 = fmul float %620, %620, !dbg !61
672
+ %629 = fmul float %609, %628, !dbg !62
673
+ %630 = fmul float %624, %629, !dbg !63
674
+ %631 = fadd float %627, %630, !dbg !64
675
+ %632 = fsub float %381, %626, !dbg !50
676
+ %633 = fadd float %508, %621, !dbg !54
677
+ %634 = fcmp oeq float %633, 0.000000e+00, !dbg !55
678
+ %635 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %508, float %633) #6, !dbg !56
679
+ %636 = select i1 %634, float 0.000000e+00, float %635, !dbg !57
680
+ %637 = fmul float %636, %632, !dbg !58
681
+ %638 = fadd float %626, %637, !dbg !59
682
+ %639 = fadd float %429, %631, !dbg !60
683
+ %640 = fmul float %632, %632, !dbg !61
684
+ %641 = fmul float %621, %640, !dbg !62
685
+ %642 = fmul float %636, %641, !dbg !63
686
+ %643 = fadd float %639, %642, !dbg !64
687
+ %644 = fsub float %382, %638, !dbg !50
688
+ %645 = fadd float %509, %633, !dbg !54
689
+ %646 = fcmp oeq float %645, 0.000000e+00, !dbg !55
690
+ %647 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %509, float %645) #6, !dbg !56
691
+ %648 = select i1 %646, float 0.000000e+00, float %647, !dbg !57
692
+ %649 = fmul float %648, %644, !dbg !58
693
+ %650 = fadd float %638, %649, !dbg !59
694
+ %651 = fadd float %430, %643, !dbg !60
695
+ %652 = fmul float %644, %644, !dbg !61
696
+ %653 = fmul float %633, %652, !dbg !62
697
+ %654 = fmul float %648, %653, !dbg !63
698
+ %655 = fadd float %651, %654, !dbg !64
699
+ %656 = fsub float %383, %650, !dbg !50
700
+ %657 = fadd float %510, %645, !dbg !54
701
+ %658 = fcmp oeq float %657, 0.000000e+00, !dbg !55
702
+ %659 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %510, float %657) #6, !dbg !56
703
+ %660 = select i1 %658, float 0.000000e+00, float %659, !dbg !57
704
+ %661 = fmul float %660, %656, !dbg !58
705
+ %662 = fadd float %650, %661, !dbg !59
706
+ %663 = fadd float %431, %655, !dbg !60
707
+ %664 = fmul float %656, %656, !dbg !61
708
+ %665 = fmul float %645, %664, !dbg !62
709
+ %666 = fmul float %660, %665, !dbg !63
710
+ %667 = fadd float %663, %666, !dbg !64
711
+ %668 = fsub float %384, %662, !dbg !50
712
+ %669 = fadd float %511, %657, !dbg !54
713
+ %670 = fcmp oeq float %669, 0.000000e+00, !dbg !55
714
+ %671 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %511, float %669) #6, !dbg !56
715
+ %672 = select i1 %670, float 0.000000e+00, float %671, !dbg !57
716
+ %673 = fmul float %672, %668, !dbg !58
717
+ %674 = fadd float %662, %673, !dbg !59
718
+ %675 = fadd float %432, %667, !dbg !60
719
+ %676 = fmul float %668, %668, !dbg !61
720
+ %677 = fmul float %657, %676, !dbg !62
721
+ %678 = fmul float %672, %677, !dbg !63
722
+ %679 = fadd float %675, %678, !dbg !64
723
+ %680 = bitcast float %590 to i32, !dbg !65
724
+ %681 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %680, i32 4, i32 31), !dbg !65
725
+ %682 = bitcast i32 %681 to float, !dbg !65
726
+ %683 = bitcast float %595 to i32, !dbg !65
727
+ %684 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %683, i32 4, i32 31), !dbg !65
728
+ %685 = bitcast i32 %684 to float, !dbg !65
729
+ %686 = bitcast float %585 to i32, !dbg !65
730
+ %687 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %686, i32 4, i32 31), !dbg !65
731
+ %688 = bitcast i32 %687 to float, !dbg !65
732
+ %689 = fsub float %682, %590, !dbg !50
733
+ %690 = fadd float %585, %688, !dbg !54
734
+ %691 = fcmp oeq float %690, 0.000000e+00, !dbg !55
735
+ %692 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %688, float %690) #6, !dbg !56
736
+ %693 = select i1 %691, float 0.000000e+00, float %692, !dbg !57
737
+ %694 = fmul float %693, %689, !dbg !58
738
+ %695 = fadd float %590, %694, !dbg !59
739
+ %696 = fadd float %595, %685, !dbg !60
740
+ %697 = fmul float %689, %689, !dbg !61
741
+ %698 = fmul float %585, %697, !dbg !62
742
+ %699 = fmul float %693, %698, !dbg !63
743
+ %700 = fadd float %696, %699, !dbg !64
744
+ %701 = bitcast float %695 to i32, !dbg !65
745
+ %702 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %701, i32 2, i32 31), !dbg !65
746
+ %703 = bitcast i32 %702 to float, !dbg !65
747
+ %704 = bitcast float %700 to i32, !dbg !65
748
+ %705 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %704, i32 2, i32 31), !dbg !65
749
+ %706 = bitcast i32 %705 to float, !dbg !65
750
+ %707 = bitcast float %690 to i32, !dbg !65
751
+ %708 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %707, i32 2, i32 31), !dbg !65
752
+ %709 = bitcast i32 %708 to float, !dbg !65
753
+ %710 = fsub float %703, %695, !dbg !50
754
+ %711 = fadd float %690, %709, !dbg !54
755
+ %712 = fcmp oeq float %711, 0.000000e+00, !dbg !55
756
+ %713 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %709, float %711) #6, !dbg !56
757
+ %714 = select i1 %712, float 0.000000e+00, float %713, !dbg !57
758
+ %715 = fmul float %714, %710, !dbg !58
759
+ %716 = fadd float %695, %715, !dbg !59
760
+ %717 = fadd float %700, %706, !dbg !60
761
+ %718 = fmul float %710, %710, !dbg !61
762
+ %719 = fmul float %690, %718, !dbg !62
763
+ %720 = fmul float %714, %719, !dbg !63
764
+ %721 = fadd float %717, %720, !dbg !64
765
+ %722 = bitcast float %716 to i32, !dbg !65
766
+ %723 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %722, i32 1, i32 31), !dbg !65
767
+ %724 = bitcast i32 %723 to float, !dbg !65
768
+ %725 = bitcast float %721 to i32, !dbg !65
769
+ %726 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %725, i32 1, i32 31), !dbg !65
770
+ %727 = bitcast i32 %726 to float, !dbg !65
771
+ %728 = bitcast float %711 to i32, !dbg !65
772
+ %729 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %728, i32 1, i32 31), !dbg !65
773
+ %730 = bitcast i32 %729 to float, !dbg !65
774
+ %731 = fsub float %724, %716, !dbg !50
775
+ %732 = fadd float %711, %730, !dbg !54
776
+ %733 = fcmp oeq float %732, 0.000000e+00, !dbg !55
777
+ %734 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %730, float %732) #6, !dbg !56
778
+ %735 = select i1 %733, float 0.000000e+00, float %734, !dbg !57
779
+ %736 = fmul float %731, %735, !dbg !58
780
+ %737 = fadd float %716, %736, !dbg !59
781
+ %738 = fadd float %721, %727, !dbg !60
782
+ %739 = fmul float %731, %731, !dbg !61
783
+ %740 = fmul float %711, %739, !dbg !62
784
+ %741 = fmul float %735, %740, !dbg !63
785
+ %742 = fadd float %738, %741, !dbg !64
786
+ %743 = bitcast float %674 to i32, !dbg !65
787
+ %744 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %743, i32 4, i32 31), !dbg !65
788
+ %745 = bitcast i32 %744 to float, !dbg !65
789
+ %746 = bitcast float %679 to i32, !dbg !65
790
+ %747 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %746, i32 4, i32 31), !dbg !65
791
+ %748 = bitcast i32 %747 to float, !dbg !65
792
+ %749 = bitcast float %669 to i32, !dbg !65
793
+ %750 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %749, i32 4, i32 31), !dbg !65
794
+ %751 = bitcast i32 %750 to float, !dbg !65
795
+ %752 = fsub float %745, %674, !dbg !50
796
+ %753 = fadd float %669, %751, !dbg !54
797
+ %754 = fcmp oeq float %753, 0.000000e+00, !dbg !55
798
+ %755 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %751, float %753) #6, !dbg !56
799
+ %756 = select i1 %754, float 0.000000e+00, float %755, !dbg !57
800
+ %757 = fmul float %752, %756, !dbg !58
801
+ %758 = fadd float %674, %757, !dbg !59
802
+ %759 = fadd float %679, %748, !dbg !60
803
+ %760 = fmul float %752, %752, !dbg !61
804
+ %761 = fmul float %669, %760, !dbg !62
805
+ %762 = fmul float %761, %756, !dbg !63
806
+ %763 = fadd float %759, %762, !dbg !64
807
+ %764 = bitcast float %758 to i32, !dbg !65
808
+ %765 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %764, i32 2, i32 31), !dbg !65
809
+ %766 = bitcast i32 %765 to float, !dbg !65
810
+ %767 = bitcast float %763 to i32, !dbg !65
811
+ %768 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %767, i32 2, i32 31), !dbg !65
812
+ %769 = bitcast i32 %768 to float, !dbg !65
813
+ %770 = bitcast float %753 to i32, !dbg !65
814
+ %771 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %770, i32 2, i32 31), !dbg !65
815
+ %772 = bitcast i32 %771 to float, !dbg !65
816
+ %773 = fsub float %766, %758, !dbg !50
817
+ %774 = fadd float %753, %772, !dbg !54
818
+ %775 = fcmp oeq float %774, 0.000000e+00, !dbg !55
819
+ %776 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %772, float %774) #6, !dbg !56
820
+ %777 = select i1 %775, float 0.000000e+00, float %776, !dbg !57
821
+ %778 = fmul float %773, %777, !dbg !58
822
+ %779 = fadd float %758, %778, !dbg !59
823
+ %780 = fadd float %763, %769, !dbg !60
824
+ %781 = fmul float %773, %773, !dbg !61
825
+ %782 = fmul float %753, %781, !dbg !62
826
+ %783 = fmul float %777, %782, !dbg !63
827
+ %784 = fadd float %780, %783, !dbg !64
828
+ %785 = bitcast float %779 to i32, !dbg !65
829
+ %786 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %785, i32 1, i32 31), !dbg !65
830
+ %787 = bitcast i32 %786 to float, !dbg !65
831
+ %788 = bitcast float %784 to i32, !dbg !65
832
+ %789 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %788, i32 1, i32 31), !dbg !65
833
+ %790 = bitcast i32 %789 to float, !dbg !65
834
+ %791 = bitcast float %774 to i32, !dbg !65
835
+ %792 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %791, i32 1, i32 31), !dbg !65
836
+ %793 = bitcast i32 %792 to float, !dbg !65
837
+ %794 = fsub float %787, %779, !dbg !50
838
+ %795 = fadd float %774, %793, !dbg !54
839
+ %796 = fcmp oeq float %795, 0.000000e+00, !dbg !55
840
+ %797 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %793, float %795) #6, !dbg !56
841
+ %798 = select i1 %796, float 0.000000e+00, float %797, !dbg !57
842
+ %799 = fmul float %794, %798, !dbg !58
843
+ %800 = fadd float %779, %799, !dbg !59
844
+ %801 = fadd float %784, %790, !dbg !60
845
+ %802 = fmul float %794, %794, !dbg !61
846
+ %803 = fmul float %774, %802, !dbg !62
847
+ %804 = fmul float %798, %803, !dbg !63
848
+ %805 = fadd float %801, %804, !dbg !64
849
+ %806 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
850
+ %807 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
851
+ %808 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
852
+ %809 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
853
+ %810 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
854
+ %811 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
855
+ %812 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
856
+ %813 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %742, float 2.560000e+02) #6, !dbg !67
857
+ %814 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
858
+ %815 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
859
+ %816 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
860
+ %817 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
861
+ %818 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
862
+ %819 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
863
+ %820 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
864
+ %821 = tail call float asm "div.full.f32 $0, $1, $2;", "=r,r,r"(float %805, float 2.560000e+02) #6, !dbg !67
865
+ %822 = fadd float %806, 0x3EE4F8B580000000, !dbg !68
866
+ %823 = fadd float %814, 0x3EE4F8B580000000, !dbg !68
867
+ br label %824, !dbg !69
868
+
869
+ 824: ; preds = %435, %__nv_rsqrtf.exit40
870
+ %825 = phi i32 [ 0, %435 ], [ %1134, %__nv_rsqrtf.exit40 ]
871
+ %826 = or i32 %825, %14, !dbg !70
872
+ %827 = or i32 %825, %15, !dbg !70
873
+ %828 = add i32 %826, %47, !dbg !71
874
+ %829 = add i32 %827, %47, !dbg !71
875
+ %830 = add i32 %826, %48, !dbg !71
876
+ %831 = add i32 %827, %48, !dbg !71
877
+ %832 = sext i32 %828 to i64, !dbg !72
878
+ %833 = getelementptr float, ptr addrspace(1) %2, i64 %832, !dbg !72
879
+ %834 = sext i32 %829 to i64, !dbg !72
880
+ %835 = getelementptr float, ptr addrspace(1) %2, i64 %834, !dbg !72
881
+ %836 = sext i32 %830 to i64, !dbg !72
882
+ %837 = getelementptr float, ptr addrspace(1) %2, i64 %836, !dbg !72
883
+ %838 = sext i32 %831 to i64, !dbg !72
884
+ %839 = getelementptr float, ptr addrspace(1) %2, i64 %838, !dbg !72
885
+ %840 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %833, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
886
+ %841 = extractvalue { i32, i32, i32, i32 } %840, 0, !dbg !73
887
+ %842 = extractvalue { i32, i32, i32, i32 } %840, 1, !dbg !73
888
+ %843 = extractvalue { i32, i32, i32, i32 } %840, 2, !dbg !73
889
+ %844 = extractvalue { i32, i32, i32, i32 } %840, 3, !dbg !73
890
+ %845 = bitcast i32 %841 to float, !dbg !73
891
+ %846 = bitcast i32 %842 to float, !dbg !73
892
+ %847 = bitcast i32 %843 to float, !dbg !73
893
+ %848 = bitcast i32 %844 to float, !dbg !73
894
+ %849 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %835, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
895
+ %850 = extractvalue { i32, i32, i32, i32 } %849, 0, !dbg !73
896
+ %851 = extractvalue { i32, i32, i32, i32 } %849, 1, !dbg !73
897
+ %852 = extractvalue { i32, i32, i32, i32 } %849, 2, !dbg !73
898
+ %853 = extractvalue { i32, i32, i32, i32 } %849, 3, !dbg !73
899
+ %854 = bitcast i32 %850 to float, !dbg !73
900
+ %855 = bitcast i32 %851 to float, !dbg !73
901
+ %856 = bitcast i32 %852 to float, !dbg !73
902
+ %857 = bitcast i32 %853 to float, !dbg !73
903
+ %858 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %837, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
904
+ %859 = extractvalue { i32, i32, i32, i32 } %858, 0, !dbg !73
905
+ %860 = extractvalue { i32, i32, i32, i32 } %858, 1, !dbg !73
906
+ %861 = extractvalue { i32, i32, i32, i32 } %858, 2, !dbg !73
907
+ %862 = extractvalue { i32, i32, i32, i32 } %858, 3, !dbg !73
908
+ %863 = bitcast i32 %859 to float, !dbg !73
909
+ %864 = bitcast i32 %860 to float, !dbg !73
910
+ %865 = bitcast i32 %861 to float, !dbg !73
911
+ %866 = bitcast i32 %862 to float, !dbg !73
912
+ %867 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %839, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !73
913
+ %868 = extractvalue { i32, i32, i32, i32 } %867, 0, !dbg !73
914
+ %869 = extractvalue { i32, i32, i32, i32 } %867, 1, !dbg !73
915
+ %870 = extractvalue { i32, i32, i32, i32 } %867, 2, !dbg !73
916
+ %871 = extractvalue { i32, i32, i32, i32 } %867, 3, !dbg !73
917
+ %872 = bitcast i32 %868 to float, !dbg !73
918
+ %873 = bitcast i32 %869 to float, !dbg !73
919
+ %874 = bitcast i32 %870 to float, !dbg !73
920
+ %875 = bitcast i32 %871 to float, !dbg !73
921
+ %876 = add i32 %826, %49, !dbg !74
922
+ %877 = add i32 %826, %50, !dbg !74
923
+ %878 = sext i32 %876 to i64, !dbg !75
924
+ %879 = getelementptr i16, ptr addrspace(1) %3, i64 %878, !dbg !75
925
+ %880 = sext i32 %877 to i64, !dbg !75
926
+ %881 = getelementptr i16, ptr addrspace(1) %3, i64 %880, !dbg !75
927
+ %882 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %879, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !76
928
+ %883 = extractvalue { i32, i32, i32, i32 } %882, 0, !dbg !76
929
+ %884 = extractvalue { i32, i32, i32, i32 } %882, 1, !dbg !76
930
+ %885 = extractvalue { i32, i32, i32, i32 } %882, 2, !dbg !76
931
+ %886 = extractvalue { i32, i32, i32, i32 } %882, 3, !dbg !76
932
+ %887 = trunc i32 %883 to i16, !dbg !76
933
+ %extelt.offset = lshr i32 %883, 16, !dbg !76
934
+ %888 = trunc i32 %extelt.offset to i16, !dbg !76
935
+ %889 = trunc i32 %884 to i16, !dbg !76
936
+ %extelt.offset2 = lshr i32 %884, 16, !dbg !76
937
+ %890 = trunc i32 %extelt.offset2 to i16, !dbg !76
938
+ %891 = trunc i32 %885 to i16, !dbg !76
939
+ %extelt.offset3 = lshr i32 %885, 16, !dbg !76
940
+ %892 = trunc i32 %extelt.offset3 to i16, !dbg !76
941
+ %893 = trunc i32 %886 to i16, !dbg !76
942
+ %extelt.offset4 = lshr i32 %886, 16, !dbg !76
943
+ %894 = trunc i32 %extelt.offset4 to i16, !dbg !76
944
+ %895 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %881, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !76
945
+ %896 = extractvalue { i32, i32, i32, i32 } %895, 0, !dbg !76
946
+ %897 = extractvalue { i32, i32, i32, i32 } %895, 1, !dbg !76
947
+ %898 = extractvalue { i32, i32, i32, i32 } %895, 2, !dbg !76
948
+ %899 = extractvalue { i32, i32, i32, i32 } %895, 3, !dbg !76
949
+ %900 = trunc i32 %896 to i16, !dbg !76
950
+ %extelt.offset5 = lshr i32 %896, 16, !dbg !76
951
+ %901 = trunc i32 %extelt.offset5 to i16, !dbg !76
952
+ %902 = trunc i32 %897 to i16, !dbg !76
953
+ %extelt.offset6 = lshr i32 %897, 16, !dbg !76
954
+ %903 = trunc i32 %extelt.offset6 to i16, !dbg !76
955
+ %904 = trunc i32 %898 to i16, !dbg !76
956
+ %extelt.offset7 = lshr i32 %898, 16, !dbg !76
957
+ %905 = trunc i32 %extelt.offset7 to i16, !dbg !76
958
+ %906 = trunc i32 %899 to i16, !dbg !76
959
+ %extelt.offset8 = lshr i32 %899, 16, !dbg !76
960
+ %907 = trunc i32 %extelt.offset8 to i16, !dbg !76
961
+ %908 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %887) #6, !dbg !77
962
+ %909 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %888) #6, !dbg !77
963
+ %910 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %889) #6, !dbg !77
964
+ %911 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %890) #6, !dbg !77
965
+ %912 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %891) #6, !dbg !77
966
+ %913 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %892) #6, !dbg !77
967
+ %914 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %893) #6, !dbg !77
968
+ %915 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %894) #6, !dbg !77
969
+ %916 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %900) #6, !dbg !77
970
+ %917 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %901) #6, !dbg !77
971
+ %918 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %902) #6, !dbg !77
972
+ %919 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %903) #6, !dbg !77
973
+ %920 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %904) #6, !dbg !77
974
+ %921 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %905) #6, !dbg !77
975
+ %922 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %906) #6, !dbg !77
976
+ %923 = tail call float asm "cvt.f32.bf16 $0, $1;", "=r,h"(i16 %907) #6, !dbg !77
977
+ %924 = zext nneg i32 %826 to i64, !dbg !78
978
+ %925 = getelementptr float, ptr addrspace(1) %4, i64 %924, !dbg !78
979
+ %926 = zext nneg i32 %827 to i64, !dbg !78
980
+ %927 = getelementptr float, ptr addrspace(1) %4, i64 %926, !dbg !78
981
+ %928 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %925, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !79
982
+ %929 = extractvalue { i32, i32, i32, i32 } %928, 0, !dbg !79
983
+ %930 = extractvalue { i32, i32, i32, i32 } %928, 1, !dbg !79
984
+ %931 = extractvalue { i32, i32, i32, i32 } %928, 2, !dbg !79
985
+ %932 = extractvalue { i32, i32, i32, i32 } %928, 3, !dbg !79
986
+ %933 = bitcast i32 %929 to float, !dbg !79
987
+ %934 = bitcast i32 %930 to float, !dbg !79
988
+ %935 = bitcast i32 %931 to float, !dbg !79
989
+ %936 = bitcast i32 %932 to float, !dbg !79
990
+ %937 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_last.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %927, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !79
991
+ %938 = extractvalue { i32, i32, i32, i32 } %937, 0, !dbg !79
992
+ %939 = extractvalue { i32, i32, i32, i32 } %937, 1, !dbg !79
993
+ %940 = extractvalue { i32, i32, i32, i32 } %937, 2, !dbg !79
994
+ %941 = extractvalue { i32, i32, i32, i32 } %937, 3, !dbg !79
995
+ %942 = bitcast i32 %938 to float, !dbg !79
996
+ %943 = bitcast i32 %939 to float, !dbg !79
997
+ %944 = bitcast i32 %940 to float, !dbg !79
998
+ %945 = bitcast i32 %941 to float, !dbg !79
999
+ br i1 %56, label %946, label %947, !dbg !80
1000
+
1001
+ 946: ; preds = %824
1002
+ tail call void @__assertfail(ptr nonnull @assertMessage_1, ptr nonnull @assertFile_1, i32 883, ptr nonnull @assertFunc_1, i64 1), !dbg !80
1003
+ br label %947, !dbg !80
1004
+
1005
+ 947: ; preds = %946, %824
1006
+ %948 = getelementptr float, ptr addrspace(1) %63, i64 %924, !dbg !81
1007
+ %949 = getelementptr float, ptr addrspace(1) %63, i64 %926, !dbg !81
1008
+ %950 = getelementptr float, ptr addrspace(1) %64, i64 %924, !dbg !81
1009
+ %951 = getelementptr float, ptr addrspace(1) %64, i64 %926, !dbg !81
1010
+ %952 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %948, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
1011
+ %953 = extractvalue { i32, i32, i32, i32 } %952, 0, !dbg !82
1012
+ %954 = extractvalue { i32, i32, i32, i32 } %952, 1, !dbg !82
1013
+ %955 = extractvalue { i32, i32, i32, i32 } %952, 2, !dbg !82
1014
+ %956 = extractvalue { i32, i32, i32, i32 } %952, 3, !dbg !82
1015
+ %957 = bitcast i32 %953 to float, !dbg !82
1016
+ %958 = bitcast i32 %954 to float, !dbg !82
1017
+ %959 = bitcast i32 %955 to float, !dbg !82
1018
+ %960 = bitcast i32 %956 to float, !dbg !82
1019
+ %961 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %949, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
1020
+ %962 = extractvalue { i32, i32, i32, i32 } %961, 0, !dbg !82
1021
+ %963 = extractvalue { i32, i32, i32, i32 } %961, 1, !dbg !82
1022
+ %964 = extractvalue { i32, i32, i32, i32 } %961, 2, !dbg !82
1023
+ %965 = extractvalue { i32, i32, i32, i32 } %961, 3, !dbg !82
1024
+ %966 = bitcast i32 %962 to float, !dbg !82
1025
+ %967 = bitcast i32 %963 to float, !dbg !82
1026
+ %968 = bitcast i32 %964 to float, !dbg !82
1027
+ %969 = bitcast i32 %965 to float, !dbg !82
1028
+ %970 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %950, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
1029
+ %971 = extractvalue { i32, i32, i32, i32 } %970, 0, !dbg !82
1030
+ %972 = extractvalue { i32, i32, i32, i32 } %970, 1, !dbg !82
1031
+ %973 = extractvalue { i32, i32, i32, i32 } %970, 2, !dbg !82
1032
+ %974 = extractvalue { i32, i32, i32, i32 } %970, 3, !dbg !82
1033
+ %975 = bitcast i32 %971 to float, !dbg !82
1034
+ %976 = bitcast i32 %972 to float, !dbg !82
1035
+ %977 = bitcast i32 %973 to float, !dbg !82
1036
+ %978 = bitcast i32 %974 to float, !dbg !82
1037
+ %979 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, 0x0;\0A\09mov.u32 $1, 0x0;\0A\09mov.u32 $2, 0x0;\0A\09mov.u32 $3, 0x0;\0A\09@$5 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $4 + 0 ];\0A\09@!$7 mov.u32 $0, $6;\0A\09@!$9 mov.u32 $1, $8;\0A\09@!$11 mov.u32 $2, $10;\0A\09@!$13 mov.u32 $3, $12;", "=r,=r,=r,=r,l,b,r,b,r,b,r,b,r,b"(ptr addrspace(1) %951, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true, i32 0, i1 true) #6, !dbg !82
1038
+ %980 = extractvalue { i32, i32, i32, i32 } %979, 0, !dbg !82
1039
+ %981 = extractvalue { i32, i32, i32, i32 } %979, 1, !dbg !82
1040
+ %982 = extractvalue { i32, i32, i32, i32 } %979, 2, !dbg !82
1041
+ %983 = extractvalue { i32, i32, i32, i32 } %979, 3, !dbg !82
1042
+ %984 = bitcast i32 %980 to float, !dbg !82
1043
+ %985 = bitcast i32 %981 to float, !dbg !82
1044
+ %986 = bitcast i32 %982 to float, !dbg !82
1045
+ %987 = bitcast i32 %983 to float, !dbg !82
1046
+ %988 = fadd float %845, %957, !dbg !83
1047
+ %989 = fadd float %846, %958, !dbg !83
1048
+ %990 = fadd float %847, %959, !dbg !83
1049
+ %991 = fadd float %848, %960, !dbg !83
1050
+ %992 = fadd float %854, %966, !dbg !83
1051
+ %993 = fadd float %855, %967, !dbg !83
1052
+ %994 = fadd float %856, %968, !dbg !83
1053
+ %995 = fadd float %857, %969, !dbg !83
1054
+ %996 = fadd float %863, %975, !dbg !83
1055
+ %997 = fadd float %864, %976, !dbg !83
1056
+ %998 = fadd float %865, %977, !dbg !83
1057
+ %999 = fadd float %866, %978, !dbg !83
1058
+ %1000 = fadd float %872, %984, !dbg !83
1059
+ %1001 = fadd float %873, %985, !dbg !83
1060
+ %1002 = fadd float %874, %986, !dbg !83
1061
+ %1003 = fadd float %875, %987, !dbg !83
1062
+ %1004 = fadd float %908, %988, !dbg !84
1063
+ %1005 = fadd float %909, %989, !dbg !84
1064
+ %1006 = fadd float %910, %990, !dbg !84
1065
+ %1007 = fadd float %911, %991, !dbg !84
1066
+ %1008 = fadd float %912, %992, !dbg !84
1067
+ %1009 = fadd float %913, %993, !dbg !84
1068
+ %1010 = fadd float %914, %994, !dbg !84
1069
+ %1011 = fadd float %915, %995, !dbg !84
1070
+ %1012 = fadd float %916, %996, !dbg !84
1071
+ %1013 = fadd float %917, %997, !dbg !84
1072
+ %1014 = fadd float %918, %998, !dbg !84
1073
+ %1015 = fadd float %919, %999, !dbg !84
1074
+ %1016 = fadd float %920, %1000, !dbg !84
1075
+ %1017 = fadd float %921, %1001, !dbg !84
1076
+ %1018 = fadd float %922, %1002, !dbg !84
1077
+ %1019 = fadd float %923, %1003, !dbg !84
1078
+ %1020 = fsub float %1004, %737, !dbg !85
1079
+ %1021 = fsub float %1005, %737, !dbg !85
1080
+ %1022 = fsub float %1006, %737, !dbg !85
1081
+ %1023 = fsub float %1007, %737, !dbg !85
1082
+ %1024 = fsub float %1008, %737, !dbg !85
1083
+ %1025 = fsub float %1009, %737, !dbg !85
1084
+ %1026 = fsub float %1010, %737, !dbg !85
1085
+ %1027 = fsub float %1011, %737, !dbg !85
1086
+ %1028 = fsub float %1012, %800, !dbg !85
1087
+ %1029 = fsub float %1013, %800, !dbg !85
1088
+ %1030 = fsub float %1014, %800, !dbg !85
1089
+ %1031 = fsub float %1015, %800, !dbg !85
1090
+ %1032 = fsub float %1016, %800, !dbg !85
1091
+ %1033 = fsub float %1017, %800, !dbg !85
1092
+ %1034 = fsub float %1018, %800, !dbg !85
1093
+ %1035 = fsub float %1019, %800, !dbg !85
1094
+ %1036 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1095
+ %.not.i = icmp eq i32 %1036, 0, !dbg !86
1096
+ br i1 %.not.i, label %1039, label %1037, !dbg !86
1097
+
1098
+ 1037: ; preds = %947
1099
+ %1038 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %822), !dbg !86
1100
+ br label %__nv_rsqrtf.exit, !dbg !86
1101
+
1102
+ 1039: ; preds = %947
1103
+ %1040 = tail call float @llvm.nvvm.rsqrt.approx.f(float %822), !dbg !86
1104
+ br label %__nv_rsqrtf.exit, !dbg !86
1105
+
1106
+ __nv_rsqrtf.exit: ; preds = %1037, %1039
1107
+ %.0.i = phi float [ %1038, %1037 ], [ %1040, %1039 ], !dbg !86
1108
+ %1041 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1109
+ %1042 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1110
+ %1043 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1111
+ %1044 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1112
+ %1045 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1113
+ %1046 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1114
+ %1047 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1115
+ %1048 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1116
+ %.not.i38 = icmp eq i32 %1048, 0, !dbg !86
1117
+ br i1 %.not.i38, label %1051, label %1049, !dbg !86
1118
+
1119
+ 1049: ; preds = %__nv_rsqrtf.exit
1120
+ %1050 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %823), !dbg !86
1121
+ br label %__nv_rsqrtf.exit40, !dbg !86
1122
+
1123
+ 1051: ; preds = %__nv_rsqrtf.exit
1124
+ %1052 = tail call float @llvm.nvvm.rsqrt.approx.f(float %823), !dbg !86
1125
+ br label %__nv_rsqrtf.exit40, !dbg !86
1126
+
1127
+ __nv_rsqrtf.exit40: ; preds = %1049, %1051
1128
+ %.0.i39 = phi float [ %1050, %1049 ], [ %1052, %1051 ], !dbg !86
1129
+ %1053 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1130
+ %1054 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1131
+ %1055 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1132
+ %1056 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1133
+ %1057 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1134
+ %1058 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1135
+ %1059 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6, !dbg !86
1136
+ %1060 = fmul float %1020, %.0.i, !dbg !87
1137
+ %1061 = fmul float %1021, %.0.i, !dbg !87
1138
+ %1062 = fmul float %1022, %.0.i, !dbg !87
1139
+ %1063 = fmul float %1023, %.0.i, !dbg !87
1140
+ %1064 = fmul float %1024, %.0.i, !dbg !87
1141
+ %1065 = fmul float %1025, %.0.i, !dbg !87
1142
+ %1066 = fmul float %1026, %.0.i, !dbg !87
1143
+ %1067 = fmul float %1027, %.0.i, !dbg !87
1144
+ %1068 = fmul float %1028, %.0.i39, !dbg !87
1145
+ %1069 = fmul float %1029, %.0.i39, !dbg !87
1146
+ %1070 = fmul float %1030, %.0.i39, !dbg !87
1147
+ %1071 = fmul float %1031, %.0.i39, !dbg !87
1148
+ %1072 = fmul float %1032, %.0.i39, !dbg !87
1149
+ %1073 = fmul float %1033, %.0.i39, !dbg !87
1150
+ %1074 = fmul float %1034, %.0.i39, !dbg !87
1151
+ %1075 = fmul float %1035, %.0.i39, !dbg !87
1152
+ %1076 = fmul float %1060, %933, !dbg !88
1153
+ %1077 = fmul float %1061, %934, !dbg !88
1154
+ %1078 = fmul float %1062, %935, !dbg !88
1155
+ %1079 = fmul float %1063, %936, !dbg !88
1156
+ %1080 = fmul float %1064, %942, !dbg !88
1157
+ %1081 = fmul float %1065, %943, !dbg !88
1158
+ %1082 = fmul float %1066, %944, !dbg !88
1159
+ %1083 = fmul float %1067, %945, !dbg !88
1160
+ %1084 = fmul float %1068, %933, !dbg !88
1161
+ %1085 = fmul float %1069, %934, !dbg !88
1162
+ %1086 = fmul float %1070, %935, !dbg !88
1163
+ %1087 = fmul float %1071, %936, !dbg !88
1164
+ %1088 = fmul float %1072, %942, !dbg !88
1165
+ %1089 = fmul float %1073, %943, !dbg !88
1166
+ %1090 = fmul float %1074, %944, !dbg !88
1167
+ %1091 = fmul float %1075, %945, !dbg !88
1168
+ %1092 = getelementptr i16, ptr addrspace(1) %5, i64 %878, !dbg !89
1169
+ %1093 = getelementptr i16, ptr addrspace(1) %5, i64 %880, !dbg !89
1170
+ %1094 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1076) #6, !dbg !90
1171
+ %1095 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1077) #6, !dbg !90
1172
+ %1096 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1078) #6, !dbg !90
1173
+ %1097 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1079) #6, !dbg !90
1174
+ %1098 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1080) #6, !dbg !90
1175
+ %1099 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1081) #6, !dbg !90
1176
+ %1100 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1082) #6, !dbg !90
1177
+ %1101 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1083) #6, !dbg !90
1178
+ %1102 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1084) #6, !dbg !90
1179
+ %1103 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1085) #6, !dbg !90
1180
+ %1104 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1086) #6, !dbg !90
1181
+ %1105 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1087) #6, !dbg !90
1182
+ %1106 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1088) #6, !dbg !90
1183
+ %1107 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1089) #6, !dbg !90
1184
+ %1108 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1090) #6, !dbg !90
1185
+ %1109 = tail call i16 asm "cvt.rn.bf16.f32 $0, $1;", "=h,r"(float %1091) #6, !dbg !90
1186
+ %1110 = insertelement <2 x i16> undef, i16 %1094, i64 0, !dbg !90
1187
+ %1111 = insertelement <2 x i16> %1110, i16 %1095, i64 1, !dbg !90
1188
+ %1112 = bitcast <2 x i16> %1111 to i32, !dbg !90
1189
+ %1113 = insertelement <2 x i16> undef, i16 %1096, i64 0, !dbg !90
1190
+ %1114 = insertelement <2 x i16> %1113, i16 %1097, i64 1, !dbg !90
1191
+ %1115 = bitcast <2 x i16> %1114 to i32, !dbg !90
1192
+ %1116 = insertelement <2 x i16> undef, i16 %1098, i64 0, !dbg !90
1193
+ %1117 = insertelement <2 x i16> %1116, i16 %1099, i64 1, !dbg !90
1194
+ %1118 = bitcast <2 x i16> %1117 to i32, !dbg !90
1195
+ %1119 = insertelement <2 x i16> undef, i16 %1100, i64 0, !dbg !90
1196
+ %1120 = insertelement <2 x i16> %1119, i16 %1101, i64 1, !dbg !90
1197
+ %1121 = bitcast <2 x i16> %1120 to i32, !dbg !90
1198
+ tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %1112, i32 %1115, i32 %1118, i32 %1121, ptr addrspace(1) %1092, i1 true) #6, !dbg !90
1199
+ %1122 = insertelement <2 x i16> undef, i16 %1102, i64 0, !dbg !90
1200
+ %1123 = insertelement <2 x i16> %1122, i16 %1103, i64 1, !dbg !90
1201
+ %1124 = bitcast <2 x i16> %1123 to i32, !dbg !90
1202
+ %1125 = insertelement <2 x i16> undef, i16 %1104, i64 0, !dbg !90
1203
+ %1126 = insertelement <2 x i16> %1125, i16 %1105, i64 1, !dbg !90
1204
+ %1127 = bitcast <2 x i16> %1126 to i32, !dbg !90
1205
+ %1128 = insertelement <2 x i16> undef, i16 %1106, i64 0, !dbg !90
1206
+ %1129 = insertelement <2 x i16> %1128, i16 %1107, i64 1, !dbg !90
1207
+ %1130 = bitcast <2 x i16> %1129 to i32, !dbg !90
1208
+ %1131 = insertelement <2 x i16> undef, i16 %1108, i64 0, !dbg !90
1209
+ %1132 = insertelement <2 x i16> %1131, i16 %1109, i64 1, !dbg !90
1210
+ %1133 = bitcast <2 x i16> %1132 to i32, !dbg !90
1211
+ tail call void asm sideeffect "@$5 st.global.v4.b32 [ $4 + 0 ], { $0, $1, $2, $3 };", "r,r,r,r,l,b"(i32 %1124, i32 %1127, i32 %1130, i32 %1133, ptr addrspace(1) %1093, i1 true) #6, !dbg !90
1212
+ %1134 = add nuw nsw i32 %825, 64, !dbg !69
1213
+ %1135 = icmp ult i32 %825, 192, !dbg !69
1214
+ br i1 %1135, label %824, label %1136, !dbg !69
1215
+
1216
+ 1136: ; preds = %__nv_rsqrtf.exit40
1217
+ ret void, !dbg !91
1218
+ }
1219
+
1220
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
1221
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
1222
+
1223
+ ; Function Attrs: convergent nocallback nounwind
1224
+ declare void @llvm.nvvm.barrier0() #1
1225
+
1226
+ ; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
1227
+ declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #2
1228
+
1229
+ ; Function Attrs: alwaysinline nounwind
1230
+ define float @__nv_rsqrtf(float %x) local_unnamed_addr #3 {
1231
+ %1 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #6
1232
+ %.not = icmp eq i32 %1, 0
1233
+ br i1 %.not, label %4, label %2
1234
+
1235
+ 2: ; preds = %0
1236
+ %3 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %x)
1237
+ br label %6
1238
+
1239
+ 4: ; preds = %0
1240
+ %5 = tail call float @llvm.nvvm.rsqrt.approx.f(float %x)
1241
+ br label %6
1242
+
1243
+ 6: ; preds = %4, %2
1244
+ %.0 = phi float [ %3, %2 ], [ %5, %4 ]
1245
+ ret float %.0
1246
+ }
1247
+
1248
+ declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
1249
+
1250
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
1251
+ declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #5
1252
+
1253
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
1254
+ declare float @llvm.nvvm.rsqrt.approx.f(float) #5
1255
+
1256
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
1257
+ attributes #1 = { convergent nocallback nounwind }
1258
+ attributes #2 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
1259
+ attributes #3 = { alwaysinline nounwind "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
1260
+ attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
1261
+ attributes #5 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
1262
+ attributes #6 = { nounwind }
1263
+
1264
+ !llvm.module.flags = !{!0, !1}
1265
+ !llvm.dbg.cu = !{!2}
1266
+ !nvvm.annotations = !{!4, !5, !5, !4}
1267
+ !llvm.ident = !{!6}
1268
+
1269
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
1270
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
1271
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: FullDebug)
1272
+ !3 = !DIFile(filename: "ccig6fki6p4lxrdmgg6eudahiexcvueeol2p4qp532pvve2y463y.py", directory: "/tmp/torchinductor_root/ci")
1273
+ !4 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"kernel", i32 1}
1274
+ !5 = !{ptr @triton__0d1d2d3d4d5d6de7de, !"maxntidx", i32 256}
1275
+ !6 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
1276
+ !7 = distinct !DISubprogram(name: "triton__0d1d2d3d4d5d6de7de", linkageName: "triton__0d1d2d3d4d5d6de7de", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
1277
+ !8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
1278
+ !9 = !{}
1279
+ !10 = !DILocation(line: 22, column: 44, scope: !7)
1280
+ !11 = !DILocation(line: 24, column: 33, scope: !7)
1281
+ !12 = !DILocation(line: 31, column: 36, scope: !7)
1282
+ !13 = !DILocation(line: 21, column: 28, scope: !7)
1283
+ !14 = !DILocation(line: 21, column: 33, scope: !7)
1284
+ !15 = !DILocation(line: 22, column: 23, scope: !7)
1285
+ !16 = !DILocation(line: 26, column: 30, scope: !7)
1286
+ !17 = !DILocation(line: 26, column: 35, scope: !7)
1287
+ !18 = !DILocation(line: 27, column: 18, scope: !7)
1288
+ !19 = !DILocation(line: 35, column: 44, scope: !7)
1289
+ !20 = !DILocation(line: 36, column: 44, scope: !7)
1290
+ !21 = !DILocation(line: 37, column: 22, scope: !7)
1291
+ !22 = !DILocation(line: 38, column: 22, scope: !7)
1292
+ !23 = !DILocation(line: 39, column: 36, scope: !7)
1293
+ !24 = !DILocation(line: 40, column: 40, scope: !7)
1294
+ !25 = !DILocation(line: 41, column: 44, scope: !7)
1295
+ !26 = !DILocation(line: 32, column: 27, scope: !7)
1296
+ !27 = !DILocation(line: 35, column: 40, scope: !7)
1297
+ !28 = !DILocation(line: 35, column: 34, scope: !7)
1298
+ !29 = !DILocation(line: 35, column: 50, scope: !7)
1299
+ !30 = !DILocation(line: 36, column: 40, scope: !7)
1300
+ !31 = !DILocation(line: 36, column: 34, scope: !7)
1301
+ !32 = !DILocation(line: 36, column: 50, scope: !7)
1302
+ !33 = !DILocation(line: 36, column: 101, scope: !7)
1303
+ !34 = !DILocation(line: 40, column: 55, scope: !7)
1304
+ !35 = !DILocation(line: 41, column: 40, scope: !7)
1305
+ !36 = !DILocation(line: 41, column: 34, scope: !7)
1306
+ !37 = !DILocation(line: 41, column: 52, scope: !7)
1307
+ !38 = !DILocation(line: 42, column: 22, scope: !7)
1308
+ !39 = !DILocation(line: 44, column: 22, scope: !7)
1309
+ !40 = !DILocation(line: 96, column: 20, scope: !41, inlinedAt: !43)
1310
+ !41 = distinct !DILexicalBlockFile(scope: !7, file: !42, discriminator: 0)
1311
+ !42 = !DIFile(filename: "triton_helpers.py", directory: "/usr/local/lib/python3.10/dist-packages/torch/_inductor")
1312
+ !43 = !DILocation(line: 47, column: 41, scope: !41)
1313
+ !44 = !DILocation(line: 97, column: 26, scope: !41, inlinedAt: !43)
1314
+ !45 = !DILocation(line: 98, column: 30, scope: !41, inlinedAt: !43)
1315
+ !46 = !DILocation(line: 98, column: 22, scope: !41, inlinedAt: !43)
1316
+ !47 = !DILocation(line: 101, column: 30, scope: !41, inlinedAt: !43)
1317
+ !48 = !DILocation(line: 101, column: 22, scope: !41, inlinedAt: !43)
1318
+ !49 = !DILocation(line: 50, column: 50, scope: !7)
1319
+ !50 = !DILocation(line: 108, column: 21, scope: !51, inlinedAt: !52)
1320
+ !51 = distinct !DILexicalBlockFile(scope: !41, file: !42, discriminator: 0)
1321
+ !52 = !DILocation(line: 120, column: 46, scope: !51, inlinedAt: !53)
1322
+ !53 = !DILocation(line: 53, column: 44, scope: !51)
1323
+ !54 = !DILocation(line: 109, column: 28, scope: !51, inlinedAt: !52)
1324
+ !55 = !DILocation(line: 110, column: 39, scope: !51, inlinedAt: !52)
1325
+ !56 = !DILocation(line: 110, column: 60, scope: !51, inlinedAt: !52)
1326
+ !57 = !DILocation(line: 110, column: 49, scope: !51, inlinedAt: !52)
1327
+ !58 = !DILocation(line: 112, column: 25, scope: !51, inlinedAt: !52)
1328
+ !59 = !DILocation(line: 112, column: 17, scope: !51, inlinedAt: !52)
1329
+ !60 = !DILocation(line: 113, column: 15, scope: !51, inlinedAt: !52)
1330
+ !61 = !DILocation(line: 113, column: 30, scope: !51, inlinedAt: !52)
1331
+ !62 = !DILocation(line: 113, column: 38, scope: !51, inlinedAt: !52)
1332
+ !63 = !DILocation(line: 113, column: 49, scope: !51, inlinedAt: !52)
1333
+ !64 = !DILocation(line: 113, column: 22, scope: !51, inlinedAt: !52)
1334
+ !65 = !DILocation(line: 120, column: 46, scope: !41, inlinedAt: !66)
1335
+ !66 = !DILocation(line: 53, column: 44, scope: !41)
1336
+ !67 = !DILocation(line: 75, column: 24, scope: !7)
1337
+ !68 = !DILocation(line: 77, column: 24, scope: !7)
1338
+ !69 = !DILocation(line: 58, column: 36, scope: !7)
1339
+ !70 = !DILocation(line: 59, column: 27, scope: !7)
1340
+ !71 = !DILocation(line: 62, column: 41, scope: !7)
1341
+ !72 = !DILocation(line: 62, column: 35, scope: !7)
1342
+ !73 = !DILocation(line: 62, column: 51, scope: !7)
1343
+ !74 = !DILocation(line: 63, column: 41, scope: !7)
1344
+ !75 = !DILocation(line: 63, column: 35, scope: !7)
1345
+ !76 = !DILocation(line: 63, column: 51, scope: !7)
1346
+ !77 = !DILocation(line: 63, column: 103, scope: !7)
1347
+ !78 = !DILocation(line: 64, column: 35, scope: !7)
1348
+ !79 = !DILocation(line: 64, column: 40, scope: !7)
1349
+ !80 = !DILocation(line: 68, column: 57, scope: !7)
1350
+ !81 = !DILocation(line: 69, column: 35, scope: !7)
1351
+ !82 = !DILocation(line: 69, column: 54, scope: !7)
1352
+ !83 = !DILocation(line: 70, column: 24, scope: !7)
1353
+ !84 = !DILocation(line: 72, column: 24, scope: !7)
1354
+ !85 = !DILocation(line: 73, column: 24, scope: !7)
1355
+ !86 = !DILocation(line: 78, column: 30, scope: !7)
1356
+ !87 = !DILocation(line: 79, column: 24, scope: !7)
1357
+ !88 = !DILocation(line: 80, column: 24, scope: !7)
1358
+ !89 = !DILocation(line: 82, column: 29, scope: !7)
1359
+ !90 = !DILocation(line: 82, column: 52, scope: !7)
1360
+ !91 = !DILocation(line: 58, column: 4, scope: !7)
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ptx ADDED
@@ -0,0 +1,2004 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.2
6
+ .target sm_89
7
+ .address_size 64
8
+
9
+ // .globl triton__0d1d2d3d4d5d6de7de
10
+ .extern .func __assertfail
11
+ (
12
+ .param .b64 __assertfail_param_0,
13
+ .param .b64 __assertfail_param_1,
14
+ .param .b32 __assertfail_param_2,
15
+ .param .b64 __assertfail_param_3,
16
+ .param .b64 __assertfail_param_4
17
+ )
18
+ ;
19
+ .global .align 1 .b8 assertFunc_1[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
20
+ .global .align 1 .b8 assertFile_1[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
21
+ .global .align 1 .b8 assertMessage_1[39] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 54, 32, 60, 32, 53, 48, 50, 53, 55};
22
+ .global .align 1 .b8 assertFunc_0[25] = {95, 99, 97, 108, 108, 95, 119, 105, 116, 104, 95, 102, 114, 97, 109, 101, 115, 95, 114, 101, 109, 111, 118, 101, 100};
23
+ .global .align 1 .b8 assertFile_0[38] = {60, 102, 114, 111, 122, 101, 110, 32, 105, 109, 112, 111, 114, 116, 108, 105, 98, 46, 95, 98, 111, 111, 116, 115, 116, 114, 97, 112, 95, 101, 120, 116, 101, 114, 110, 97, 108, 62};
24
+ .global .align 1 .b8 assertMessage_0[38] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 51, 32, 60, 32, 53, 48, 50, 53, 55};
25
+ .extern .shared .align 1 .b8 global_smem[];
26
+ .global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90, 0};
27
+
28
+ .visible .entry triton__0d1d2d3d4d5d6de7de(
29
+ .param .u64 triton__0d1d2d3d4d5d6de7de_param_0,
30
+ .param .u64 triton__0d1d2d3d4d5d6de7de_param_1,
31
+ .param .u64 triton__0d1d2d3d4d5d6de7de_param_2,
32
+ .param .u64 triton__0d1d2d3d4d5d6de7de_param_3,
33
+ .param .u64 triton__0d1d2d3d4d5d6de7de_param_4,
34
+ .param .u64 triton__0d1d2d3d4d5d6de7de_param_5,
35
+ .param .u32 triton__0d1d2d3d4d5d6de7de_param_6,
36
+ .param .u32 triton__0d1d2d3d4d5d6de7de_param_7
37
+ )
38
+ .maxntid 256, 1, 1
39
+ {
40
+ .reg .pred %p<157>;
41
+ .reg .b16 %rs<49>;
42
+ .reg .b32 %r<474>;
43
+ .reg .f32 %f<678>;
44
+ .reg .b64 %rd<118>;
45
+ .loc 1 18 0
46
+ $L__func_begin0:
47
+ .loc 1 18 0
48
+
49
+ ld.param.u64 %rd16, [triton__0d1d2d3d4d5d6de7de_param_5];
50
+ ld.param.u64 %rd15, [triton__0d1d2d3d4d5d6de7de_param_4];
51
+ ld.param.u64 %rd14, [triton__0d1d2d3d4d5d6de7de_param_3];
52
+ ld.param.u64 %rd52, [triton__0d1d2d3d4d5d6de7de_param_0];
53
+ ld.param.u64 %rd53, [triton__0d1d2d3d4d5d6de7de_param_1];
54
+ $L__tmp0:
55
+ .loc 1 22 44
56
+ mov.u32 %r12, %tid.x;
57
+ ld.param.u64 %rd54, [triton__0d1d2d3d4d5d6de7de_param_2];
58
+ bfe.u32 %r1, %r12, 3, 5;
59
+ and.b32 %r2, %r12, 63;
60
+ .loc 1 24 33
61
+ shl.b32 %r13, %r12, 3;
62
+ and.b32 %r3, %r13, 56;
63
+ .loc 1 31 36
64
+ shr.u32 %r4, %r12, 6;
65
+ .loc 1 21 28
66
+ mov.u32 %r10, %ctaid.x;
67
+ .loc 1 21 33
68
+ shl.b32 %r14, %r10, 6;
69
+ .loc 1 22 23
70
+ or.b32 %r15, %r14, %r1;
71
+ or.b32 %r16, %r15, 32;
72
+ or.b32 %r17, %r14, %r2;
73
+ .loc 1 26 30
74
+ mul.wide.s32 %rd55, %r15, 8;
75
+ add.s64 %rd18, %rd52, %rd55;
76
+ add.s64 %rd34, %rd18, 256;
77
+ mul.wide.s32 %rd56, %r17, 8;
78
+ add.s64 %rd50, %rd52, %rd56;
79
+ mov.pred %p1, -1;
80
+ .loc 1 26 35
81
+ mov.u64 %rd17, 0x0;
82
+ @%p1 ld.global.L1::evict_last.b64 { %rd17 }, [ %rd18 + 0 ];
83
+ mov.u64 %rd19, 0x0;
84
+ @%p1 ld.global.L1::evict_last.b64 { %rd19 }, [ %rd18 + 0 ];
85
+ mov.u64 %rd21, 0x0;
86
+ @%p1 ld.global.L1::evict_last.b64 { %rd21 }, [ %rd18 + 0 ];
87
+ mov.u64 %rd23, 0x0;
88
+ @%p1 ld.global.L1::evict_last.b64 { %rd23 }, [ %rd18 + 0 ];
89
+ mov.u64 %rd25, 0x0;
90
+ @%p1 ld.global.L1::evict_last.b64 { %rd25 }, [ %rd18 + 0 ];
91
+ mov.u64 %rd27, 0x0;
92
+ @%p1 ld.global.L1::evict_last.b64 { %rd27 }, [ %rd18 + 0 ];
93
+ mov.u64 %rd29, 0x0;
94
+ @%p1 ld.global.L1::evict_last.b64 { %rd29 }, [ %rd18 + 0 ];
95
+ mov.u64 %rd31, 0x0;
96
+ @%p1 ld.global.L1::evict_last.b64 { %rd31 }, [ %rd18 + 0 ];
97
+ mov.u64 %rd33, 0x0;
98
+ @%p1 ld.global.L1::evict_last.b64 { %rd33 }, [ %rd34 + 0 ];
99
+ mov.u64 %rd35, 0x0;
100
+ @%p1 ld.global.L1::evict_last.b64 { %rd35 }, [ %rd34 + 0 ];
101
+ mov.u64 %rd37, 0x0;
102
+ @%p1 ld.global.L1::evict_last.b64 { %rd37 }, [ %rd34 + 0 ];
103
+ mov.u64 %rd39, 0x0;
104
+ @%p1 ld.global.L1::evict_last.b64 { %rd39 }, [ %rd34 + 0 ];
105
+ mov.u64 %rd41, 0x0;
106
+ @%p1 ld.global.L1::evict_last.b64 { %rd41 }, [ %rd34 + 0 ];
107
+ mov.u64 %rd43, 0x0;
108
+ @%p1 ld.global.L1::evict_last.b64 { %rd43 }, [ %rd34 + 0 ];
109
+ mov.u64 %rd45, 0x0;
110
+ @%p1 ld.global.L1::evict_last.b64 { %rd45 }, [ %rd34 + 0 ];
111
+ mov.u64 %rd47, 0x0;
112
+ @%p1 ld.global.L1::evict_last.b64 { %rd47 }, [ %rd34 + 0 ];
113
+ mov.u64 %rd49, 0x0;
114
+ @%p1 ld.global.L1::evict_last.b64 { %rd49 }, [ %rd50 + 0 ];
115
+ .loc 1 27 18
116
+ bfe.s32 %r18, %r10, 25, 1;
117
+ shr.u32 %r19, %r18, 23;
118
+ add.s32 %r20, %r15, %r19;
119
+ and.b32 %r21, %r20, 16776704;
120
+ sub.s32 %r22, %r15, %r21;
121
+ add.s32 %r23, %r16, %r19;
122
+ and.b32 %r24, %r23, 16776704;
123
+ sub.s32 %r25, %r16, %r24;
124
+ .loc 1 35 44
125
+ shl.b32 %r26, %r22, 8;
126
+ shl.b32 %r27, %r25, 8;
127
+ .loc 1 37 22
128
+ add.s64 %rd57, %rd49, 50257;
129
+ .loc 1 38 22
130
+ setp.lt.s64 %p18, %rd17, 0;
131
+ setp.lt.s64 %p19, %rd33, 0;
132
+ setp.lt.s64 %p20, %rd49, 0;
133
+ .loc 1 39 36
134
+ selp.b64 %rd1, %rd57, %rd49, %p20;
135
+ .loc 1 41 44
136
+ shl.b64 %rd58, %rd17, 8;
137
+ add.s64 %rd59, %rd58, 12865792;
138
+ selp.b64 %rd60, %rd59, %rd58, %p18;
139
+ shl.b64 %rd61, %rd33, 8;
140
+ add.s64 %rd62, %rd61, 12865792;
141
+ selp.b64 %rd63, %rd62, %rd61, %p19;
142
+ .loc 1 31 36
143
+ and.b32 %r28, %r12, 7;
144
+ mul.wide.u32 %rd2, %r28, 32;
145
+ shl.b64 %rd64, %rd63, 2;
146
+ or.b64 %rd65, %rd2, %rd64;
147
+ add.s64 %rd3, %rd53, %rd65;
148
+ shl.b64 %rd66, %rd60, 2;
149
+ or.b64 %rd67, %rd2, %rd66;
150
+ add.s64 %rd4, %rd53, %rd67;
151
+ or.b32 %r29, %r27, %r3;
152
+ mul.wide.s32 %rd68, %r29, 4;
153
+ add.s64 %rd5, %rd54, %rd68;
154
+ or.b32 %r30, %r26, %r3;
155
+ mul.wide.s32 %rd69, %r30, 4;
156
+ add.s64 %rd6, %rd54, %rd69;
157
+ shl.b32 %r31, %r10, 14;
158
+ shl.b32 %r32, %r1, 8;
159
+ or.b32 %r33, %r31, %r32;
160
+ or.b32 %r5, %r33, %r3;
161
+ mov.f32 %f614, 0f00000000;
162
+ mov.u64 %rd116, 0;
163
+ mov.b32 %r472, -64;
164
+ mov.f32 %f615, %f614;
165
+ mov.f32 %f616, %f614;
166
+ mov.f32 %f617, %f614;
167
+ mov.f32 %f618, %f614;
168
+ mov.f32 %f619, %f614;
169
+ mov.f32 %f620, %f614;
170
+ mov.f32 %f621, %f614;
171
+ mov.f32 %f622, %f614;
172
+ mov.f32 %f623, %f614;
173
+ mov.f32 %f624, %f614;
174
+ mov.f32 %f625, %f614;
175
+ mov.f32 %f626, %f614;
176
+ mov.f32 %f627, %f614;
177
+ mov.f32 %f628, %f614;
178
+ mov.f32 %f629, %f614;
179
+ mov.f32 %f630, %f614;
180
+ mov.f32 %f631, %f614;
181
+ mov.f32 %f632, %f614;
182
+ mov.f32 %f633, %f614;
183
+ mov.f32 %f634, %f614;
184
+ mov.f32 %f635, %f614;
185
+ mov.f32 %f636, %f614;
186
+ mov.f32 %f637, %f614;
187
+ mov.f32 %f638, %f614;
188
+ mov.f32 %f639, %f614;
189
+ mov.f32 %f640, %f614;
190
+ mov.f32 %f641, %f614;
191
+ mov.f32 %f642, %f614;
192
+ mov.f32 %f643, %f614;
193
+ mov.f32 %f644, %f614;
194
+ mov.f32 %f645, %f614;
195
+ mov.f32 %f646, %f614;
196
+ mov.f32 %f647, %f614;
197
+ mov.f32 %f648, %f614;
198
+ mov.f32 %f649, %f614;
199
+ mov.f32 %f650, %f614;
200
+ mov.f32 %f651, %f614;
201
+ mov.f32 %f652, %f614;
202
+ mov.f32 %f653, %f614;
203
+ mov.f32 %f654, %f614;
204
+ mov.f32 %f655, %f614;
205
+ mov.f32 %f656, %f614;
206
+ mov.f32 %f657, %f614;
207
+ mov.f32 %f658, %f614;
208
+ mov.f32 %f659, %f614;
209
+ mov.f32 %f660, %f614;
210
+ mov.f32 %f661, %f614;
211
+ mov.f32 %f662, %f614;
212
+ mov.f32 %f663, %f614;
213
+ mov.f32 %f664, %f614;
214
+ mov.f32 %f665, %f614;
215
+ mov.f32 %f666, %f614;
216
+ mov.f32 %f667, %f614;
217
+ mov.f32 %f668, %f614;
218
+ mov.f32 %f669, %f614;
219
+ mov.f32 %f670, %f614;
220
+ mov.f32 %f671, %f614;
221
+ mov.f32 %f672, %f614;
222
+ mov.f32 %f673, %f614;
223
+ mov.f32 %f674, %f614;
224
+ mov.f32 %f675, %f614;
225
+ mov.f32 %f676, %f614;
226
+ mov.f32 %f677, %f614;
227
+ bra.uni $L__BB0_1;
228
+ $L__BB0_3:
229
+ .loc 1 41 40
230
+ add.s64 %rd85, %rd4, %rd116;
231
+ .loc 1 41 34
232
+ add.s64 %rd86, %rd85, 16;
233
+ add.s64 %rd87, %rd3, %rd116;
234
+ .loc 1 41 52
235
+ add.s64 %rd88, %rd87, 16;
236
+ mov.u32 %r102, 0x0;
237
+ mov.u32 %r103, 0x0;
238
+ mov.u32 %r104, 0x0;
239
+ mov.u32 %r105, 0x0;
240
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r102, %r103, %r104, %r105 }, [ %rd85 + 0 ];
241
+ @!%p1 mov.u32 %r102, %r411;
242
+ @!%p1 mov.u32 %r103, %r411;
243
+ @!%p1 mov.u32 %r104, %r411;
244
+ @!%p1 mov.u32 %r105, %r411;
245
+ mov.b32 %f206, %r102;
246
+ mov.b32 %f207, %r103;
247
+ mov.b32 %f208, %r104;
248
+ mov.b32 %f209, %r105;
249
+ mov.u32 %r110, 0x0;
250
+ mov.u32 %r111, 0x0;
251
+ mov.u32 %r112, 0x0;
252
+ mov.u32 %r113, 0x0;
253
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r110, %r111, %r112, %r113 }, [ %rd86 + 0 ];
254
+ @!%p1 mov.u32 %r110, %r411;
255
+ @!%p1 mov.u32 %r111, %r411;
256
+ @!%p1 mov.u32 %r112, %r411;
257
+ @!%p1 mov.u32 %r113, %r411;
258
+ mov.b32 %f210, %r110;
259
+ mov.b32 %f211, %r111;
260
+ mov.b32 %f212, %r112;
261
+ mov.b32 %f213, %r113;
262
+ mov.u32 %r118, 0x0;
263
+ mov.u32 %r119, 0x0;
264
+ mov.u32 %r120, 0x0;
265
+ mov.u32 %r121, 0x0;
266
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r118, %r119, %r120, %r121 }, [ %rd87 + 0 ];
267
+ @!%p1 mov.u32 %r118, %r411;
268
+ @!%p1 mov.u32 %r119, %r411;
269
+ @!%p1 mov.u32 %r120, %r411;
270
+ @!%p1 mov.u32 %r121, %r411;
271
+ mov.b32 %f214, %r118;
272
+ mov.b32 %f215, %r119;
273
+ mov.b32 %f216, %r120;
274
+ mov.b32 %f217, %r121;
275
+ mov.u32 %r126, 0x0;
276
+ mov.u32 %r127, 0x0;
277
+ mov.u32 %r128, 0x0;
278
+ mov.u32 %r129, 0x0;
279
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r126, %r127, %r128, %r129 }, [ %rd88 + 0 ];
280
+ @!%p1 mov.u32 %r126, %r411;
281
+ @!%p1 mov.u32 %r127, %r411;
282
+ @!%p1 mov.u32 %r128, %r411;
283
+ @!%p1 mov.u32 %r129, %r411;
284
+ mov.b32 %f218, %r126;
285
+ mov.b32 %f219, %r127;
286
+ mov.b32 %f220, %r128;
287
+ mov.b32 %f221, %r129;
288
+ .loc 1 42 22
289
+ add.f32 %f222, %f65, %f206;
290
+ add.f32 %f223, %f66, %f207;
291
+ add.f32 %f224, %f67, %f208;
292
+ add.f32 %f225, %f68, %f209;
293
+ add.f32 %f226, %f69, %f210;
294
+ add.f32 %f227, %f70, %f211;
295
+ add.f32 %f228, %f71, %f212;
296
+ add.f32 %f229, %f72, %f213;
297
+ add.f32 %f230, %f73, %f214;
298
+ add.f32 %f231, %f74, %f215;
299
+ add.f32 %f232, %f75, %f216;
300
+ add.f32 %f233, %f76, %f217;
301
+ add.f32 %f234, %f77, %f218;
302
+ add.f32 %f235, %f78, %f219;
303
+ add.f32 %f236, %f79, %f220;
304
+ add.f32 %f237, %f80, %f221;
305
+ .loc 1 44 22
306
+ add.f32 %f238, %f81, %f222;
307
+ add.f32 %f239, %f82, %f223;
308
+ add.f32 %f240, %f83, %f224;
309
+ add.f32 %f241, %f84, %f225;
310
+ add.f32 %f242, %f85, %f226;
311
+ add.f32 %f243, %f86, %f227;
312
+ add.f32 %f244, %f87, %f228;
313
+ add.f32 %f245, %f88, %f229;
314
+ add.f32 %f246, %f89, %f230;
315
+ add.f32 %f247, %f90, %f231;
316
+ add.f32 %f248, %f91, %f232;
317
+ add.f32 %f249, %f92, %f233;
318
+ add.f32 %f250, %f93, %f234;
319
+ add.f32 %f251, %f94, %f235;
320
+ add.f32 %f252, %f95, %f236;
321
+ add.f32 %f253, %f96, %f237;
322
+ $L__tmp1:
323
+ .loc 2 96 20
324
+ sub.f32 %f254, %f238, %f662;
325
+ sub.f32 %f255, %f239, %f663;
326
+ sub.f32 %f256, %f240, %f664;
327
+ sub.f32 %f257, %f241, %f665;
328
+ sub.f32 %f258, %f242, %f666;
329
+ sub.f32 %f259, %f243, %f667;
330
+ sub.f32 %f260, %f244, %f668;
331
+ sub.f32 %f261, %f245, %f669;
332
+ sub.f32 %f262, %f246, %f670;
333
+ sub.f32 %f263, %f247, %f671;
334
+ sub.f32 %f264, %f248, %f672;
335
+ sub.f32 %f265, %f249, %f673;
336
+ sub.f32 %f266, %f250, %f674;
337
+ sub.f32 %f267, %f251, %f675;
338
+ sub.f32 %f268, %f252, %f676;
339
+ sub.f32 %f269, %f253, %f677;
340
+ .loc 2 97 26
341
+ add.f32 %f614, %f614, 0f3F800000;
342
+ add.f32 %f615, %f615, 0f3F800000;
343
+ add.f32 %f616, %f616, 0f3F800000;
344
+ add.f32 %f617, %f617, 0f3F800000;
345
+ add.f32 %f618, %f618, 0f3F800000;
346
+ add.f32 %f619, %f619, 0f3F800000;
347
+ add.f32 %f620, %f620, 0f3F800000;
348
+ add.f32 %f621, %f621, 0f3F800000;
349
+ add.f32 %f622, %f622, 0f3F800000;
350
+ add.f32 %f623, %f623, 0f3F800000;
351
+ add.f32 %f624, %f624, 0f3F800000;
352
+ add.f32 %f625, %f625, 0f3F800000;
353
+ add.f32 %f626, %f626, 0f3F800000;
354
+ add.f32 %f627, %f627, 0f3F800000;
355
+ add.f32 %f628, %f628, 0f3F800000;
356
+ add.f32 %f629, %f629, 0f3F800000;
357
+ add.f32 %f630, %f630, 0f3F800000;
358
+ add.f32 %f631, %f631, 0f3F800000;
359
+ add.f32 %f632, %f632, 0f3F800000;
360
+ add.f32 %f633, %f633, 0f3F800000;
361
+ add.f32 %f634, %f634, 0f3F800000;
362
+ add.f32 %f635, %f635, 0f3F800000;
363
+ add.f32 %f636, %f636, 0f3F800000;
364
+ add.f32 %f637, %f637, 0f3F800000;
365
+ add.f32 %f638, %f638, 0f3F800000;
366
+ add.f32 %f639, %f639, 0f3F800000;
367
+ add.f32 %f640, %f640, 0f3F800000;
368
+ add.f32 %f641, %f641, 0f3F800000;
369
+ add.f32 %f642, %f642, 0f3F800000;
370
+ add.f32 %f643, %f643, 0f3F800000;
371
+ add.f32 %f644, %f644, 0f3F800000;
372
+ add.f32 %f645, %f645, 0f3F800000;
373
+ .loc 2 98 30
374
+ mov.b32 %r135, %f254;
375
+ mov.b32 %r136, %f614;
376
+ div.full.f32 %r134, %r135, %r136;
377
+ mov.b32 %f270, %r134;
378
+ mov.b32 %r138, %f255;
379
+ mov.b32 %r139, %f615;
380
+ div.full.f32 %r137, %r138, %r139;
381
+ mov.b32 %f271, %r137;
382
+ mov.b32 %r141, %f256;
383
+ mov.b32 %r142, %f616;
384
+ div.full.f32 %r140, %r141, %r142;
385
+ mov.b32 %f272, %r140;
386
+ mov.b32 %r144, %f257;
387
+ mov.b32 %r145, %f617;
388
+ div.full.f32 %r143, %r144, %r145;
389
+ mov.b32 %f273, %r143;
390
+ mov.b32 %r147, %f258;
391
+ mov.b32 %r148, %f618;
392
+ div.full.f32 %r146, %r147, %r148;
393
+ mov.b32 %f274, %r146;
394
+ mov.b32 %r150, %f259;
395
+ mov.b32 %r151, %f619;
396
+ div.full.f32 %r149, %r150, %r151;
397
+ mov.b32 %f275, %r149;
398
+ mov.b32 %r153, %f260;
399
+ mov.b32 %r154, %f620;
400
+ div.full.f32 %r152, %r153, %r154;
401
+ mov.b32 %f276, %r152;
402
+ mov.b32 %r156, %f261;
403
+ mov.b32 %r157, %f621;
404
+ div.full.f32 %r155, %r156, %r157;
405
+ mov.b32 %f277, %r155;
406
+ mov.b32 %r159, %f262;
407
+ mov.b32 %r160, %f622;
408
+ div.full.f32 %r158, %r159, %r160;
409
+ mov.b32 %f278, %r158;
410
+ mov.b32 %r162, %f263;
411
+ mov.b32 %r163, %f623;
412
+ div.full.f32 %r161, %r162, %r163;
413
+ mov.b32 %f279, %r161;
414
+ mov.b32 %r165, %f264;
415
+ mov.b32 %r166, %f624;
416
+ div.full.f32 %r164, %r165, %r166;
417
+ mov.b32 %f280, %r164;
418
+ mov.b32 %r168, %f265;
419
+ mov.b32 %r169, %f625;
420
+ div.full.f32 %r167, %r168, %r169;
421
+ mov.b32 %f281, %r167;
422
+ mov.b32 %r171, %f266;
423
+ mov.b32 %r172, %f626;
424
+ div.full.f32 %r170, %r171, %r172;
425
+ mov.b32 %f282, %r170;
426
+ mov.b32 %r174, %f267;
427
+ mov.b32 %r175, %f627;
428
+ div.full.f32 %r173, %r174, %r175;
429
+ mov.b32 %f283, %r173;
430
+ mov.b32 %r177, %f268;
431
+ mov.b32 %r178, %f628;
432
+ div.full.f32 %r176, %r177, %r178;
433
+ mov.b32 %f284, %r176;
434
+ mov.b32 %r180, %f269;
435
+ mov.b32 %r181, %f629;
436
+ div.full.f32 %r179, %r180, %r181;
437
+ mov.b32 %f285, %r179;
438
+ .loc 2 98 22
439
+ add.f32 %f662, %f662, %f270;
440
+ add.f32 %f663, %f663, %f271;
441
+ add.f32 %f664, %f664, %f272;
442
+ add.f32 %f665, %f665, %f273;
443
+ add.f32 %f666, %f666, %f274;
444
+ add.f32 %f667, %f667, %f275;
445
+ add.f32 %f668, %f668, %f276;
446
+ add.f32 %f669, %f669, %f277;
447
+ add.f32 %f670, %f670, %f278;
448
+ add.f32 %f671, %f671, %f279;
449
+ add.f32 %f672, %f672, %f280;
450
+ add.f32 %f673, %f673, %f281;
451
+ add.f32 %f674, %f674, %f282;
452
+ add.f32 %f675, %f675, %f283;
453
+ add.f32 %f676, %f676, %f284;
454
+ add.f32 %f677, %f677, %f285;
455
+ .loc 2 101 30
456
+ sub.f32 %f286, %f238, %f662;
457
+ sub.f32 %f287, %f239, %f663;
458
+ sub.f32 %f288, %f240, %f664;
459
+ sub.f32 %f289, %f241, %f665;
460
+ sub.f32 %f290, %f242, %f666;
461
+ sub.f32 %f291, %f243, %f667;
462
+ sub.f32 %f292, %f244, %f668;
463
+ sub.f32 %f293, %f245, %f669;
464
+ sub.f32 %f294, %f246, %f670;
465
+ sub.f32 %f295, %f247, %f671;
466
+ sub.f32 %f296, %f248, %f672;
467
+ sub.f32 %f297, %f249, %f673;
468
+ sub.f32 %f298, %f250, %f674;
469
+ sub.f32 %f299, %f251, %f675;
470
+ sub.f32 %f300, %f252, %f676;
471
+ sub.f32 %f301, %f253, %f677;
472
+ $L__tmp2:
473
+ .loc 1 50 50
474
+ fma.rn.f32 %f646, %f254, %f286, %f646;
475
+ fma.rn.f32 %f647, %f255, %f287, %f647;
476
+ fma.rn.f32 %f648, %f256, %f288, %f648;
477
+ fma.rn.f32 %f649, %f257, %f289, %f649;
478
+ fma.rn.f32 %f650, %f258, %f290, %f650;
479
+ fma.rn.f32 %f651, %f259, %f291, %f651;
480
+ fma.rn.f32 %f652, %f260, %f292, %f652;
481
+ fma.rn.f32 %f653, %f261, %f293, %f653;
482
+ fma.rn.f32 %f654, %f262, %f294, %f654;
483
+ fma.rn.f32 %f655, %f263, %f295, %f655;
484
+ fma.rn.f32 %f656, %f264, %f296, %f656;
485
+ fma.rn.f32 %f657, %f265, %f297, %f657;
486
+ fma.rn.f32 %f658, %f266, %f298, %f658;
487
+ fma.rn.f32 %f659, %f267, %f299, %f659;
488
+ fma.rn.f32 %f660, %f268, %f300, %f660;
489
+ fma.rn.f32 %f661, %f269, %f301, %f661;
490
+ .loc 1 31 36
491
+ add.s64 %rd116, %rd116, 256;
492
+ add.s32 %r472, %r472, 64;
493
+ setp.lt.u32 %p72, %r472, 192;
494
+ @%p72 bra $L__BB0_1;
495
+ bra.uni $L__BB0_4;
496
+ $L__BB0_1:
497
+ .loc 1 40 40
498
+ setp.lt.u64 %p51, %rd1, 50257;
499
+ .loc 1 35 34
500
+ add.s64 %rd70, %rd6, %rd116;
501
+ add.s64 %rd71, %rd70, 16;
502
+ add.s64 %rd72, %rd5, %rd116;
503
+ .loc 1 35 50
504
+ add.s64 %rd73, %rd72, 16;
505
+ mov.b32 %r411, 0;
506
+ mov.u32 %r34, 0x0;
507
+ mov.u32 %r35, 0x0;
508
+ mov.u32 %r36, 0x0;
509
+ mov.u32 %r37, 0x0;
510
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r34, %r35, %r36, %r37 }, [ %rd70 + 0 ];
511
+ @!%p1 mov.u32 %r34, %r411;
512
+ @!%p1 mov.u32 %r35, %r411;
513
+ @!%p1 mov.u32 %r36, %r411;
514
+ @!%p1 mov.u32 %r37, %r411;
515
+ mov.b32 %f65, %r34;
516
+ mov.b32 %f66, %r35;
517
+ mov.b32 %f67, %r36;
518
+ mov.b32 %f68, %r37;
519
+ mov.u32 %r42, 0x0;
520
+ mov.u32 %r43, 0x0;
521
+ mov.u32 %r44, 0x0;
522
+ mov.u32 %r45, 0x0;
523
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r42, %r43, %r44, %r45 }, [ %rd71 + 0 ];
524
+ @!%p1 mov.u32 %r42, %r411;
525
+ @!%p1 mov.u32 %r43, %r411;
526
+ @!%p1 mov.u32 %r44, %r411;
527
+ @!%p1 mov.u32 %r45, %r411;
528
+ mov.b32 %f69, %r42;
529
+ mov.b32 %f70, %r43;
530
+ mov.b32 %f71, %r44;
531
+ mov.b32 %f72, %r45;
532
+ mov.u32 %r50, 0x0;
533
+ mov.u32 %r51, 0x0;
534
+ mov.u32 %r52, 0x0;
535
+ mov.u32 %r53, 0x0;
536
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r50, %r51, %r52, %r53 }, [ %rd72 + 0 ];
537
+ @!%p1 mov.u32 %r50, %r411;
538
+ @!%p1 mov.u32 %r51, %r411;
539
+ @!%p1 mov.u32 %r52, %r411;
540
+ @!%p1 mov.u32 %r53, %r411;
541
+ mov.b32 %f73, %r50;
542
+ mov.b32 %f74, %r51;
543
+ mov.b32 %f75, %r52;
544
+ mov.b32 %f76, %r53;
545
+ mov.u32 %r58, 0x0;
546
+ mov.u32 %r59, 0x0;
547
+ mov.u32 %r60, 0x0;
548
+ mov.u32 %r61, 0x0;
549
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r58, %r59, %r60, %r61 }, [ %rd73 + 0 ];
550
+ @!%p1 mov.u32 %r58, %r411;
551
+ @!%p1 mov.u32 %r59, %r411;
552
+ @!%p1 mov.u32 %r60, %r411;
553
+ @!%p1 mov.u32 %r61, %r411;
554
+ mov.b32 %f77, %r58;
555
+ mov.b32 %f78, %r59;
556
+ mov.b32 %f79, %r60;
557
+ mov.b32 %f80, %r61;
558
+ .loc 1 36 40
559
+ add.s32 %r98, %r5, %r472;
560
+ add.s32 %r99, %r98, 64;
561
+ .loc 1 36 34
562
+ add.s32 %r100, %r98, 8256;
563
+ mul.wide.s32 %rd76, %r99, 2;
564
+ add.s64 %rd74, %rd14, %rd76;
565
+ mul.wide.s32 %rd77, %r100, 2;
566
+ add.s64 %rd75, %rd14, %rd77;
567
+ .loc 1 36 50
568
+ mov.u32 %r66, 0x0;
569
+ mov.u32 %r67, 0x0;
570
+ mov.u32 %r68, 0x0;
571
+ mov.u32 %r69, 0x0;
572
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r66, %r67, %r68, %r69 }, [ %rd74 + 0 ];
573
+ @!%p1 mov.u32 %r66, %r411;
574
+ @!%p1 mov.u32 %r67, %r411;
575
+ @!%p1 mov.u32 %r68, %r411;
576
+ @!%p1 mov.u32 %r69, %r411;
577
+ cvt.u16.u32 %rs1, %r66;
578
+ { .reg .b16 tmp; mov.b32 {tmp, %rs2}, %r66; }
579
+ cvt.u16.u32 %rs3, %r67;
580
+ { .reg .b16 tmp; mov.b32 {tmp, %rs4}, %r67; }
581
+ cvt.u16.u32 %rs5, %r68;
582
+ { .reg .b16 tmp; mov.b32 {tmp, %rs6}, %r68; }
583
+ cvt.u16.u32 %rs7, %r69;
584
+ { .reg .b16 tmp; mov.b32 {tmp, %rs8}, %r69; }
585
+ mov.u32 %r74, 0x0;
586
+ mov.u32 %r75, 0x0;
587
+ mov.u32 %r76, 0x0;
588
+ mov.u32 %r77, 0x0;
589
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r74, %r75, %r76, %r77 }, [ %rd75 + 0 ];
590
+ @!%p1 mov.u32 %r74, %r411;
591
+ @!%p1 mov.u32 %r75, %r411;
592
+ @!%p1 mov.u32 %r76, %r411;
593
+ @!%p1 mov.u32 %r77, %r411;
594
+ cvt.u16.u32 %rs9, %r74;
595
+ { .reg .b16 tmp; mov.b32 {tmp, %rs10}, %r74; }
596
+ cvt.u16.u32 %rs11, %r75;
597
+ { .reg .b16 tmp; mov.b32 {tmp, %rs12}, %r75; }
598
+ cvt.u16.u32 %rs13, %r76;
599
+ { .reg .b16 tmp; mov.b32 {tmp, %rs14}, %r76; }
600
+ cvt.u16.u32 %rs15, %r77;
601
+ { .reg .b16 tmp; mov.b32 {tmp, %rs16}, %r77; }
602
+ .loc 1 36 101
603
+ cvt.f32.bf16 %r82, %rs1;
604
+ mov.b32 %f81, %r82;
605
+ cvt.f32.bf16 %r83, %rs2;
606
+ mov.b32 %f82, %r83;
607
+ cvt.f32.bf16 %r84, %rs3;
608
+ mov.b32 %f83, %r84;
609
+ cvt.f32.bf16 %r85, %rs4;
610
+ mov.b32 %f84, %r85;
611
+ cvt.f32.bf16 %r86, %rs5;
612
+ mov.b32 %f85, %r86;
613
+ cvt.f32.bf16 %r87, %rs6;
614
+ mov.b32 %f86, %r87;
615
+ cvt.f32.bf16 %r88, %rs7;
616
+ mov.b32 %f87, %r88;
617
+ cvt.f32.bf16 %r89, %rs8;
618
+ mov.b32 %f88, %r89;
619
+ cvt.f32.bf16 %r90, %rs9;
620
+ mov.b32 %f89, %r90;
621
+ cvt.f32.bf16 %r91, %rs10;
622
+ mov.b32 %f90, %r91;
623
+ cvt.f32.bf16 %r92, %rs11;
624
+ mov.b32 %f91, %r92;
625
+ cvt.f32.bf16 %r93, %rs12;
626
+ mov.b32 %f92, %r93;
627
+ cvt.f32.bf16 %r94, %rs13;
628
+ mov.b32 %f93, %r94;
629
+ cvt.f32.bf16 %r95, %rs14;
630
+ mov.b32 %f94, %r95;
631
+ cvt.f32.bf16 %r96, %rs15;
632
+ mov.b32 %f95, %r96;
633
+ cvt.f32.bf16 %r97, %rs16;
634
+ mov.b32 %f96, %r97;
635
+ mov.b32 %r471, 883;
636
+ mov.u64 %rd115, 1;
637
+ .loc 1 40 55
638
+ @%p51 bra $L__BB0_3;
639
+ mov.u64 %rd78, assertMessage_0;
640
+ cvta.global.u64 %rd79, %rd78;
641
+ mov.u64 %rd80, assertFile_0;
642
+ cvta.global.u64 %rd81, %rd80;
643
+ mov.u64 %rd82, assertFunc_0;
644
+ cvta.global.u64 %rd83, %rd82;
645
+ { // callseq 6, 0
646
+ .reg .b32 temp_param_reg;
647
+ .param .b64 param0;
648
+ st.param.b64 [param0+0], %rd79;
649
+ .param .b64 param1;
650
+ st.param.b64 [param1+0], %rd81;
651
+ .param .b32 param2;
652
+ st.param.b32 [param2+0], %r471;
653
+ .param .b64 param3;
654
+ st.param.b64 [param3+0], %rd83;
655
+ .param .b64 param4;
656
+ st.param.b64 [param4+0], %rd115;
657
+ call.uni
658
+ __assertfail,
659
+ (
660
+ param0,
661
+ param1,
662
+ param2,
663
+ param3,
664
+ param4
665
+ );
666
+ } // callseq 6
667
+ bra.uni $L__BB0_3;
668
+ $L__BB0_4:
669
+ .loc 1 31 36
670
+ and.b32 %r291, %r4, 3;
671
+ mad.lo.s32 %r292, %r291, 72, %r2;
672
+ shl.b32 %r293, %r292, 2;
673
+ mov.u32 %r294, global_smem;
674
+ add.s32 %r295, %r294, %r293;
675
+ st.shared.f32 [%r295], %f630;
676
+ st.shared.f32 [%r295+1152], %f631;
677
+ st.shared.f32 [%r295+2304], %f632;
678
+ st.shared.f32 [%r295+3456], %f633;
679
+ st.shared.f32 [%r295+4608], %f634;
680
+ st.shared.f32 [%r295+5760], %f635;
681
+ st.shared.f32 [%r295+6912], %f636;
682
+ st.shared.f32 [%r295+8064], %f637;
683
+ bar.sync 0;
684
+ mad.lo.s32 %r296, %r1, 72, %r3;
685
+ shl.b32 %r297, %r296, 2;
686
+ add.s32 %r298, %r294, %r297;
687
+ ld.shared.v4.f32 {%f302, %f303, %f304, %f305}, [%r298];
688
+ ld.shared.v4.f32 {%f306, %f307, %f308, %f309}, [%r298+16];
689
+ bar.sync 0;
690
+ st.shared.f32 [%r295], %f638;
691
+ st.shared.f32 [%r295+1152], %f639;
692
+ st.shared.f32 [%r295+2304], %f640;
693
+ st.shared.f32 [%r295+3456], %f641;
694
+ st.shared.f32 [%r295+4608], %f642;
695
+ st.shared.f32 [%r295+5760], %f643;
696
+ st.shared.f32 [%r295+6912], %f644;
697
+ st.shared.f32 [%r295+8064], %f645;
698
+ bar.sync 0;
699
+ ld.shared.v4.f32 {%f310, %f311, %f312, %f313}, [%r298];
700
+ ld.shared.v4.f32 {%f314, %f315, %f316, %f317}, [%r298+16];
701
+ $L__tmp3:
702
+ .loc 2 108 21
703
+ sub.f32 %f318, %f663, %f662;
704
+ .loc 2 109 28
705
+ add.f32 %f319, %f302, %f303;
706
+ .loc 2 110 39
707
+ setp.eq.f32 %p73, %f319, 0f00000000;
708
+ .loc 2 110 60
709
+ mov.b32 %r183, %f303;
710
+ mov.b32 %r184, %f319;
711
+ div.full.f32 %r182, %r183, %r184;
712
+ mov.b32 %f320, %r182;
713
+ .loc 2 110 49
714
+ selp.f32 %f321, 0f00000000, %f320, %p73;
715
+ .loc 2 112 17
716
+ fma.rn.f32 %f322, %f318, %f321, %f662;
717
+ .loc 2 113 15
718
+ add.f32 %f323, %f646, %f647;
719
+ .loc 2 113 30
720
+ mul.f32 %f324, %f318, %f318;
721
+ .loc 2 113 38
722
+ mul.f32 %f325, %f324, %f302;
723
+ .loc 2 113 22
724
+ fma.rn.f32 %f326, %f325, %f321, %f323;
725
+ .loc 2 108 21
726
+ sub.f32 %f327, %f664, %f322;
727
+ .loc 2 109 28
728
+ add.f32 %f328, %f304, %f319;
729
+ .loc 2 110 39
730
+ setp.eq.f32 %p74, %f328, 0f00000000;
731
+ .loc 2 110 60
732
+ mov.b32 %r187, %f328;
733
+ mov.b32 %r186, %f304;
734
+ div.full.f32 %r185, %r186, %r187;
735
+ mov.b32 %f329, %r185;
736
+ .loc 2 110 49
737
+ selp.f32 %f330, 0f00000000, %f329, %p74;
738
+ .loc 2 112 17
739
+ fma.rn.f32 %f331, %f330, %f327, %f322;
740
+ .loc 2 113 15
741
+ add.f32 %f332, %f648, %f326;
742
+ .loc 2 113 30
743
+ mul.f32 %f333, %f327, %f327;
744
+ .loc 2 113 38
745
+ mul.f32 %f334, %f319, %f333;
746
+ .loc 2 113 22
747
+ fma.rn.f32 %f335, %f330, %f334, %f332;
748
+ .loc 2 108 21
749
+ sub.f32 %f336, %f665, %f331;
750
+ .loc 2 109 28
751
+ add.f32 %f337, %f305, %f328;
752
+ .loc 2 110 39
753
+ setp.eq.f32 %p75, %f337, 0f00000000;
754
+ .loc 2 110 60
755
+ mov.b32 %r190, %f337;
756
+ mov.b32 %r189, %f305;
757
+ div.full.f32 %r188, %r189, %r190;
758
+ mov.b32 %f338, %r188;
759
+ .loc 2 110 49
760
+ selp.f32 %f339, 0f00000000, %f338, %p75;
761
+ .loc 2 112 17
762
+ fma.rn.f32 %f340, %f339, %f336, %f331;
763
+ .loc 2 113 15
764
+ add.f32 %f341, %f649, %f335;
765
+ .loc 2 113 30
766
+ mul.f32 %f342, %f336, %f336;
767
+ .loc 2 113 38
768
+ mul.f32 %f343, %f328, %f342;
769
+ .loc 2 113 22
770
+ fma.rn.f32 %f344, %f339, %f343, %f341;
771
+ .loc 2 108 21
772
+ sub.f32 %f345, %f666, %f340;
773
+ .loc 2 109 28
774
+ add.f32 %f346, %f306, %f337;
775
+ .loc 2 110 39
776
+ setp.eq.f32 %p76, %f346, 0f00000000;
777
+ .loc 2 110 60
778
+ mov.b32 %r193, %f346;
779
+ mov.b32 %r192, %f306;
780
+ div.full.f32 %r191, %r192, %r193;
781
+ mov.b32 %f347, %r191;
782
+ .loc 2 110 49
783
+ selp.f32 %f348, 0f00000000, %f347, %p76;
784
+ .loc 2 112 17
785
+ fma.rn.f32 %f349, %f348, %f345, %f340;
786
+ .loc 2 113 15
787
+ add.f32 %f350, %f650, %f344;
788
+ .loc 2 113 30
789
+ mul.f32 %f351, %f345, %f345;
790
+ .loc 2 113 38
791
+ mul.f32 %f352, %f337, %f351;
792
+ .loc 2 113 22
793
+ fma.rn.f32 %f353, %f348, %f352, %f350;
794
+ .loc 2 108 21
795
+ sub.f32 %f354, %f667, %f349;
796
+ .loc 2 109 28
797
+ add.f32 %f355, %f307, %f346;
798
+ .loc 2 110 39
799
+ setp.eq.f32 %p77, %f355, 0f00000000;
800
+ .loc 2 110 60
801
+ mov.b32 %r196, %f355;
802
+ mov.b32 %r195, %f307;
803
+ div.full.f32 %r194, %r195, %r196;
804
+ mov.b32 %f356, %r194;
805
+ .loc 2 110 49
806
+ selp.f32 %f357, 0f00000000, %f356, %p77;
807
+ .loc 2 112 17
808
+ fma.rn.f32 %f358, %f357, %f354, %f349;
809
+ .loc 2 113 15
810
+ add.f32 %f359, %f651, %f353;
811
+ .loc 2 113 30
812
+ mul.f32 %f360, %f354, %f354;
813
+ .loc 2 113 38
814
+ mul.f32 %f361, %f346, %f360;
815
+ .loc 2 113 22
816
+ fma.rn.f32 %f362, %f357, %f361, %f359;
817
+ .loc 2 108 21
818
+ sub.f32 %f363, %f668, %f358;
819
+ .loc 2 109 28
820
+ add.f32 %f364, %f308, %f355;
821
+ .loc 2 110 39
822
+ setp.eq.f32 %p78, %f364, 0f00000000;
823
+ .loc 2 110 60
824
+ mov.b32 %r199, %f364;
825
+ mov.b32 %r198, %f308;
826
+ div.full.f32 %r197, %r198, %r199;
827
+ mov.b32 %f365, %r197;
828
+ .loc 2 110 49
829
+ selp.f32 %f366, 0f00000000, %f365, %p78;
830
+ .loc 2 112 17
831
+ fma.rn.f32 %f367, %f366, %f363, %f358;
832
+ .loc 2 113 15
833
+ add.f32 %f368, %f652, %f362;
834
+ .loc 2 113 30
835
+ mul.f32 %f369, %f363, %f363;
836
+ .loc 2 113 38
837
+ mul.f32 %f370, %f355, %f369;
838
+ .loc 2 113 22
839
+ fma.rn.f32 %f371, %f366, %f370, %f368;
840
+ .loc 2 108 21
841
+ sub.f32 %f372, %f669, %f367;
842
+ .loc 2 109 28
843
+ add.f32 %f373, %f309, %f364;
844
+ .loc 2 110 39
845
+ setp.eq.f32 %p79, %f373, 0f00000000;
846
+ .loc 2 110 60
847
+ mov.b32 %r202, %f373;
848
+ mov.b32 %r201, %f309;
849
+ div.full.f32 %r200, %r201, %r202;
850
+ mov.b32 %f374, %r200;
851
+ .loc 2 110 49
852
+ selp.f32 %f375, 0f00000000, %f374, %p79;
853
+ .loc 2 112 17
854
+ fma.rn.f32 %f376, %f375, %f372, %f367;
855
+ .loc 2 113 15
856
+ add.f32 %f377, %f653, %f371;
857
+ .loc 2 113 30
858
+ mul.f32 %f378, %f372, %f372;
859
+ .loc 2 113 38
860
+ mul.f32 %f379, %f364, %f378;
861
+ .loc 2 113 22
862
+ fma.rn.f32 %f380, %f375, %f379, %f377;
863
+ .loc 2 108 21
864
+ sub.f32 %f381, %f671, %f670;
865
+ .loc 2 109 28
866
+ add.f32 %f382, %f310, %f311;
867
+ .loc 2 110 39
868
+ setp.eq.f32 %p80, %f382, 0f00000000;
869
+ .loc 2 110 60
870
+ mov.b32 %r204, %f311;
871
+ mov.b32 %r205, %f382;
872
+ div.full.f32 %r203, %r204, %r205;
873
+ mov.b32 %f383, %r203;
874
+ .loc 2 110 49
875
+ selp.f32 %f384, 0f00000000, %f383, %p80;
876
+ .loc 2 112 17
877
+ fma.rn.f32 %f385, %f381, %f384, %f670;
878
+ .loc 2 113 15
879
+ add.f32 %f386, %f654, %f655;
880
+ .loc 2 113 30
881
+ mul.f32 %f387, %f381, %f381;
882
+ .loc 2 113 38
883
+ mul.f32 %f388, %f387, %f310;
884
+ .loc 2 113 22
885
+ fma.rn.f32 %f389, %f388, %f384, %f386;
886
+ .loc 2 108 21
887
+ sub.f32 %f390, %f672, %f385;
888
+ .loc 2 109 28
889
+ add.f32 %f391, %f312, %f382;
890
+ .loc 2 110 39
891
+ setp.eq.f32 %p81, %f391, 0f00000000;
892
+ .loc 2 110 60
893
+ mov.b32 %r208, %f391;
894
+ mov.b32 %r207, %f312;
895
+ div.full.f32 %r206, %r207, %r208;
896
+ mov.b32 %f392, %r206;
897
+ .loc 2 110 49
898
+ selp.f32 %f393, 0f00000000, %f392, %p81;
899
+ .loc 2 112 17
900
+ fma.rn.f32 %f394, %f393, %f390, %f385;
901
+ .loc 2 113 15
902
+ add.f32 %f395, %f656, %f389;
903
+ .loc 2 113 30
904
+ mul.f32 %f396, %f390, %f390;
905
+ .loc 2 113 38
906
+ mul.f32 %f397, %f382, %f396;
907
+ .loc 2 113 22
908
+ fma.rn.f32 %f398, %f393, %f397, %f395;
909
+ .loc 2 108 21
910
+ sub.f32 %f399, %f673, %f394;
911
+ .loc 2 109 28
912
+ add.f32 %f400, %f313, %f391;
913
+ .loc 2 110 39
914
+ setp.eq.f32 %p82, %f400, 0f00000000;
915
+ .loc 2 110 60
916
+ mov.b32 %r211, %f400;
917
+ mov.b32 %r210, %f313;
918
+ div.full.f32 %r209, %r210, %r211;
919
+ mov.b32 %f401, %r209;
920
+ .loc 2 110 49
921
+ selp.f32 %f402, 0f00000000, %f401, %p82;
922
+ .loc 2 112 17
923
+ fma.rn.f32 %f403, %f402, %f399, %f394;
924
+ .loc 2 113 15
925
+ add.f32 %f404, %f657, %f398;
926
+ .loc 2 113 30
927
+ mul.f32 %f405, %f399, %f399;
928
+ .loc 2 113 38
929
+ mul.f32 %f406, %f391, %f405;
930
+ .loc 2 113 22
931
+ fma.rn.f32 %f407, %f402, %f406, %f404;
932
+ .loc 2 108 21
933
+ sub.f32 %f408, %f674, %f403;
934
+ .loc 2 109 28
935
+ add.f32 %f409, %f314, %f400;
936
+ .loc 2 110 39
937
+ setp.eq.f32 %p83, %f409, 0f00000000;
938
+ .loc 2 110 60
939
+ mov.b32 %r214, %f409;
940
+ mov.b32 %r213, %f314;
941
+ div.full.f32 %r212, %r213, %r214;
942
+ mov.b32 %f410, %r212;
943
+ .loc 2 110 49
944
+ selp.f32 %f411, 0f00000000, %f410, %p83;
945
+ .loc 2 112 17
946
+ fma.rn.f32 %f412, %f411, %f408, %f403;
947
+ .loc 2 113 15
948
+ add.f32 %f413, %f658, %f407;
949
+ .loc 2 113 30
950
+ mul.f32 %f414, %f408, %f408;
951
+ .loc 2 113 38
952
+ mul.f32 %f415, %f400, %f414;
953
+ .loc 2 113 22
954
+ fma.rn.f32 %f416, %f411, %f415, %f413;
955
+ .loc 2 108 21
956
+ sub.f32 %f417, %f675, %f412;
957
+ .loc 2 109 28
958
+ add.f32 %f418, %f315, %f409;
959
+ .loc 2 110 39
960
+ setp.eq.f32 %p84, %f418, 0f00000000;
961
+ .loc 2 110 60
962
+ mov.b32 %r217, %f418;
963
+ mov.b32 %r216, %f315;
964
+ div.full.f32 %r215, %r216, %r217;
965
+ mov.b32 %f419, %r215;
966
+ .loc 2 110 49
967
+ selp.f32 %f420, 0f00000000, %f419, %p84;
968
+ .loc 2 112 17
969
+ fma.rn.f32 %f421, %f420, %f417, %f412;
970
+ .loc 2 113 15
971
+ add.f32 %f422, %f659, %f416;
972
+ .loc 2 113 30
973
+ mul.f32 %f423, %f417, %f417;
974
+ .loc 2 113 38
975
+ mul.f32 %f424, %f409, %f423;
976
+ .loc 2 113 22
977
+ fma.rn.f32 %f425, %f420, %f424, %f422;
978
+ .loc 2 108 21
979
+ sub.f32 %f426, %f676, %f421;
980
+ .loc 2 109 28
981
+ add.f32 %f427, %f316, %f418;
982
+ .loc 2 110 39
983
+ setp.eq.f32 %p85, %f427, 0f00000000;
984
+ .loc 2 110 60
985
+ mov.b32 %r220, %f427;
986
+ mov.b32 %r219, %f316;
987
+ div.full.f32 %r218, %r219, %r220;
988
+ mov.b32 %f428, %r218;
989
+ .loc 2 110 49
990
+ selp.f32 %f429, 0f00000000, %f428, %p85;
991
+ .loc 2 112 17
992
+ fma.rn.f32 %f430, %f429, %f426, %f421;
993
+ .loc 2 113 15
994
+ add.f32 %f431, %f660, %f425;
995
+ .loc 2 113 30
996
+ mul.f32 %f432, %f426, %f426;
997
+ .loc 2 113 38
998
+ mul.f32 %f433, %f418, %f432;
999
+ .loc 2 113 22
1000
+ fma.rn.f32 %f434, %f429, %f433, %f431;
1001
+ .loc 2 108 21
1002
+ sub.f32 %f435, %f677, %f430;
1003
+ .loc 2 109 28
1004
+ add.f32 %f436, %f317, %f427;
1005
+ .loc 2 110 39
1006
+ setp.eq.f32 %p86, %f436, 0f00000000;
1007
+ .loc 2 110 60
1008
+ mov.b32 %r223, %f436;
1009
+ mov.b32 %r222, %f317;
1010
+ div.full.f32 %r221, %r222, %r223;
1011
+ mov.b32 %f437, %r221;
1012
+ .loc 2 110 49
1013
+ selp.f32 %f438, 0f00000000, %f437, %p86;
1014
+ .loc 2 112 17
1015
+ fma.rn.f32 %f439, %f438, %f435, %f430;
1016
+ .loc 2 113 15
1017
+ add.f32 %f440, %f661, %f434;
1018
+ .loc 2 113 30
1019
+ mul.f32 %f441, %f435, %f435;
1020
+ .loc 2 113 38
1021
+ mul.f32 %f442, %f427, %f441;
1022
+ .loc 2 113 22
1023
+ fma.rn.f32 %f443, %f438, %f442, %f440;
1024
+ $L__tmp4:
1025
+ .loc 2 120 46
1026
+ mov.b32 %r299, %f376;
1027
+ shfl.sync.bfly.b32 %r300, %r299, 4, 31, -1;
1028
+ mov.b32 %f444, %r300;
1029
+ mov.b32 %r301, %f380;
1030
+ shfl.sync.bfly.b32 %r302, %r301, 4, 31, -1;
1031
+ mov.b32 %f445, %r302;
1032
+ shfl.sync.bfly.b32 %r225, %r202, 4, 31, -1;
1033
+ mov.b32 %f446, %r225;
1034
+ $L__tmp5:
1035
+ .loc 2 108 21
1036
+ sub.f32 %f447, %f444, %f376;
1037
+ .loc 2 109 28
1038
+ add.f32 %f448, %f373, %f446;
1039
+ .loc 2 110 39
1040
+ setp.eq.f32 %p87, %f448, 0f00000000;
1041
+ .loc 2 110 60
1042
+ mov.b32 %r226, %f448;
1043
+ div.full.f32 %r224, %r225, %r226;
1044
+ mov.b32 %f449, %r224;
1045
+ .loc 2 110 49
1046
+ selp.f32 %f450, 0f00000000, %f449, %p87;
1047
+ .loc 2 112 17
1048
+ fma.rn.f32 %f451, %f450, %f447, %f376;
1049
+ .loc 2 113 15
1050
+ add.f32 %f452, %f380, %f445;
1051
+ .loc 2 113 30
1052
+ mul.f32 %f453, %f447, %f447;
1053
+ .loc 2 113 38
1054
+ mul.f32 %f454, %f373, %f453;
1055
+ .loc 2 113 22
1056
+ fma.rn.f32 %f455, %f450, %f454, %f452;
1057
+ $L__tmp6:
1058
+ .loc 2 120 46
1059
+ mov.b32 %r303, %f451;
1060
+ shfl.sync.bfly.b32 %r304, %r303, 2, 31, -1;
1061
+ mov.b32 %f456, %r304;
1062
+ mov.b32 %r305, %f455;
1063
+ shfl.sync.bfly.b32 %r306, %r305, 2, 31, -1;
1064
+ mov.b32 %f457, %r306;
1065
+ shfl.sync.bfly.b32 %r228, %r226, 2, 31, -1;
1066
+ mov.b32 %f458, %r228;
1067
+ $L__tmp7:
1068
+ .loc 2 108 21
1069
+ sub.f32 %f459, %f456, %f451;
1070
+ .loc 2 109 28
1071
+ add.f32 %f460, %f448, %f458;
1072
+ .loc 2 110 39
1073
+ setp.eq.f32 %p88, %f460, 0f00000000;
1074
+ .loc 2 110 60
1075
+ mov.b32 %r229, %f460;
1076
+ div.full.f32 %r227, %r228, %r229;
1077
+ mov.b32 %f461, %r227;
1078
+ .loc 2 110 49
1079
+ selp.f32 %f462, 0f00000000, %f461, %p88;
1080
+ .loc 2 112 17
1081
+ fma.rn.f32 %f463, %f462, %f459, %f451;
1082
+ .loc 2 113 15
1083
+ add.f32 %f464, %f455, %f457;
1084
+ .loc 2 113 30
1085
+ mul.f32 %f465, %f459, %f459;
1086
+ .loc 2 113 38
1087
+ mul.f32 %f466, %f448, %f465;
1088
+ .loc 2 113 22
1089
+ fma.rn.f32 %f467, %f462, %f466, %f464;
1090
+ $L__tmp8:
1091
+ .loc 2 120 46
1092
+ mov.b32 %r307, %f463;
1093
+ shfl.sync.bfly.b32 %r308, %r307, 1, 31, -1;
1094
+ mov.b32 %f468, %r308;
1095
+ mov.b32 %r309, %f467;
1096
+ shfl.sync.bfly.b32 %r310, %r309, 1, 31, -1;
1097
+ mov.b32 %f469, %r310;
1098
+ shfl.sync.bfly.b32 %r231, %r229, 1, 31, -1;
1099
+ mov.b32 %f470, %r231;
1100
+ $L__tmp9:
1101
+ .loc 2 108 21
1102
+ sub.f32 %f471, %f468, %f463;
1103
+ .loc 2 109 28
1104
+ add.f32 %f472, %f460, %f470;
1105
+ .loc 2 110 39
1106
+ setp.eq.f32 %p89, %f472, 0f00000000;
1107
+ .loc 2 110 60
1108
+ mov.b32 %r232, %f472;
1109
+ div.full.f32 %r230, %r231, %r232;
1110
+ mov.b32 %f473, %r230;
1111
+ .loc 2 110 49
1112
+ selp.f32 %f474, 0f00000000, %f473, %p89;
1113
+ .loc 2 112 17
1114
+ fma.rn.f32 %f161, %f471, %f474, %f463;
1115
+ .loc 2 113 15
1116
+ add.f32 %f475, %f467, %f469;
1117
+ .loc 2 113 30
1118
+ mul.f32 %f476, %f471, %f471;
1119
+ .loc 2 113 38
1120
+ mul.f32 %f477, %f460, %f476;
1121
+ .loc 2 113 22
1122
+ fma.rn.f32 %f478, %f474, %f477, %f475;
1123
+ $L__tmp10:
1124
+ .loc 2 120 46
1125
+ mov.b32 %r311, %f439;
1126
+ shfl.sync.bfly.b32 %r312, %r311, 4, 31, -1;
1127
+ mov.b32 %f479, %r312;
1128
+ mov.b32 %r313, %f443;
1129
+ shfl.sync.bfly.b32 %r314, %r313, 4, 31, -1;
1130
+ mov.b32 %f480, %r314;
1131
+ shfl.sync.bfly.b32 %r234, %r223, 4, 31, -1;
1132
+ mov.b32 %f481, %r234;
1133
+ $L__tmp11:
1134
+ .loc 2 108 21
1135
+ sub.f32 %f482, %f479, %f439;
1136
+ .loc 2 109 28
1137
+ add.f32 %f483, %f436, %f481;
1138
+ .loc 2 110 39
1139
+ setp.eq.f32 %p90, %f483, 0f00000000;
1140
+ .loc 2 110 60
1141
+ mov.b32 %r235, %f483;
1142
+ div.full.f32 %r233, %r234, %r235;
1143
+ mov.b32 %f484, %r233;
1144
+ .loc 2 110 49
1145
+ selp.f32 %f485, 0f00000000, %f484, %p90;
1146
+ .loc 2 112 17
1147
+ fma.rn.f32 %f486, %f482, %f485, %f439;
1148
+ .loc 2 113 15
1149
+ add.f32 %f487, %f443, %f480;
1150
+ .loc 2 113 30
1151
+ mul.f32 %f488, %f482, %f482;
1152
+ .loc 2 113 38
1153
+ mul.f32 %f489, %f436, %f488;
1154
+ .loc 2 113 22
1155
+ fma.rn.f32 %f490, %f489, %f485, %f487;
1156
+ $L__tmp12:
1157
+ .loc 2 120 46
1158
+ mov.b32 %r315, %f486;
1159
+ shfl.sync.bfly.b32 %r316, %r315, 2, 31, -1;
1160
+ mov.b32 %f491, %r316;
1161
+ mov.b32 %r317, %f490;
1162
+ shfl.sync.bfly.b32 %r318, %r317, 2, 31, -1;
1163
+ mov.b32 %f492, %r318;
1164
+ shfl.sync.bfly.b32 %r237, %r235, 2, 31, -1;
1165
+ mov.b32 %f493, %r237;
1166
+ $L__tmp13:
1167
+ .loc 2 108 21
1168
+ sub.f32 %f494, %f491, %f486;
1169
+ .loc 2 109 28
1170
+ add.f32 %f495, %f483, %f493;
1171
+ .loc 2 110 39
1172
+ setp.eq.f32 %p91, %f495, 0f00000000;
1173
+ .loc 2 110 60
1174
+ mov.b32 %r238, %f495;
1175
+ div.full.f32 %r236, %r237, %r238;
1176
+ mov.b32 %f496, %r236;
1177
+ .loc 2 110 49
1178
+ selp.f32 %f497, 0f00000000, %f496, %p91;
1179
+ .loc 2 112 17
1180
+ fma.rn.f32 %f498, %f494, %f497, %f486;
1181
+ .loc 2 113 15
1182
+ add.f32 %f499, %f490, %f492;
1183
+ .loc 2 113 30
1184
+ mul.f32 %f500, %f494, %f494;
1185
+ .loc 2 113 38
1186
+ mul.f32 %f501, %f483, %f500;
1187
+ .loc 2 113 22
1188
+ fma.rn.f32 %f502, %f497, %f501, %f499;
1189
+ $L__tmp14:
1190
+ .loc 2 120 46
1191
+ mov.b32 %r319, %f498;
1192
+ shfl.sync.bfly.b32 %r320, %r319, 1, 31, -1;
1193
+ mov.b32 %f503, %r320;
1194
+ mov.b32 %r321, %f502;
1195
+ shfl.sync.bfly.b32 %r322, %r321, 1, 31, -1;
1196
+ mov.b32 %f504, %r322;
1197
+ shfl.sync.bfly.b32 %r240, %r238, 1, 31, -1;
1198
+ mov.b32 %f505, %r240;
1199
+ $L__tmp15:
1200
+ .loc 2 108 21
1201
+ sub.f32 %f506, %f503, %f498;
1202
+ .loc 2 109 28
1203
+ add.f32 %f507, %f495, %f505;
1204
+ .loc 2 110 39
1205
+ setp.eq.f32 %p92, %f507, 0f00000000;
1206
+ .loc 2 110 60
1207
+ mov.b32 %r241, %f507;
1208
+ div.full.f32 %r239, %r240, %r241;
1209
+ mov.b32 %f508, %r239;
1210
+ .loc 2 110 49
1211
+ selp.f32 %f509, 0f00000000, %f508, %p92;
1212
+ .loc 2 112 17
1213
+ fma.rn.f32 %f162, %f506, %f509, %f498;
1214
+ .loc 2 113 15
1215
+ add.f32 %f510, %f502, %f504;
1216
+ .loc 2 113 30
1217
+ mul.f32 %f511, %f506, %f506;
1218
+ .loc 2 113 38
1219
+ mul.f32 %f512, %f495, %f511;
1220
+ .loc 2 113 22
1221
+ fma.rn.f32 %f513, %f509, %f512, %f510;
1222
+ $L__tmp16:
1223
+ .loc 1 75 24
1224
+ mov.b32 %r243, %f478;
1225
+ mov.b32 %r244, 1132462080;
1226
+ div.full.f32 %r242, %r243, %r244;
1227
+ mov.b32 %f514, %r242;
1228
+ mov.b32 %r267, %f513;
1229
+ div.full.f32 %r266, %r267, %r244;
1230
+ mov.b32 %f515, %r266;
1231
+ .loc 1 77 24
1232
+ add.f32 %f163, %f514, 0f3727C5AC;
1233
+ add.f32 %f164, %f515, 0f3727C5AC;
1234
+ .loc 1 58 36
1235
+ add.s64 %rd9, %rd15, %rd2;
1236
+ mov.u64 %rd117, 0;
1237
+ mov.b32 %r473, -64;
1238
+ rsqrt.approx.ftz.f32 %f580, %f163;
1239
+ rsqrt.approx.ftz.f32 %f581, %f164;
1240
+ bra.uni $L__BB0_5;
1241
+ $L__BB0_7:
1242
+ .loc 1 69 35
1243
+ add.s64 %rd107, %rd4, %rd117;
1244
+ add.s64 %rd108, %rd107, 16;
1245
+ add.s64 %rd109, %rd3, %rd117;
1246
+ .loc 1 69 54
1247
+ add.s64 %rd110, %rd109, 16;
1248
+ mov.u32 %r407, 0x0;
1249
+ mov.u32 %r408, 0x0;
1250
+ mov.u32 %r409, 0x0;
1251
+ mov.u32 %r410, 0x0;
1252
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r407, %r408, %r409, %r410 }, [ %rd107 + 0 ];
1253
+ @!%p1 mov.u32 %r407, %r411;
1254
+ @!%p1 mov.u32 %r408, %r411;
1255
+ @!%p1 mov.u32 %r409, %r411;
1256
+ @!%p1 mov.u32 %r410, %r411;
1257
+ mov.b32 %f516, %r407;
1258
+ mov.b32 %f517, %r408;
1259
+ mov.b32 %f518, %r409;
1260
+ mov.b32 %f519, %r410;
1261
+ mov.u32 %r415, 0x0;
1262
+ mov.u32 %r416, 0x0;
1263
+ mov.u32 %r417, 0x0;
1264
+ mov.u32 %r418, 0x0;
1265
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r415, %r416, %r417, %r418 }, [ %rd108 + 0 ];
1266
+ @!%p1 mov.u32 %r415, %r411;
1267
+ @!%p1 mov.u32 %r416, %r411;
1268
+ @!%p1 mov.u32 %r417, %r411;
1269
+ @!%p1 mov.u32 %r418, %r411;
1270
+ mov.b32 %f520, %r415;
1271
+ mov.b32 %f521, %r416;
1272
+ mov.b32 %f522, %r417;
1273
+ mov.b32 %f523, %r418;
1274
+ mov.u32 %r423, 0x0;
1275
+ mov.u32 %r424, 0x0;
1276
+ mov.u32 %r425, 0x0;
1277
+ mov.u32 %r426, 0x0;
1278
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r423, %r424, %r425, %r426 }, [ %rd109 + 0 ];
1279
+ @!%p1 mov.u32 %r423, %r411;
1280
+ @!%p1 mov.u32 %r424, %r411;
1281
+ @!%p1 mov.u32 %r425, %r411;
1282
+ @!%p1 mov.u32 %r426, %r411;
1283
+ mov.b32 %f524, %r423;
1284
+ mov.b32 %f525, %r424;
1285
+ mov.b32 %f526, %r425;
1286
+ mov.b32 %f527, %r426;
1287
+ mov.u32 %r431, 0x0;
1288
+ mov.u32 %r432, 0x0;
1289
+ mov.u32 %r433, 0x0;
1290
+ mov.u32 %r434, 0x0;
1291
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r431, %r432, %r433, %r434 }, [ %rd110 + 0 ];
1292
+ @!%p1 mov.u32 %r431, %r411;
1293
+ @!%p1 mov.u32 %r432, %r411;
1294
+ @!%p1 mov.u32 %r433, %r411;
1295
+ @!%p1 mov.u32 %r434, %r411;
1296
+ mov.b32 %f528, %r431;
1297
+ mov.b32 %f529, %r432;
1298
+ mov.b32 %f530, %r433;
1299
+ mov.b32 %f531, %r434;
1300
+ .loc 1 70 24
1301
+ add.f32 %f532, %f165, %f516;
1302
+ add.f32 %f533, %f166, %f517;
1303
+ add.f32 %f534, %f167, %f518;
1304
+ add.f32 %f535, %f168, %f519;
1305
+ add.f32 %f536, %f169, %f520;
1306
+ add.f32 %f537, %f170, %f521;
1307
+ add.f32 %f538, %f171, %f522;
1308
+ add.f32 %f539, %f172, %f523;
1309
+ add.f32 %f540, %f173, %f524;
1310
+ add.f32 %f541, %f174, %f525;
1311
+ add.f32 %f542, %f175, %f526;
1312
+ add.f32 %f543, %f176, %f527;
1313
+ add.f32 %f544, %f177, %f528;
1314
+ add.f32 %f545, %f178, %f529;
1315
+ add.f32 %f546, %f179, %f530;
1316
+ add.f32 %f547, %f180, %f531;
1317
+ .loc 1 72 24
1318
+ add.f32 %f548, %f181, %f532;
1319
+ add.f32 %f549, %f182, %f533;
1320
+ add.f32 %f550, %f183, %f534;
1321
+ add.f32 %f551, %f184, %f535;
1322
+ add.f32 %f552, %f185, %f536;
1323
+ add.f32 %f553, %f186, %f537;
1324
+ add.f32 %f554, %f187, %f538;
1325
+ add.f32 %f555, %f188, %f539;
1326
+ add.f32 %f556, %f189, %f540;
1327
+ add.f32 %f557, %f190, %f541;
1328
+ add.f32 %f558, %f191, %f542;
1329
+ add.f32 %f559, %f192, %f543;
1330
+ add.f32 %f560, %f193, %f544;
1331
+ add.f32 %f561, %f194, %f545;
1332
+ add.f32 %f562, %f195, %f546;
1333
+ add.f32 %f563, %f196, %f547;
1334
+ .loc 1 73 24
1335
+ sub.f32 %f564, %f548, %f161;
1336
+ sub.f32 %f565, %f549, %f161;
1337
+ sub.f32 %f566, %f550, %f161;
1338
+ sub.f32 %f567, %f551, %f161;
1339
+ sub.f32 %f568, %f552, %f161;
1340
+ sub.f32 %f569, %f553, %f161;
1341
+ sub.f32 %f570, %f554, %f161;
1342
+ sub.f32 %f571, %f555, %f161;
1343
+ sub.f32 %f572, %f556, %f162;
1344
+ sub.f32 %f573, %f557, %f162;
1345
+ sub.f32 %f574, %f558, %f162;
1346
+ sub.f32 %f575, %f559, %f162;
1347
+ sub.f32 %f576, %f560, %f162;
1348
+ sub.f32 %f577, %f561, %f162;
1349
+ sub.f32 %f578, %f562, %f162;
1350
+ sub.f32 %f579, %f563, %f162;
1351
+ .loc 1 79 24
1352
+ mul.f32 %f582, %f564, %f580;
1353
+ mul.f32 %f583, %f565, %f580;
1354
+ mul.f32 %f584, %f566, %f580;
1355
+ mul.f32 %f585, %f567, %f580;
1356
+ mul.f32 %f586, %f568, %f580;
1357
+ mul.f32 %f587, %f569, %f580;
1358
+ mul.f32 %f588, %f570, %f580;
1359
+ mul.f32 %f589, %f571, %f580;
1360
+ mul.f32 %f590, %f572, %f581;
1361
+ mul.f32 %f591, %f573, %f581;
1362
+ mul.f32 %f592, %f574, %f581;
1363
+ mul.f32 %f593, %f575, %f581;
1364
+ mul.f32 %f594, %f576, %f581;
1365
+ mul.f32 %f595, %f577, %f581;
1366
+ mul.f32 %f596, %f578, %f581;
1367
+ mul.f32 %f597, %f579, %f581;
1368
+ .loc 1 80 24
1369
+ mul.f32 %f598, %f582, %f197;
1370
+ mul.f32 %f599, %f583, %f198;
1371
+ mul.f32 %f600, %f584, %f199;
1372
+ mul.f32 %f601, %f585, %f200;
1373
+ mul.f32 %f602, %f586, %f201;
1374
+ mul.f32 %f603, %f587, %f202;
1375
+ mul.f32 %f604, %f588, %f203;
1376
+ mul.f32 %f605, %f589, %f204;
1377
+ mul.f32 %f606, %f590, %f197;
1378
+ mul.f32 %f607, %f591, %f198;
1379
+ mul.f32 %f608, %f592, %f199;
1380
+ mul.f32 %f609, %f593, %f200;
1381
+ mul.f32 %f610, %f594, %f201;
1382
+ mul.f32 %f611, %f595, %f202;
1383
+ mul.f32 %f612, %f596, %f203;
1384
+ mul.f32 %f613, %f597, %f204;
1385
+ .loc 1 82 29
1386
+ shl.b64 %rd113, %rd11, 1;
1387
+ add.s64 %rd111, %rd16, %rd113;
1388
+ shl.b64 %rd114, %rd12, 1;
1389
+ add.s64 %rd112, %rd16, %rd114;
1390
+ .loc 1 82 52
1391
+ mov.b32 %r439, %f598;
1392
+ cvt.rn.bf16.f32 %rs33, %r439;
1393
+ mov.b32 %r440, %f599;
1394
+ cvt.rn.bf16.f32 %rs34, %r440;
1395
+ mov.b32 %r441, %f600;
1396
+ cvt.rn.bf16.f32 %rs35, %r441;
1397
+ mov.b32 %r442, %f601;
1398
+ cvt.rn.bf16.f32 %rs36, %r442;
1399
+ mov.b32 %r443, %f602;
1400
+ cvt.rn.bf16.f32 %rs37, %r443;
1401
+ mov.b32 %r444, %f603;
1402
+ cvt.rn.bf16.f32 %rs38, %r444;
1403
+ mov.b32 %r445, %f604;
1404
+ cvt.rn.bf16.f32 %rs39, %r445;
1405
+ mov.b32 %r446, %f605;
1406
+ cvt.rn.bf16.f32 %rs40, %r446;
1407
+ mov.b32 %r447, %f606;
1408
+ cvt.rn.bf16.f32 %rs41, %r447;
1409
+ mov.b32 %r448, %f607;
1410
+ cvt.rn.bf16.f32 %rs42, %r448;
1411
+ mov.b32 %r449, %f608;
1412
+ cvt.rn.bf16.f32 %rs43, %r449;
1413
+ mov.b32 %r450, %f609;
1414
+ cvt.rn.bf16.f32 %rs44, %r450;
1415
+ mov.b32 %r451, %f610;
1416
+ cvt.rn.bf16.f32 %rs45, %r451;
1417
+ mov.b32 %r452, %f611;
1418
+ cvt.rn.bf16.f32 %rs46, %r452;
1419
+ mov.b32 %r453, %f612;
1420
+ cvt.rn.bf16.f32 %rs47, %r453;
1421
+ mov.b32 %r454, %f613;
1422
+ cvt.rn.bf16.f32 %rs48, %r454;
1423
+ mov.b32 %r463, {%rs33, %rs34};
1424
+ mov.b32 %r464, {%rs35, %rs36};
1425
+ mov.b32 %r465, {%rs37, %rs38};
1426
+ mov.b32 %r466, {%rs39, %rs40};
1427
+ @%p1 st.global.v4.b32 [ %rd111 + 0 ], { %r463, %r464, %r465, %r466 };
1428
+ mov.b32 %r467, {%rs41, %rs42};
1429
+ mov.b32 %r468, {%rs43, %rs44};
1430
+ mov.b32 %r469, {%rs45, %rs46};
1431
+ mov.b32 %r470, {%rs47, %rs48};
1432
+ @%p1 st.global.v4.b32 [ %rd112 + 0 ], { %r467, %r468, %r469, %r470 };
1433
+ .loc 1 58 36
1434
+ add.s64 %rd117, %rd117, 256;
1435
+ add.s32 %r473, %r473, 64;
1436
+ setp.lt.u32 %p156, %r473, 192;
1437
+ @%p156 bra $L__BB0_5;
1438
+ bra.uni $L__BB0_8;
1439
+ $L__BB0_5:
1440
+ .loc 1 62 35
1441
+ add.s64 %rd90, %rd6, %rd117;
1442
+ add.s64 %rd91, %rd90, 16;
1443
+ add.s64 %rd92, %rd5, %rd117;
1444
+ .loc 1 62 51
1445
+ add.s64 %rd93, %rd92, 16;
1446
+ mov.u32 %r323, 0x0;
1447
+ mov.u32 %r324, 0x0;
1448
+ mov.u32 %r325, 0x0;
1449
+ mov.u32 %r326, 0x0;
1450
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r323, %r324, %r325, %r326 }, [ %rd90 + 0 ];
1451
+ @!%p1 mov.u32 %r323, %r411;
1452
+ @!%p1 mov.u32 %r324, %r411;
1453
+ @!%p1 mov.u32 %r325, %r411;
1454
+ @!%p1 mov.u32 %r326, %r411;
1455
+ mov.b32 %f165, %r323;
1456
+ mov.b32 %f166, %r324;
1457
+ mov.b32 %f167, %r325;
1458
+ mov.b32 %f168, %r326;
1459
+ mov.u32 %r331, 0x0;
1460
+ mov.u32 %r332, 0x0;
1461
+ mov.u32 %r333, 0x0;
1462
+ mov.u32 %r334, 0x0;
1463
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r331, %r332, %r333, %r334 }, [ %rd91 + 0 ];
1464
+ @!%p1 mov.u32 %r331, %r411;
1465
+ @!%p1 mov.u32 %r332, %r411;
1466
+ @!%p1 mov.u32 %r333, %r411;
1467
+ @!%p1 mov.u32 %r334, %r411;
1468
+ mov.b32 %f169, %r331;
1469
+ mov.b32 %f170, %r332;
1470
+ mov.b32 %f171, %r333;
1471
+ mov.b32 %f172, %r334;
1472
+ mov.u32 %r339, 0x0;
1473
+ mov.u32 %r340, 0x0;
1474
+ mov.u32 %r341, 0x0;
1475
+ mov.u32 %r342, 0x0;
1476
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r339, %r340, %r341, %r342 }, [ %rd92 + 0 ];
1477
+ @!%p1 mov.u32 %r339, %r411;
1478
+ @!%p1 mov.u32 %r340, %r411;
1479
+ @!%p1 mov.u32 %r341, %r411;
1480
+ @!%p1 mov.u32 %r342, %r411;
1481
+ mov.b32 %f173, %r339;
1482
+ mov.b32 %f174, %r340;
1483
+ mov.b32 %f175, %r341;
1484
+ mov.b32 %f176, %r342;
1485
+ mov.u32 %r347, 0x0;
1486
+ mov.u32 %r348, 0x0;
1487
+ mov.u32 %r349, 0x0;
1488
+ mov.u32 %r350, 0x0;
1489
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r347, %r348, %r349, %r350 }, [ %rd93 + 0 ];
1490
+ @!%p1 mov.u32 %r347, %r411;
1491
+ @!%p1 mov.u32 %r348, %r411;
1492
+ @!%p1 mov.u32 %r349, %r411;
1493
+ @!%p1 mov.u32 %r350, %r411;
1494
+ mov.b32 %f177, %r347;
1495
+ mov.b32 %f178, %r348;
1496
+ mov.b32 %f179, %r349;
1497
+ mov.b32 %f180, %r350;
1498
+ .loc 1 63 41
1499
+ add.s32 %r403, %r5, %r473;
1500
+ add.s32 %r404, %r403, 64;
1501
+ .loc 1 63 35
1502
+ add.s32 %r405, %r403, 8256;
1503
+ cvt.s64.s32 %rd11, %r404;
1504
+ mul.wide.s32 %rd98, %r404, 2;
1505
+ add.s64 %rd94, %rd14, %rd98;
1506
+ cvt.s64.s32 %rd12, %r405;
1507
+ mul.wide.s32 %rd99, %r405, 2;
1508
+ add.s64 %rd95, %rd14, %rd99;
1509
+ .loc 1 63 51
1510
+ mov.u32 %r355, 0x0;
1511
+ mov.u32 %r356, 0x0;
1512
+ mov.u32 %r357, 0x0;
1513
+ mov.u32 %r358, 0x0;
1514
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r355, %r356, %r357, %r358 }, [ %rd94 + 0 ];
1515
+ @!%p1 mov.u32 %r355, %r411;
1516
+ @!%p1 mov.u32 %r356, %r411;
1517
+ @!%p1 mov.u32 %r357, %r411;
1518
+ @!%p1 mov.u32 %r358, %r411;
1519
+ cvt.u16.u32 %rs17, %r355;
1520
+ { .reg .b16 tmp; mov.b32 {tmp, %rs18}, %r355; }
1521
+ cvt.u16.u32 %rs19, %r356;
1522
+ { .reg .b16 tmp; mov.b32 {tmp, %rs20}, %r356; }
1523
+ cvt.u16.u32 %rs21, %r357;
1524
+ { .reg .b16 tmp; mov.b32 {tmp, %rs22}, %r357; }
1525
+ cvt.u16.u32 %rs23, %r358;
1526
+ { .reg .b16 tmp; mov.b32 {tmp, %rs24}, %r358; }
1527
+ mov.u32 %r363, 0x0;
1528
+ mov.u32 %r364, 0x0;
1529
+ mov.u32 %r365, 0x0;
1530
+ mov.u32 %r366, 0x0;
1531
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r363, %r364, %r365, %r366 }, [ %rd95 + 0 ];
1532
+ @!%p1 mov.u32 %r363, %r411;
1533
+ @!%p1 mov.u32 %r364, %r411;
1534
+ @!%p1 mov.u32 %r365, %r411;
1535
+ @!%p1 mov.u32 %r366, %r411;
1536
+ cvt.u16.u32 %rs25, %r363;
1537
+ { .reg .b16 tmp; mov.b32 {tmp, %rs26}, %r363; }
1538
+ cvt.u16.u32 %rs27, %r364;
1539
+ { .reg .b16 tmp; mov.b32 {tmp, %rs28}, %r364; }
1540
+ cvt.u16.u32 %rs29, %r365;
1541
+ { .reg .b16 tmp; mov.b32 {tmp, %rs30}, %r365; }
1542
+ cvt.u16.u32 %rs31, %r366;
1543
+ { .reg .b16 tmp; mov.b32 {tmp, %rs32}, %r366; }
1544
+ .loc 1 63 103
1545
+ cvt.f32.bf16 %r371, %rs17;
1546
+ mov.b32 %f181, %r371;
1547
+ cvt.f32.bf16 %r372, %rs18;
1548
+ mov.b32 %f182, %r372;
1549
+ cvt.f32.bf16 %r373, %rs19;
1550
+ mov.b32 %f183, %r373;
1551
+ cvt.f32.bf16 %r374, %rs20;
1552
+ mov.b32 %f184, %r374;
1553
+ cvt.f32.bf16 %r375, %rs21;
1554
+ mov.b32 %f185, %r375;
1555
+ cvt.f32.bf16 %r376, %rs22;
1556
+ mov.b32 %f186, %r376;
1557
+ cvt.f32.bf16 %r377, %rs23;
1558
+ mov.b32 %f187, %r377;
1559
+ cvt.f32.bf16 %r378, %rs24;
1560
+ mov.b32 %f188, %r378;
1561
+ cvt.f32.bf16 %r379, %rs25;
1562
+ mov.b32 %f189, %r379;
1563
+ cvt.f32.bf16 %r380, %rs26;
1564
+ mov.b32 %f190, %r380;
1565
+ cvt.f32.bf16 %r381, %rs27;
1566
+ mov.b32 %f191, %r381;
1567
+ cvt.f32.bf16 %r382, %rs28;
1568
+ mov.b32 %f192, %r382;
1569
+ cvt.f32.bf16 %r383, %rs29;
1570
+ mov.b32 %f193, %r383;
1571
+ cvt.f32.bf16 %r384, %rs30;
1572
+ mov.b32 %f194, %r384;
1573
+ cvt.f32.bf16 %r385, %rs31;
1574
+ mov.b32 %f195, %r385;
1575
+ cvt.f32.bf16 %r386, %rs32;
1576
+ mov.b32 %f196, %r386;
1577
+ .loc 1 64 35
1578
+ add.s64 %rd96, %rd9, %rd117;
1579
+ .loc 1 64 40
1580
+ add.s64 %rd97, %rd96, 16;
1581
+ mov.u32 %r387, 0x0;
1582
+ mov.u32 %r388, 0x0;
1583
+ mov.u32 %r389, 0x0;
1584
+ mov.u32 %r390, 0x0;
1585
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r387, %r388, %r389, %r390 }, [ %rd96 + 0 ];
1586
+ @!%p1 mov.u32 %r387, %r411;
1587
+ @!%p1 mov.u32 %r388, %r411;
1588
+ @!%p1 mov.u32 %r389, %r411;
1589
+ @!%p1 mov.u32 %r390, %r411;
1590
+ mov.b32 %f197, %r387;
1591
+ mov.b32 %f198, %r388;
1592
+ mov.b32 %f199, %r389;
1593
+ mov.b32 %f200, %r390;
1594
+ mov.u32 %r395, 0x0;
1595
+ mov.u32 %r396, 0x0;
1596
+ mov.u32 %r397, 0x0;
1597
+ mov.u32 %r398, 0x0;
1598
+ @%p1 ld.global.L1::evict_last.v4.b32 { %r395, %r396, %r397, %r398 }, [ %rd97 + 0 ];
1599
+ @!%p1 mov.u32 %r395, %r411;
1600
+ @!%p1 mov.u32 %r396, %r411;
1601
+ @!%p1 mov.u32 %r397, %r411;
1602
+ @!%p1 mov.u32 %r398, %r411;
1603
+ mov.b32 %f201, %r395;
1604
+ mov.b32 %f202, %r396;
1605
+ mov.b32 %f203, %r397;
1606
+ mov.b32 %f204, %r398;
1607
+ .loc 1 68 57
1608
+ @%p51 bra $L__BB0_7;
1609
+ mov.u64 %rd100, assertMessage_1;
1610
+ cvta.global.u64 %rd101, %rd100;
1611
+ mov.u64 %rd102, assertFile_1;
1612
+ cvta.global.u64 %rd103, %rd102;
1613
+ mov.u64 %rd104, assertFunc_1;
1614
+ cvta.global.u64 %rd105, %rd104;
1615
+ { // callseq 7, 0
1616
+ .reg .b32 temp_param_reg;
1617
+ .param .b64 param0;
1618
+ st.param.b64 [param0+0], %rd101;
1619
+ .param .b64 param1;
1620
+ st.param.b64 [param1+0], %rd103;
1621
+ .param .b32 param2;
1622
+ st.param.b32 [param2+0], %r471;
1623
+ .param .b64 param3;
1624
+ st.param.b64 [param3+0], %rd105;
1625
+ .param .b64 param4;
1626
+ st.param.b64 [param4+0], %rd115;
1627
+ call.uni
1628
+ __assertfail,
1629
+ (
1630
+ param0,
1631
+ param1,
1632
+ param2,
1633
+ param3,
1634
+ param4
1635
+ );
1636
+ } // callseq 7
1637
+ bra.uni $L__BB0_7;
1638
+ $L__BB0_8:
1639
+ .loc 1 58 4
1640
+ ret;
1641
+ $L__tmp17:
1642
+ $L__func_end0:
1643
+
1644
+ }
1645
+ // .globl __nv_rsqrtf
1646
+ .visible .func (.param .b32 func_retval0) __nv_rsqrtf(
1647
+ .param .b32 __nv_rsqrtf_param_0
1648
+ )
1649
+ {
1650
+ .reg .f32 %f<3>;
1651
+ $L__func_begin1:
1652
+
1653
+ ld.param.f32 %f1, [__nv_rsqrtf_param_0];
1654
+ rsqrt.approx.ftz.f32 %f2, %f1;
1655
+ st.param.f32 [func_retval0+0], %f2;
1656
+ ret;
1657
+ $L__func_end1:
1658
+
1659
+ }
1660
+ .file 1 "/tmp/torchinductor_root/ci/ccig6fki6p4lxrdmgg6eudahiexcvueeol2p4qp532pvve2y463y.py"
1661
+ .file 2 "/usr/local/lib/python3.10/dist-packages/torch/_inductor/triton_helpers.py"
1662
+ .section .debug_abbrev
1663
+ {
1664
+ .b8 1
1665
+ .b8 17
1666
+ .b8 1
1667
+ .b8 37
1668
+ .b8 8
1669
+ .b8 19
1670
+ .b8 5
1671
+ .b8 3
1672
+ .b8 8
1673
+ .b8 16
1674
+ .b8 6
1675
+ .b8 27
1676
+ .b8 8
1677
+ .b8 180
1678
+ .b8 66
1679
+ .b8 12
1680
+ .b8 17
1681
+ .b8 1
1682
+ .b8 18
1683
+ .b8 1
1684
+ .b8 0
1685
+ .b8 0
1686
+ .b8 2
1687
+ .b8 46
1688
+ .b8 0
1689
+ .b8 135
1690
+ .b8 64
1691
+ .b8 8
1692
+ .b8 3
1693
+ .b8 8
1694
+ .b8 58
1695
+ .b8 11
1696
+ .b8 59
1697
+ .b8 11
1698
+ .b8 63
1699
+ .b8 12
1700
+ .b8 32
1701
+ .b8 11
1702
+ .b8 0
1703
+ .b8 0
1704
+ .b8 3
1705
+ .b8 46
1706
+ .b8 1
1707
+ .b8 17
1708
+ .b8 1
1709
+ .b8 18
1710
+ .b8 1
1711
+ .b8 64
1712
+ .b8 10
1713
+ .b8 49
1714
+ .b8 19
1715
+ .b8 0
1716
+ .b8 0
1717
+ .b8 4
1718
+ .b8 29
1719
+ .b8 0
1720
+ .b8 49
1721
+ .b8 19
1722
+ .b8 17
1723
+ .b8 1
1724
+ .b8 18
1725
+ .b8 1
1726
+ .b8 88
1727
+ .b8 11
1728
+ .b8 89
1729
+ .b8 11
1730
+ .b8 87
1731
+ .b8 11
1732
+ .b8 0
1733
+ .b8 0
1734
+ .b8 5
1735
+ .b8 29
1736
+ .b8 1
1737
+ .b8 49
1738
+ .b8 19
1739
+ .b8 17
1740
+ .b8 1
1741
+ .b8 18
1742
+ .b8 1
1743
+ .b8 88
1744
+ .b8 11
1745
+ .b8 89
1746
+ .b8 11
1747
+ .b8 87
1748
+ .b8 11
1749
+ .b8 0
1750
+ .b8 0
1751
+ .b8 0
1752
+ }
1753
+ .section .debug_info
1754
+ {
1755
+ .b32 302
1756
+ .b8 2
1757
+ .b8 0
1758
+ .b32 .debug_abbrev
1759
+ .b8 8
1760
+ .b8 1
1761
+ .b8 116
1762
+ .b8 114
1763
+ .b8 105
1764
+ .b8 116
1765
+ .b8 111
1766
+ .b8 110
1767
+ .b8 0
1768
+ .b8 2
1769
+ .b8 0
1770
+ .b8 99
1771
+ .b8 99
1772
+ .b8 105
1773
+ .b8 103
1774
+ .b8 54
1775
+ .b8 102
1776
+ .b8 107
1777
+ .b8 105
1778
+ .b8 54
1779
+ .b8 112
1780
+ .b8 52
1781
+ .b8 108
1782
+ .b8 120
1783
+ .b8 114
1784
+ .b8 100
1785
+ .b8 109
1786
+ .b8 103
1787
+ .b8 103
1788
+ .b8 54
1789
+ .b8 101
1790
+ .b8 117
1791
+ .b8 100
1792
+ .b8 97
1793
+ .b8 104
1794
+ .b8 105
1795
+ .b8 101
1796
+ .b8 120
1797
+ .b8 99
1798
+ .b8 118
1799
+ .b8 117
1800
+ .b8 101
1801
+ .b8 101
1802
+ .b8 111
1803
+ .b8 108
1804
+ .b8 50
1805
+ .b8 112
1806
+ .b8 52
1807
+ .b8 113
1808
+ .b8 112
1809
+ .b8 53
1810
+ .b8 51
1811
+ .b8 50
1812
+ .b8 112
1813
+ .b8 118
1814
+ .b8 118
1815
+ .b8 101
1816
+ .b8 50
1817
+ .b8 121
1818
+ .b8 52
1819
+ .b8 54
1820
+ .b8 51
1821
+ .b8 121
1822
+ .b8 46
1823
+ .b8 112
1824
+ .b8 121
1825
+ .b8 0
1826
+ .b32 .debug_line
1827
+ .b8 47
1828
+ .b8 116
1829
+ .b8 109
1830
+ .b8 112
1831
+ .b8 47
1832
+ .b8 116
1833
+ .b8 111
1834
+ .b8 114
1835
+ .b8 99
1836
+ .b8 104
1837
+ .b8 105
1838
+ .b8 110
1839
+ .b8 100
1840
+ .b8 117
1841
+ .b8 99
1842
+ .b8 116
1843
+ .b8 111
1844
+ .b8 114
1845
+ .b8 95
1846
+ .b8 114
1847
+ .b8 111
1848
+ .b8 111
1849
+ .b8 116
1850
+ .b8 47
1851
+ .b8 99
1852
+ .b8 105
1853
+ .b8 0
1854
+ .b8 1
1855
+ .b64 $L__func_begin0
1856
+ .b64 $L__func_end0
1857
+ .b8 2
1858
+ .b8 116
1859
+ .b8 114
1860
+ .b8 105
1861
+ .b8 116
1862
+ .b8 111
1863
+ .b8 110
1864
+ .b8 95
1865
+ .b8 95
1866
+ .b8 48
1867
+ .b8 100
1868
+ .b8 49
1869
+ .b8 100
1870
+ .b8 50
1871
+ .b8 100
1872
+ .b8 51
1873
+ .b8 100
1874
+ .b8 52
1875
+ .b8 100
1876
+ .b8 53
1877
+ .b8 100
1878
+ .b8 54
1879
+ .b8 100
1880
+ .b8 101
1881
+ .b8 55
1882
+ .b8 100
1883
+ .b8 101
1884
+ .b8 0
1885
+ .b8 116
1886
+ .b8 114
1887
+ .b8 105
1888
+ .b8 116
1889
+ .b8 111
1890
+ .b8 110
1891
+ .b8 95
1892
+ .b8 95
1893
+ .b8 48
1894
+ .b8 100
1895
+ .b8 49
1896
+ .b8 100
1897
+ .b8 50
1898
+ .b8 100
1899
+ .b8 51
1900
+ .b8 100
1901
+ .b8 52
1902
+ .b8 100
1903
+ .b8 53
1904
+ .b8 100
1905
+ .b8 54
1906
+ .b8 100
1907
+ .b8 101
1908
+ .b8 55
1909
+ .b8 100
1910
+ .b8 101
1911
+ .b8 0
1912
+ .b8 1
1913
+ .b8 18
1914
+ .b8 1
1915
+ .b8 1
1916
+ .b8 3
1917
+ .b64 $L__func_begin0
1918
+ .b64 $L__func_end0
1919
+ .b8 1
1920
+ .b8 156
1921
+ .b32 125
1922
+ .b8 4
1923
+ .b32 125
1924
+ .b64 $L__tmp1
1925
+ .b64 $L__tmp2
1926
+ .b8 2
1927
+ .b8 47
1928
+ .b8 41
1929
+ .b8 5
1930
+ .b32 125
1931
+ .b64 $L__tmp3
1932
+ .b64 $L__tmp16
1933
+ .b8 2
1934
+ .b8 53
1935
+ .b8 44
1936
+ .b8 4
1937
+ .b32 125
1938
+ .b64 $L__tmp3
1939
+ .b64 $L__tmp16
1940
+ .b8 2
1941
+ .b8 120
1942
+ .b8 46
1943
+ .b8 0
1944
+ .b8 4
1945
+ .b32 125
1946
+ .b64 $L__tmp4
1947
+ .b64 $L__tmp15
1948
+ .b8 2
1949
+ .b8 53
1950
+ .b8 44
1951
+ .b8 0
1952
+ .b8 0
1953
+ }
1954
+ .section .debug_pubnames
1955
+ {
1956
+ .b32 $L__pubNames_end0-$L__pubNames_start0
1957
+ $L__pubNames_start0:
1958
+ .b8 2
1959
+ .b8 0
1960
+ .b32 .debug_info
1961
+ .b32 306
1962
+ .b32 125
1963
+ .b8 116
1964
+ .b8 114
1965
+ .b8 105
1966
+ .b8 116
1967
+ .b8 111
1968
+ .b8 110
1969
+ .b8 95
1970
+ .b8 95
1971
+ .b8 48
1972
+ .b8 100
1973
+ .b8 49
1974
+ .b8 100
1975
+ .b8 50
1976
+ .b8 100
1977
+ .b8 51
1978
+ .b8 100
1979
+ .b8 52
1980
+ .b8 100
1981
+ .b8 53
1982
+ .b8 100
1983
+ .b8 54
1984
+ .b8 100
1985
+ .b8 101
1986
+ .b8 55
1987
+ .b8 100
1988
+ .b8 101
1989
+ .b8 0
1990
+ .b32 0
1991
+ $L__pubNames_end0:
1992
+ }
1993
+ .section .debug_pubtypes
1994
+ {
1995
+ .b32 $L__pubTypes_end0-$L__pubTypes_start0
1996
+ $L__pubTypes_start0:
1997
+ .b8 2
1998
+ .b8 0
1999
+ .b32 .debug_info
2000
+ .b32 306
2001
+ .b32 0
2002
+ $L__pubTypes_end0:
2003
+ }
2004
+ .section .debug_loc { }
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttgir ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [1, 8], threadsPerWarp = [4, 8], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
2
+ #blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [8, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
3
+ #blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [4, 2], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
4
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 8 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
5
+ tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
6
+ %cst = arith.constant dense<512> : tensor<64x1xi32, #blocked>
7
+ %cst_0 = arith.constant dense<256> : tensor<1x64xi32, #blocked>
8
+ %cst_1 = arith.constant dense<256> : tensor<64x1xi32, #blocked>
9
+ %cst_2 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked>
10
+ %cst_3 = arith.constant dense<0.000000e+00> : tensor<1x64xf32, #blocked>
11
+ %cst_4 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked>
12
+ %cst_5 = arith.constant dense<256> : tensor<64x1xi64, #blocked>
13
+ %cst_6 = arith.constant dense<0> : tensor<64x1xi64, #blocked>
14
+ %cst_7 = arith.constant dense<50257> : tensor<64x1xi64, #blocked>
15
+ %cst_8 = arith.constant dense<50257> : tensor<64x1xi64, #blocked1>
16
+ %cst_9 = arith.constant dense<0> : tensor<64x1xi64, #blocked1>
17
+ %c0_i32 = arith.constant 0 : i32
18
+ %c64_i32 = arith.constant 64 : i32
19
+ %c256_i32 = arith.constant 256 : i32
20
+ %cst_10 = arith.constant dense<1.000000e+00> : tensor<64x64xf32, #blocked2>
21
+ %cst_11 = arith.constant 0.000000e+00 : f32
22
+ %cst_12 = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked2>
23
+ %cst_13 = arith.constant dense<256> : tensor<1x64xi32, #blocked2>
24
+ %cst_14 = arith.constant dense<9.99999974E-6> : tensor<64x1xf32, #blocked>
25
+ %cst_15 = arith.constant dense<2.560000e+02> : tensor<64x1xf32, #blocked>
26
+ %cst_16 = arith.constant dense<0.000000e+00> : tensor<64x64xbf16, #blocked>
27
+ %0 = tt.get_program_id x : i32
28
+ %1 = arith.muli %0, %c64_i32 : i32
29
+ %2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
30
+ %3 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
31
+ %4 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xi32, #blocked>
32
+ %5 = tt.expand_dims %3 {axis = 1 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<64x1xi32, #blocked1>
33
+ %6 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked>
34
+ %7 = tt.splat %1 : (i32) -> tensor<64x1xi32, #blocked1>
35
+ %8 = arith.addi %6, %4 : tensor<64x1xi32, #blocked>
36
+ %9 = arith.addi %7, %5 : tensor<64x1xi32, #blocked1>
37
+ %10 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
38
+ %11 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>
39
+ %12 = tt.expand_dims %10 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x64xi32, #blocked>
40
+ %13 = tt.expand_dims %11 {axis = 0 : i32} : (tensor<64xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>) -> tensor<1x64xi32, #blocked2>
41
+ %14 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked>
42
+ %15 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>, #blocked1>
43
+ %16 = tt.addptr %14, %8 : tensor<64x1x!tt.ptr<i64, 1>, #blocked>, tensor<64x1xi32, #blocked>
44
+ %17 = tt.addptr %15, %9 : tensor<64x1x!tt.ptr<i64, 1>, #blocked1>, tensor<64x1xi32, #blocked1>
45
+ %18 = tt.load %16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked>
46
+ %19 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64, #blocked1>
47
+ %20 = arith.remsi %8, %cst : tensor<64x1xi32, #blocked>
48
+ %21 = arith.muli %20, %cst_1 : tensor<64x1xi32, #blocked>
49
+ %22 = tt.broadcast %21 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
50
+ %23 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
51
+ %24 = arith.muli %8, %cst_1 : tensor<64x1xi32, #blocked>
52
+ %25 = tt.broadcast %24 : (tensor<64x1xi32, #blocked>) -> tensor<64x64xi32, #blocked>
53
+ %26 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>, #blocked>
54
+ %27 = arith.addi %18, %cst_7 : tensor<64x1xi64, #blocked>
55
+ %28 = arith.addi %19, %cst_8 : tensor<64x1xi64, #blocked1>
56
+ %29 = arith.cmpi slt, %18, %cst_6 : tensor<64x1xi64, #blocked>
57
+ %30 = arith.cmpi slt, %19, %cst_9 : tensor<64x1xi64, #blocked1>
58
+ %31 = arith.select %29, %27, %18 : tensor<64x1xi1, #blocked>, tensor<64x1xi64, #blocked>
59
+ %32 = arith.select %30, %28, %19 : tensor<64x1xi1, #blocked1>, tensor<64x1xi64, #blocked1>
60
+ %33 = arith.cmpi sge, %32, %cst_9 : tensor<64x1xi64, #blocked1>
61
+ %34 = arith.cmpi slt, %32, %cst_8 : tensor<64x1xi64, #blocked1>
62
+ %35 = arith.andi %33, %34 : tensor<64x1xi1, #blocked1>
63
+ %36 = arith.muli %31, %cst_5 : tensor<64x1xi64, #blocked>
64
+ %37 = tt.broadcast %36 : (tensor<64x1xi64, #blocked>) -> tensor<64x64xi64, #blocked>
65
+ %38 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>, #blocked>
66
+ %39:4 = scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 iter_args(%arg9 = %cst_2, %arg10 = %cst_2, %arg11 = %cst_12, %arg12 = %cst_2) -> (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>) : i32 {
67
+ %49 = tt.splat %arg8 : (i32) -> tensor<1x64xi32, #blocked>
68
+ %50 = tt.splat %arg8 : (i32) -> tensor<1x64xi32, #blocked2>
69
+ %51 = arith.addi %49, %12 : tensor<1x64xi32, #blocked>
70
+ %52 = arith.addi %50, %13 : tensor<1x64xi32, #blocked2>
71
+ %53 = arith.cmpi slt, %51, %cst_0 : tensor<1x64xi32, #blocked>
72
+ %54 = arith.cmpi slt, %52, %cst_13 : tensor<1x64xi32, #blocked2>
73
+ %55 = tt.broadcast %51 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
74
+ %56 = arith.addi %55, %22 : tensor<64x64xi32, #blocked>
75
+ %57 = tt.addptr %23, %56 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
76
+ %58 = tt.broadcast %53 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
77
+ %59 = tt.broadcast %54 : (tensor<1x64xi1, #blocked2>) -> tensor<64x64xi1, #blocked2>
78
+ %60 = tt.load %57, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
79
+ %61 = arith.addi %55, %25 : tensor<64x64xi32, #blocked>
80
+ %62 = tt.addptr %26, %61 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
81
+ %63 = tt.load %62, %58, %cst_16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xbf16, #blocked>
82
+ %64 = arith.extf %63 : tensor<64x64xbf16, #blocked> to tensor<64x64xf32, #blocked>
83
+ tt.assert %35, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
84
+ %65 = arith.extsi %51 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
85
+ %66 = tt.broadcast %65 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
86
+ %67 = arith.addi %66, %37 : tensor<64x64xi64, #blocked>
87
+ %68 = tt.addptr %38, %67 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
88
+ %69 = tt.load %68, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
89
+ %70 = arith.addf %69, %60 : tensor<64x64xf32, #blocked>
90
+ %71 = arith.addf %70, %64 : tensor<64x64xf32, #blocked>
91
+ %72 = arith.subf %71, %arg9 : tensor<64x64xf32, #blocked>
92
+ %73 = arith.addf %arg12, %cst_4 : tensor<64x64xf32, #blocked>
93
+ %74 = arith.addf %arg11, %cst_10 : tensor<64x64xf32, #blocked2>
94
+ %75 = arith.divf %72, %73 : tensor<64x64xf32, #blocked>
95
+ %76 = arith.addf %arg9, %75 : tensor<64x64xf32, #blocked>
96
+ %77 = arith.subf %71, %76 : tensor<64x64xf32, #blocked>
97
+ %78 = arith.mulf %72, %77 : tensor<64x64xf32, #blocked>
98
+ %79 = arith.addf %arg10, %78 : tensor<64x64xf32, #blocked>
99
+ %80 = arith.select %58, %76, %arg9 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
100
+ %81 = arith.select %58, %79, %arg10 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
101
+ %82 = arith.select %58, %73, %arg12 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked>
102
+ %83 = arith.select %59, %74, %arg11 : tensor<64x64xi1, #blocked2>, tensor<64x64xf32, #blocked2>
103
+ scf.yield %80, %81, %83, %82 : tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked2>, tensor<64x64xf32, #blocked>
104
+ }
105
+ %40 = triton_gpu.convert_layout %39#2 : (tensor<64x64xf32, #blocked2>) -> tensor<64x64xf32, #blocked>
106
+ %41:3 = "tt.reduce"(%39#0, %39#1, %40) <{axis = 1 : i32}> ({
107
+ ^bb0(%arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32, %arg13: f32):
108
+ %49 = arith.subf %arg11, %arg8 : f32
109
+ %50 = arith.addf %arg10, %arg13 : f32
110
+ %51 = arith.cmpf oeq, %50, %cst_11 : f32
111
+ %52 = arith.divf %arg13, %50 : f32
112
+ %53 = arith.select %51, %cst_11, %52 : f32
113
+ %54 = arith.mulf %49, %53 : f32
114
+ %55 = arith.addf %arg8, %54 : f32
115
+ %56 = arith.addf %arg9, %arg12 : f32
116
+ %57 = arith.mulf %49, %49 : f32
117
+ %58 = arith.mulf %57, %arg10 : f32
118
+ %59 = arith.mulf %58, %53 : f32
119
+ %60 = arith.addf %56, %59 : f32
120
+ tt.reduce.return %55, %60, %50 : f32, f32, f32
121
+ }) : (tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>, tensor<64x64xf32, #blocked>) -> (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>)
122
+ %42 = tt.expand_dims %41#0 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
123
+ %43 = tt.expand_dims %41#1 {axis = 1 : i32} : (tensor<64xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<64x1xf32, #blocked>
124
+ %44 = tt.splat %arg4 : (!tt.ptr<f32, 1>) -> tensor<1x64x!tt.ptr<f32, 1>, #blocked>
125
+ %45 = tt.broadcast %42 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
126
+ %46 = arith.divf %43, %cst_15 : tensor<64x1xf32, #blocked>
127
+ %47 = arith.addf %46, %cst_14 : tensor<64x1xf32, #blocked>
128
+ %48 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>, #blocked>
129
+ scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 : i32 {
130
+ %49 = tt.splat %arg8 : (i32) -> tensor<1x64xi32, #blocked>
131
+ %50 = arith.addi %49, %12 : tensor<1x64xi32, #blocked>
132
+ %51 = arith.cmpi slt, %50, %cst_0 : tensor<1x64xi32, #blocked>
133
+ %52 = tt.broadcast %50 : (tensor<1x64xi32, #blocked>) -> tensor<64x64xi32, #blocked>
134
+ %53 = arith.addi %52, %22 : tensor<64x64xi32, #blocked>
135
+ %54 = tt.addptr %23, %53 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi32, #blocked>
136
+ %55 = tt.broadcast %51 : (tensor<1x64xi1, #blocked>) -> tensor<64x64xi1, #blocked>
137
+ %56 = tt.load %54, %55, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
138
+ %57 = arith.addi %52, %25 : tensor<64x64xi32, #blocked>
139
+ %58 = tt.addptr %26, %57 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
140
+ %59 = tt.load %58, %55, %cst_16 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xbf16, #blocked>
141
+ %60 = arith.extf %59 : tensor<64x64xbf16, #blocked> to tensor<64x64xf32, #blocked>
142
+ %61 = tt.addptr %44, %50 : tensor<1x64x!tt.ptr<f32, 1>, #blocked>, tensor<1x64xi32, #blocked>
143
+ %62 = tt.load %61, %51, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x64xf32, #blocked>
144
+ tt.assert %35, "index out of bounds: 0 <= tmp16 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1, #blocked1>
145
+ %63 = arith.extsi %50 : tensor<1x64xi32, #blocked> to tensor<1x64xi64, #blocked>
146
+ %64 = tt.broadcast %63 : (tensor<1x64xi64, #blocked>) -> tensor<64x64xi64, #blocked>
147
+ %65 = arith.addi %64, %37 : tensor<64x64xi64, #blocked>
148
+ %66 = tt.addptr %38, %65 : tensor<64x64x!tt.ptr<f32, 1>, #blocked>, tensor<64x64xi64, #blocked>
149
+ %67 = tt.load %66, %55, %cst_2 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xf32, #blocked>
150
+ %68 = arith.addf %67, %56 : tensor<64x64xf32, #blocked>
151
+ %69 = arith.addf %68, %60 : tensor<64x64xf32, #blocked>
152
+ %70 = arith.subf %69, %45 : tensor<64x64xf32, #blocked>
153
+ %71 = tt.extern_elementwise %47 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<64x1xf32, #blocked>) -> tensor<64x1xf32, #blocked>
154
+ %72 = tt.broadcast %71 : (tensor<64x1xf32, #blocked>) -> tensor<64x64xf32, #blocked>
155
+ %73 = arith.mulf %70, %72 : tensor<64x64xf32, #blocked>
156
+ %74 = tt.broadcast %62 : (tensor<1x64xf32, #blocked>) -> tensor<64x64xf32, #blocked>
157
+ %75 = arith.mulf %73, %74 : tensor<64x64xf32, #blocked>
158
+ %76 = tt.addptr %48, %57 : tensor<64x64x!tt.ptr<bf16, 1>, #blocked>, tensor<64x64xi32, #blocked>
159
+ %77 = arith.truncf %75 : tensor<64x64xf32, #blocked> to tensor<64x64xbf16, #blocked>
160
+ tt.store %76, %77, %55 {cache = 1 : i32, evict = 1 : i32} : tensor<64x64xbf16, #blocked>
161
+ }
162
+ tt.return
163
+ }
164
+ }
.triton/dump/53075505618c3af0ef6ce61f3300cdcb/triton_.ttir ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ module {
2
+ tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
3
+ %cst = arith.constant dense<0.000000e+00> : tensor<64x64xbf16>
4
+ %cst_0 = arith.constant 0.000000e+00 : f32
5
+ %cst_1 = arith.constant dense<1.000000e+00> : tensor<64x64xf32>
6
+ %c256_i32 = arith.constant 256 : i32
7
+ %c64_i32 = arith.constant 64 : i32
8
+ %c0_i32 = arith.constant 0 : i32
9
+ %cst_2 = arith.constant dense<256> : tensor<64x1xi64>
10
+ %cst_3 = arith.constant dense<0> : tensor<64x1xi64>
11
+ %cst_4 = arith.constant dense<50257> : tensor<64x1xi64>
12
+ %cst_5 = arith.constant dense<9.99999974E-6> : tensor<64x1xf32>
13
+ %cst_6 = arith.constant dense<2.560000e+02> : tensor<64x1xf32>
14
+ %cst_7 = arith.constant dense<0.000000e+00> : tensor<1x64xf32>
15
+ %cst_8 = arith.constant dense<0.000000e+00> : tensor<64x64xf32>
16
+ %cst_9 = arith.constant dense<256> : tensor<64x1xi32>
17
+ %cst_10 = arith.constant dense<256> : tensor<1x64xi32>
18
+ %cst_11 = arith.constant dense<512> : tensor<64x1xi32>
19
+ %0 = tt.get_program_id x : i32
20
+ %1 = arith.muli %0, %c64_i32 : i32
21
+ %2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32>
22
+ %3 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<64xi32>) -> tensor<64x1xi32>
23
+ %4 = tt.splat %1 : (i32) -> tensor<64x1xi32>
24
+ %5 = arith.addi %4, %3 : tensor<64x1xi32>
25
+ %6 = tt.expand_dims %2 {axis = 0 : i32} : (tensor<64xi32>) -> tensor<1x64xi32>
26
+ %7 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<64x1x!tt.ptr<i64, 1>>
27
+ %8 = tt.addptr %7, %5 : tensor<64x1x!tt.ptr<i64, 1>>, tensor<64x1xi32>
28
+ %9 = tt.load %8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x1xi64>
29
+ %10 = arith.remsi %5, %cst_11 : tensor<64x1xi32>
30
+ %11 = arith.muli %10, %cst_9 : tensor<64x1xi32>
31
+ %12 = tt.broadcast %11 : (tensor<64x1xi32>) -> tensor<64x64xi32>
32
+ %13 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
33
+ %14 = arith.muli %5, %cst_9 : tensor<64x1xi32>
34
+ %15 = tt.broadcast %14 : (tensor<64x1xi32>) -> tensor<64x64xi32>
35
+ %16 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>>
36
+ %17 = arith.addi %9, %cst_4 : tensor<64x1xi64>
37
+ %18 = arith.cmpi slt, %9, %cst_3 : tensor<64x1xi64>
38
+ %19 = arith.select %18, %17, %9 : tensor<64x1xi1>, tensor<64x1xi64>
39
+ %20 = arith.cmpi sge, %19, %cst_3 : tensor<64x1xi64>
40
+ %21 = arith.cmpi slt, %19, %cst_4 : tensor<64x1xi64>
41
+ %22 = arith.andi %20, %21 : tensor<64x1xi1>
42
+ %23 = arith.muli %19, %cst_2 : tensor<64x1xi64>
43
+ %24 = tt.broadcast %23 : (tensor<64x1xi64>) -> tensor<64x64xi64>
44
+ %25 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
45
+ %26:3 = scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 iter_args(%arg9 = %cst_8, %arg10 = %cst_8, %arg11 = %cst_8) -> (tensor<64x64xf32>, tensor<64x64xf32>, tensor<64x64xf32>) : i32 {
46
+ %50 = tt.splat %arg8 : (i32) -> tensor<1x64xi32>
47
+ %51 = arith.addi %50, %6 : tensor<1x64xi32>
48
+ %52 = arith.cmpi slt, %51, %cst_10 : tensor<1x64xi32>
49
+ %53 = tt.broadcast %51 : (tensor<1x64xi32>) -> tensor<64x64xi32>
50
+ %54 = arith.addi %53, %12 : tensor<64x64xi32>
51
+ %55 = tt.addptr %13, %54 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi32>
52
+ %56 = tt.broadcast %52 : (tensor<1x64xi1>) -> tensor<64x64xi1>
53
+ %57 = tt.load %55, %56, %cst_8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32>
54
+ %58 = arith.addi %53, %15 : tensor<64x64xi32>
55
+ %59 = tt.addptr %16, %58 : tensor<64x64x!tt.ptr<bf16, 1>>, tensor<64x64xi32>
56
+ %60 = tt.load %59, %56, %cst {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xbf16>
57
+ %61 = arith.extf %60 : tensor<64x64xbf16> to tensor<64x64xf32>
58
+ tt.assert %22, "index out of bounds: 0 <= tmp3 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1>
59
+ %62 = arith.extsi %51 : tensor<1x64xi32> to tensor<1x64xi64>
60
+ %63 = tt.broadcast %62 : (tensor<1x64xi64>) -> tensor<64x64xi64>
61
+ %64 = arith.addi %63, %24 : tensor<64x64xi64>
62
+ %65 = tt.addptr %25, %64 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi64>
63
+ %66 = tt.load %65, %56, %cst_8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32>
64
+ %67 = arith.addf %66, %57 : tensor<64x64xf32>
65
+ %68 = arith.addf %67, %61 : tensor<64x64xf32>
66
+ %69 = arith.subf %68, %arg9 : tensor<64x64xf32>
67
+ %70 = arith.addf %arg11, %cst_1 : tensor<64x64xf32>
68
+ %71 = arith.divf %69, %70 : tensor<64x64xf32>
69
+ %72 = arith.addf %arg9, %71 : tensor<64x64xf32>
70
+ %73 = arith.subf %68, %72 : tensor<64x64xf32>
71
+ %74 = arith.mulf %69, %73 : tensor<64x64xf32>
72
+ %75 = arith.addf %arg10, %74 : tensor<64x64xf32>
73
+ %76 = arith.select %56, %72, %arg9 : tensor<64x64xi1>, tensor<64x64xf32>
74
+ %77 = arith.select %56, %75, %arg10 : tensor<64x64xi1>, tensor<64x64xf32>
75
+ %78 = arith.select %56, %70, %arg11 : tensor<64x64xi1>, tensor<64x64xf32>
76
+ scf.yield %76, %77, %78 : tensor<64x64xf32>, tensor<64x64xf32>, tensor<64x64xf32>
77
+ }
78
+ %27:3 = "tt.reduce"(%26#0, %26#1, %26#2) <{axis = 1 : i32}> ({
79
+ ^bb0(%arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32, %arg13: f32):
80
+ %50 = arith.subf %arg11, %arg8 : f32
81
+ %51 = arith.addf %arg10, %arg13 : f32
82
+ %52 = arith.cmpf oeq, %51, %cst_0 : f32
83
+ %53 = arith.divf %arg13, %51 : f32
84
+ %54 = arith.select %52, %cst_0, %53 : f32
85
+ %55 = arith.mulf %50, %54 : f32
86
+ %56 = arith.addf %arg8, %55 : f32
87
+ %57 = arith.addf %arg9, %arg12 : f32
88
+ %58 = arith.mulf %50, %50 : f32
89
+ %59 = arith.mulf %58, %arg10 : f32
90
+ %60 = arith.mulf %59, %54 : f32
91
+ %61 = arith.addf %57, %60 : f32
92
+ tt.reduce.return %56, %61, %51 : f32, f32, f32
93
+ }) : (tensor<64x64xf32>, tensor<64x64xf32>, tensor<64x64xf32>) -> (tensor<64xf32>, tensor<64xf32>, tensor<64xf32>)
94
+ %28 = tt.expand_dims %27#0 {axis = 1 : i32} : (tensor<64xf32>) -> tensor<64x1xf32>
95
+ %29 = tt.expand_dims %27#1 {axis = 1 : i32} : (tensor<64xf32>) -> tensor<64x1xf32>
96
+ %30 = arith.muli %10, %cst_9 : tensor<64x1xi32>
97
+ %31 = tt.broadcast %30 : (tensor<64x1xi32>) -> tensor<64x64xi32>
98
+ %32 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
99
+ %33 = arith.muli %5, %cst_9 : tensor<64x1xi32>
100
+ %34 = tt.broadcast %33 : (tensor<64x1xi32>) -> tensor<64x64xi32>
101
+ %35 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>>
102
+ %36 = tt.splat %arg4 : (!tt.ptr<f32, 1>) -> tensor<1x64x!tt.ptr<f32, 1>>
103
+ %37 = arith.addi %9, %cst_4 : tensor<64x1xi64>
104
+ %38 = arith.cmpi slt, %9, %cst_3 : tensor<64x1xi64>
105
+ %39 = arith.select %38, %37, %9 : tensor<64x1xi1>, tensor<64x1xi64>
106
+ %40 = arith.cmpi sge, %39, %cst_3 : tensor<64x1xi64>
107
+ %41 = arith.cmpi slt, %39, %cst_4 : tensor<64x1xi64>
108
+ %42 = arith.andi %40, %41 : tensor<64x1xi1>
109
+ %43 = arith.muli %39, %cst_2 : tensor<64x1xi64>
110
+ %44 = tt.broadcast %43 : (tensor<64x1xi64>) -> tensor<64x64xi64>
111
+ %45 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<64x64x!tt.ptr<f32, 1>>
112
+ %46 = tt.broadcast %28 : (tensor<64x1xf32>) -> tensor<64x64xf32>
113
+ %47 = arith.divf %29, %cst_6 : tensor<64x1xf32>
114
+ %48 = arith.addf %47, %cst_5 : tensor<64x1xf32>
115
+ %49 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<64x64x!tt.ptr<bf16, 1>>
116
+ scf.for %arg8 = %c0_i32 to %c256_i32 step %c64_i32 : i32 {
117
+ %50 = tt.splat %arg8 : (i32) -> tensor<1x64xi32>
118
+ %51 = arith.addi %50, %6 : tensor<1x64xi32>
119
+ %52 = arith.cmpi slt, %51, %cst_10 : tensor<1x64xi32>
120
+ %53 = tt.broadcast %51 : (tensor<1x64xi32>) -> tensor<64x64xi32>
121
+ %54 = arith.addi %53, %31 : tensor<64x64xi32>
122
+ %55 = tt.addptr %32, %54 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi32>
123
+ %56 = tt.broadcast %52 : (tensor<1x64xi1>) -> tensor<64x64xi1>
124
+ %57 = tt.load %55, %56, %cst_8 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<64x64xf32>
125
+ %58 = arith.addi %53, %34 : tensor<64x64xi32>
126
+ %59 = tt.addptr %35, %58 : tensor<64x64x!tt.ptr<bf16, 1>>, tensor<64x64xi32>
127
+ %60 = tt.load %59, %56, %cst {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xbf16>
128
+ %61 = arith.extf %60 : tensor<64x64xbf16> to tensor<64x64xf32>
129
+ %62 = tt.addptr %36, %51 : tensor<1x64x!tt.ptr<f32, 1>>, tensor<1x64xi32>
130
+ %63 = tt.load %62, %52, %cst_7 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x64xf32>
131
+ tt.assert %42, "index out of bounds: 0 <= tmp16 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<64x1xi1>
132
+ %64 = arith.extsi %51 : tensor<1x64xi32> to tensor<1x64xi64>
133
+ %65 = tt.broadcast %64 : (tensor<1x64xi64>) -> tensor<64x64xi64>
134
+ %66 = arith.addi %65, %44 : tensor<64x64xi64>
135
+ %67 = tt.addptr %45, %66 : tensor<64x64x!tt.ptr<f32, 1>>, tensor<64x64xi64>
136
+ %68 = tt.load %67, %56, %cst_8 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<64x64xf32>
137
+ %69 = arith.addf %68, %57 : tensor<64x64xf32>
138
+ %70 = arith.addf %69, %61 : tensor<64x64xf32>
139
+ %71 = arith.subf %70, %46 : tensor<64x64xf32>
140
+ %72 = tt.extern_elementwise %48 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<64x1xf32>) -> tensor<64x1xf32>
141
+ %73 = tt.broadcast %72 : (tensor<64x1xf32>) -> tensor<64x64xf32>
142
+ %74 = arith.mulf %71, %73 : tensor<64x64xf32>
143
+ %75 = tt.broadcast %63 : (tensor<1x64xf32>) -> tensor<64x64xf32>
144
+ %76 = arith.mulf %74, %75 : tensor<64x64xf32>
145
+ %77 = tt.addptr %49, %58 : tensor<64x64x!tt.ptr<bf16, 1>>, tensor<64x64xi32>
146
+ %78 = arith.truncf %76 : tensor<64x64xf32> to tensor<64x64xbf16>
147
+ tt.store %77, %78, %56 {cache = 1 : i32, evict = 1 : i32} : tensor<64x64xbf16>
148
+ }
149
+ tt.return
150
+ }
151
+ }
.triton/dump/76fb48b96c75cb8e388c291a18ef9b02/triton_.ttgir ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [1, 2], threadsPerWarp = [1, 32], warpsPerCTA = [2, 2], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
2
+ #blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [4, 1], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
3
+ #blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 4], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
4
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 4 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
5
+ tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
6
+ %cst = arith.constant dense<512> : tensor<2x1xi32, #blocked>
7
+ %cst_0 = arith.constant dense<256> : tensor<1x128xi32, #blocked>
8
+ %cst_1 = arith.constant dense<256> : tensor<2x1xi32, #blocked>
9
+ %cst_2 = arith.constant dense<0.000000e+00> : tensor<2x128xf32, #blocked>
10
+ %cst_3 = arith.constant dense<1.000000e+00> : tensor<2x128xf32, #blocked>
11
+ %cst_4 = arith.constant dense<256> : tensor<2x1xi64, #blocked>
12
+ %cst_5 = arith.constant dense<0> : tensor<2x1xi64, #blocked>
13
+ %cst_6 = arith.constant dense<50257> : tensor<2x1xi64, #blocked>
14
+ %cst_7 = arith.constant dense<50257> : tensor<2x1xi64, #blocked1>
15
+ %cst_8 = arith.constant dense<0> : tensor<2x1xi64, #blocked1>
16
+ %c0_i32 = arith.constant 0 : i32
17
+ %c128_i32 = arith.constant 128 : i32
18
+ %c256_i32 = arith.constant 256 : i32
19
+ %cst_9 = arith.constant dense<1.000000e+00> : tensor<2x128xf32, #blocked2>
20
+ %cst_10 = arith.constant 0.000000e+00 : f32
21
+ %cst_11 = arith.constant dense<0.000000e+00> : tensor<1x128xf32, #blocked2>
22
+ %cst_12 = arith.constant dense<0.000000e+00> : tensor<2x128xf32, #blocked2>
23
+ %cst_13 = arith.constant dense<256> : tensor<1x128xi32, #blocked2>
24
+ %cst_14 = arith.constant dense<9.99999974E-6> : tensor<2x1xf32, #blocked>
25
+ %cst_15 = arith.constant dense<2.560000e+02> : tensor<2x1xf32, #blocked>
26
+ %cst_16 = arith.constant dense<0.000000e+00> : tensor<2x128xbf16, #blocked>
27
+ %c2_i32 = arith.constant 2 : i32
28
+ %0 = tt.get_program_id x : i32
29
+ %1 = arith.muli %0, %c2_i32 : i32
30
+ %2 = tt.make_range {end = 2 : i32, start = 0 : i32} : tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
31
+ %3 = tt.make_range {end = 2 : i32, start = 0 : i32} : tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
32
+ %4 = tt.expand_dims %2 {axis = 1 : i32} : (tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<2x1xi32, #blocked>
33
+ %5 = tt.expand_dims %3 {axis = 1 : i32} : (tensor<2xi32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<2x1xi32, #blocked1>
34
+ %6 = tt.splat %1 : (i32) -> tensor<2x1xi32, #blocked>
35
+ %7 = tt.splat %1 : (i32) -> tensor<2x1xi32, #blocked1>
36
+ %8 = arith.addi %6, %4 : tensor<2x1xi32, #blocked>
37
+ %9 = arith.addi %7, %5 : tensor<2x1xi32, #blocked1>
38
+ %10 = tt.make_range {end = 128 : i32, start = 0 : i32} : tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
39
+ %11 = tt.make_range {end = 128 : i32, start = 0 : i32} : tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>
40
+ %12 = tt.expand_dims %10 {axis = 0 : i32} : (tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x128xi32, #blocked>
41
+ %13 = tt.expand_dims %11 {axis = 0 : i32} : (tensor<128xi32, #triton_gpu.slice<{dim = 0, parent = #blocked2}>>) -> tensor<1x128xi32, #blocked2>
42
+ %14 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<2x1x!tt.ptr<i64, 1>, #blocked>
43
+ %15 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<2x1x!tt.ptr<i64, 1>, #blocked1>
44
+ %16 = tt.addptr %14, %8 : tensor<2x1x!tt.ptr<i64, 1>, #blocked>, tensor<2x1xi32, #blocked>
45
+ %17 = tt.addptr %15, %9 : tensor<2x1x!tt.ptr<i64, 1>, #blocked1>, tensor<2x1xi32, #blocked1>
46
+ %18 = tt.load %16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x1xi64, #blocked>
47
+ %19 = tt.load %17 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x1xi64, #blocked1>
48
+ %20 = arith.remsi %8, %cst : tensor<2x1xi32, #blocked>
49
+ %21 = arith.muli %20, %cst_1 : tensor<2x1xi32, #blocked>
50
+ %22 = tt.broadcast %21 : (tensor<2x1xi32, #blocked>) -> tensor<2x128xi32, #blocked>
51
+ %23 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<2x128x!tt.ptr<f32, 1>, #blocked>
52
+ %24 = arith.muli %8, %cst_1 : tensor<2x1xi32, #blocked>
53
+ %25 = tt.broadcast %24 : (tensor<2x1xi32, #blocked>) -> tensor<2x128xi32, #blocked>
54
+ %26 = tt.splat %arg3 : (!tt.ptr<bf16, 1>) -> tensor<2x128x!tt.ptr<bf16, 1>, #blocked>
55
+ %27 = arith.addi %18, %cst_6 : tensor<2x1xi64, #blocked>
56
+ %28 = arith.addi %19, %cst_7 : tensor<2x1xi64, #blocked1>
57
+ %29 = arith.cmpi slt, %18, %cst_5 : tensor<2x1xi64, #blocked>
58
+ %30 = arith.cmpi slt, %19, %cst_8 : tensor<2x1xi64, #blocked1>
59
+ %31 = arith.select %29, %27, %18 : tensor<2x1xi1, #blocked>, tensor<2x1xi64, #blocked>
60
+ %32 = arith.select %30, %28, %19 : tensor<2x1xi1, #blocked1>, tensor<2x1xi64, #blocked1>
61
+ %33 = arith.cmpi sge, %32, %cst_8 : tensor<2x1xi64, #blocked1>
62
+ %34 = arith.cmpi slt, %32, %cst_7 : tensor<2x1xi64, #blocked1>
63
+ %35 = arith.andi %33, %34 : tensor<2x1xi1, #blocked1>
64
+ %36 = arith.muli %31, %cst_4 : tensor<2x1xi64, #blocked>
65
+ %37 = tt.broadcast %36 : (tensor<2x1xi64, #blocked>) -> tensor<2x128xi64, #blocked>
66
+ %38 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<2x128x!tt.ptr<f32, 1>, #blocked>
67
+ %39:4 = scf.for %arg8 = %c0_i32 to %c256_i32 step %c128_i32 iter_args(%arg9 = %cst_2, %arg10 = %cst_2, %arg11 = %cst_12, %arg12 = %cst_2) -> (tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked2>, tensor<2x128xf32, #blocked>) : i32 {
68
+ %49 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked>
69
+ %50 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked2>
70
+ %51 = arith.addi %49, %12 : tensor<1x128xi32, #blocked>
71
+ %52 = arith.addi %50, %13 : tensor<1x128xi32, #blocked2>
72
+ %53 = arith.cmpi slt, %51, %cst_0 : tensor<1x128xi32, #blocked>
73
+ %54 = arith.cmpi slt, %52, %cst_13 : tensor<1x128xi32, #blocked2>
74
+ %55 = tt.broadcast %51 : (tensor<1x128xi32, #blocked>) -> tensor<2x128xi32, #blocked>
75
+ %56 = arith.addi %55, %22 : tensor<2x128xi32, #blocked>
76
+ %57 = tt.addptr %23, %56 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi32, #blocked>
77
+ %58 = tt.broadcast %53 : (tensor<1x128xi1, #blocked>) -> tensor<2x128xi1, #blocked>
78
+ %59 = tt.broadcast %54 : (tensor<1x128xi1, #blocked2>) -> tensor<2x128xi1, #blocked2>
79
+ %60 = tt.load %57, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
80
+ %61 = arith.addi %55, %25 : tensor<2x128xi32, #blocked>
81
+ %62 = tt.addptr %26, %61 : tensor<2x128x!tt.ptr<bf16, 1>, #blocked>, tensor<2x128xi32, #blocked>
82
+ %63 = tt.load %62, %58, %cst_16 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xbf16, #blocked>
83
+ %64 = arith.extf %63 : tensor<2x128xbf16, #blocked> to tensor<2x128xf32, #blocked>
84
+ tt.assert %35, "index out of bounds: 0 <= tmp3 < 50257", "/usr/local/lib/python3.10/dist-packages/torch/_inductor/codecache.py", "<module>", 1892 : tensor<2x1xi1, #blocked1>
85
+ %65 = arith.extsi %51 : tensor<1x128xi32, #blocked> to tensor<1x128xi64, #blocked>
86
+ %66 = tt.broadcast %65 : (tensor<1x128xi64, #blocked>) -> tensor<2x128xi64, #blocked>
87
+ %67 = arith.addi %66, %37 : tensor<2x128xi64, #blocked>
88
+ %68 = tt.addptr %38, %67 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi64, #blocked>
89
+ %69 = tt.load %68, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
90
+ %70 = arith.addf %69, %60 : tensor<2x128xf32, #blocked>
91
+ %71 = arith.addf %70, %64 : tensor<2x128xf32, #blocked>
92
+ %72 = arith.subf %71, %arg9 : tensor<2x128xf32, #blocked>
93
+ %73 = arith.addf %arg12, %cst_3 : tensor<2x128xf32, #blocked>
94
+ %74 = arith.addf %arg11, %cst_9 : tensor<2x128xf32, #blocked2>
95
+ %75 = arith.divf %72, %73 : tensor<2x128xf32, #blocked>
96
+ %76 = arith.addf %arg9, %75 : tensor<2x128xf32, #blocked>
97
+ %77 = arith.subf %71, %76 : tensor<2x128xf32, #blocked>
98
+ %78 = arith.mulf %72, %77 : tensor<2x128xf32, #blocked>
99
+ %79 = arith.addf %arg10, %78 : tensor<2x128xf32, #blocked>
100
+ %80 = arith.select %58, %76, %arg9 : tensor<2x128xi1, #blocked>, tensor<2x128xf32, #blocked>
101
+ %81 = arith.select %58, %79, %arg10 : tensor<2x128xi1, #blocked>, tensor<2x128xf32, #blocked>
102
+ %82 = arith.select %58, %73, %arg12 : tensor<2x128xi1, #blocked>, tensor<2x128xf32, #blocked>
103
+ %83 = arith.select %59, %74, %arg11 : tensor<2x128xi1, #blocked2>, tensor<2x128xf32, #blocked2>
104
+ scf.yield %80, %81, %83, %82 : tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked2>, tensor<2x128xf32, #blocked>
105
+ }
106
+ %40 = triton_gpu.convert_layout %39#2 : (tensor<2x128xf32, #blocked2>) -> tensor<2x128xf32, #blocked>
107
+ %41:3 = "tt.reduce"(%39#0, %39#1, %40) <{axis = 1 : i32}> ({
108
+ ^bb0(%arg8: f32, %arg9: f32, %arg10: f32, %arg11: f32, %arg12: f32, %arg13: f32):
109
+ %49 = arith.subf %arg11, %arg8 : f32
110
+ %50 = arith.addf %arg10, %arg13 : f32
111
+ %51 = arith.cmpf oeq, %50, %cst_10 : f32
112
+ %52 = arith.divf %arg13, %50 : f32
113
+ %53 = arith.select %51, %cst_10, %52 : f32
114
+ %54 = arith.mulf %49, %53 : f32
115
+ %55 = arith.addf %arg8, %54 : f32
116
+ %56 = arith.addf %arg9, %arg12 : f32
117
+ %57 = arith.mulf %49, %49 : f32
118
+ %58 = arith.mulf %57, %arg10 : f32
119
+ %59 = arith.mulf %58, %53 : f32
120
+ %60 = arith.addf %56, %59 : f32
121
+ tt.reduce.return %55, %60, %50 : f32, f32, f32
122
+ }) : (tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>, tensor<2x128xf32, #blocked>) -> (tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>, tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>)
123
+ %42 = tt.expand_dims %41#0 {axis = 1 : i32} : (tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<2x1xf32, #blocked>
124
+ %43 = tt.expand_dims %41#1 {axis = 1 : i32} : (tensor<2xf32, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<2x1xf32, #blocked>
125
+ %44 = tt.splat %arg4 : (!tt.ptr<f32, 1>) -> tensor<1x128x!tt.ptr<f32, 1>, #blocked2>
126
+ %45 = tt.broadcast %42 : (tensor<2x1xf32, #blocked>) -> tensor<2x128xf32, #blocked>
127
+ %46 = arith.divf %43, %cst_15 : tensor<2x1xf32, #blocked>
128
+ %47 = arith.addf %46, %cst_14 : tensor<2x1xf32, #blocked>
129
+ %48 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<2x128x!tt.ptr<bf16, 1>, #blocked>
130
+ scf.for %arg8 = %c0_i32 to %c256_i32 step %c128_i32 : i32 {
131
+ %49 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked>
132
+ %50 = tt.splat %arg8 : (i32) -> tensor<1x128xi32, #blocked2>
133
+ %51 = arith.addi %49, %12 : tensor<1x128xi32, #blocked>
134
+ %52 = arith.addi %50, %13 : tensor<1x128xi32, #blocked2>
135
+ %53 = arith.cmpi slt, %51, %cst_0 : tensor<1x128xi32, #blocked>
136
+ %54 = arith.cmpi slt, %52, %cst_13 : tensor<1x128xi32, #blocked2>
137
+ %55 = tt.broadcast %51 : (tensor<1x128xi32, #blocked>) -> tensor<2x128xi32, #blocked>
138
+ %56 = arith.addi %55, %22 : tensor<2x128xi32, #blocked>
139
+ %57 = tt.addptr %23, %56 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi32, #blocked>
140
+ %58 = tt.broadcast %53 : (tensor<1x128xi1, #blocked>) -> tensor<2x128xi1, #blocked>
141
+ %59 = tt.load %57, %58, %cst_2 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
142
+ %60 = arith.addi %55, %25 : tensor<2x128xi32, #blocked>
143
+ %61 = tt.addptr %26, %60 : tensor<2x128x!tt.ptr<bf16, 1>, #blocked>, tensor<2x128xi32, #blocked>
144
+ %62 = tt.load %61, %58, %cst_16 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<2x128xbf16, #blocked>
145
+ %63 = arith.extf %62 : tensor<2x128xbf16, #blocked> to tensor<2x128xf32, #blocked>
146
+ %64 = tt.addptr %44, %52 : tensor<1x128x!tt.ptr<f32, 1>, #blocked2>, tensor<1x128xi32, #blocked2>
147
+ %65 = tt.load %64, %54, %cst_11 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x128xf32, #blocked2>
148
+ tt.assert %35, "index out of bounds: 0 <= tmp16 < 50257", "/usr/local/lib/python3.10/dist-packages/torch/_inductor/codecache.py", "<module>", 1892 : tensor<2x1xi1, #blocked1>
149
+ %66 = arith.extsi %51 : tensor<1x128xi32, #blocked> to tensor<1x128xi64, #blocked>
150
+ %67 = tt.broadcast %66 : (tensor<1x128xi64, #blocked>) -> tensor<2x128xi64, #blocked>
151
+ %68 = arith.addi %67, %37 : tensor<2x128xi64, #blocked>
152
+ %69 = tt.addptr %38, %68 : tensor<2x128x!tt.ptr<f32, 1>, #blocked>, tensor<2x128xi64, #blocked>
153
+ %70 = tt.load %69, %58, %cst_2 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<2x128xf32, #blocked>
154
+ %71 = arith.addf %70, %59 : tensor<2x128xf32, #blocked>
155
+ %72 = arith.addf %71, %63 : tensor<2x128xf32, #blocked>
156
+ %73 = arith.subf %72, %45 : tensor<2x128xf32, #blocked>
157
+ %74 = tt.extern_elementwise %47 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (tensor<2x1xf32, #blocked>) -> tensor<2x1xf32, #blocked>
158
+ %75 = tt.broadcast %74 : (tensor<2x1xf32, #blocked>) -> tensor<2x128xf32, #blocked>
159
+ %76 = arith.mulf %73, %75 : tensor<2x128xf32, #blocked>
160
+ %77 = triton_gpu.convert_layout %65 : (tensor<1x128xf32, #blocked2>) -> tensor<1x128xf32, #blocked>
161
+ %78 = tt.broadcast %77 : (tensor<1x128xf32, #blocked>) -> tensor<2x128xf32, #blocked>
162
+ %79 = arith.mulf %76, %78 : tensor<2x128xf32, #blocked>
163
+ %80 = tt.addptr %48, %60 : tensor<2x128x!tt.ptr<bf16, 1>, #blocked>, tensor<2x128xi32, #blocked>
164
+ %81 = arith.truncf %79 : tensor<2x128xf32, #blocked> to tensor<2x128xbf16, #blocked>
165
+ tt.store %80, %81, %58 {cache = 1 : i32, evict = 1 : i32} : tensor<2x128xbf16, #blocked>
166
+ }
167
+ tt.return
168
+ }
169
+ }
.triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttgir ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [1, 4], threadsPerWarp = [1, 32], warpsPerCTA = [1, 8], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
2
+ #blocked1 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 8], order = [0, 1], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
3
+ #blocked2 = #triton_gpu.blocked<{sizePerThread = [1, 1], threadsPerWarp = [1, 32], warpsPerCTA = [1, 8], order = [1, 0], CTAsPerCGA = [1, 1], CTASplitNum = [1, 1], CTAOrder = [1, 0]}>
4
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 8 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
5
+ tt.func public @triton__0d1d2d3d4d5d6e7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg6: i64 {tt.max_divisibility = 8 : i32}, %arg7: i64 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
6
+ %cst = arith.constant dense<7680> : tensor<1x2048xi64, #blocked>
7
+ %cst_0 = arith.constant dense<7680> : tensor<1x2048xi64, #blocked1>
8
+ %cst_1 = arith.constant dense<50257> : tensor<1x2048xi64, #blocked>
9
+ %c385973760_i64 = arith.constant 385973760 : i64
10
+ %c7680_i64 = arith.constant 7680 : i64
11
+ %c8_i64 = arith.constant 8 : i64
12
+ %cst_2 = arith.constant dense<-1> : tensor<1x2048xi64, #blocked>
13
+ %cst_3 = arith.constant dense<0> : tensor<1x2048xi64, #blocked>
14
+ %cst_4 = arith.constant dense<0.000000e+00> : tensor<1x2048xf32, #blocked1>
15
+ %cst_5 = arith.constant dense<0.000000e+00> : tensor<1x2048xf32, #blocked>
16
+ %cst_6 = arith.constant dense<0.000000e+00> : tensor<1x2048xbf16, #blocked1>
17
+ %c0_i32 = arith.constant 0 : i32
18
+ %c7680_i32 = arith.constant 7680 : i32
19
+ %c2048_i32 = arith.constant 2048 : i32
20
+ %0 = tt.get_program_id x : i32
21
+ %1 = arith.extsi %0 : i32 to i64
22
+ %2 = arith.cmpi slt, %1, %c8_i64 : i64
23
+ %3 = tt.make_range {end = 2048 : i32, start = 0 : i32} : tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>
24
+ %4 = tt.make_range {end = 2048 : i32, start = 0 : i32} : tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked1}>>
25
+ %5 = tt.expand_dims %3 {axis = 0 : i32} : (tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked}>>) -> tensor<1x2048xi32, #blocked>
26
+ %6 = tt.expand_dims %4 {axis = 0 : i32} : (tensor<2048xi32, #triton_gpu.slice<{dim = 0, parent = #blocked1}>>) -> tensor<1x2048xi32, #blocked1>
27
+ %7 = arith.extsi %5 : tensor<1x2048xi32, #blocked> to tensor<1x2048xi64, #blocked>
28
+ %8 = arith.extsi %6 : tensor<1x2048xi32, #blocked1> to tensor<1x2048xi64, #blocked1>
29
+ %9 = arith.muli %1, %c7680_i64 : i64
30
+ %10 = tt.splat %9 : (i64) -> tensor<1x2048xi64, #blocked>
31
+ %11 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<1x2048x!tt.ptr<i64, 1>, #blocked>
32
+ %12 = tt.splat %2 : (i1) -> tensor<1x2048xi1, #blocked>
33
+ %13 = tt.splat %2 : (i1) -> tensor<1x2048xi1, #blocked1>
34
+ %14 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>, #blocked>
35
+ %15 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>, #blocked>
36
+ %16 = arith.muli %1, %c385973760_i64 : i64
37
+ %17 = tt.splat %16 : (i64) -> tensor<1x2048xi64, #blocked>
38
+ %18 = tt.splat %arg1 : (!tt.ptr<bf16, 1>) -> tensor<1x2048x!tt.ptr<bf16, 1>, #blocked>
39
+ %19:2 = scf.for %arg8 = %c0_i32 to %c7680_i32 step %c2048_i32 iter_args(%arg9 = %cst_4, %arg10 = %cst_3) -> (tensor<1x2048xf32, #blocked1>, tensor<1x2048xi64, #blocked>) : i32 {
40
+ %30 = arith.extsi %arg8 : i32 to i64
41
+ %31 = tt.splat %30 : (i64) -> tensor<1x2048xi64, #blocked>
42
+ %32 = tt.splat %30 : (i64) -> tensor<1x2048xi64, #blocked1>
43
+ %33 = arith.addi %31, %7 : tensor<1x2048xi64, #blocked>
44
+ %34 = arith.addi %32, %8 : tensor<1x2048xi64, #blocked1>
45
+ %35 = arith.cmpi slt, %33, %cst : tensor<1x2048xi64, #blocked>
46
+ %36 = arith.cmpi slt, %34, %cst_0 : tensor<1x2048xi64, #blocked1>
47
+ %37 = arith.addi %33, %10 : tensor<1x2048xi64, #blocked>
48
+ %38 = tt.addptr %11, %37 : tensor<1x2048x!tt.ptr<i64, 1>, #blocked>, tensor<1x2048xi64, #blocked>
49
+ %39 = arith.andi %35, %12 : tensor<1x2048xi1, #blocked>
50
+ %40 = arith.andi %36, %13 : tensor<1x2048xi1, #blocked1>
51
+ %41 = tt.load %38, %39, %cst_3 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xi64, #blocked>
52
+ %42 = tt.addptr %14, %37 : tensor<1x2048x!tt.ptr<f32, 1>, #blocked>, tensor<1x2048xi64, #blocked>
53
+ %43 = tt.load %42, %39, %cst_5 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32, #blocked>
54
+ %44 = triton_gpu.convert_layout %43 : (tensor<1x2048xf32, #blocked>) -> tensor<1x2048xf32, #blocked1>
55
+ %45 = tt.addptr %15, %37 : tensor<1x2048x!tt.ptr<f32, 1>, #blocked>, tensor<1x2048xi64, #blocked>
56
+ %46 = tt.load %45, %39, %cst_5 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32, #blocked>
57
+ %47 = arith.cmpi ne, %41, %cst_2 : tensor<1x2048xi64, #blocked>
58
+ %48 = triton_gpu.convert_layout %47 : (tensor<1x2048xi1, #blocked>) -> tensor<1x2048xi1, #blocked1>
59
+ %49 = arith.select %47, %41, %cst_3 : tensor<1x2048xi1, #blocked>, tensor<1x2048xi64, #blocked>
60
+ %50 = arith.addi %49, %cst_1 : tensor<1x2048xi64, #blocked>
61
+ %51 = arith.cmpi slt, %49, %cst_3 : tensor<1x2048xi64, #blocked>
62
+ %52 = arith.select %51, %50, %49 : tensor<1x2048xi1, #blocked>, tensor<1x2048xi64, #blocked>
63
+ %53 = arith.cmpi sge, %52, %cst_3 : tensor<1x2048xi64, #blocked>
64
+ %54 = arith.cmpi slt, %52, %cst_1 : tensor<1x2048xi64, #blocked>
65
+ %55 = arith.andi %53, %54 : tensor<1x2048xi1, #blocked>
66
+ %56 = triton_gpu.convert_layout %55 : (tensor<1x2048xi1, #blocked>) -> tensor<1x2048xi1, #blocked2>
67
+ tt.assert %56, "index out of bounds: 0 <= tmp7 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<1x2048xi1, #blocked2>
68
+ %57 = arith.muli %33, %cst_1 : tensor<1x2048xi64, #blocked>
69
+ %58 = arith.addi %52, %57 : tensor<1x2048xi64, #blocked>
70
+ %59 = arith.addi %58, %17 : tensor<1x2048xi64, #blocked>
71
+ %60 = tt.addptr %18, %59 : tensor<1x2048x!tt.ptr<bf16, 1>, #blocked>, tensor<1x2048xi64, #blocked>
72
+ %61 = triton_gpu.convert_layout %60 : (tensor<1x2048x!tt.ptr<bf16, 1>, #blocked>) -> tensor<1x2048x!tt.ptr<bf16, 1>, #blocked1>
73
+ %62 = tt.load %61, %40, %cst_6 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x2048xbf16, #blocked1>
74
+ %63 = arith.extf %62 : tensor<1x2048xbf16, #blocked1> to tensor<1x2048xf32, #blocked1>
75
+ %64 = arith.subf %63, %44 : tensor<1x2048xf32, #blocked1>
76
+ %65 = math.log %46 : tensor<1x2048xf32, #blocked>
77
+ %66 = triton_gpu.convert_layout %65 : (tensor<1x2048xf32, #blocked>) -> tensor<1x2048xf32, #blocked1>
78
+ %67 = arith.subf %64, %66 : tensor<1x2048xf32, #blocked1>
79
+ %68 = arith.subf %cst_4, %67 : tensor<1x2048xf32, #blocked1>
80
+ %69 = arith.select %48, %68, %cst_4 : tensor<1x2048xi1, #blocked1>, tensor<1x2048xf32, #blocked1>
81
+ %70 = arith.addf %arg9, %69 : tensor<1x2048xf32, #blocked1>
82
+ %71 = arith.select %40, %70, %arg9 : tensor<1x2048xi1, #blocked1>, tensor<1x2048xf32, #blocked1>
83
+ %72 = arith.extui %47 : tensor<1x2048xi1, #blocked> to tensor<1x2048xi64, #blocked>
84
+ %73 = arith.addi %arg10, %72 : tensor<1x2048xi64, #blocked>
85
+ %74 = arith.select %39, %73, %arg10 : tensor<1x2048xi1, #blocked>, tensor<1x2048xi64, #blocked>
86
+ scf.yield %71, %74 : tensor<1x2048xf32, #blocked1>, tensor<1x2048xi64, #blocked>
87
+ }
88
+ %20 = "tt.reduce"(%19#0) <{axis = 1 : i32}> ({
89
+ ^bb0(%arg8: f32, %arg9: f32):
90
+ %30 = arith.addf %arg8, %arg9 : f32
91
+ tt.reduce.return %30 : f32
92
+ }) : (tensor<1x2048xf32, #blocked1>) -> tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
93
+ %21 = tt.expand_dims %20 {axis = 1 : i32} : (tensor<1xf32, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<1x1xf32, #blocked1>
94
+ %22 = tt.addptr %arg4, %1 : !tt.ptr<f32, 1>, i64
95
+ %23 = tt.splat %22 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>, #blocked1>
96
+ %24 = tt.splat %2 : (i1) -> tensor<1x1xi1, #blocked1>
97
+ tt.store %23, %21, %24 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32, #blocked1>
98
+ %25 = "tt.reduce"(%19#1) <{axis = 1 : i32}> ({
99
+ ^bb0(%arg8: i64, %arg9: i64):
100
+ %30 = arith.addi %arg8, %arg9 : i64
101
+ tt.reduce.return %30 : i64
102
+ }) : (tensor<1x2048xi64, #blocked>) -> tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>
103
+ %26 = triton_gpu.convert_layout %25 : (tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked}>>) -> tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>
104
+ %27 = tt.expand_dims %26 {axis = 1 : i32} : (tensor<1xi64, #triton_gpu.slice<{dim = 1, parent = #blocked1}>>) -> tensor<1x1xi64, #blocked1>
105
+ %28 = tt.addptr %arg5, %1 : !tt.ptr<i64, 1>, i64
106
+ %29 = tt.splat %28 : (!tt.ptr<i64, 1>) -> tensor<1x1x!tt.ptr<i64, 1>, #blocked1>
107
+ tt.store %29, %27, %24 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xi64, #blocked1>
108
+ tt.return
109
+ }
110
+ }
.triton/dump/89f8cc1079aa03024e56dc2aee42813a/triton_.ttir ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ module {
2
+ tt.func public @triton__0d1d2d3d4d5d6e7de(%arg0: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<i64, 1> {tt.divisibility = 16 : i32}, %arg6: i64 {tt.max_divisibility = 8 : i32}, %arg7: i64 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
3
+ %c8_i64 = arith.constant 8 : i64
4
+ %c7680_i64 = arith.constant 7680 : i64
5
+ %c385973760_i64 = arith.constant 385973760 : i64
6
+ %cst = arith.constant dense<0.000000e+00> : tensor<1x2048xbf16>
7
+ %cst_0 = arith.constant dense<50257> : tensor<1x2048xi64>
8
+ %cst_1 = arith.constant dense<7680> : tensor<1x2048xi64>
9
+ %c2048_i32 = arith.constant 2048 : i32
10
+ %c7680_i32 = arith.constant 7680 : i32
11
+ %c0_i32 = arith.constant 0 : i32
12
+ %cst_2 = arith.constant dense<-1> : tensor<1x2048xi64>
13
+ %cst_3 = arith.constant dense<0> : tensor<1x2048xi64>
14
+ %cst_4 = arith.constant dense<0.000000e+00> : tensor<1x2048xf32>
15
+ %0 = tt.get_program_id x : i32
16
+ %1 = arith.extsi %0 : i32 to i64
17
+ %2 = arith.cmpi slt, %1, %c8_i64 : i64
18
+ %3 = tt.splat %2 : (i1) -> tensor<1x1xi1>
19
+ %4 = tt.make_range {end = 2048 : i32, start = 0 : i32} : tensor<2048xi32>
20
+ %5 = tt.expand_dims %4 {axis = 0 : i32} : (tensor<2048xi32>) -> tensor<1x2048xi32>
21
+ %6 = arith.extsi %5 : tensor<1x2048xi32> to tensor<1x2048xi64>
22
+ %7 = arith.muli %1, %c7680_i64 : i64
23
+ %8 = tt.splat %7 : (i64) -> tensor<1x2048xi64>
24
+ %9 = tt.splat %arg0 : (!tt.ptr<i64, 1>) -> tensor<1x2048x!tt.ptr<i64, 1>>
25
+ %10 = tt.splat %2 : (i1) -> tensor<1x2048xi1>
26
+ %11 = tt.splat %arg2 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>>
27
+ %12 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<1x2048x!tt.ptr<f32, 1>>
28
+ %13 = arith.muli %1, %c385973760_i64 : i64
29
+ %14 = tt.splat %13 : (i64) -> tensor<1x2048xi64>
30
+ %15 = tt.splat %arg1 : (!tt.ptr<bf16, 1>) -> tensor<1x2048x!tt.ptr<bf16, 1>>
31
+ %16:2 = scf.for %arg8 = %c0_i32 to %c7680_i32 step %c2048_i32 iter_args(%arg9 = %cst_4, %arg10 = %cst_3) -> (tensor<1x2048xf32>, tensor<1x2048xi64>) : i32 {
32
+ %25 = arith.extsi %arg8 : i32 to i64
33
+ %26 = tt.splat %25 : (i64) -> tensor<1x2048xi64>
34
+ %27 = arith.addi %26, %6 : tensor<1x2048xi64>
35
+ %28 = arith.cmpi slt, %27, %cst_1 : tensor<1x2048xi64>
36
+ %29 = arith.addi %27, %8 : tensor<1x2048xi64>
37
+ %30 = tt.addptr %9, %29 : tensor<1x2048x!tt.ptr<i64, 1>>, tensor<1x2048xi64>
38
+ %31 = arith.andi %28, %10 : tensor<1x2048xi1>
39
+ %32 = tt.load %30, %31, %cst_3 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xi64>
40
+ %33 = tt.addptr %11, %29 : tensor<1x2048x!tt.ptr<f32, 1>>, tensor<1x2048xi64>
41
+ %34 = tt.load %33, %31, %cst_4 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32>
42
+ %35 = tt.addptr %12, %29 : tensor<1x2048x!tt.ptr<f32, 1>>, tensor<1x2048xi64>
43
+ %36 = tt.load %35, %31, %cst_4 {cache = 1 : i32, evict = 2 : i32, isVolatile = false} : tensor<1x2048xf32>
44
+ %37 = arith.cmpi ne, %32, %cst_2 : tensor<1x2048xi64>
45
+ %38 = arith.select %37, %32, %cst_3 : tensor<1x2048xi1>, tensor<1x2048xi64>
46
+ %39 = arith.addi %38, %cst_0 : tensor<1x2048xi64>
47
+ %40 = arith.cmpi slt, %38, %cst_3 : tensor<1x2048xi64>
48
+ %41 = arith.select %40, %39, %38 : tensor<1x2048xi1>, tensor<1x2048xi64>
49
+ %42 = arith.cmpi sge, %41, %cst_3 : tensor<1x2048xi64>
50
+ %43 = arith.cmpi slt, %41, %cst_0 : tensor<1x2048xi64>
51
+ %44 = arith.andi %42, %43 : tensor<1x2048xi1>
52
+ tt.assert %44, "index out of bounds: 0 <= tmp7 < 50257", "<frozen importlib._bootstrap_external>", "_call_with_frames_removed", 883 : tensor<1x2048xi1>
53
+ %45 = arith.muli %27, %cst_0 : tensor<1x2048xi64>
54
+ %46 = arith.addi %41, %45 : tensor<1x2048xi64>
55
+ %47 = arith.addi %46, %14 : tensor<1x2048xi64>
56
+ %48 = tt.addptr %15, %47 : tensor<1x2048x!tt.ptr<bf16, 1>>, tensor<1x2048xi64>
57
+ %49 = tt.load %48, %31, %cst {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<1x2048xbf16>
58
+ %50 = arith.extf %49 : tensor<1x2048xbf16> to tensor<1x2048xf32>
59
+ %51 = arith.subf %50, %34 : tensor<1x2048xf32>
60
+ %52 = math.log %36 : tensor<1x2048xf32>
61
+ %53 = arith.subf %51, %52 : tensor<1x2048xf32>
62
+ %54 = arith.subf %cst_4, %53 : tensor<1x2048xf32>
63
+ %55 = arith.select %37, %54, %cst_4 : tensor<1x2048xi1>, tensor<1x2048xf32>
64
+ %56 = arith.addf %arg9, %55 : tensor<1x2048xf32>
65
+ %57 = arith.select %31, %56, %arg9 : tensor<1x2048xi1>, tensor<1x2048xf32>
66
+ %58 = arith.extui %37 : tensor<1x2048xi1> to tensor<1x2048xi64>
67
+ %59 = arith.addi %arg10, %58 : tensor<1x2048xi64>
68
+ %60 = arith.select %31, %59, %arg10 : tensor<1x2048xi1>, tensor<1x2048xi64>
69
+ scf.yield %57, %60 : tensor<1x2048xf32>, tensor<1x2048xi64>
70
+ }
71
+ %17 = "tt.reduce"(%16#0) <{axis = 1 : i32}> ({
72
+ ^bb0(%arg8: f32, %arg9: f32):
73
+ %25 = arith.addf %arg8, %arg9 : f32
74
+ tt.reduce.return %25 : f32
75
+ }) : (tensor<1x2048xf32>) -> tensor<1xf32>
76
+ %18 = tt.expand_dims %17 {axis = 1 : i32} : (tensor<1xf32>) -> tensor<1x1xf32>
77
+ %19 = tt.addptr %arg4, %1 : !tt.ptr<f32, 1>, i64
78
+ %20 = tt.splat %19 : (!tt.ptr<f32, 1>) -> tensor<1x1x!tt.ptr<f32, 1>>
79
+ tt.store %20, %18, %3 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xf32>
80
+ %21 = "tt.reduce"(%16#1) <{axis = 1 : i32}> ({
81
+ ^bb0(%arg8: i64, %arg9: i64):
82
+ %25 = arith.addi %arg8, %arg9 : i64
83
+ tt.reduce.return %25 : i64
84
+ }) : (tensor<1x2048xi64>) -> tensor<1xi64>
85
+ %22 = tt.expand_dims %21 {axis = 1 : i32} : (tensor<1xi64>) -> tensor<1x1xi64>
86
+ %23 = tt.addptr %arg5, %1 : !tt.ptr<i64, 1>, i64
87
+ %24 = tt.splat %23 : (!tt.ptr<i64, 1>) -> tensor<1x1x!tt.ptr<i64, 1>>
88
+ tt.store %24, %22, %3 {cache = 1 : i32, evict = 1 : i32} : tensor<1x1xi64>
89
+ tt.return
90
+ }
91
+ }
.triton/dump/ab89f85e55d5ddd9676325b49df9419f/triton_.ttgir ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #triton_gpu.blocked<{sizePerThread = [4], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
2
+ #blocked1 = #triton_gpu.blocked<{sizePerThread = [1], threadsPerWarp = [32], warpsPerCTA = [2], order = [0], CTAsPerCGA = [1], CTASplitNum = [1], CTAOrder = [0]}>
3
+ module attributes {"triton_gpu.compute-capability" = 89 : i32, "triton_gpu.num-ctas" = 1 : i32, "triton_gpu.num-warps" = 2 : i32, "triton_gpu.threads-per-warp" = 32 : i32} {
4
+ tt.func public @triton__0d1d2d3d4d5d6de7de(%arg0: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg1: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg2: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg3: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg4: !tt.ptr<f32, 1> {tt.divisibility = 16 : i32}, %arg5: !tt.ptr<bf16, 1> {tt.divisibility = 16 : i32}, %arg6: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}, %arg7: i32 {tt.divisibility = 16 : i32, tt.max_divisibility = 16 : i32}) attributes {noinline = false} {
5
+ %cst = arith.constant dense<256> : tensor<256xi32, #blocked>
6
+ %cst_0 = arith.constant 9.99999974E-6 : f32
7
+ %cst_1 = arith.constant 2.560000e+02 : f32
8
+ %cst_2 = arith.constant 0.000000e+00 : f32
9
+ %c256_i32 = arith.constant 256 : i32
10
+ %cst_3 = arith.constant dense<0.000000e+00> : tensor<256xf32, #blocked>
11
+ %cst_4 = arith.constant dense<0.000000e+00> : tensor<256xbf16, #blocked>
12
+ %0 = tt.get_program_id x : i32
13
+ %1 = tt.make_range {end = 256 : i32, start = 0 : i32} : tensor<256xi32, #blocked>
14
+ %2 = arith.cmpi slt, %1, %cst : tensor<256xi32, #blocked>
15
+ %3 = arith.muli %0, %c256_i32 : i32
16
+ %4 = tt.splat %3 : (i32) -> tensor<256xi32, #blocked>
17
+ %5 = arith.addi %1, %4 : tensor<256xi32, #blocked>
18
+ %6 = tt.splat %arg1 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
19
+ %7 = tt.addptr %6, %5 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
20
+ %8 = tt.load %7, %2, %cst_3 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xf32, #blocked>
21
+ %9 = tt.splat %arg2 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
22
+ %10 = tt.addptr %9, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
23
+ %11 = tt.load %10, %2, %cst_4 {cache = 1 : i32, evict = 1 : i32, isVolatile = false} : tensor<256xbf16, #blocked>
24
+ %12 = arith.extf %11 : tensor<256xbf16, #blocked> to tensor<256xf32, #blocked>
25
+ %13 = tt.splat %arg3 : (!tt.ptr<f32, 1>) -> tensor<256x!tt.ptr<f32, 1>, #blocked>
26
+ %14 = tt.addptr %13, %1 : tensor<256x!tt.ptr<f32, 1>, #blocked>, tensor<256xi32, #blocked>
27
+ %15 = tt.load %14, %2, %cst_3 {cache = 1 : i32, evict = 3 : i32, isVolatile = false} : tensor<256xf32, #blocked>
28
+ %16 = arith.addf %8, %12 : tensor<256xf32, #blocked>
29
+ %17 = arith.select %2, %16, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
30
+ %18 = "tt.reduce"(%17) <{axis = 0 : i32}> ({
31
+ ^bb0(%arg8: f32, %arg9: f32):
32
+ %42 = arith.addf %arg8, %arg9 : f32
33
+ tt.reduce.return %42 : f32
34
+ }) : (tensor<256xf32, #blocked>) -> f32
35
+ %19 = arith.addf %18, %cst_2 : f32
36
+ %20 = arith.divf %19, %cst_1 : f32
37
+ %21 = tt.splat %20 : (f32) -> tensor<1xf32, #blocked1>
38
+ %22 = tt.splat %20 : (f32) -> tensor<256xf32, #blocked>
39
+ %23 = arith.subf %16, %22 : tensor<256xf32, #blocked>
40
+ %24 = arith.mulf %23, %23 : tensor<256xf32, #blocked>
41
+ %25 = arith.select %2, %24, %cst_3 : tensor<256xi1, #blocked>, tensor<256xf32, #blocked>
42
+ %26 = "tt.reduce"(%25) <{axis = 0 : i32}> ({
43
+ ^bb0(%arg8: f32, %arg9: f32):
44
+ %42 = arith.addf %arg8, %arg9 : f32
45
+ tt.reduce.return %42 : f32
46
+ }) : (tensor<256xf32, #blocked>) -> f32
47
+ %27 = arith.addf %26, %cst_2 : f32
48
+ %28 = arith.divf %27, %cst_1 : f32
49
+ %29 = arith.addf %28, %cst_0 : f32
50
+ %30 = tt.extern_elementwise %29 {libname = "libdevice", libpath = "/usr/local/lib/python3.10/dist-packages/triton/language/../third_party/cuda/lib/libdevice.10.bc", pure = true, symbol = "__nv_rsqrtf"} : (f32) -> f32
51
+ %31 = tt.splat %30 : (f32) -> tensor<1xf32, #blocked1>
52
+ %32 = tt.splat %30 : (f32) -> tensor<256xf32, #blocked>
53
+ %33 = arith.mulf %23, %32 : tensor<256xf32, #blocked>
54
+ %34 = arith.mulf %33, %15 : tensor<256xf32, #blocked>
55
+ gpu.barrier
56
+ %35 = tt.addptr %arg0, %0 : !tt.ptr<f32, 1>, i32
57
+ %36 = tt.splat %35 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
58
+ tt.store %36, %31 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
59
+ %37 = tt.splat %arg5 : (!tt.ptr<bf16, 1>) -> tensor<256x!tt.ptr<bf16, 1>, #blocked>
60
+ %38 = tt.addptr %37, %5 : tensor<256x!tt.ptr<bf16, 1>, #blocked>, tensor<256xi32, #blocked>
61
+ %39 = arith.truncf %34 : tensor<256xf32, #blocked> to tensor<256xbf16, #blocked>
62
+ tt.store %38, %39, %2 {cache = 1 : i32, evict = 1 : i32} : tensor<256xbf16, #blocked>
63
+ %40 = tt.addptr %arg4, %0 : !tt.ptr<f32, 1>, i32
64
+ %41 = tt.splat %40 : (!tt.ptr<f32, 1>) -> tensor<1x!tt.ptr<f32, 1>, #blocked1>
65
+ tt.store %41, %21 {cache = 1 : i32, evict = 1 : i32} : tensor<1xf32, #blocked1>
66
+ tt.return
67
+ }
68
+ }