nouamanetazi HF staff commited on
Commit
5d1a52a
·
1 Parent(s): cb40edf
assets/images/profile_trace_annotated.png ADDED

Git LFS Details

  • SHA256: e1806f717e427febe26bfa45135d45d76adc9808c8a92553f7f7e0bb9faa80ae
  • Pointer size: 131 Bytes
  • Size of remote file: 995 kB
dist/assets/images/profile_trace_annotated.png ADDED

Git LFS Details

  • SHA256: e1806f717e427febe26bfa45135d45d76adc9808c8a92553f7f7e0bb9faa80ae
  • Pointer size: 131 Bytes
  • Size of remote file: 995 kB
dist/index.html CHANGED
@@ -534,7 +534,7 @@
534
  <li>Kernel execution times and memory allocation</li>
535
  </ul>
536
 
537
- <p><img alt="image.png" src="/assets/images/placeholder.png" /></p>
538
  <p>Figure: Example trace showing CPU thread launching kernels asynchronously to GPU, with compute kernels and communication happening in parallel across different CUDA streams</p>
539
 
540
  <p>The trace helps identify bottlenecks like:</p>
 
534
  <li>Kernel execution times and memory allocation</li>
535
  </ul>
536
 
537
+ <p><img alt="profile_trace_annotated.png" src="/assets/images/profile_trace_annotated.png" /></p>
538
  <p>Figure: Example trace showing CPU thread launching kernels asynchronously to GPU, with compute kernels and communication happening in parallel across different CUDA streams</p>
539
 
540
  <p>The trace helps identify bottlenecks like:</p>
src/index.html CHANGED
@@ -534,7 +534,7 @@
534
  <li>Kernel execution times and memory allocation</li>
535
  </ul>
536
 
537
- <p><img alt="image.png" src="/assets/images/placeholder.png" /></p>
538
  <p>Figure: Example trace showing CPU thread launching kernels asynchronously to GPU, with compute kernels and communication happening in parallel across different CUDA streams</p>
539
 
540
  <p>The trace helps identify bottlenecks like:</p>
 
534
  <li>Kernel execution times and memory allocation</li>
535
  </ul>
536
 
537
+ <p><img alt="profile_trace_annotated.png" src="/assets/images/profile_trace_annotated.png" /></p>
538
  <p>Figure: Example trace showing CPU thread launching kernels asynchronously to GPU, with compute kernels and communication happening in parallel across different CUDA streams</p>
539
 
540
  <p>The trace helps identify bottlenecks like:</p>