| | .. _faqs: |
| |
|
| | FAQs |
| | ==== |
| |
|
| | General |
| | |
| |
|
| | **Are the DSLs replacing C++ templates?** |
| |
|
| | TL;DR: No - but also yes. The CUTLASS 4.0 release (CuTe DSL), along with all |
| | future extensions to our Python-native programming models, does not come at the |
| | expense of CUTLASS C++. CUTLASS 2.x and 3.x C++ APIs are both going to continue |
| | receiving fixes and updates for the architectures we support them for. However, |
| | CUTLASS 4.x CuTe DSL is fully isomorphic in its programming model and performance |
| | with CuTe C++ for Blackwell, and it is our hope that the community embraces this |
| | for much easier while still equally performant custom kernel development. This is |
| | why we are releasing CuTe DSL with support for all architectures starting with the |
| | NVIDIA Ampere Architecture. |
| |
|
| | **What is the difference between CuTe DSL, CUTLASS Python, and CUTLASS DSLs?** |
| |
|
| | CUTLASS Python was the Python interface for instantiating C++ kernels via a Python |
| | frontend. This is now deprecated with the release of CUTLASS 4.0. CUTLASS DSLs are |
| | a family of Python DSLs for native device programming in Python. Currently, this is |
| | limited to our initial release of CuTe DSL, but future versions will include higher-level |
| | abstractions that gradually trade off control for convenience. |
| |
|
| | **What should I learn, CUTLASS C++ or the Python DSLs?** |
| |
|
| | We believe the Python DSLs will significantly improve the learning curve and recommend starting |
| | with them for all newcomers, as they eliminate the inherent complexity of learning C++ |
| | metaprogramming for GPU kernel programming. Since CuTe C++ and CuTe DSL share fully isomorphic |
| | programming models and patterns, any knowledge gained can eventually be applied to C++. |
| |
|
| | **Where will the code live? PIP wheel or GitHub repo? Do I have to build it myself?** |
| |
|
| | This is a major change compared to CUTLASS C++ and Python DSLs. Going forward, |
| | the GitHub code only exists as a way for users to file issues and pull requests against. |
| | While it can be used with the pip wheel, we do not recommend most users do so unless they are |
| | hacking on the DSL itself. For all other users, we recommend they |
| | simply ``pip install nvidia-cutlas-dsl`` and use the pip wheel as the single source |
| | of truth for the dialect compiler and DSL implementation. CUTLASS GitHub repository will |
| | contain a ``requirements.txt`` file pinning the version of the wheel consistent with the state |
| | of the OSS repository (please see :doc:`quick_start`). This means getting started with |
| | CUTLASS is easier than ever: no more CMake command lines to learn and no more builds to kick |
| | off. Simply install the pip wheel and start running the examples. |
| |
|
| | Migration |
| | |
| |
|
| | **Should I port my code from C++ templates to Python?** |
| |
|
| | Almost certainly not, unless you need extremely fast JIT times for your kernel and C++ compile times |
| | are a blocker for you. The 2.x and 3.x APIs will continue to be supported, and Nvidia's Hopper and |
| | Blackwell architectures 3.x will continue to improve in terms of features |
| | and performance. |
| |
|
| | **Are portability promises different with Python?** |
| |
|
| | For the initial release while the DSL is still in beta, we do not promise any portability |
| | as we may make changes to the DSL itself. While we do not expect any changes to the CuTe operations, |
| | the DSL utilities, decorators, helper classes like pipelines and schedulers may change as we refine them |
| | with community feedback. We encourage users to file issues and discussions on GitHub during this |
| | beta period with their feedback! |
| |
|
| | In the long term, we plan to continue to treat the OSS community with care. |
| | Just like the prior history of CUTLASS, we plan not to break users unless necessary, |
| | but we reserve the right to make limited breaking changes in case we believe it is a |
| | net benefit to the community and project. These will be announced ahead of time and/or |
| | clearly highlighted in the CHANGELOG of each release. |
| |
|
| | Technical |
| | |
| | **What NVIDIA architectures will it support?** |
| |
|
| | CuTe DSL will support all NVIDIA GPU architectures starting with NVIDIA Ampere Architecture (SM80). |
| |
|
| | **Will it be compatible with DL frameworks (e.g., PyTorch, JAX)?** |
| |
|
| | Yes, we will provide utilities to convert from DLPack-supported tensor formats |
| | to ``cute.Tensor``. This should allow a user to never have to leave Python |
| | when writing model code in their framework of choice. Our JAX interoperability story is not |
| | as strong as PyTorch's today, however, we are actively working on improving it |
| | and welcome contributions in this space. |
| |
|
| | **Does it compile to PTX or SASS?** |
| |
|
| | CuTe DSL compiles the program down to PTX. After that, we currently use the PTX compiler that |
| | ships with the CUDA toolkit to compile the PTX down to SASS. We plan to remove |
| | this limitation in the future and allow the use of the PTX JIT that is included in the |
| | CUDA driver in case a user does not have a CUDA toolkit installed. |
| |
|
| | **Do I need to use NVCC or NVRTC?** |
| |
|
| | No, the ``nvidia-cutlass-dsl`` wheel packages is everything needed to generate GPU kernels. It |
| | shares the driver requirements of the 12.9 toolkit which can be found |
| | `here <https://developer.nvidia.com/cuda-toolkit-archive>`__. |
| |
|
| | **How would one debug the code?** |
| |
|
| | Since CuTe DSL is not native python and an embedded DSL instead, tools like `pdb` |
| | cannot be used. However, if you have experience with GPU kernel programming, the debugging |
| | techniques will be nearly identical. Typically, compile time and runtime printing |
| | of types and values are the most expedient. Please see `documentation on printing <https://github.com/NVIDIA/cutlass/tree/main/examples/python/CuTeDSL/notebooks/print.ipynb>`__ |
| | to learn how to print types and values at both compile time and runtime. |
| | You can also use ``cuda-gdb`` to set breakpoints in the program and step through the execution |
| | or use tools such as ``compute-sanitizer`` to detect and triage bugs in your program. As the DSL |
| | matures, our source location tracking from Python user programs will also improve to provide |
| | more helpful source-level mapping when setting breakpoints and using other tools such as nsight. |
| |
|
| | **How would one implement warp specialization in CuTe DSL?** |
| |
|
| | Exactly the same way you would in C++ but in a Python-native syntax instead. |
| | Consult our :doc:`cute_dsl_general/dsl_control_flow` and |
| | `"Blackwell kernel example" <https://github.com/NVIDIA/cutlass/tree/main/examples/python/CuTeDSL/blackwell/dense_gemm_persistent.py>`__ |
| | for a detailed how-to guide. |
| |
|
| | **Can I call functions from other functions or use OOP?** |
| |
|
| | Yes. We frequently call functions from one another and set up class |
| | hierarchies to organize and modularize our code for pipelines and schedulers. |
| | Consult the :doc:`cute_dsl_general/dsl_introduction` documentation or our examples for more details. |
| |
|
| | License |
| | |
| | **What is the license for CuTe DSL and the associated GitHub samples?** |
| |
|
| | CuTe DSL components available `on Github <https://github.com/NVIDIA/cutlass/tree/main/python/CuTeDSL>`__ and via the nvidia-cutlass-dsl Python pip wheel |
| | are released under the `"NVIDIA Software End User License Agreement (EULA)" <https://github.com/NVIDIA/cutlass/tree/main/EULA.txt>`__. |
| | Because the pip package includes a compiler that shares several components with the CUDA Toolkit, |
| | it is subject to usage terms and restrictions similar to those of the CUDA SDK. Please refer to the EULA for specific terms of use. |
| |
|
| | CuTe DSL samples and Jupyter notbooks, released `on GitHub <https://github.com/NVIDIA/cutlass/tree/main/examples/python/CuTeDSL>`__ are provided under |
| | the BSD 3-Clause License and may be used and redistributed under those terms. This distinction ensures that developers have flexibility |
| | when using or modifying the code samples, independent of the compiler and runtime components governed by the EULA. |
| |
|
| | If you have any questions or need clarification, feel free to contact us. |
| |
|