nvidia-cudnn-frontend1.17.0
Published
CUDNN FrontEnd python library
pip install nvidia-cudnn-frontend
Package Downloads
Authors
Project URLs
Requires Python
>=3.9
cuDNN FrontEnd(FE)
cuDNN FE is the modern, open-source entry point to the NVIDIA cuDNN library and high performance open-source kernels. It provides a C++ header-only library and a Python interface to access the powerful cuDNN Graph API and open-source kernels.
🚀 Embracing Open Source
We will begin open-sourcing kernels based on customer needs, with the goal to educate developers and enable them to customize as needed.
We are now shipping OSS kernels, allowing you to inspect, modify, and contribute to the core logic. Check out our latest implementations:
- GEMM + Amax: Optimized FP8 matrix multiplication with absolute maximum calculation.
- GEMM + SwiGLU: High-performance implementation of the SwiGLU activation fused with GEMM.
- NSA: Native Sparse attention as described in the Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention.
Key Features
- Unified Graph API: Create reusable, persistent
cudnn_frontend::graph::Graphobjects to describe complex subgraphs. - Ease of Use: Simplified C++ and Python bindings (via
pybind11) that abstract away the boilerplate of the backend API. - Performance: Built-in autotuning and support for the latest NVIDIA GPU architectures.
Installation
🐍 Python
The easiest way to get started is via pip:
pip install nvidia_cudnn_frontend
Requirements:
- Python 3.8+
- NVIDIA driver and CUDA Toolkit
⚙️ C++ (Header Only)
Since the C++ API is header-only, integration is seamless. Simply include the header in your compilation unit:
#include <cudnn_frontend.h>
Ensure your include path points to the include/ directory of this repository.
Building from Source
If you want to build the Python bindings from source or run the C++ samples:
1. Dependencies
python-dev(e.g.,apt-get install python-dev)- Dependencies listed in
requirements.txt(pip install -r requirements.txt)
2. Python Source Build
pip install -v git+https://github.com/NVIDIA/cudnn-frontend.git
Environment variables CUDAToolkit_ROOT and CUDNN_PATH can be used to override default paths.
3. C++ Samples Build
mkdir build && cd build
cmake -DCUDNN_PATH=/path/to/cudnn -DCUDAToolkit_ROOT=/path/to/cuda ../
cmake --build . -j16
./bin/samples
Documentation & Examples
- Developer Guide: Official NVIDIA Documentation
- C++ Samples: See
samples/cppfor comprehensive usage examples. - Python Samples: See
samples/pythonfor pythonic implementations.
🤝 Contributing
We strictly welcome contributions! Whether you are fixing a bug, improving documentation, or optimizing one of our new OSS kernels, your help makes cuDNN better for everyone.
- Check the Contribution Guide for details.
- Fork the repo and create your branch.
- Submit a Pull Request.
Debugging
To view the execution flow and debug issues, you can enable logging via environment variables:
# Log to stdout
export CUDNN_FRONTEND_LOG_INFO=1
export CUDNN_FRONTEND_LOG_FILE=stdout
# Log to a file
export CUDNN_FRONTEND_LOG_INFO=1
export CUDNN_FRONTEND_LOG_FILE=execution_log.txt
Alternatively, you can control logging programmatically via cudnn_frontend::isLoggingEnabled().
License
This project is licensed under the MIT License.