Instructions to use kernels-community/flash-mla with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Kernels
How to use kernels-community/flash-mla with Kernels:
# !pip install kernels from kernels import get_kernel kernel = get_kernel("kernels-community/flash-mla") - Notebooks
- Google Colab
- Kaggle
| library_name: kernels | |
| license: mit | |
| This is the repository card of kernels-community/flash-mla that has been pushed on the Hub. It was built to be used with the [`kernels` library](https://github.com/huggingface/kernels). This card was automatically generated. | |
| ## How to use | |
| ```python | |
| # make sure `kernels` is installed: `pip install -U kernels` | |
| from kernels import get_kernel | |
| kernel_module = get_kernel("kernels-community/flash-mla") | |
| __version__ = kernel_module.__version__ | |
| __version__(...) | |
| ``` | |
| ## Available functions | |
| - `__version__` | |
| - `FlashMLASchedMeta` | |
| - `get_mla_metadata` | |
| - `flash_mla_with_kvcache` | |
| - `flash_attn_varlen_func` | |
| - `flash_attn_varlen_qkvpacked_func` | |
| - `flash_attn_varlen_kvpacked_func` | |
| - `flash_mla_sparse_fwd` | |
| ## Benchmarks | |
| Benchmarking script is available for this kernel. Run `kernels benchmark kernels-community/flash-mla`. | |