#
cutlass
Here are 4 public repositories matching this topic...
Performance of the C++ interface of flash attention and flash attention v2 in large language model (LLM) inference scenarios.
gpu cuda inference nvidia cutlass mha multi-head-attention llm tensor-core large-language-model flash-attention flash-attention-2
-
Updated
Feb 27, 2025 - C++
Multiple GEMM operators are constructed with cutlass to support LLM inference.
-
Updated
Sep 27, 2024 - C++
pytorch implements block sparse
-
Updated
May 13, 2023 - C++
Improve this page
Add a description, image, and links to the cutlass topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the cutlass topic, visit your repo's landing page and select "manage topics."