File tree Expand file tree Collapse file tree 1 file changed +9
-5
lines changed
Expand file tree Collapse file tree 1 file changed +9
-5
lines changed Original file line number Diff line number Diff line change @@ -235,18 +235,22 @@ This project is licensed under the BSD 3-Clause License. See [LICENSE](LICENSE)
235235If you use Flash-DMA in your research, please cite:
236236
237237``` bibtex
238- @misc{flash_dma_2025,
239- title={Trainable Dynamic Mask Sparse Attention},
240- author={Jingze Shi and Yifan Wu and Bingheng Wu and Yiran Peng and Yuyu Luo},
241- year={2025},
242- url={https://github.com/SmallDoges/flash-dmattn}
238+ @misc{shi2025trainabledynamicmasksparse,
239+ title={Trainable Dynamic Mask Sparse Attention},
240+ author={Jingze Shi and Yifan Wu and Bingheng Wu and Yiran Peng and Liangdong Wang and Guang Liu and Yuyu Luo},
241+ year={2025},
242+ eprint={2508.02124},
243+ archivePrefix={arXiv},
244+ primaryClass={cs.AI},
245+ url={https://arxiv.org/abs/2508.02124},
243246}
244247```
245248
246249## Acknowledgments
247250
248251This project builds upon and integrates several excellent works:
249252
253+ - ** [ OpenSeek] ( https://github.com/FlagAI-Open/OpenSeek ) ** - Kernel development support
250254- ** [ Flash-Attention] ( https://github.com/Dao-AILab/flash-attention ) ** - Memory-efficient attention computation
251255- ** [ NVIDIA CUTLASS] ( https://github.com/NVIDIA/cutlass ) ** - High-performance matrix operations library
252256
You can’t perform that action at this time.
0 commit comments