FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Tri Dao, Daniel Y. Fu, Stefano Ermon, Atri Rudra, Christopher Ré
Paper: https://arxiv.org/abs/2205.14135
추후 작성하기
반응형
LIST
'Artificial Intelligence' 카테고리의 다른 글
ChatGPT 사용시 프롬프트 엔지니어링 7요소 (0) | 2024.06.07 |
---|---|
GAN(Generative adversarial network) 이란? (0) | 2023.01.07 |
Top 10 Performance Tuning Practices for Pytorch (0) | 2022.09.21 |
ImageNet Dataset 압축해제 및 처리 방법 (0) | 2022.08.10 |
ZeRO-2 & DeepSpeed: Shattering barriers of deep learning speed & scale (0) | 2022.07.29 |