Mar 26, 2023 Paper Summary #8 - FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness Oct 10, 2022 Paper Summary #7 - Efficient Transformers: A Survey Apr 18, 2021 Paper Summary #1 - Attention Is All You Need