Papers Read on AI

Papers Read on AI header image 1
May 25, 2022  

Self-attention Does Not Need O(n2) Memory

May 25, 2022

We provide a practical implementation for accelerators that requires O( √ n) memory, is numerically stable, and is within a few percent of the runtime of the standard implementation of attention. We also demonstrate how to differentiate the function while remaining memory-efficient.

2021: Markus N. Rabe, Charles Staats