lime-cakes / memory_efficient_attention.pytorch Goto Github PK
View Code? Open in Web Editor NEWThis project forked from moskomule/memory_efficient_attention.pytorch
A human-readable PyTorch implementation of "Self-attention Does Not Need O(n^2) Memory" (Rabe&Staats'21).
License: Apache License 2.0