Implementation of Axial attention - attending to multi-dimensional data efficiently
MIT License
Fully featured implementation of Routing Transformer
Reformer, the efficient Transformer, in Pytorch
CCNet: Criss-Cross Attention for Semantic Segmentation (TPAMI 2020 & ICCV 2019).
Axial Positional Embedding for Pytorch
Transformer based on a variant of attention that is linear complexity in respect to sequence length