Pytorch implementation of Compressive Transformers, from Deepmind
MIT License
DALL·E Mini - Generate images from a text prompt
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch
Implementation of Memorizing Transformers (ICLR 2022), attention net augmented with indexing and ...
Implementation of Phenaki Video, which uses Mask GIT to produce text guided videos of up to 2 min...
An implementation of local windowed attention for language modeling
A simple but complete full-attention transformer with a set of promising experimental features fr...
Implementation of SE3-Transformers for Equivariant Self-Attention, in Pytorch. This specific repo...
Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Re...
Implementation of E(n)-Transformer, which incorporates attention mechanisms into Welling's E(n)-E...
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with ...
Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch
Implementation of Block Recurrent Transformer - Pytorch
An implementation of Performer, a linear attention-based transformer, in Pytorch
A variant of Transformer-XL where the memory is updated not with a queue, but with attention
Implementation of the Equiformer, SE3/E3 equivariant attention network that reaches new SOTA, and...