Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT
MIT License
Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Re...
A simple but complete full-attention transformer with a set of promising experimental features fr...
Implementation of 🌻 Mirasol, SOTA Multimodal Autoregressive model out of Google Deepmind, in Pytorch
Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch
An implementation of local windowed attention for language modeling
Implementation of Parti, Google's pure attention-based text-to-image neural network, in Pytorch
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch
Unofficial implementation of iTransformer - SOTA Time Series Forecasting using Attention networks...
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architectu...
An implementation of Performer, a linear attention-based transformer, in Pytorch
Implementation of E(n)-Transformer, which incorporates attention mechanisms into Welling's E(n)-E...
Implementation of Phenaki Video, which uses Mask GIT to produce text guided videos of up to 2 min...
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch
Implementation of the Equiformer, SE3/E3 equivariant attention network that reaches new SOTA, and...