Implementation of Metaformer, but in an autoregressive manner
MIT License
Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT
Implementation of Parti, Google's pure attention-based text-to-image neural network, in Pytorch
Implementation of ETSformer, state of the art time-series Transformer, in Pytorch
Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Py...
Implementation and replication of ProGen, Language Modeling for Protein Generation, in Jax
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch
Implementation of Graph Transformer in Pytorch, for potential use in replicating Alphafold2
Unofficial implementation of iTransformer - SOTA Time Series Forecasting using Attention networks...
Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch
Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
Implementation of Fast Transformer in Pytorch
Meta-Transformer for Unified Multimodal Learning
Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the se...
Implementation of Memformer, a Memory-augmented Transformer, in Pytorch
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch