Minimal Decision Transformer Implementation written in Jax (Flax).
MIT License
Statistics for this project are still being loaded, please check back later.
Implementation of Long-Short Transformer, combining local and global inductive biases for attenti...
An implementation of Performer, a linear attention-based transformer, in Pytorch
Implementation of Feedback Transformer in Pytorch
GPT implementation in Flax
Pytorch implementation of Compressive Transformers, from Deepmind
Implementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Implementation of a Transformer, but completely in Triton
PyTorch code for Vision Transformers training with the Self-Supervised learning method DINO
Implementation of RT1 (Robotic Transformer) in Pytorch
Implementation of Fast Transformer in Pytorch
Implementation of TabTransformer, attention network for tabular data, in Pytorch
Official codebase for Decision Transformer: Reinforcement Learning via Sequence Modeling.
Generalized Decision Transformer for Offline Hindsight Information Matching (ICLR2022)