Pytorch implementation of the hamburger module from the ICLR 2021 paper "Is Attention Better Than Matrix Decomposition"
MIT License
An implementation of (Induced) Set Attention Block, from the Set Transformers paper
Implementation of the 😇 Attention layer from the paper, Scaling Local Self-Attention For Paramete...
Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI
Implementation of ResMLP, an all MLP solution to image classification, in Pytorch
Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less...
Pytorch reimplementation of Molecule Attention Transformer, which uses a transformer to tackle th...
PyTorch extensions for fast R&D prototyping and Kaggle farming
Implementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
Implementation of Deformable Attention in Pytorch from the paper "Vision Transformer with Deforma...
Implementation of the Point Transformer layer, in Pytorch
Implementation of gMLP, an all-MLP replacement for Transformers, in Pytorch
Implementation of Hierarchical Transformer Memory (HTM) for Pytorch
An attempt at the implementation of Glom, Geoffrey Hinton's new idea that integrates concepts fro...
Implementation of Segformer, Attention + MLP neural network for segmentation, in Pytorch
Implementation of the Hybrid Perception Block and Dual-Pruned Self-Attention block from the ITTR ...