Tools for building attention networks for Tensorflow Mesh
MIT License
Implementation of Hierarchical Transformer Memory (HTM) for Pytorch
Implementation of TabTransformer, attention network for tabular data, in Pytorch
Implementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
Implementation of Infini-Transformer in Pytorch
Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI
Graph neural network message passing reframed as a Transformer with local attention
Pytorch reimplementation of Molecule Attention Transformer, which uses a transformer to tackle th...
Implementation of Feedback Transformer in Pytorch
Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/ab...
Implementation of Fast Transformer in Pytorch
Tencent Advertisement Algorithm Competition 2020 / 2020 腾讯广告算法大赛 TOP12
Implementation of Transformer in Transformer, pixel level attention paired with patch level atten...
Implementation of the Point Transformer layer, in Pytorch
An implementation of masked language modeling for Pytorch, made as concise and simple as possible
(Unofficial) Implementation of dilated attention from "LongNet: Scaling Transformers to 1,000,000...