Transformer: PyTorch Implementation of "Attention Is All You Need"
Reformer, the efficient Transformer, in Pytorch
PyTorch and TensorFlow/Keras image models with automatic weight conversions and equal API/impleme...
TensorFlow 2.X reimplementation of CvT: Introducing Convolutions to Vision Transformers, Haiping ...
Transformer from Scratch using PyTorch
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
Sinkhorn Transformer - Practical implementation of Sparse Sinkhorn Attention
A PyTorch implementation of the Transformer model in "Attention is All You Need".
This repository contains demos I made with the Transformers library by HuggingFace.
Transformer based on a variant of attention that is linear complexity in respect to sequence length
Explainability for Vision Transformers
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating poin...
🍀 Pytorch implementation of various Attention Mechanisms, MLP, Re-parameter, Convolution, which i...