A simple cross attention that updates both the source and target in one step
MIT License
Implementation of TimeSformer from Facebook AI, a pure attention-based solution for video classif...
Implementation of Kronecker Attention in Pytorch
Summary of related papers on visual attention. Related code will be released based on Jittor grad...
Implementation of Spear-TTS - multi-speaker text-to-speech attention network, in Pytorch
Implementation of Flash Attention in Jax
Implementation of Perceiver, General Perception with Iterative Attention, in Pytorch
Implementation of Perceiver AR, Deepmind's new long-context attention network based on Perceiver ...
An implementation of local windowed attention for language modeling
Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attenti...
Implementation of Deformable Attention in Pytorch from the paper "Vision Transformer with Deforma...
Implementation of Memory-Compressed Attention, from the paper "Generating Wikipedia By Summarizin...
A Pytorch implementation of Attention on Attention module (both self and guided variants), for Vi...
Implementation of Perceiver, General Perception with Iterative Attention in TensorFlow
Implementation of OmniNet, Omnidirectional Representations from Transformers, in Pytorch
Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch