Fast and memory-efficient exact attention
BSD-3-CLAUSE License
Bot releases are hidden (Show)
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architectu...
Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)
Implementation of Make-A-Video, new SOTA text to video generator from Meta AI, in Pytorch
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch
Implementation of Flash Attention in Jax
LongLLaMA is a large language model capable of handling long contexts. It is based on OpenLLaMA a...
Implementation of GigaGAN, new SOTA GAN out of Adobe. Culmination of nearly a decade of research ...
An implementation of Performer, a linear attention-based transformer, in Pytorch
Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attenti...
Implementation of AudioLM, a SOTA Language Modeling Approach to Audio Generation out of Google Re...
Implementation of Block Recurrent Transformer - Pytorch
Implementation of the Equiformer, SE3/E3 equivariant attention network that reaches new SOTA, and...
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch
Implementation of the Transformer variant proposed in "Transformer Quality in Linear Time"
Implementation of Spear-TTS - multi-speaker text-to-speech attention network, in Pytorch