My own attempt at a long context genomics model, leveraging recent advances in long context attention modeling (Flash Attention + other hierarchical methods)
MIT License
Implementation of Alphafold 3 in Pytorch
Unofficial implementation of iTransformer - SOTA Time Series Forecasting using Attention networks...
Implementation of Spear-TTS - multi-speaker text-to-speech attention network, in Pytorch
Fast and memory-efficient exact attention
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architectu...
Implementation of Enformer, Deepmind's attention network for predicting gene expression, in Pytorch
Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Py...
Implementation of Natural Speech 2, Zero-shot Speech and Singing Synthesizer, in Pytorch
Implementation of Voicebox, new SOTA Text-to-speech network from MetaAI, in Pytorch
To eventually become an unofficial Pytorch implementation / replication of Alphafold2, as details...
Implementation of E(n)-Transformer, which incorporates attention mechanisms into Welling's E(n)-E...
A repository with exploration into using transformers to predict DNA ↔ transcription factor binding
Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT
Memory optimization and training recipes to extrapolate language models' context length to 1 mill...
Implementation of MagViT2 Tokenizer in Pytorch