Usable implementation of Mogrifier, a circuit for enhancing LSTMs and potentially other networks, from Deepmind
MIT License
A simple to use pytorch wrapper for contrastive self-supervised learning on any neural network
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch
Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composi...
Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch
Implementation of N-Grammer, augmenting Transformers with latent n-grams, in Pytorch
GPT, but made only out of MLPs
Implementation for MatMul-free LM.
Official repository of the xLSTM.
Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Py...
Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI
Implementation of Segformer, Attention + MLP neural network for segmentation, in Pytorch
Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper
Implementation of Memformer, a Memory-augmented Transformer, in Pytorch
Naively combining transformers and Kolmogorov-Arnold Networks to learn and experiment
A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the param...