Tutel MoE: An Optimized Mixture-of-Experts Implementation
MIT License
AICI: Prompts as (Wasm) Programs
Community for applying LLMs to robotics and a robot simulator with ChatGPT integration
MSCCL++: A GPU-driven communication stack for scalable AI applications
To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention,...
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
Generation of protein sequences and evolutionary alignments via discrete diffusion models
A Python package for generating concise, high-quality summaries of a probability distribution
Subseasonal forecasting models
Repo for WWW 2022 paper: Progressively Optimized Bi-Granular Document Representation for Scalable...
Table Transformer (TATR) is a deep learning model for extracting tables from unstructured documen...
Building modular LMs with parameter-efficient fine-tuning.
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using S...
Foundation Architecture for (M)LLMs