GRadient-INformed MoE
OTHER License
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
Tutel MoE: An Optimized Mixture-of-Experts Implementation
Repo for WWW 2022 paper: Progressively Optimized Bi-Granular Document Representation for Scalable...
Official Codebase for MEGAVERSE: (published in ACL: NAACL 2024)
Generic accelerator for Build your own copilot
Workshop "Build a Retail Copilot Code-First on Azure AI": attendee and presenter resources.
A metric suite leveraging the logical inference capabilities of LLMs, for radiology report genera...
Large Language Models for Supply Chain Optimization
A framework for standardizing evaluations of large foundation models, beyond single-score reporti...