This is the implementation of the paper AdaMix: Mixture-of-Adaptations for Parameter-efficient Model Tuning (https://arxiv.org/abs/2205.12410).
APACHE-2.0 License
An efficient implementation of the popular sequence models for text generation, summarization, an...
CodeBERT
JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf
MASS: Masked Sequence to Sequence Pre-training for Language Generation
Foundation Architecture for (M)LLMs
MMdnn is a set of tools to help users inter-operate among different deep learning frameworks. E.g...
This repository contains resources for accessing the official benchmarks, codes, and checkpoints ...
Large-scale pretraining for dialogue
Dedicated to building industrial foundation models for universal data intelligence across industr...
Unified-Modal Speech-Text Pre-Training for Spoken Language Processing
The implementation of DeBERTa
AI-powered ab initio biomolecular dynamics simulation
A Multi-Task Dataset for Simulated Humanoid Control
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabil...