OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM
MIT License
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
CodeBERT
This repository contains resources for accessing the official benchmarks, codes, and checkpoints ...
Large-scale pretraining for dialogue
An efficient implementation of the popular sequence models for text generation, summarization, an...
[CVPR 2023] Official Implementation of X-Decoder for generalized decoding for pixel, image and la...
Dedicated to building industrial foundation models for universal data intelligence across industr...
Unified-Modal Speech-Text Pre-Training for Spoken Language Processing
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Grounded Language-Image Pre-training
The implementation of DeBERTa
AI-powered ab initio biomolecular dynamics simulation
MMdnn is a set of tools to help users inter-operate among different deep learning frameworks. E.g...
JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf
A flexible and efficient deep neural network (DNN) compiler that generates high-performance execu...