An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks. https://arxiv.org/pdf/2106.04718.pdf
MIT License
MASS: Masked Sequence to Sequence Pre-training for Language Generation
JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf
[CVPR 2023] Official Implementation of X-Decoder for generalized decoding for pixel, image and la...
Unified-Modal Speech-Text Pre-Training for Spoken Language Processing
This repository contains resources for accessing the official benchmarks, codes, and checkpoints ...
Foundation Architecture for (M)LLMs
TensorFlow 2 library implementing Graph Neural Networks
Large-scale pretraining for dialogue
AI-powered ab initio biomolecular dynamics simulation
A flexible and efficient deep neural network (DNN) compiler that generates high-performance execu...
The implementation of DeBERTa
Dedicated to building industrial foundation models for universal data intelligence across industr...
CodeBERT
This is the implementation of the paper AdaMix: Mixture-of-Adaptations for Parameter-efficient Mo...