Implementation of "SpikeGPT: Generative Pre-trained Language Model with Spiking Neural Networks"
BSD-2-CLAUSE License
Statistics for this project are still being loaded, please check back later.
中文对话0.2B小模型(ChatLM-Chinese-0.2B),开源所有数据集来源、数据清洗、tokenizer训练、模型预训练、SFT指令微调、RLHF优化等流程的全部代码。支持下游任务sf...
Official implementation code of the paper <AnyText: Multilingual Visual Text Generation And Editing>
Ongoing research training transformer models at scale
utilities for decoding deep representations (like sentence embeddings) back to text
[ACL 2022] LinkBERT: A Knowledgeable Language Model 😎 Pretrained with Document Links
GLM (General Language Model)
Source code for Grounded Adaptation for Zero-shot Executable Semantic Parsing
Public release of the TransCoder research project https://arxiv.org/pdf/2006.03511.pdf
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Home of StarCoder2!
Ongoing research training transformer language models at scale, including: BERT & GPT-2
TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones
Home of StarCoder: fine-tuning & inference!
ktrain is a Python library that makes deep learning and AI more accessible and easier to apply
Code for training and evaluation of the model from "Language Generation with Recurrent Generative...