Code for Limbacher, T. and Legenstein, R. (2020). H-Mem: Harnessing synaptic plasticity with Hebbian Memory Networks
GPL-3.0 License
MambaOut: Do We Really Need Mamba for Vision?
Ongoing research training transformer models at scale
中文分词 词性标注 命名实体识别 依存句法分析 成分句法分析 语义依存分析 语义角色标注 指代消解 风格转换 语义相似度 新词发现 关键词短语提取 自动摘要 文本分类聚类 拼音简繁转换 自然语言处理
Mamba state-space model
Incorporating Learnable Membrane Time Constant to Enhance Learning of Spiking Neural Networks
Re-implementation: Ask Me Anything: Dynamic Memory Networks for Natural Language Processing
A variant of Transformer-XL where the memory is updated not with a queue, but with attention
CVPR2023 - Activating More Pixels in Image Super-Resolution Transformer Arxiv - HAT: Hybrid Atten...
Continual Hyperparameter Selection Framework. Compares 11 state-of-the-art Lifelong Learning meth...
Implementation of Hierarchical Transformer Memory (HTM) for Pytorch
Pytorch code of for our CVPR 2018 paper "Neural Baby Talk"
GLM (General Language Model)
[CAAI AIR'24] Bilateral Reference for High-Resolution Dichotomous Image Segmentation
Official implementation for "Multimodal Chain-of-Thought Reasoning in Language Models" (stay tune...
Home of StarCoder: fine-tuning & inference!