Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackable.
APACHE-2.0 License
torchao: PyTorch Architecture Optimization (AO). Performant kernels that work with PyTorch.
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reprod...
Burn is a new comprehensive dynamic Deep Learning Framework built using Rust with extreme flexibi...
A JavaScript library like PyTorch, with GPU acceleration.
Run PyTorch LLMs locally on servers, desktop and mobile
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating poin...
Large Language Model Text Generation Inference
The Incredible PyTorch: a curated list of tutorials, papers, projects, communities and more relat...
In this repository, I will share some useful notes and references about deploying deep learning-b...
An unofficial styleguide and best practices summary for PyTorch
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
This repository contains demos I made with the Transformers library by HuggingFace.
Transformers4Rec is a flexible and efficient library for sequential and session-based recommendat...
Serve, optimize and scale PyTorch models in production
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on...