An easy to use, high performant CUDA powered LLM inference library.
BSD-3-CLAUSE License
Published by jarroddavis68 5 months ago
The following CUDA & llama.cpp runtime DLLs must be present on your target device for Infero to function properly. Please ensure they are placed in the same directory as your Infero executable file.
Computer vision library with focus on heterogeneous systems
HPC solver for nonlinear optimization problems
A curated list of awesome GPGPU (CUDA/OpenCL/Vulkan) resources
LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!
Object detection for video surveillance
Cross-platform, customizable multimedia/video processing framework. With strong GPU acceleration...
A high-performance inference system for large language models, designed for production environments.
A highly optimised C++ library for mathematical applications and neural networks.
CUDA C++ Core Libraries
This repository lists some awesome public Rust projects, Videos, Blogs and Jobs.
Classes enabling finmath-lib to run its Monte-Carlo models on Cuda GPUs
Real-time large scale dense visual SLAM system
Real-time dense visual SLAM system
An architecture for LLMs' continual-learning and long-term memories
3D Gaussian Splatting, reimagined: Unleashing unmatched speed with C++ and CUDA from the ground up!