Tags: #hardware-acceleration
Machine Learning Inference Library
Rust
2.2k
pykeio/ort
A high-performance Rust interface for hardware-accelerated machine learning inference and training with ONNX models, leveraging ONNX Runtime and pure-Rust backends.
AI Optimization Library
python
1.0k
intel/auto-round
AutoRound is an advanced quantization toolkit for Large Language Models (LLMs) and Vision-Language Models (VLMs), enabling high-accuracy, ultra-low-bit inference across diverse hardware.
CLI Tool / Local AI Inference Platform
llama.cpp
2.8k
janhq/cortex.cpp
A local AI API platform designed to run various AI models (vision, speech, language) on local hardware with an OpenAI-compatible API.
Replaces:
Details