Tags: #hardware-acceleration

Machine Learning Inference Library
Rust
2.2k

pykeio/ort

A high-performance Rust interface for hardware-accelerated machine learning inference and training with ONNX models, leveraging ONNX Runtime and pure-Rust backends.

AI Optimization Library
python
1.0k

intel/auto-round

AutoRound is an advanced quantization toolkit for Large Language Models (LLMs) and Vision-Language Models (VLMs), enabling high-accuracy, ultra-low-bit inference across diverse hardware.

CLI Tool / Local AI Inference Platform
llama.cpp
2.8k

janhq/cortex.cpp

A local AI API platform designed to run various AI models (vision, speech, language) on local hardware with an OpenAI-compatible API.

OSS Alternative

Explore the best open source alternatives to commercial software.

© 2026 OSS Alternative. hotgithub.com - All rights reserved.