Ecosystem & Stack: llamacpp
ollama/ollama
Easily run, manage, and interact with open-source large language models locally on your machine.
xorbitsai/inference
A unified, production-ready inference API for effortlessly deploying and serving open-source language, speech, and multimodal AI models across various environments.
AAswordman/Operit
Operit AI is the first fully-featured, independently running AI assistant application for Android, offering powerful tool-calling, deep search, and automation capabilities.
RunanywhereAI/runanywhere-sdks
A production-ready toolkit enabling developers to integrate private, offline, and fast on-device AI capabilities like LLMs, speech-to-text, and text-to-speech into their applications across various platforms.
Mobile-Artificial-Intelligence/maid
A free and open-source Android application for local and remote interaction with various large language models, emphasizing speed and user-friendliness.
ymcui/Chinese-LLaMA-Alpaca
An open-source project providing Chinese LLaMA and Alpaca large language models, enhanced with Chinese vocabulary and data for improved understanding and local deployment on CPU/GPU.
ymcui/Chinese-LLaMA-Alpaca-2
An open-source project providing Chinese LLaMA-2 and Alpaca-2 large language models with enhanced Chinese capabilities and support for ultra-long contexts up to 64K.
OpenMOSS/MOSS-TTS
An open-source AI model family for high-fidelity, expressive speech and sound generation across diverse real-world applications.
withcatai/node-llama-cpp
A Node.js library providing bindings for llama.cpp, enabling local AI model inference with advanced features like JSON schema enforcement and function calling.
edwko/OuteTTS
A versatile interface for OuteTTS models, providing flexible text-to-speech generation capabilities across various AI inference backends and hardware platforms.
janhq/cortex.cpp
A local AI API platform designed to run various AI models (vision, speech, language) on local hardware with an OpenAI-compatible API.
SciSharp/LLamaSharp
A cross-platform C#/.NET library for efficient local inference of large language models (LLMs) like LLaMA and LLAVA.
johnbean393/Sidekick
A native macOS app that allows users to chat with a local LLM, responding with information from files, folders, and websites on your Mac, ensuring privacy and offline functionality.