Tags: #ai-infrastructure
MemoriLabs/Memori
Memori provides SQL-native, LLM-agnostic memory infrastructure for AI agents, transforming execution and conversation into structured, persistent state.
inclusionAI/AReaL
AReaL is a scalable and flexible asynchronous reinforcement learning infrastructure designed to bridge foundation model training with modern LLM-based agent applications.
Tencent/AI-Infra-Guard
A full-stack AI Red Teaming platform for comprehensive security risk self-examination across AI ecosystems, including LLM jailbreak evaluation and infrastructure vulnerability scanning.
InternLM/xtuner
XTuner V1 is a next-generation training engine specifically designed for ultra-large-scale Mixture-of-Experts (MoE) models, offering superior efficiency and scalability.
LMCache/LMCache
LMCache is an LLM serving engine extension designed to significantly reduce Time-To-First-Token (TTFT) and boost throughput, especially for long-context scenarios, by intelligently reusing KV caches.
dstackai/dstack
A vendor-agnostic unified control plane for GPU provisioning and orchestration across clouds, Kubernetes, and on-prem for AI/ML workloads.
predibase/lorax
A multi-LoRA inference server designed to efficiently serve thousands of fine-tuned Large Language Models on a single GPU, drastically cutting serving costs while maintaining high throughput and low latency.
instill-ai/instill-core
An end-to-end AI infrastructure platform for data, model, and pipeline orchestration, designed to streamline the development of versatile AI-first applications.