Run AI models locally on your hardware. Privacy-first, offline-capable, open-source tools for self-hosted inference.
Run LLMs locally with one command — the easiest way to get AI running on your machine.
The C/C++ engine powering local AI — lightning-fast inference that Ollama and LM Studio build on.
Beautiful desktop app for running LLMs locally — discover, download, and chat with AI models.
Self-hosted ChatGPT-style interface for Ollama and OpenAI-compatible APIs.
High-throughput LLM serving engine — the production standard for GPU inference at scale.
Open-source ChatGPT alternative that runs 100% offline on your computer.
The Swiss Army knife of local AI — Gradio interface supporting every model format and backend.
Free, local, privacy-aware AI — run chatbots on consumer hardware with no GPU required.
Self-hosted OpenAI-compatible API — drop-in replacement for cloud AI in your infrastructure.
AI-powered creative writing suite — the go-to tool for interactive fiction and storytelling.
1–10 of 10 tools