🏠

Local LLM Hub

Run LLMs on your own hardware. Find the right launcher, engine, and configuration for your setup.

NVIDIA-first. Mac-strong. Pick your GPU, get your stack.

Quick Start: Mac (16GB RAM (M1/M2/M3 Base))

Beginnergguf
ollama + llama.cpp
Quant: Q4_K_M
7B models (Q4) run smoothly. Metal supported
GUIgguf
lm-studio + llama.cpp
Quant: Q4_K_M
Mac native. Easy to use
Apple Nativemlx
mlx-community + mlx
Quant: 4bit
Apple optimized. Limited model support
8 local LLM tools
NameRoleBackendsFormatsScoreInstall
Text Generation WebUI
Gradio web UI for running Large Language Models
Launchercuda, metal, rocm...gguf, gptq...97 (A+)🍎🐧🪟
KoboldCpp
Easy-to-use AI text generation with llama.cpp backend
Launchercuda, metal, rocm...gguf, ggml92 (A+)🍎🐧🪟
Ollama
Get up and running with large language models locally
Launchercuda, metal, rocm...gguf90 (A+)🍎🐧🪟
Jan
Open-source ChatGPT alternative that runs offline
Launchercuda, metal, cpu...gguf87 (A)🍎🐧🪟
LM Studio
Discover, download, and run local LLMs with a beautiful GUI
Launchercuda, metal, cpu...gguf87 (A)🍎🐧🪟
LocalAI
Free, open-source OpenAI alternative with local inference
Launchercuda, metal, rocm...gguf, safetensors77 (B+)🍎🐧
GPT4All
Free-to-use, locally running, privacy-aware chatbot
Launchercuda, metal, cpugguf72 (B)🍎🐧🪟
Open WebUI
User-friendly WebUI for LLMs with Ollama/OpenAI support
Launchercuda, metal, rocm...gguf62 (C+)🍎🐧