🏠

Local LLM Hub

Run LLMs on your own hardware. Find the right launcher, engine, and configuration for your setup.

NVIDIA-first. Mac-strong. Pick your GPU, get your stack.

Quick Start: NVIDIA (8GB VRAM (RTX 3060/3070, RTX 4060))

Beginnergguf
ollama + llama.cpp
Quant: Q4_K_M
7B models (Q4) run smoothly. 13B is challenging
GUIgguf
lm-studio + llama.cpp
Quant: Q4_K_M
Easy GUI. Simple model management
Powergguf, gptq
text-generation-webui + llama.cpp
Quant: Q4_K_M or GPTQ-4bit
For users who need fine-grained control
1 local LLM tools
NameRoleBackendsFormatsScoreInstall
LLM (Python CLI)
Access large language models from the command-line
Toolcuda, metal, cpugguf52 (C-)🍎🐧