AI Tools

AI launchers and inference engines for local LLM deployment.

6 items
NameRoleTypeExecLanguagesScoreCold StartMemory
llama.cpp
LLM inference in C/C++ with minimal dependencies
EngineengineaotC, C++C+100ms50MB
llamafile
Distribute and run LLMs with a single file
EngineengineaotC, C++C-500ms100MB
Candle
Minimalist ML framework for Rust with GPU support
EngineenginejitRustD300ms200MB
ExLlamaV2
Fast inference library for running LLMs locally on NVIDIA GPUs
EngineengineaotPython, C++, CUDAD1000ms300MB
MLX
Apple's array framework for machine learning on Apple Silicon
EngineenginejitPython, C++, SwiftD500ms200MB
CTransformers
Python bindings for GGML models with GPU acceleration
EngineenginehybridPython, C++D800ms200MB