1) Detect hardware
Reads RAM, CPU cores, GPU backend, and VRAM so recommendations match your machine.
ForeLLM reads your RAM, CPU, and GPU, scores hundreds of models by fit and speed, and tells you which will run well locally. TUI, CLI, REST API, and an optional desktop GUI.
Three steps from hardware to running models.
Reads RAM, CPU cores, GPU backend, and VRAM so recommendations match your machine.
Ranks by quality, speed, fit, and context so the list is useful, not generic.
Use top picks with Ollama, llama.cpp, or MLX—copy the command and go.
What you get out of the box.
forellm serve for integrations and automation.Copy the command for your platform and run it.
scoop install forellm
brew install forellm
docker build -t forellm . && docker run --rm -it forellm
From repo root. Run CLI: docker run --rm forellm fit --json
curl -fsSL https://raw.githubusercontent.com/emireln/forellm/main/install.sh | sh
git clone https://github.com/emireln/forellm.git && cd forellm && cargo build --release
Requires Rust. Binary: target/release/forellm