Everything you need to run AI-powered n8n workflows locally.
Run n8n and local AI models entirely on your machine, free of charge.
Your workflows and data never leave your computer.
No manual Docker commands, no config files to edit — just download and run.
Ships with Ollama pre-configured, so AI-powered workflows work out of the box.
Pull any model from Ollama's library or connect to LM Studio, llama.cpp, or a host Ollama instance.
Runs the official n8n Docker image with no integrations removed or disabled.
n8n diagnostics and version notifications are disabled to match the local-first philosophy.
Local LLMs ready to power your workflows.