Running low-cost automations locally using n8n and Ollama

Derek BlankDerek Blank|June 3, 2025

We wanted to take the cloud-based workflows we explored in the Doing More with Less series and see if we could run them entirely locally. By "local," we mean everything runs on your own machine: no third-party servers, no cloud dependencies, and complete control over the data. And most importantly, local usually means free. This post details how we did just that using n8n and a locally-hosted LLM, all running privately on our own machines.


The Goal: Run an LLM automation workflow locally

We used a real-world example for this flow. The idea was to make creating a monthly email newsletter easier by summarising recent blog posts from an organisation's website. The workflow should:

  • Fetch all published WordPress posts from the previous month.
  • Summarise each one into 2-3 engaging, human-like sentences.
  • Compile it into styled HTML, and output the formatted email-ready markup.
  • Keep all data processing local. No third-party or cloud LLM services.

Tools We Used

  • n8n: Automation platform, self-hosted in Docker.
  • Ollama: Lightweight LLM runtime for running models locally on your own machine.
  • Llama 3: Meta's free large language model, great for summarising text.
  • Docker: Container platform used to host n8n and isolate services locally.

How the Automation Works

The workflow running n8n locally

The workflow running n8n locally

The core of this workflow uses n8n to automate the process. It's pretty simple: retrieve blog posts from WordPress, summarise them with an LLM, and then format the results into email-ready HTML. This flow can be scheduled to run at any interval (like the first day of every month), and the HTML output could be sent directly via email or another service for even further automation.


Why Local AI?

Running LLMs locally with Ollama offers major benefits:

  • Privacy: No external service sees our content.
  • Control: We can fine-tune prompts or swap models as needed.
  • Zero cost: Ollama is free, open-source software, while n8n is free to self-host for internal or personal use.

It should be noted that running an LLM locally requires a machine with sufficient hardware capabilities. Here's what to consider:

  • CPU: Modern multi-core processors can handle smaller models, but performance may be slower compared to GPUs.
  • GPU: A dedicated GPU with at least 8GB of VRAM is recommended for running larger models like Llama 3 efficiently. NVIDIA GPUs with CUDA support are commonly used.
  • RAM: At least 16GB of system memory is ideal, though more may be needed for larger models or multitasking.
  • Storage: Model files themselves can range from a few gigabytes to over 100GB depending on the model size.

MacBook M1/M2 laptops with 8GB RAM can also run local models, but 16GB or more is recommended for better results. Performance depends on model complexity and input size. Techniques like quantisation (reducing model precision) and batching (processing multiple inputs at once) can optimize speed and efficiency, even on mid-range hardware.


What's Next?

This particular system started as a way to automate one workflow, but the process itself is also valuable for rapidly prototyping workflows before putting them into production.

Once the idea of a workflow is validated, we can:

  • Run it in production on our own Docker host for full control
  • Alternatively, use n8n Cloud for hands-off hosting and autoscaling
  • Swap in a hosted LLM API like OpenAI if we later want better performance or language quality

The workflow itself is a simple one, but it's an example of how a repetitive time-consuming task can be automated quickly.


Want to build automated workflows for your organisation?

Get in touch to chat about your comms challenges, or follow us on LinkedIn to learn more about our work with automation.