Running LLMs Locally Just Got Way Better - Ollama + MCP

Tech With Tim
Tech With TimApr 12, 2026

Why It Matters

Local, tool‑enabled LLMs give enterprises privacy and cost control while unlocking the same automation capabilities previously limited to cloud services.

Key Takeaways

  • Install Ollama, then pull a tool‑calling LLM like Gwendolyn 3.5.
  • Verify your GPU/Unified memory can support model size before downloading.
  • Use Zapier MCP server to expose 8,000+ integrations to the local model.
  • Select models with tool‑calling capability; older LLMs lack actionable functions.
  • Bridge Ollama with MXGP client to enable real‑time tool calls securely.

Summary

The video walks viewers through setting up a private, locally‑run large language model using Ollama and connecting it to external services via the Zapier MCP server. It emphasizes that the combination delivers cloud‑level functionality—such as accessing Google, Notion, or Facebook Ads—without exposing data to third‑party APIs.

Key technical points include checking GPU or unified memory capacity, choosing a model that supports tool‑calling, and balancing parameter count against available RAM. The presenter demonstrates downloading the Gwendolyn 3.5 model (35 billion parameters) and explains why older LLMs without tool‑calling are unsuitable for automation tasks.

A practical example shows the MXGP client acting as a bridge between Ollama and Zapier’s MCP server, enabling real‑time calls to over 8,000 integrations. The speaker notes that each Zap counts as a credit on a Zapier plan, but the free tier typically suffices for modest usage, and highlights the speed differences when running 27‑billion versus 35‑billion‑parameter models on an M2 Max Mac.

The overall implication is that businesses and developers can now run powerful, privacy‑preserving AI agents on commodity hardware, cutting cloud costs while retaining the ability to automate workflows across a vast ecosystem of tools.

Original Description

Use the Zapier MCP server to connect to over 8000 applications/tools: https://bit.ly/4vn0jrC
If you want to run a local model that is free, private, and capable of connecting to all of your external tools, then keep watching this video. I'm going to show you how to run a capable local model on your own machine, and connect it to external services.
🎞 Video Resources 🎞
Ollama Download: https://ollama.com
🚀 Tools I Use
Get 10% off with code techwithtim
⏳ Timestamps ⏳
00:00 | Overview
00:34 | What is an Agent
02:25 | Ollama Setup
03:36 | Understanding Model Selection
09:01 | Running the Model
11:40 | Connecting Integrations (Zapier MCP)
14:37 | Running MCP Connector
19:51 | Running in Code (Langchain)
Hashtags
#Ollama #Zapier #Langchain
UAE Media License Number: 3635141

Comments

Want to join the conversation?

Loading comments...