Sitemap
Tech AI Chat

Chat on Technology Archive and Insights

Member-only story

Learning Generative AI

Make Local MCP and Run on Local LLM

Experimenting Generative AI Without The Need of Internet

8 min readJun 9, 2025

--

Photo by Domenico Loia on Unsplash

We explored ChatGPT, a powerful large language model (LLM) hosted in the cloud. We also learned about MCP (Model Context Protocol) — the next-generation API for AI, often dubbed the “USB-C of AI.” Like ChatGPT, MCP services are also cloud-based.

But what if you want to understand the concept, experiment with it, and build your own solutions — all without relying on paid cloud access?

How can you connect your own API, and create a local MCP setup that works with your locally running LLM?

Yes — this article shows you exactly how to do that.

Setup local LLM

An LLM (Large Language Model) is simply a model that allows you to interact with a computer using natural human language — and receive responses in the same way. ChatGPT is one example of this.

If you have an LLM running locally, you can carry out these natural language conversations entirely offline. This is made possible with tools like Ollama, which lets you download and serve models such as DeepSeek and LLaMA 3 directly on your machine.

--

--

Elye - A One Eye Dev By His Grace
Elye - A One Eye Dev By His Grace

Written by Elye - A One Eye Dev By His Grace

Sharing Software, Life and Faith Journey. Follow me on Twitter/X to access to my article free

Responses (1)