DevDuniya
May 16, 2025
In the rapidly growing world of AI and large language models (LLMs), many developers and enthusiasts are looking for local alternatives to cloud-based AI tools like ChatGPT or Bard. Enter Ollama – a fantastic way to run open-source LLMs like LLaMA, Mistral, and others on your own computer.
This blog is a complete beginner’s guide to:
✅ What is Ollama
✅ Why use it
✅ How to install Ollama on Windows, macOS, or Linux
✅ How to run AI models locally
✅ Useful Ollama commands
✅ Creating your own custom models (Modelfile)
✅ Integrating Ollama with other tools like LangChain or Python
Ollama is an open-source tool that allows you to run large language models locally on your computer with a simple command-line interface (CLI).
Think of it as your personal, offline ChatGPT that runs fully on your machine.
brew install ollama
Or, download directly from: https://ollama.com/download
.exe
file and follow the prompts.ollama --version
If installed correctly, it will show the version.
curl -fsSL https://ollama.com/install.sh | sh
Then run:
ollama run llama2
Once installed, open your terminal and simply run:
ollama run llama2
It will:
Example:
> ollama run llama2
>>> What is Laravel?
Laravel is a PHP web application framework...
You can explore models at: https://ollama.com/library
Popular models:
Model | Command | Description |
---|---|---|
LLaMA2 | ollama run llama2 |
Meta's LLM |
Mistral | ollama run mistral |
Lightweight & fast |
Gemma | ollama run gemma |
Google’s open-source LLM |
Code LLaMA | ollama run codellama |
Optimized for coding |
Phi | ollama run phi |
Lightweight, Microsoft model |
Want to try a different model?
ollama run mistral
You can run as many models as you want — just one at a time in a session.
Here are some useful commands:
Command | Description |
---|---|
ollama run [model] |
Run a model |
ollama pull [model] |
Download a model without running |
ollama list |
See downloaded models |
ollama rm [model] |
Remove a model |
ollama create [model-name] |
Create a custom model using a Modelfile |
ollama serve |
Run the Ollama server for API use |
ollama help |
Show help information |
ollama pull mistral
You can customize a model’s behavior with a Modelfile
.
Modelfile
FROM mistral
SYSTEM "You are an assistant that helps write Laravel PHP code."
Then create the model:
ollama create laravel-assistant -f Modelfile
Run it:
ollama run laravel-assistant
Now it will always guide users on Laravel-related questions!
Start the Ollama server:
ollama serve
Then send a request via curl:
curl http://localhost:11434/api/generate -d '{
"model": "mistral",
"prompt": "Explain Docker in simple terms"
}'
Or with Python:
import requests
response = requests.post(
'http://localhost:11434/api/generate',
json={
'model': 'mistral',
'prompt': 'What is Laravel?',
}
)
print(response.json()['response'])
Ollama works well with LangChain, enabling advanced workflows and chaining.
from langchain.llms import Ollama
llm = Ollama(model="mistral")
print(llm("Tell me a joke"))
If you want to free up space:
ollama rm llama2
ollama rm -a
On macOS/Linux:
~/.ollama/models
On Windows:
C:\Users\YourUsername\.ollama\models
Task | Command |
---|---|
Run a model | ollama run mistral |
List models | ollama list |
Remove model | ollama rm mistral |
Create custom model | ollama create my-model -f Modelfile |
Use API | ollama serve and call /api/generate |
Download model only | ollama pull model-name |
Ollama is a game-changer for developers, educators, and AI enthusiasts who want local control, privacy, and speed when working with powerful language models.
Whether you’re building an app, exploring AI, or just want your own offline ChatGPT — Ollama is the perfect starting point.
Drop them in the comments, or connect on GitHub or Twitter. Happy hacking! 💻✨
Would you like this blog in Markdown, PDF, or as a Notion page? I can format and export it for you.