Ollama Local LLM Guide
5 posts in this series
Getting Started with Ollama: Your First Step to Running LLMs Locally
Want to run large language models on your own machine? This guide walks you through installing and configuring Ollama from scratch, covering multi-platform setup, model management, GPU acceleration, and API integration
Complete Guide to Ollama Model Management: Download, Switch, Delete & Version Control
Master Ollama model management with pull, run, list, rm commands. Learn version selection, batch deletion scripts, disk space optimization. Perfect for AI developers and OpenClaw deployers managing local LLM libraries.
Ollama Modelfile Parameters Explained: A Complete Guide to Creating Custom Models
A detailed guide to Ollama Modelfile's 10 core parameters, including optimization tips for temperature, num_ctx, and more. Includes 4 ready-to-use practical templates to help you create your own custom models.
Ollama API Calls: From curl to OpenAI SDK Compatible Interface
Learn two ways to call Ollama API: native REST API (curl) and OpenAI SDK compatible interface. Includes complete code examples, streaming response handling, and best practices guide
Ollama + Open WebUI: Build Your Own Local ChatGPT Interface (Complete Guide)
Step-by-step guide to setting up a ChatGPT-style AI interface locally with Ollama and Open WebUI. Covers installation, model selection, RAG knowledge base, API integration, and performance tuning. Get your local AI assistant running in 30 minutes.