<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"><channel><title>local-llm.net</title><description>The definitive guide to deploying AI locally. Guides, tutorials, comparisons, and news.</description><link>https://local-llm.net/</link><language>en-us</language><item><title>Building an Offline AI Phone App with Llamafu and Flutter</title><link>https://local-llm.net/blog/building-offline-app-llamafu/</link><guid isPermaLink="true">https://local-llm.net/blog/building-offline-app-llamafu/</guid><description>A step-by-step tutorial for building a fully offline AI assistant app for Android and iOS using Llamafu for on-device inference and Flutter for the UI. No internet required.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Fine-Tuning a Customer Support Model with Unsloth in 4 Hours</title><link>https://local-llm.net/blog/fine-tuning-unsloth-4-hours/</link><guid isPermaLink="true">https://local-llm.net/blog/fine-tuning-unsloth-4-hours/</guid><description>An end-to-end walkthrough of fine-tuning a language model for customer support using Unsloth and QLoRA. From dataset preparation to GGUF export and Ollama deployment — all on a single consumer GPU.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Why We Built local-llm.net: A Community Hub for Everyone Running AI Locally</title><link>https://local-llm.net/blog/introducing-local-llm-net/</link><guid isPermaLink="true">https://local-llm.net/blog/introducing-local-llm-net/</guid><description>Announcing local-llm.net — the community-driven guide to deploying AI locally. We cover the entire ecosystem because the local AI movement belongs to everyone.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Local AI for Creative Writers: KoboldCpp, SillyTavern, and Uncensored Models</title><link>https://local-llm.net/blog/local-ai-creative-writers/</link><guid isPermaLink="true">https://local-llm.net/blog/local-ai-creative-writers/</guid><description>A non-technical guide for fiction writers who want AI tools that respect creative freedom and privacy. How to set up KoboldCpp, SillyTavern, and uncensored models for brainstorming, worldbuilding, and prose generation.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Local AI Privacy Audit: What Data Do These Tools Actually Collect?</title><link>https://local-llm.net/blog/local-ai-privacy-audit/</link><guid isPermaLink="true">https://local-llm.net/blog/local-ai-privacy-audit/</guid><description>An honest audit of telemetry, data collection, and privacy practices across Ollama, LM Studio, Jan, GPT4All, and Open WebUI. What runs locally does not always stay local.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>The 2026 Local AI Model Tier List: Every Model Ranked by Use Case</title><link>https://local-llm.net/blog/model-tier-list-2026/</link><guid isPermaLink="true">https://local-llm.net/blog/model-tier-list-2026/</guid><description>An opinionated S/A/B/C/D tier ranking of every major local AI model across six categories: chat, coding, reasoning, creative writing, vision, and embeddings. Updated quarterly.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>The RAG Stack That Actually Works: Ollama + ChromaDB + LangChain</title><link>https://local-llm.net/blog/rag-stack-that-works/</link><guid isPermaLink="true">https://local-llm.net/blog/rag-stack-that-works/</guid><description>A step-by-step guide to building a Retrieval-Augmented Generation pipeline that actually works in production. Ollama for inference, ChromaDB for vectors, LangChain for orchestration — with solutions to every common pain point.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>I Replaced ChatGPT with a Fully Local Stack — Here&apos;s What Happened</title><link>https://local-llm.net/blog/replaced-chatgpt-local/</link><guid isPermaLink="true">https://local-llm.net/blog/replaced-chatgpt-local/</guid><description>A 30-day experiment ditching ChatGPT Plus for a fully local AI setup. What worked, what failed, exact costs, and an honest verdict on whether local AI is ready for daily use.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>RTX 3090 in 2026: Still the Best Value GPU for Local AI</title><link>https://local-llm.net/blog/rtx-3090-best-value-2026/</link><guid isPermaLink="true">https://local-llm.net/blog/rtx-3090-best-value-2026/</guid><description>Benchmarking the used RTX 3090 against the RTX 4090 and RTX 5090 for local AI inference. The 3090&apos;s 24GB VRAM at $500-800 used makes it the unbeatable value pick for running large language models locally.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Self-Hosted ChatGPT for Your Team: Open WebUI + Ollama Deployment Guide</title><link>https://local-llm.net/blog/self-hosted-chatgpt-team/</link><guid isPermaLink="true">https://local-llm.net/blog/self-hosted-chatgpt-team/</guid><description>Deploy a multi-user ChatGPT alternative for your team using Open WebUI and Ollama. Complete guide covering Docker, HTTPS, authentication, model management, and cost analysis vs ChatGPT Team.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>The State of Local AI in 2026: Everything Has Changed</title><link>https://local-llm.net/blog/state-of-local-ai-2026/</link><guid isPermaLink="true">https://local-llm.net/blog/state-of-local-ai-2026/</guid><description>A comprehensive look at where local AI stands in 2026 — from 70B models on consumer GPUs to mobile inference, and everything that shifted since the chaotic early days of 2024.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>How to Choose the Right Local LLM for Your Use Case</title><link>https://local-llm.net/guides/choose-right-local-llm/</link><guid isPermaLink="true">https://local-llm.net/guides/choose-right-local-llm/</guid><description>A decision framework for selecting the best local LLM based on your task, hardware, and requirements. Includes model comparisons, quantization guide, and VRAM recommendations.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Docker and Kubernetes for Local AI: Container Deployment Guide</title><link>https://local-llm.net/guides/docker-kubernetes-local-ai/</link><guid isPermaLink="true">https://local-llm.net/guides/docker-kubernetes-local-ai/</guid><description>Deploy local LLMs with Docker and Kubernetes. Covers Ollama Docker setup, Open WebUI compose stacks, NVIDIA Container Toolkit, GPU passthrough, and Kubernetes GPU scheduling.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Enterprise Local AI: Deploying LLMs for Your Organization</title><link>https://local-llm.net/guides/enterprise-local-ai-deployment/</link><guid isPermaLink="true">https://local-llm.net/guides/enterprise-local-ai-deployment/</guid><description>Deploy local LLMs for enterprise use. Covers architecture patterns, vLLM with NVIDIA GPUs, multi-user interfaces with LibreChat, security hardening, compliance considerations, and cost analysis.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Fine-Tuning Your Own Local Model: From Data to Deployment</title><link>https://local-llm.net/guides/fine-tuning-local-model/</link><guid isPermaLink="true">https://local-llm.net/guides/fine-tuning-local-model/</guid><description>Learn when and how to fine-tune a local LLM. Covers dataset preparation, QLoRA training with Unsloth, evaluation, GGUF export, and deployment with Ollama.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Your First Local AI in 5 Minutes: Ollama Quickstart</title><link>https://local-llm.net/guides/first-local-ai-5-minutes/</link><guid isPermaLink="true">https://local-llm.net/guides/first-local-ai-5-minutes/</guid><description>Install Ollama and start chatting with an AI model on your own machine in under 5 minutes. No GPU required. Works on Mac, Windows, and Linux.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Local AI Code Assistant: Setting Up Copilot Without the Cloud</title><link>https://local-llm.net/guides/local-ai-code-assistant/</link><guid isPermaLink="true">https://local-llm.net/guides/local-ai-code-assistant/</guid><description>Set up a fully local AI code assistant using Continue with Ollama in VS Code, Tabby for self-hosted completions, and Aider for terminal-based coding. Includes model benchmarks and configuration.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Local AI Hardware Guide: What Do You Need to Run LLMs?</title><link>https://local-llm.net/guides/local-ai-hardware-guide/</link><guid isPermaLink="true">https://local-llm.net/guides/local-ai-hardware-guide/</guid><description>Complete hardware guide for running local LLMs. GPU recommendations by budget, VRAM requirements, CPU-only setups, Apple Silicon performance, and used GPU buying advice.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>The Local AI Stack: Choosing Your Engine, UI, and Framework</title><link>https://local-llm.net/guides/local-ai-stack/</link><guid isPermaLink="true">https://local-llm.net/guides/local-ai-stack/</guid><description>Understand the three-layer local AI architecture and choose the right inference engine, user interface, and application framework for your needs. Includes 5 reference stacks for common scenarios.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Local Image Generation: Stable Diffusion, FLUX, and ComfyUI Guide</title><link>https://local-llm.net/guides/local-image-generation/</link><guid isPermaLink="true">https://local-llm.net/guides/local-image-generation/</guid><description>Generate images locally with Stable Diffusion, FLUX, and ComfyUI. Covers setup, ControlNet, LoRAs, VRAM management, prompt engineering, and workflow optimization.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Running Local LLMs on Linux: Ubuntu, Fedora, and Arch Guide</title><link>https://local-llm.net/guides/local-llm-linux/</link><guid isPermaLink="true">https://local-llm.net/guides/local-llm-linux/</guid><description>Complete guide to running local LLMs on Linux. Covers NVIDIA CUDA and AMD ROCm setup, Ollama installation, building llama.cpp from source, systemd services, and performance tuning for Ubuntu, Fedora, and Arch.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Running Local LLMs on macOS: Apple Silicon Optimization Guide</title><link>https://local-llm.net/guides/local-llm-macos-apple-silicon/</link><guid isPermaLink="true">https://local-llm.net/guides/local-llm-macos-apple-silicon/</guid><description>Complete guide to running local LLMs on macOS with Apple Silicon. Covers Ollama, MLX, LM Studio, unified memory optimization, and model recommendations for M1 through M4 chips.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Running Local LLMs on Windows: Complete Setup Guide</title><link>https://local-llm.net/guides/local-llm-windows/</link><guid isPermaLink="true">https://local-llm.net/guides/local-llm-windows/</guid><description>Step-by-step guide to running local LLMs on Windows with NVIDIA and AMD GPUs. Covers Ollama, LM Studio, WSL2, CUDA setup, and troubleshooting common issues.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Building a Local RAG Chatbot: Documents, Embeddings, and Retrieval</title><link>https://local-llm.net/guides/local-rag-chatbot/</link><guid isPermaLink="true">https://local-llm.net/guides/local-rag-chatbot/</guid><description>Build a fully local RAG (Retrieval-Augmented Generation) chatbot that answers questions about your documents. Covers architecture, chunking strategies, embedding models, vector databases, and prompt engineering.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Building a Local Voice Assistant: Whisper + LLM + TTS</title><link>https://local-llm.net/guides/local-voice-assistant/</link><guid isPermaLink="true">https://local-llm.net/guides/local-voice-assistant/</guid><description>Build a fully local voice assistant pipeline with speech-to-text (Whisper.cpp), an LLM for processing (Ollama), and text-to-speech (Piper/Kokoro). Includes latency optimization and wake word detection.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>Open WebUI + Ollama: Self-Hosted ChatGPT in 10 Minutes</title><link>https://local-llm.net/guides/open-webui-ollama-setup/</link><guid isPermaLink="true">https://local-llm.net/guides/open-webui-ollama-setup/</guid><description>Set up Open WebUI with Ollama for a self-hosted ChatGPT-like experience. Covers Docker Compose installation, model configuration, user management, plugins, and HTTPS reverse proxy setup.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item><item><title>What Is a Local LLM? The Complete Beginner&apos;s Guide</title><link>https://local-llm.net/guides/what-is-a-local-llm/</link><guid isPermaLink="true">https://local-llm.net/guides/what-is-a-local-llm/</guid><description>Learn what a local LLM is, how it works, what hardware you need, and why running AI on your own machine matters for privacy, cost, and control.</description><pubDate>Wed, 08 Apr 2026 00:00:00 GMT</pubDate></item></channel></rss>