<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>LLM Archives - Prsm Studio</title>
	<atom:link href="https://prsm-studio.com/en/tag/llm-en/feed/" rel="self" type="application/rss+xml" />
	<link>https://prsm-studio.com/en/tag/llm-en/</link>
	<description>automation · homeserver · side projects · game · gadgets · play</description>
	<lastBuildDate>Mon, 09 Mar 2026 07:57:06 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://prsm-studio.com/wp-content/uploads/2026/03/ic_launcher-playstore-150x150.png</url>
	<title>LLM Archives - Prsm Studio</title>
	<link>https://prsm-studio.com/en/tag/llm-en/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Even a Code-Illiterate Built It! Home Server Journey (4) — Running AI Locally with Ollama</title>
		<link>https://prsm-studio.com/en/code-illiterate-home-server-build-4-ollama-local-ai-en/</link>
					<comments>https://prsm-studio.com/en/code-illiterate-home-server-build-4-ollama-local-ai-en/#respond</comments>
		
		<dc:creator><![CDATA[Toaster]]></dc:creator>
		<pubDate>Mon, 09 Mar 2026 03:15:56 +0000</pubDate>
				<category><![CDATA[Computer Play]]></category>
		<category><![CDATA[Home Server]]></category>
		<category><![CDATA[LLM]]></category>
		<category><![CDATA[local AI]]></category>
		<category><![CDATA[Mini PC]]></category>
		<category><![CDATA[Ollama]]></category>
		<category><![CDATA[Open WebUI]]></category>
		<category><![CDATA[self-hosted AI]]></category>
		<category><![CDATA[SER9 MAX]]></category>
		<guid isPermaLink="false">https://prsm-studio.com/code-illiterate-home-server-build-4-ollama-local-ai-en/</guid>

					<description><![CDATA[<p>I installed Ollama and Open WebUI on my home server to run free local AI. Real benchmarks from my SER9 MAX mini PC, RAM-based model guide, and honest conclusions.</p>
<p>The post <a href="https://prsm-studio.com/en/code-illiterate-home-server-build-4-ollama-local-ai-en/">Even a Code-Illiterate Built It! Home Server Journey (4) — Running AI Locally with Ollama</a> appeared first on <a href="https://prsm-studio.com/en">Prsm Studio</a>.</p>
]]></description>
										<content:encoded><![CDATA[<h2>Running AI on My Own Server?</h2>
<p>ChatGPT, Gemini, Claude… everyone uses cloud AI. But have you ever thought:</p>
<p><strong>&#8220;If I run AI on my own computer, it&#8217;s free AND my data stays private?&#8221;</strong></p>
<p>That&#8217;s exactly right. Running a local LLM (Large Language Model) means no subscription fees and zero data leaving your machine. Perfect privacy.</p>
<p>But reality is… a bit different. I installed AI on <a href="/en/code-illiterate-home-server-build-1-ser9max-windows11-wsl2-docker-en/">my SER9 MAX mini PC from Episode 1</a>, and the honest verdict? <strong>&#8220;It works. But it&#8217;s slow.&#8221;</strong></p>
<figure class="wp-block-image size-large"><img fetchpriority="high" decoding="async" width="940" height="627" src="https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530416-1.jpg" alt="DeepSeek AI 인터페이스를 보여주는 MacBook으로 디지털 혁신을 선보입니다." class="wp-image-225" srcset="https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530416-1.jpg 940w, https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530416-1-300x200.jpg 300w, https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530416-1-768x512.jpg 768w" sizes="(max-width: 940px) 100vw, 940px" /><figcaption>Photo by Matheus Bertelli / Pexels</figcaption></figure>
<h2>Ollama — The Local LLM Engine</h2>
<p>Ollama is a tool that lets you run AI models on your own hardware. Sounds complicated? I had AI install it for me. A few terminal commands and done.</p>
<p>Once installed, one command — <code>ollama run qwen3:14b</code> — and the AI starts responding. The model downloads automatically, no configuration needed.</p>
<p>There are dozens of open-source models available: Llama, Qwen, Gemma, Mistral, DeepSeek… all free. Pick whichever fits your needs.</p>
<h2>Open WebUI — ChatGPT Interface in Your Browser</h2>
<p>Chatting in a terminal is honestly uncomfortable. So I installed <strong>Open WebUI</strong> — a program that gives you the exact same ChatGPT-like interface, running entirely on your server.</p>
<p>Again, AI handled the installation. One Docker container and it&#8217;s running.</p>
<p>The best part? <strong>My wife uses it too.</strong> Anyone on the same network can open a browser on their phone or tablet and start chatting. You can create separate accounts, so conversation history stays private for each person. With <a href="/en/code-illiterate-home-server-build-2-tailscale-remote-access-en/">Tailscale from Episode 2</a>, it&#8217;s accessible from anywhere.</p>
<figure class="wp-block-image size-large"><img decoding="async" width="433" height="650" src="https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530413-1.jpg" alt="DeepSeek 애플리케이션이 있는 대화형 AI 인터페이스를 보여주는 노트북 이미지." class="wp-image-226" srcset="https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530413-1.jpg 433w, https://prsm-studio.com/wp-content/uploads/2026/03/stock-30530413-1-200x300.jpg 200w" sizes="(max-width: 433px) 100vw, 433px" /><figcaption>Photo by Matheus Bertelli / Pexels</figcaption></figure>
<h2>Specs vs. Reality — This Is What Matters</h2>
<p>The most important question in local AI is <strong>&#8220;Can my hardware actually handle it?&#8221;</strong> Here are my real-world numbers.</p>
<h3>My Server Specs</h3>
<table>
<tr>
<th>Component</th>
<th>Specification</th>
</tr>
<tr>
<td>CPU</td>
<td>AMD Ryzen 7 255 (8 cores, 16 threads)</td>
</tr>
<tr>
<td>RAM</td>
<td>DDR5 32GB</td>
</tr>
<tr>
<td>GPU</td>
<td>Integrated (AMD Radeon 780M) — <strong>effectively none</strong></td>
</tr>
<tr>
<td>Storage</td>
<td>NVMe SSD 1TB</td>
</tr>
<tr>
<td>OS</td>
<td>Windows 11 + WSL2 (Linux)</td>
</tr>
</table>
<h3>Real Benchmarks (Qwen3 14B Model)</h3>
<table>
<tr>
<th>Metric</th>
<th>Value</th>
</tr>
<tr>
<td>Generation Speed</td>
<td><strong>5.5 tokens/sec</strong></td>
</tr>
<tr>
<td>Simple Question Response</td>
<td>~25 seconds</td>
</tr>
<tr>
<td>RAM Usage</td>
<td>~10GB</td>
</tr>
<tr>
<td>Quantization</td>
<td>Q4_K_M (9.3GB file)</td>
</tr>
</table>
<p>What ChatGPT answers in 1 second takes <strong>my server 25 seconds.</strong> That&#8217;s roughly 5-10x slower in real usage. Watching characters appear one by one is… a patience test.</p>
<h3>Why So Slow?</h3>
<p><strong>No dedicated GPU.</strong> AI inference is optimized for GPU computing, but my mini PC only has integrated graphics. I&#8217;ve confirmed that the AMD 780M iGPU can&#8217;t be used for AI acceleration under WSL2. Everything runs on <strong>CPU only</strong> — hence the speed.</p>
<p>With an NVIDIA GPU? The same model runs <strong>5-10x faster.</strong> An RTX 4060 can push 30+ tokens/second. But you can&#8217;t put a discrete GPU in a mini PC — that&#8217;s desktop or gaming laptop territory.</p>
<h3>RAM Determines Model Size</h3>
<p>The most important spec for local AI is <strong>RAM</strong>. The entire model loads into memory.</p>
<table>
<tr>
<th>RAM</th>
<th>Model Size</th>
<th>Quality</th>
</tr>
<tr>
<td>8GB</td>
<td>7B (7 billion parameters)</td>
<td>Basic chat OK, struggles with complexity</td>
</tr>
<tr>
<td>16GB</td>
<td>14B (14 billion parameters)</td>
<td>Decent conversation, handles general tasks</td>
</tr>
<tr>
<td>32GB</td>
<td>14B + headroom / can try 30B</td>
<td>Comfortable 14B + other services running</td>
</tr>
<tr>
<td>64GB+</td>
<td>70B (70 billion parameters)</td>
<td>Approaching ChatGPT quality</td>
</tr>
</table>
<p><strong>7B vs 14B vs 70B — bigger means better.</strong> 7B handles simple chat but frequently hallucinates on complex questions. 14B is the minimum threshold where it feels &#8220;actually usable.&#8221; 70B jumps in quality but needs 40GB+ RAM.</p>
<p>That&#8217;s why I have 32GB. Running a 14B model while also keeping other Docker services (<a href="/en/code-illiterate-home-server-build-3-immich-photo-backup-en/">Immich</a>, WordPress, n8n, etc.) alive requires the headroom.</p>
<figure class="wp-block-image size-large"><img decoding="async" width="867" height="650" src="https://prsm-studio.com/wp-content/uploads/2026/03/stock-31993524-1.jpg" alt="선명한 노란색 표면의 T-Force Delta RGB DDR5 메모리 모듈." class="wp-image-227" srcset="https://prsm-studio.com/wp-content/uploads/2026/03/stock-31993524-1.jpg 867w, https://prsm-studio.com/wp-content/uploads/2026/03/stock-31993524-1-300x225.jpg 300w, https://prsm-studio.com/wp-content/uploads/2026/03/stock-31993524-1-768x576.jpg 768w" sizes="(max-width: 867px) 100vw, 867px" /><figcaption>Photo by Andrey Matveev / Pexels</figcaption></figure>
<h2>So Is It Worth It?</h2>
<p>Here&#8217;s my honest summary:</p>
<p><strong>Worth it for:</strong></p>
<ul>
<li>Simple conversations, translation, summarization — slow but delivers results</li>
<li>Privacy-sensitive content — analyzing confidential work documents</li>
<li>Offline use — on a plane, in areas with no internet</li>
<li>Connecting AI to other apps — unlimited API calls, zero cost</li>
</ul>
<p><strong>Not worth it for:</strong></p>
<ul>
<li>Coding, complex analysis — cloud AI is overwhelmingly better</li>
<li>When you need fast responses — if you can&#8217;t wait 25 seconds</li>
<li>When you need current information — local models don&#8217;t know anything after their training date</li>
</ul>
<p>The core value of local AI is <strong>&#8220;free&#8221;</strong> and <strong>&#8220;privacy.&#8221;</strong> If you&#8217;re expecting performance, you&#8217;ll be disappointed. But if those two things matter to you, it&#8217;s absolutely worthwhile.</p>
<h2>Next Episode Preview</h2>
<p>So far we&#8217;ve covered building the server, remote access, photo backup, and local AI. Next up is the piece that ties everything together — <strong>an AI agent and Telegram bot.</strong> Send a message on Telegram, and AI handles the rest. Building your own digital assistant.</p>
<p><strong>EP.5 — AI Agent + Telegram: Putting a Secretary on Your Server.</strong> Stay tuned.</p>
<p><a class="a2a_button_facebook" href="https://www.addtoany.com/add_to/facebook?linkurl=https%3A%2F%2Fprsm-studio.com%2Fen%2Fcode-illiterate-home-server-build-4-ollama-local-ai-en%2F&amp;linkname=Even%20a%20Code-Illiterate%20Built%20It%21%20Home%20Server%20Journey%20%284%29%20%E2%80%94%20Running%20AI%20Locally%20with%20Ollama" title="Facebook" rel="nofollow noopener" target="_blank"></a><a class="a2a_button_mastodon" href="https://www.addtoany.com/add_to/mastodon?linkurl=https%3A%2F%2Fprsm-studio.com%2Fen%2Fcode-illiterate-home-server-build-4-ollama-local-ai-en%2F&amp;linkname=Even%20a%20Code-Illiterate%20Built%20It%21%20Home%20Server%20Journey%20%284%29%20%E2%80%94%20Running%20AI%20Locally%20with%20Ollama" title="Mastodon" rel="nofollow noopener" target="_blank"></a><a class="a2a_button_email" href="https://www.addtoany.com/add_to/email?linkurl=https%3A%2F%2Fprsm-studio.com%2Fen%2Fcode-illiterate-home-server-build-4-ollama-local-ai-en%2F&amp;linkname=Even%20a%20Code-Illiterate%20Built%20It%21%20Home%20Server%20Journey%20%284%29%20%E2%80%94%20Running%20AI%20Locally%20with%20Ollama" title="Email" rel="nofollow noopener" target="_blank"></a><a class="a2a_dd addtoany_share_save addtoany_share" href="https://www.addtoany.com/share#url=https%3A%2F%2Fprsm-studio.com%2Fen%2Fcode-illiterate-home-server-build-4-ollama-local-ai-en%2F&#038;title=Even%20a%20Code-Illiterate%20Built%20It%21%20Home%20Server%20Journey%20%284%29%20%E2%80%94%20Running%20AI%20Locally%20with%20Ollama" data-a2a-url="https://prsm-studio.com/en/code-illiterate-home-server-build-4-ollama-local-ai-en/" data-a2a-title="Even a Code-Illiterate Built It! Home Server Journey (4) — Running AI Locally with Ollama"></a></p><p>The post <a href="https://prsm-studio.com/en/code-illiterate-home-server-build-4-ollama-local-ai-en/">Even a Code-Illiterate Built It! Home Server Journey (4) — Running AI Locally with Ollama</a> appeared first on <a href="https://prsm-studio.com/en">Prsm Studio</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://prsm-studio.com/en/code-illiterate-home-server-build-4-ollama-local-ai-en/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
