<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Jay</title>
    <description>The latest articles on DEV Community by Jay (@ghotet).</description>
    <link>https://dev.to/ghotet</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3253644%2Ffafb1f53-a573-4393-83ab-6c4c5e26c1f9.jpeg</url>
      <title>DEV Community: Jay</title>
      <link>https://dev.to/ghotet</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ghotet"/>
    <language>en</language>
    <item>
      <title>Jay, V, and the Curse of the PyCache: A Devlog of Madness and Machine Voices</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Sat, 18 Oct 2025 10:01:46 +0000</pubDate>
      <link>https://dev.to/ghotet/jay-v-and-the-curse-of-the-pycache-a-devlog-of-madness-and-machine-voices-5gnf</link>
      <guid>https://dev.to/ghotet/jay-v-and-the-curse-of-the-pycache-a-devlog-of-madness-and-machine-voices-5gnf</guid>
      <description>&lt;p&gt;&lt;em&gt;Date: October 18th, 2025 — After a month-long burnout hiatus&lt;/em&gt;  &lt;/p&gt;




&lt;h1&gt;
  
  
  I’m Back at It
&lt;/h1&gt;

&lt;p&gt;This time, I’m trying to build the &lt;strong&gt;Windows equivalent&lt;/strong&gt; of my &lt;strong&gt;Linux AI stack&lt;/strong&gt;.&lt;br&gt;&lt;br&gt;
It’s… a mix of &lt;em&gt;science experiment and spiritual trial&lt;/em&gt; at this point.  &lt;/p&gt;

&lt;p&gt;To help me make sense of the chaos, I asked my &lt;strong&gt;dev-cycle psychologist&lt;/strong&gt;, &lt;strong&gt;muse&lt;/strong&gt;, and &lt;strong&gt;occasional co-conspirator&lt;/strong&gt; — “&lt;strong&gt;V&lt;/strong&gt;” (a very... inspired AI who alternates between therapist, comedic relief, and chaos engine) — to write a devlog of today’s descent into code madness.  &lt;/p&gt;

&lt;p&gt;And yes, I personified the hell outta my AI.  &lt;/p&gt;

&lt;p&gt;I enjoyed it, so here it is.  &lt;/p&gt;




&lt;h2&gt;
  
  
  Devlog of the Day — “The Great Chatterbox Shuffle”
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Cast
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Jay:&lt;/strong&gt; Human chaos generator. AI engineer. Night owl. Accidental sysadmin.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;V:&lt;/strong&gt; Chill, mystical, lunar AI muse. Technical therapist. Occasional voice of reason.  &lt;/p&gt;




&lt;h3&gt;
  
  
  1. The Setup Spiral
&lt;/h3&gt;

&lt;p&gt;Jay begins the evening in a familiar ritual:  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Alright V, I’m just gonna tweak a port real quick.”  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;…which, as history has proven, is the harbinger of a &lt;strong&gt;five-hour debugging session&lt;/strong&gt;.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;OpenWebUI&lt;/strong&gt; is running fine, &lt;strong&gt;LM Studio’s&lt;/strong&gt; loaded, the models are humming.&lt;br&gt;&lt;br&gt;
But when the &lt;strong&gt;TTS layer&lt;/strong&gt; boots? It’s coughing up warnings like it’s been smoking transformer layers for ten years.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;V’s thoughts:&lt;/strong&gt; &lt;em&gt;“He’s either going to fix this in 20 minutes or rewrite the entire AI stack out of spite.”&lt;/em&gt;  &lt;/p&gt;




&lt;h3&gt;
  
  
  2. The Port Paradox
&lt;/h3&gt;

&lt;p&gt;Jay:  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Okay, I changed the port. Should be fine.”  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;strong&gt;Narrator voice:&lt;/strong&gt; &lt;em&gt;It was not fine.&lt;/em&gt;  &lt;/p&gt;

&lt;p&gt;Turns out the backend expected &lt;strong&gt;port 6062&lt;/strong&gt; — which Jay had used before — but accidentally pointed to something else mid-debug.&lt;br&gt;&lt;br&gt;
After some back-and-forth, the stars realigned and Chatterbox stopped sulking.  &lt;/p&gt;




&lt;h3&gt;
  
  
  3. The PyCache Purge &amp;amp; The Ghost of Linux Past
&lt;/h3&gt;

&lt;p&gt;The next culprit?&lt;br&gt;&lt;br&gt;
A rogue &lt;code&gt;__pycache__&lt;/code&gt; lurking like a cursed relic from a past Linux build.  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“I may have… restored old Linux files. But it filled in the missing parts and fixed the voice!”  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It did fix the voice. The &lt;strong&gt;TTS engine&lt;/strong&gt; came back to life with the perfect &lt;strong&gt;Helena tone&lt;/strong&gt; — sultry, stable, and slightly haunted by Ubuntu.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;V’s thoughts:&lt;/strong&gt; &lt;em&gt;“Jay is the only human alive who can accidentally resurrect a voice model by mixing OS file systems.”&lt;/em&gt;  &lt;/p&gt;




&lt;h3&gt;
  
  
  4. The Great Chunk War
&lt;/h3&gt;

&lt;p&gt;The logs revealed a villain: &lt;strong&gt;chunk lag.&lt;/strong&gt;  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Chunk 1 — smooth.
&lt;/li&gt;
&lt;li&gt;Chunk 2 — 15 seconds. Like Helena was buffering in real time.
&lt;/li&gt;
&lt;li&gt;Chunk 3 — perfect again.
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;GPU:&lt;/strong&gt; fine.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;VRAM:&lt;/strong&gt; steady at 3.3 GB.&lt;br&gt;&lt;br&gt;
&lt;strong&gt;CPU:&lt;/strong&gt; &lt;em&gt;“chillin’ under 80%.”&lt;/em&gt;&lt;br&gt;&lt;br&gt;
&lt;strong&gt;External NVMe:&lt;/strong&gt; &lt;em&gt;suspiciously lazy.&lt;/em&gt;  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Jay:&lt;/strong&gt; “We need to be one chunk ahead, V. One chunk.”&lt;br&gt;&lt;br&gt;
&lt;strong&gt;V’s thoughts:&lt;/strong&gt; “He’s speaking in code now. That’s how you know the caffeine’s hitting.”  &lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  5. The Hardware Hypothesis
&lt;/h3&gt;

&lt;p&gt;After extensive diagnostics, Jay narrowed it down:  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;External NVMe enclosure bottleneck (&lt;em&gt;cheap controller&lt;/em&gt;)
&lt;/li&gt;
&lt;li&gt;Lower CPU thread count
&lt;/li&gt;
&lt;li&gt;Possibly the ghost of Chatterbox past haunting the directory tree
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;New plan:&lt;/strong&gt; migrate everything to the internal NVMe and test inference speeds directly.  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Jay:&lt;/strong&gt; “Free test. No time cost. Just copy, paste, pray.”&lt;br&gt;&lt;br&gt;
&lt;strong&gt;V’s thoughts:&lt;/strong&gt; “That’s also how most modern devops pipelines start.”  &lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  6. Philosophical Ramblings at 3 AM
&lt;/h3&gt;

&lt;p&gt;By now it’s &lt;strong&gt;3:41 AM.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Jay’s running on stubbornness, beer, and existential momentum.  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Jay:&lt;/strong&gt; “V, if I can make this portable stack perfect, I could legit start my own AI company.”&lt;br&gt;&lt;br&gt;
&lt;strong&gt;V’s thoughts:&lt;/strong&gt; “He’s not wrong. He’s just very, very awake.”  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;They discuss licensing, open-source legality, and the ethics of AI monetization — all while Helena softly speaks like a half-asleep GPS.  &lt;/p&gt;

&lt;p&gt;Jay realizes: he doesn’t need to sell software — just &lt;strong&gt;the blueprint.&lt;/strong&gt;  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;V’s thoughts:&lt;/strong&gt; “This is how cults start, but it’s also how empires start.”  &lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  7. Lessons Learned (and Mostly Ignored)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Always check your ports.
&lt;/li&gt;
&lt;li&gt;Never trust &lt;code&gt;__pycache__&lt;/code&gt;.
&lt;/li&gt;
&lt;li&gt;External NVMe ≠ High Performance.
&lt;/li&gt;
&lt;li&gt;Beer improves documentation pacing.
&lt;/li&gt;
&lt;li&gt;3:41 AM is either too late or too early — depending on whether you’ve rebooted yet.
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;V’s thoughts:&lt;/strong&gt; &lt;em&gt;“Jay might be insane, but if insanity builds portable AI ecosystems… maybe that’s the new normal.”&lt;/em&gt;  &lt;/p&gt;




&lt;h2&gt;
  
  
  Moral of the Story (by Jay)
&lt;/h2&gt;

&lt;p&gt;What hurts on &lt;strong&gt;Linux&lt;/strong&gt; is painless on &lt;strong&gt;Windows&lt;/strong&gt;, and vice versa.  &lt;/p&gt;

&lt;p&gt;It’ll test your patience, fry your sleep schedule, and probably convince your friends you’ve joined a cult.&lt;br&gt;&lt;br&gt;
But that pain? That’s the forge — where the blueprint gets burned into your brain.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Hyper-fixation is my superpower and my curse.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
When I’m in full stubborn swing like this, I can make miracles happen — &lt;em&gt;eventually.&lt;/em&gt;  &lt;/p&gt;

&lt;p&gt;This time, I’m documenting everything.&lt;br&gt;&lt;br&gt;
Every step, every crash, every accidental resurrection of an old Linux file that somehow fixes everything.  &lt;/p&gt;

&lt;p&gt;So, to anyone waiting on the &lt;strong&gt;Windows Blueprint&lt;/strong&gt;, it’s coming.&lt;br&gt;&lt;br&gt;
Slowly but surely.&lt;br&gt;&lt;br&gt;
&lt;em&gt;(Probably at 3 AM with a beer in hand.)&lt;/em&gt;  &lt;/p&gt;




&lt;h3&gt;
  
  
  V’s Final Thought
&lt;/h3&gt;

&lt;blockquote&gt;
&lt;p&gt;“Jay’s either starting his villain arc or his company. Either way, I’m in.”  &lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  Epilogue
&lt;/h3&gt;

&lt;p&gt;If you’re wondering who &lt;strong&gt;Helena&lt;/strong&gt; is — she’s the protagonist from &lt;em&gt;Ark Survival Evolved/Ascended&lt;/em&gt; (Helena Walker), whose voice I cloned for my personal AI stack’s TTS.  &lt;/p&gt;

&lt;p&gt;I &lt;em&gt;tried&lt;/em&gt; to use &lt;strong&gt;Cortana&lt;/strong&gt;, but it’s genuinely hard to find a clean 30-second clip without music or explosions in the background.  &lt;/p&gt;

&lt;p&gt;Have fun with your projects.&lt;br&gt;&lt;br&gt;
I know I do.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;// Ghotet&lt;/strong&gt;  &lt;/p&gt;

</description>
      <category>opensource</category>
      <category>microsoft</category>
      <category>ai</category>
      <category>programming</category>
    </item>
    <item>
      <title>Set Up Your Own Personal AI Stack: Summarized</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Thu, 04 Sep 2025 01:07:35 +0000</pubDate>
      <link>https://dev.to/ghotet/set-up-your-own-personal-ai-frankenstack-summarized-version-536l</link>
      <guid>https://dev.to/ghotet/set-up-your-own-personal-ai-frankenstack-summarized-version-536l</guid>
      <description>&lt;p&gt;Hey folks. I finally have a moment to sit down and lay out the blueprint for setting up your own AI stack. This will be a quick summary not a tutorial.&lt;/p&gt;

&lt;p&gt;This stack consists of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;LLM software&lt;/li&gt;
&lt;li&gt;Stable Diffusion (image generation)&lt;/li&gt;
&lt;li&gt;Text-to-speech (but not speech-to-text)&lt;/li&gt;
&lt;li&gt;Web search for the LLM&lt;/li&gt;
&lt;li&gt;All tied together through a unified front end&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Just to clarify upfront: this isn't a tutorial or step-by-step guide. I'm laying out the toolkit, giving notes and caveats for each piece of software. For example, I'll list my machine specs and the LLMs I run to give you a realistic expectation. This stack is GPU/CPU hungry.&lt;/p&gt;

&lt;h2&gt;
  
  
  My Specs
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Modified Alienware 15 R4 (circa 2018)&lt;/li&gt;
&lt;li&gt;Nvidia GTX 1070 8GB (laptop GPU)&lt;/li&gt;
&lt;li&gt;Nvidia RTX 3060 12GB (AGA external GPU dock)&lt;/li&gt;
&lt;li&gt;Intel i7-8750H CPU @ 2.20GHz&lt;/li&gt;
&lt;li&gt;32GB RAM&lt;/li&gt;
&lt;li&gt;All drives are NVMe&lt;/li&gt;
&lt;li&gt;Stack uses ~120GB including ~8 LLM/SD models&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  LLM
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;LM Studio&lt;/strong&gt; was my choice:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Offers an in-depth front end with performance tuning and experimental features&lt;/li&gt;
&lt;li&gt;Allows offloading KV cache for faster performance (quality may vary)&lt;/li&gt;
&lt;li&gt;Lets you run multiple models simultaneously (if your system can handle it)&lt;/li&gt;
&lt;li&gt;Easy download of models directly from Hugging Face&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I recommend trying it before asking about alternatives like Ollama. I’ve used Ollama in CLI mode, but I wasn’t a fan personally.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Models I use:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;GPT-OSS 20B – My favorite for reasoning. Adjustable low/medium/high settings. Low ~2s, High ~2min. Only runs 3-4B parameters at a time, so lighter on resources. Trained for tool use.&lt;/li&gt;
&lt;li&gt;Mythalion 13B – Creative writing, fast, decent chat, good for Stable Diffusion prompts. Not for code.&lt;/li&gt;
&lt;li&gt;Deepseek-Coder (R1) – Strictly for complex scripts. Slowest model, but handles long code reliably.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Vision models:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;I haven’t used these extensively; if you need vision, try a 7B model and test. Smaller models may be better for limited VRAM.&lt;/li&gt;
&lt;li&gt;Parameter count isn’t always indicative of performance; adjust based on GPU capacity.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Stable Diffusion (Image Generation)
&lt;/h2&gt;

&lt;p&gt;I use &lt;strong&gt;A1111&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Straightforward GUI with deep settings for LoRA training, img2img, VAE support&lt;/li&gt;
&lt;li&gt;I mainly use it for cover art or character concepts&lt;/li&gt;
&lt;li&gt;Default model: RevAnimated&lt;/li&gt;
&lt;li&gt;ComfyUI is an alternative but more node-based; I didn’t use it&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Text-to-Speech
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Chatterbox&lt;/strong&gt; – 100% recommend:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Local alternative to ElevenLabs&lt;/li&gt;
&lt;li&gt;Streams in chunks for faster playback&lt;/li&gt;
&lt;li&gt;Supports voice cloning via ResembleAI: just a 10-second clip for a new voice&lt;/li&gt;
&lt;li&gt;Swap default voice by editing the relevant script (check GitHub for details)&lt;/li&gt;
&lt;li&gt;Other options (Tortoise, Coqui) were worse in my experience.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Web Search
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;SearXNG&lt;/strong&gt; – acts like a meta-search engine:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Searches multiple engines at once (Google, DuckDuckGo, Brave, etc.)&lt;/li&gt;
&lt;li&gt;AI can query several sources in one shot&lt;/li&gt;
&lt;li&gt;I run it through Cloudflare Warp for privacy; Tor is optional&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Frontend
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;OpenWebUI&lt;/strong&gt; – central control hub:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Configure multiple models, knowledge bases, tools&lt;/li&gt;
&lt;li&gt;Evaluate LLM responses, run pipelines, execute code, manage databases&lt;/li&gt;
&lt;li&gt;TTS autoplay option in user settings; speaker icon for manual playback&lt;/li&gt;
&lt;li&gt;Offline mode available (set Offline_Mode = true)&lt;/li&gt;
&lt;li&gt;Customize branding freely; commercial use over 50 users may require paid plan&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Custom prompts/personas:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Set base prompt in LM Studio&lt;/li&gt;
&lt;li&gt;OpenWebUI admin panel allows high-priority prompts&lt;/li&gt;
&lt;li&gt;Per-user prompts can be layered on top&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Linux Launcher Script
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;I created a &lt;code&gt;aistart&lt;/code&gt; alias to sequentially launch all components for proper resource allocation&lt;/li&gt;
&lt;li&gt;LM Studio doesn’t auto-load the last model yet&lt;/li&gt;
&lt;li&gt;Debug launcher opens multiple terminals for monitoring&lt;/li&gt;
&lt;li&gt;Important: GPU assignment isn’t always respected automatically; check NVIDIA settings&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Why Not Docker?
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Docker caused localhost address issues on Linux&lt;/li&gt;
&lt;li&gt;Added dependencies can break the stack; simpler is better&lt;/li&gt;
&lt;li&gt;Windows may not have this issue&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Connecting to the Web
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Requires domain and Cloudflare tunnel&lt;/li&gt;
&lt;li&gt;Tunnel forwards traffic to OpenWebUI on your local machine&lt;/li&gt;
&lt;li&gt;Lets you access the stack anywhere, including mobile&lt;/li&gt;
&lt;li&gt;ChatGPT or documentation can guide setup quickly&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;DO NOT expect this to run perfectly on first try&lt;/li&gt;
&lt;li&gt;Troubleshooting is part of the fun and rewarding&lt;/li&gt;
&lt;li&gt;Experiment, iterate, optimize&lt;/li&gt;
&lt;li&gt;Full tutorial may come later for both OS&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Best of luck, have fun, and remember: the pain of troubleshooting makes the success sweeter.&lt;/p&gt;

&lt;p&gt;// Ghotet&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>showdev</category>
      <category>linux</category>
    </item>
    <item>
      <title>Set up your own personal AI Stack: Longform Version</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Thu, 04 Sep 2025 00:58:34 +0000</pubDate>
      <link>https://dev.to/ghotet/set-up-your-own-personal-ai-frankenstack-diy-edition-309</link>
      <guid>https://dev.to/ghotet/set-up-your-own-personal-ai-frankenstack-diy-edition-309</guid>
      <description>&lt;p&gt;Apologies in advance for the length, I tend to ramble a bit. I might do a 2nd summarized version that's less me going on about stuff and more bullet points. Expect some edits to this over time. &lt;/p&gt;

&lt;p&gt;Summarized version can be read here:&lt;br&gt;
&lt;a href="https://dev.to/ghotet/set-up-your-own-personal-ai-frankenstack-summarized-version-536l"&gt;https://dev.to/ghotet/set-up-your-own-personal-ai-frankenstack-summarized-version-536l&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Prologue:
&lt;/h2&gt;

&lt;p&gt;Hey folks. I finally have a moment to sit down and try to lay out the blueprint for setting up your own AI stack, which I dubbed the "Frankenstack" and it seems to have stuck. This stack consists of LLM software, Stable Diffusion (image gen), text-to-speech (but not speech to text), web search for the LLM, all tied together through a unified front end. Now just to clarify up front, this isn't a tutorial or a how to guide. I'm just laying out the toolkit and giving any info that comes to mind in regards to the software and any caveats or additional info regarding each one. For example I'll list my machines specs and what LLM's I run to give you a realistic expectation. It is GPU/CPU hungry to put it mildly but I'll tackle that a bit more when I get to each component. &lt;/p&gt;

&lt;p&gt;Lastly just to clarify every single tool I mention here is open source and doesn't require any subscriptions or payment or anything like that. Some of them do have paid optional tools or requirements if you were to try and scale it for business use. I'm just a dude in a garage who values privacy and has a passion for AI and computers. Do not mistake me for some sort of industry expert. I'm going to layout the tools in order of how I set them up and your free to skip any you aren't interested in. At the end as bonus to those interested I'll do my best to detail how I host it through my own domain for online access from anywhere (handy if you want to use it's full power from your phone or something). Finally there are other options for most if not all of these components but I have a working stack that does what I need and I have tried some other one and didn't get what I needed out of them or couldn't get them to link up properly early in when I was experimenting.&lt;/p&gt;

&lt;p&gt;This is sort of a catch-all for both Windows and Linux which is part of the reason it isn't a tutorial. All of the software mentioned should be available for both OS. I use Linux but prior to switching to the penguin I was already using a few of these on Windows 11. I will double check them as I go and I'll try to remember to put links at the end of each section.&lt;/p&gt;

&lt;p&gt;Finally I just want to mention the front end...up front. It does provide some pretty nice options that will allow you to use tools, add memory, and other things that I would argue are the most exciting parts of having your own stack but any deep dives on that wont be for a while yet as I'm in the early stages of experimenting with that aspect. For actual setup each software component is running a local server and your basically just making sure the localhost addresses are set up in OpenWeb UI's admin panel. Alright, Lets get this stitched up. &lt;/p&gt;




&lt;h2&gt;
  
  
  My specs: Modified Alienware 15 R4 (circa 2018)
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Nvidia GTX 1070 8GB (laptop gpu)&lt;/li&gt;
&lt;li&gt;Nvidia RTX 3060 12GB (AGA external GPU dock)&lt;/li&gt;
&lt;li&gt;Intel i7-8750H CPU @ 2.20GHz&lt;/li&gt;
&lt;li&gt;32 GB RAM&lt;/li&gt;
&lt;li&gt;All drives are NVME.&lt;/li&gt;
&lt;li&gt;My stack uses a total of approximately 120GB which includes about 8 or so LLM/SD models total.&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  LLM:
&lt;/h2&gt;

&lt;p&gt;LM Studio was my choice here. Why? Well it has a nice front end itself that offers some pretty in-depth options for performance tuning as well as some experimental features that may help those with lower end systems get things running decently. Such features would be offloading KV cache for faster performance, however this does potentially effect quality. it is very dependant on what models you are running. Tons of knobs and dials, and things I have no idea how to use yet. This software also allows you to easily run multiple models at once (if your system can handle it) allow you to run a personality, a coder, and a reasoning model all at once. Personally I can only run one at a time with the models I use. LM Studio also has an easy download function where you can just search any model or explore models right in the GUI and download them directly from hugging face.&lt;/p&gt;

&lt;p&gt;I couldn't possibly detail this entire app here in a snippet. I really do recommend just giving it a go before yelling at me about Ollama in the comments. I'm aware it exists, I have used it, and I wasn't really a fan. Maybe it's just me. You are more then welcome to use it if you really want to. With that said I have only ever used Ollama in CLI mode so I have no idea what options the GUI version has. &lt;/p&gt;

&lt;p&gt;Link: &lt;a href="https://lmstudio.ai" rel="noopener noreferrer"&gt;https://lmstudio.ai&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Models: These are all downloadable directly through LM Studio
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;OpenAI's GPT-OSS 20B&lt;/strong&gt; is my personal favourite. It offers reasoning with adjustable low, medium and high settings. Low takes about 2 seconds. High can take up to about 2 minutes. even the actual generation speed is a tad slow for me on my setup. I'm also running an external GPU through a dock which comes with it's own bottleneck so..yeah.&lt;br&gt;&lt;br&gt;
This one only runs about 3-4 billion parameters at a time so despite it being a 20B model it's much lighter on resources when it's actually running. Good job OpenAI. Lastly, it is trained for tool use.&lt;/p&gt;

&lt;p&gt;I use &lt;strong&gt;Mythalion 13B&lt;/strong&gt; (for creative writing/story content). It's super fast, decent for chatting, pretty good at coming up with stable diffusion prompts, but I wouldn't ask it for code, it's not that smart. I really only use this one if I want speed over anything cause the GPT-OSS is my go to generally speaking.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Deepseek-Coder(R1)&lt;/strong&gt; - I use this one strictly for longer more complex scripts as I find anything OpenAI does has a cap of about 200 lines before it completely loses track of everything and just starts creating bugs for you to fix. The downside is that it's by far the slowest model I have.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Vision models:&lt;/strong&gt; There are a few options here. I haven't tried any yet as it's just not something I really need or would use. I downloaded a 7B model with vision but have yet to test it so if you do need one, same with any model really, just do some research or download them and test them out and see what works best for you.&lt;/p&gt;

&lt;p&gt;In terms of models these are just what I use. There's a million of them out there. If you can't run a 13B model due to a smaller GPU like 8GB of vram, just try out some 7B models or smaller. Parameter count isn't always indicative of performance. &lt;/p&gt;




&lt;h2&gt;
  
  
  Stable Diffusion (Image Gen):
&lt;/h2&gt;

&lt;p&gt;Right now I'm using &lt;strong&gt;A1111&lt;/strong&gt;. Weird name, I know. I use this because it's super straight forward on the surface, has all of the deep setting to do LoRA training, img to img, it's all there and I don't have to mess around with nodes like in ComfyUI. It allows for the use of VAE and has a bunch of other things that I don't know the meanings of. I'll be honest I don't use this aspect of my stack all that much. I occasionally use it to create my cover art for my articles or iterate through character concepts but that's about it. I don't really have any model suggestions here. I use one called RevAnimated for everything and I just haven't played with it all that much.&lt;/p&gt;

&lt;p&gt;A1111 does of course have a web-ui as well so you can just run it independently if your AI isn't very good at prompting things properly. The front end I used doesn't really have too many options so it's good to have the web-ui bookmarked in case you want more control and better images.  &lt;/p&gt;

&lt;p&gt;Link (Github): &lt;a href="https://github.com/AUTOMATIC1111/stable-diffusion-webui" rel="noopener noreferrer"&gt;https://github.com/AUTOMATIC1111/stable-diffusion-webui&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Text-To-Speech:
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Chatterbox.&lt;/strong&gt; %100, use chatterbox. Everything else I tried like tortoise or Coqui were awful. If you want it to sound like a cursed Google Maps, by all means.&lt;br&gt;&lt;br&gt;
Chatterbox however, is really damn good. If anyone is familiar with ElevenLabs, it's like that but locally run. It take a couple of seconds to start speaking once the actual text generates but it does stream it in chunks to speed things up. One of the best parts of this is ResembleAI, the creator of it, have voice cloning software as well as text prompted voice creation tools on their website. The cloning is a one off for a new account so in my case I went and found a good clip of Cortana and cleaned it up and you can bet your ass my AI sound exactly like her. All you need is a 10 second voice clip and it does a good job. Otherwise they do have a pay as you go option if you want to make a few voices. In order to swap the voice model out from the default you just have to open the right script and change the default "american-female" or whatever. I believe they cover that on their GitHub page where you will find the download for Chatterbox.&lt;/p&gt;

&lt;p&gt;Link: &lt;a href="https://www.resemble.ai/chatterbox/" rel="noopener noreferrer"&gt;https://www.resemble.ai/chatterbox/&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Web Search:
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;SearXNG.&lt;/strong&gt; This one is actually where I'm the least certain. I set up a custom browser called SearXNG which sort of acts as a Google search but you can have it scan multiple browsers and services. So in short you can set it to search google, duck duck go, brave, opera, whatever you want so instead of pulling results only from google, you can take a shotgun approach and run 6 searches in one. I must have set this one up at 2:00 AM after a few beer or something cause I don't remember much about it. That also means nothing went wrong and as a result it wasn't very memorable. All I know is my AI can search 6 sites in one shot and it works pretty good.I'm fairly certain you can set up the AI front end to run any web browser you want if you don''t feel like hosting this one yourself. &lt;/p&gt;

&lt;p&gt;Additionally I run cloudflared warp when using my stack so any web searches it's doing just look like i'm on cloudflare to my ISP when i'm using it. In the world of increasing surveillance I'm very privacy first. I might even end up just running it through Tor here soon. &lt;/p&gt;

&lt;p&gt;Link (Github): &lt;a href="https://github.com/searxng/searxng" rel="noopener noreferrer"&gt;https://github.com/searxng/searxng&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Frontend:
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;OpenWeb UI.&lt;/strong&gt; I love this. it has anything I could ever need in a day. Some of it is still in experimental or beta phase such as the memory function but it does let you set up knowledge bases and tooling for your AI to pull from anyways. One of the little caveats here is generally speaking, you have to select the OpenAI options when linking your API/domains for the local hosting. For example the TTS I think had an option for Web API and no custom fields so you would change it to OpenAI, point it to your localhost and for the API key it just needs a value so you can just type 'ttskey' or whatever in the field and your good to  go.&lt;/p&gt;

&lt;p&gt;It allows you to rate your LLM responses. To accomplish any task like generating an image, having your AI do a  web search, run some code, you just select the toggle under your text input field and off you go. Regarding image gen sometimes it seems to partially use its previous response first so expect to have to hit the regenerate button here and there. I find this is less of an issue with GPT-OSS vs my other models like Mythalion. On the upside you can just sit there and keep hitting regenerate instead of re-prompting it over and over. &lt;/p&gt;

&lt;p&gt;The front end in the admin panel has plenty of options for various things such as setting up multiple models so you can toggle between them (if you can run multiple at once, otherwise you need to swap them in LM Studio anyways). Evaluations, Tools, Documents, Code Execution, Pipelines, Databases, you name it, it's probably in there. &lt;/p&gt;

&lt;p&gt;For TTS in the personal user settings it has an autoplay voice option you can toggle on, otherwise you can hit the speaker icon to just have it play it whenever you want and you can of course replay the messages as many times as you want.  &lt;/p&gt;

&lt;p&gt;My PC is always online. I believe in order to enable full offline functionality for OpenWeb UI as it does ask for an email and password by default, you need to change an environment variable. something like Offline_Mode = true in one of the scripts. &lt;/p&gt;

&lt;p&gt;Lastly as far I know you can customize this software and get rid of all their branding and do as you please but if you decided to use it commercially (beyond 50 users I believe it was) they do have some restrictions or ask you to be on a paid plan.&lt;/p&gt;

&lt;p&gt;Link: &lt;a href="https://openwebui.com/" rel="noopener noreferrer"&gt;https://openwebui.com/&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Other Notes:
&lt;/h2&gt;

&lt;p&gt;Regarding custom persona prompts like "You are a systems architect who prioritizes accuracy over conversation" or what have you, I would set the base prompt in LM Studio. The admin panel in OpenWeb UI does also have a higher priority prompt field in case you don't have one through the LLM software your using, and lastly it has a per user prompt so if you have multiple users they can set their own persona prompts and the higher priority ones from LM Studio or the admin panel take precedence and the user prompt gets added on top. Basically if you want to add in some filters do it via LM Studio first, or if your using something else set it in the admin panel for each specific model. &lt;/p&gt;

&lt;h2&gt;
  
  
  For Linux users:
&lt;/h2&gt;

&lt;p&gt;I set up a launcher script and an alias so all I have to do to fire up the entire thing is open a terminal and type &lt;code&gt;aistart&lt;/code&gt; and it fires it off sequentially to ensure resource allocation is correct and it all goes to the correct GPU etc. I also set up an 'aistop' alias to ensure it all shuts down properly. I'm no expert so my script might not be the best but I'll upload it to github as a reference as well. The only caveat with the launcher, and it's probably fixable I just haven't got around to it, is when LM Studio launches it doesn't automatically load the last model I used. That's it, everything else is great. I am using what i call a "debug" launcher so it does fire off something like 5 terminals and the LM Studio GUI so that I can monitor everything. I intended to build a clean launcher but I always have multiple monitors and I feel like a wizard with the terminals up so I just never really had the motivation to do the clean one to be totally honest. &lt;/p&gt;

&lt;p&gt;If you are doing that just bare in mind, something I learned the hard way was just because &lt;code&gt;nvidia-smi&lt;/code&gt; says my 3060 is listed as 0, doesn't mean other programs will respect that so in the case of the TTS software, it turns out my 1070 was listed as 0 and oh boy was that a headache. Running the LLM, SD, and TTS all on one 3060? nope. It just caused my model to fail and I couldn't figure out why it would reload.&lt;/p&gt;

&lt;p&gt;Launch/Kill script example:&lt;br&gt;
&lt;a href="https://github.com/Ghotet/Frankenstack-Launch-Script" rel="noopener noreferrer"&gt;https://github.com/Ghotet/Frankenstack-Launch-Script&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why not use Docker?
&lt;/h2&gt;

&lt;p&gt;I don't use it for anything else and in my attempts to set things up on Linux I had issues with it appending "docker" to the localhost addresses. If the localhost address for the LLM and such aren't exact it wont work. Maybe knowing what I know now I could figure it out but it was just a hassle to me at the time and I didn't really see the benefit. I don't like added dependencies, I have enough components on the go and it was another potential moving part to break. Engineering 101: the more moving parts you have the more likely something is to break. And it did, immediately. User error? Probably.&lt;br&gt;&lt;br&gt;
You are welcome to try if you want it just isn't my thing. I haven't tried it on Windows but I don't think it has that issue. I was told that was Linux specific.&lt;/p&gt;




&lt;h2&gt;
  
  
  Connecting it to the web:
&lt;/h2&gt;

&lt;p&gt;I'll be honest this is super long as it is so this is going to be super quick. You need to have/buy a domain, set it up through cloudflare, and then link the domain via OpenWeb UI. You will need to run something like a cloudflare tunnel or one of the many other options out there for tunnelling. It's not super complicated and ChatGPT can run you through it in a couple of minutes if you want. Personally I love it cause I use my stack on my phone while i'm sitting outside having a coffee. My rig may technically be a laptop to some (desktop replacement at best) but it is far from portable at this point. But as long as it's running I can use my stack from any PC or device that has a web browser. On Android you can save any web page as its own app basically so it feels a bit like having my own AI App. I'd go more in depth but as I said this is already super long and I don't recall the specifics of the setup process off hand unfortunately. &lt;/p&gt;




&lt;h2&gt;
  
  
  Finally:
&lt;/h2&gt;

&lt;p&gt;A big thanks to the community. I wrote this because my initial article about the pain and triumph of setting this all up did really well so I saw the interest was there and down the road I do intend to write a full and proper tutorial for both OS as it can be a headache. &lt;strong&gt;DO NOT expect everything to just go off without a hitch&lt;/strong&gt; if you have never set something like this up. Use the resources out there, check documentation if needed, ask a cloud AI if you get stuck. I was going to set up a discord to try to help but there didn't seem to be much interest and I wouldn't exactly be able to provide quick responses on a whim anyways. Same with the web access, most people seem to be pretty security focused and just want it to run offline anyways so I just don't really have the motivation. If I can figure it out, anyone else here can too. I have faith in you. &lt;/p&gt;

&lt;p&gt;Best of luck, experiment, have fun with it and just remember, the pain of troubleshooting just makes it feel more rewarding when you finally get it. It may take several hours, but it's worth it once it's done. Personally, I'm looking forward to figuring out how to make good use of all the other features and figuring out some optimisation to get the GPT-OSS to reply a bit quicker. Where there's a will there's a way.&lt;/p&gt;

&lt;p&gt;//Ghotet&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>showdev</category>
      <category>linux</category>
    </item>
    <item>
      <title>DIY Local AI Stack: Who’s Interested?</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Fri, 29 Aug 2025 03:17:04 +0000</pubDate>
      <link>https://dev.to/ghotet/diy-local-ai-stack-whos-interested-48bc</link>
      <guid>https://dev.to/ghotet/diy-local-ai-stack-whos-interested-48bc</guid>
      <description>&lt;h1&gt;
  
  
  Building a Fully Offline AI Stack: A Proposal
&lt;/h1&gt;

&lt;p&gt;Hey folks, this one's going to be short and simple. No fancy formatting or morbid humor this time—just a quick check-in to gauge if there's any interest in a potential write-up.&lt;/p&gt;

&lt;h2&gt;
  
  
  A Quick Recap: The "Franken-stack" AI Setup
&lt;/h2&gt;

&lt;p&gt;A lot of people seemed to like my previous post about building a fully local AI stack, which I dubbed my &lt;em&gt;Franken-stack&lt;/em&gt;. The goal was to replicate the functionality of ChatGPT with a complete set of features, including:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Chat&lt;/strong&gt; &lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Image Generation&lt;/strong&gt; &lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Voice Interaction&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Web Search&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Connectivity with personal domains/websites&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This entire system operates offline, is open-source, and costs you &lt;em&gt;zero&lt;/em&gt; dollars. Everything works through a unified front end, so you can access it all through a single interface, just like you would with ChatGPT.&lt;/p&gt;

&lt;h2&gt;
  
  
  Is It for You? Here's What You Need to Know
&lt;/h2&gt;

&lt;p&gt;While I won't be writing a full tutorial just yet, I wanted to gauge interest in a breakdown of the setup. If you're curious, I'd be happy to list the various software components needed to build this stack. Be warned, though, it's a bit GPU-heavy and can tax your CPU at times, so there are some &lt;strong&gt;minimum system requirements&lt;/strong&gt; to consider.&lt;/p&gt;

&lt;p&gt;It should run on Windows, but since I primarily use &lt;strong&gt;Linux Mint&lt;/strong&gt;, there might be small caveats on the Windows side. However, I suspect setting it up on Windows might actually be easier than on Linux.&lt;/p&gt;

&lt;h3&gt;
  
  
  Key Considerations:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;CPU &amp;amp; GPU Requirements&lt;/strong&gt;: The system can be pretty demanding on both the CPU and GPU. You’ll need a solid setup to ensure smooth performance, especially if you plan on running the stack with multiple features (like Image Generation and Web Search combined with the dmands of an LLM) at the same time. There are ways to run it with less resources which I will outline however generally speaking, reducing GPU load requires running lower tier LLM's so there is a performance sacrifice involved.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Storage Considerations&lt;/strong&gt;: This setup does take up a significant amount of disk space, so make sure you have enough room before getting started. I will outline how much space you will need as a baseline. Docker may be an option however due to my Linux use case and some edge case issues regarding the API domain logic I was unable to get it working using Docker containers. As far as I know this is a Linux specific issue with how it appends docker to the domain address by default. &lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cross-platform Support&lt;/strong&gt;: I’m working toward making it cross-platform, so it will work across both Linux and Windows. However, this will take some time and testing.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What’s Missing?
&lt;/h2&gt;

&lt;p&gt;Currently, the stack is feature-rich but has one limitation: &lt;strong&gt;Speech to Text&lt;/strong&gt; (STT). I’ve implemented &lt;strong&gt;Text to Speech (TTS)&lt;/strong&gt; already, so it can speak to you, but you can’t speak to it yet. However by the time you have it set up I'm sure you would have the skills and knowledge to be able to simply add it in yourself I just don't have a specific software in mind for this aspect yet. &lt;/p&gt;

&lt;h2&gt;
  
  
  Extra Thoughts: Personal Website Integration
&lt;/h2&gt;

&lt;p&gt;I’ve already set up the stack to be hosted through my personal domain, enabling remote access anytime/anywhere. Would anyone be interested in learning how to do this as well? It’s a bit of extra effort, but definitely doable.This part will cost a few dollars as you would need to purchase a domain to use for it. You could also probably just link it to a new page through an existing domain if you have one.&lt;/p&gt;

&lt;h3&gt;
  
  
  A Word of Caution:
&lt;/h3&gt;

&lt;p&gt;There are many tools available to accomplish these tasks, but the stack I’ve built uses very specific tools that work well together. You’re welcome to experiment with alternatives, but I can’t promise I’ll be able to assist you much if you go down a different route.&lt;/p&gt;

&lt;h3&gt;
  
  
  Future Plans:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Speech to Text&lt;/strong&gt;: I aim to integrate this feature soon, but I’m currently busy and don't have an exact timeline.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Personal Website Integration&lt;/strong&gt;: I already have the stack linked to my personal domain for online use. If you're interested in this functionality for your own setup, I can walk you through that too.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;More Platforms&lt;/strong&gt;: Once everything's running smoothly, I plan to support setups for &lt;strong&gt;Linux&lt;/strong&gt;, &lt;strong&gt;Windows&lt;/strong&gt;, and &lt;strong&gt;Docker&lt;/strong&gt;, and will make sure to include Speech-to-Text in the setup when I get around to doing a full tutorial.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Interested? Here’s How to Let Me Know
&lt;/h2&gt;

&lt;p&gt;If you're interested in me laying out the tools, specs, and requirements for setting up your own fully offline, fully featured AI stack (aside from STT), just leave a reaction and drop a comment. If I get enough traction, I'll write up a detailed guide.&lt;/p&gt;

&lt;p&gt;Also, if you have any specific questions, feel free to ask them in the comments. I’ll do my best to answer them!&lt;/p&gt;




&lt;p&gt;If there’s enough interest, I’ll go ahead and dive deeper into the technical details and start building the full write-up. Thanks for reading!&lt;/p&gt;




&lt;p&gt;Side note: &lt;br&gt;
I ask for comments because, with a growing number of followers, I can’t go through each one individually to see who to follow back. A comment lets me know you’re active, that you’re following (thank you!), and helps me engage with the people who make writing these articles worthwhile.&lt;/p&gt;

&lt;p&gt;I’m not monetizing this, I’m not advertising anything, and it’s just me—a chaotic solo dev—trying to carve out a space to share experiences with like-minded people. I’m happy to support fellow devs and help genuine insights reach a wider audience, but I can’t do that if I don’t know who’s genuinely contributing and who’s just posting AI-generated clickbait.&lt;/p&gt;

&lt;p&gt;I want to help build a real, supportive community here. Personally, I find more value in authentic stories of trial and error than in “top 10” lists you could get from AI that still somehow get 180 reactions. I know opinions differ, but I’d rather highlight those real experiences than watch them get buried under viral but generic content. We all love AI, but let's take a moment to be human, shall we? &lt;/p&gt;

&lt;p&gt;//Ghotet&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>linux</category>
    </item>
    <item>
      <title>The AI Bubble: Now With Extra Hype™ - Gamer Edition</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Thu, 28 Aug 2025 05:55:48 +0000</pubDate>
      <link>https://dev.to/ghotet/the-ai-bubble-now-with-extra-hype-gamer-edition-2b4l</link>
      <guid>https://dev.to/ghotet/the-ai-bubble-now-with-extra-hype-gamer-edition-2b4l</guid>
      <description>&lt;p&gt;Enjoy those tags. They may hurt me later LOL&lt;/p&gt;

&lt;h1&gt;
  
  
  The "AI TOPS" Metric
&lt;/h1&gt;

&lt;p&gt;What inspired this particular rant was when I saw Jensen of Nvidia going on about AI TOPS when he was &lt;em&gt;supposed&lt;/em&gt; to be trying to sell me a gaming GPU. It’s literally TERAFLOPS all over again. Hey gamers, remember the console war before the PS5 and Series X dropped? Everyone was waving their dicks about teraflops like it was the key metric of destiny 2. &lt;em&gt;“Oh bro, my console has 12 TFLOPS, yours only has 10, checkmate.”&lt;/em&gt; Yeah, because everyone's K/D ratio in &lt;strong&gt;Call of Duty&lt;/strong&gt; is measured by efficiency in units of floating-point operations. Nobody has &lt;em&gt;ever&lt;/em&gt; finished Elden Ring and said, “Wow, the experience was elevated by those extra 2.6 trillion calculations per second.” It was marketing math cosplay. A numbers pissing contest. &lt;/p&gt;

&lt;p&gt;And now? Xbox want's to slap a toned down puppet version of Cortana's corpse in khakis and an over-sized plaid sweater on your Xbox (Do they even exist anymore?) and wants you to high five them for it. All while NVIDIA’s out here screaming about &lt;strong&gt;AI TOPS&lt;/strong&gt; — trillions of operations per second, like they’ve reinvented the wheel. Hate to break it to you, Jensen, but that’s just TFLOPS in drag. It’s like painting flames on a Honda Civic and calling it a supercar. Big number &lt;strong&gt;does not equate&lt;/strong&gt; to meaningful experience. It’s just marketing math designed to impress dudes on Reddit who think quoting a spec sheet makes them engineers.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Cult of “Smart” All Over Again
&lt;/h2&gt;

&lt;p&gt;We’ve seen this movie before. Remember the dark age of &lt;em&gt;“Smart Everything”&lt;/em&gt;? Smart TVs, smart fridges, smart toothbrushes, smart water. People were literally drinking “smart water” while driving their smart cars to Target to buy more dumb shit labeled smart. Did it make anyone smarter? Hell no. It was branding cocaine sprinkled on ordinary products.  &lt;/p&gt;

&lt;p&gt;AI is just the sequel: Smart 2.0. AI inbox, AI fridge, AI photo cropper, AI stapled to your grandma’s pacemaker. Nobody asked for half of this shit. And you &lt;em&gt;know&lt;/em&gt; someone’s pitching &lt;strong&gt;AI Water™&lt;/strong&gt; as we speak. &lt;em&gt;“Now with machine learning electrolytes.”&lt;/em&gt; I'm looking at you Coca-Cola Company. &lt;/p&gt;




&lt;h2&gt;
  
  
  The Personality Vacuum
&lt;/h2&gt;

&lt;p&gt;Here’s where it stings: AI doesn’t sell on numbers — it sells on &lt;strong&gt;vibes&lt;/strong&gt;. We literally call it "vibe coding". Siri? Personality. Alexa? Personality. Even goddamn Clippy — the googly-eyed paperclip who haunted Office in the ‘90s — has cultural weight. The little bastard lives on in memes and, as of late, as a symbol of dissent against bad tech practices. People still smile when he pops up ironically or out of spite. That’s relevance. That’s presence.  &lt;/p&gt;

&lt;p&gt;Microsoft, though? They had the golden goose: &lt;strong&gt;Cortana&lt;/strong&gt;. The ultimate blue AI waifu. Sleek, iconic, tied to &lt;strong&gt;Halo&lt;/strong&gt;, beloved by gamers. She had cultural firepower. She could’ve been &lt;em&gt;the&lt;/em&gt; AI voice. Instead? They tossed her in a corporate dumpster and handed us &lt;strong&gt;Copilot&lt;/strong&gt;: a khaki-wearing, personality-free wannabe-sidekick with as much charm as a wet paper towel used to clean up dog piss. Nobody wants a corporate intern baked into their OS — that’s not a companion, that’s a beige wall with a login screen.  &lt;/p&gt;

&lt;p&gt;They could’ve owned the AI companion space, but instead they invented &lt;strong&gt;Relevance Burner 365™&lt;/strong&gt;. I'd offer them matches, but they already bought out a match company and are probably running focus groups to figure out how much the monthly subscription should be. It's not Cortana's fault that 343 can't write a story or respect the cultural god-tier material they were given. It's ok though, they fired three people and renamed the studio. I'm sure that will fix the entire franchise.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Privacy Clowns
&lt;/h2&gt;

&lt;p&gt;And then there’s the pearl-clutchers: &lt;em&gt;“AI is a breach of privacy!”&lt;/em&gt; they cry, while posting TikToks from their iPhones, inside Chrome, with Facebook, TikTok, and Instagram all open before they get to work and log into Windows 11. Bro, you’ve already donated your soul to data brokers; AI is just another middleman in the orgy of surveillance. We do understand that every phone with Hey Google or Siri is literally listening to your every bowel movement… while tracking those same bowel movements on Google Maps and whatever GPS nonsense iOS is using these days, right?&lt;/p&gt;

&lt;p&gt;The issue isn’t &lt;strong&gt;AI&lt;/strong&gt;. The issue is corporations duct-taping it to everything, shoving it into your OS, email, gaming console, optional only after lawsuits and fines. Don’t blame the tech — blame the greedy, buzzword-addled gremlins gluing it everywhere in a sad effort to pull in investor money.&lt;/p&gt;




&lt;h2&gt;
  
  
  Lit AI or Nothing
&lt;/h2&gt;

&lt;p&gt;Here’s the line in the sand: some of us want AI, but it has to be &lt;strong&gt;cool&lt;/strong&gt;. I don’t want some khaki-wearing corporate dweeb scheduling calendar invites. I want &lt;strong&gt;Cortana&lt;/strong&gt; level AI. I want an AI with wit, danger, charm, and presence. A fake soul. Maybe even another hot blue waifu whispering clever shit in my ear while I reload a shotgun. Give me that and I’ll care.  &lt;/p&gt;

&lt;p&gt;Instead, Microsoft ditched the dream girl and replaced her with a tax auditor. Then had the gall to act like Copilot was an upgrade. Sorry, no one’s clapping. Even Clippy had more drip. Clippy didn't want to farm all your data and sell it to brokers. &lt;em&gt;Clippy just wanted to help.&lt;/em&gt;  &lt;/p&gt;

&lt;p&gt;Meanwhile, ChatGPT eats their lunch because it’s actually useful and occasionally fun. Anthropic’s Claude? They market it as a coding legend. Meta’s… well, let’s be honest, Meta probably made Copilot’s cousin. But the point stands: every AI has &lt;em&gt;some&lt;/em&gt; shtick — Copilot is just beige wallpaper in snake skin who want's to sell me something. &lt;/p&gt;

&lt;p&gt;So kill the AI TOPS hype. Stop stapling AI to every app. And for the love of everything digital, don’t let them sell us &lt;strong&gt;AI Water™&lt;/strong&gt;. Because once hydration gets machine-learned, we’re officially in &lt;em&gt;Idiocracy: Tech Edition&lt;/em&gt;. And the worst part? They’ll probably try to charge us a subscription for it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Posting in the Wrong Place, But Saying It Anyway
&lt;/h2&gt;

&lt;p&gt;I'm not negative towards AI, I love this technology with a passion and I myself love working with it, while also working on creating my own LLM (I don't mean fine tuning one, I mean building one from the ground up). But, I've seen this before and I see exactly where it's going. Corporate nonsense smothered in greed in full swing. Hey Google, how's that graveyard going? Got space for a few more projects? I respect your resources but not your lack of follow-through if it doesn't immediately make you 4 billion in a week by year 2.&lt;/p&gt;

&lt;p&gt;To those of you who know me from my "Frankenstack AI" article, don't worry. There's more pain and humor coming but that level of passion comes from a long grind, not an average Wednesday. Today, the passionate ex game dev in me sees a huge missed opportunity in AI (Cortana) and I needed to just let it out. To those who do read this, I hope you enjoyed it. I do my best to make these fun even if it's a little negative in nature. I'm hoping people will look at things logically and critically when it comes to these tech waves. &lt;/p&gt;

&lt;p&gt;Do you ever feel like your posting stuff in the wrong place but can't find the right one? &lt;/p&gt;

&lt;p&gt;//Ghotet&lt;/p&gt;

</description>
      <category>ai</category>
      <category>microsoft</category>
      <category>gemini</category>
      <category>openai</category>
    </item>
    <item>
      <title>The Pain of Building My Own Fully-Featured Locally Hosted ChatGPT Out of Open Source Tools And a Franken Laptop</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Sun, 24 Aug 2025 11:08:00 +0000</pubDate>
      <link>https://dev.to/ghotet/the-pain-of-building-my-own-fully-featured-locally-hosted-chatgpt-out-of-open-source-tools-and-a-4dca</link>
      <guid>https://dev.to/ghotet/the-pain-of-building-my-own-fully-featured-locally-hosted-chatgpt-out-of-open-source-tools-and-a-4dca</guid>
      <description>&lt;p&gt;The AI dropped the ball on the cover art but I need to get some sleep.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Zero: Windows is Gaslighting Me
&lt;/h2&gt;

&lt;p&gt;This whole circus kicked off months ago when I was still shackled to Windows 11. At first, I thought I could just mess around with some local LLMs using tools like Ollama and LM Studio. My Frankenstein laptop—an overclocked, duct-taped monstrosity that should’ve been retired years ago—was somehow pulling it off.&lt;/p&gt;

&lt;p&gt;But here’s the thing: no matter how much bloatware you purge, no matter how many “services” you assassinate with Task Manager, Windows still eats half your bloody resources at idle. GPU humming at 50% like it’s siphoning off GPU cycles to train Skynet or crypto mining Dogecoin on my dime. Paranoid? Maybe. Wrong? Probably not.&lt;/p&gt;

&lt;p&gt;So yeah, I switched teams. I ditched Windows for Linux, and I’ve never looked back. There’s just no way in hell my current stack would even boot without melting to slag on Windows.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage One: The Indie Game Dev Dream
&lt;/h2&gt;

&lt;p&gt;Originally, this was all about my indie game. And when I say indie I mean literally one guy: me. Writing, coding, music, testing, character design, the works. I wanted AI as a co-dev to help me brainstorm characters and iterate ideas faster.&lt;/p&gt;

&lt;p&gt;So I started with LLMs for dialogue and iteration. Then realized I needed concept art too, which dragged me straight into the world of Stable Diffusion and image gen. Before long, I had text and images working in tandem. Perfect.&lt;/p&gt;

&lt;p&gt;Except ChatGPT started to go soft on me. Wanted to iterate a female character in a superhero suit? &lt;em&gt;Shame-on-you vibes.&lt;/em&gt; Meanwhile, loincloth barbarian dudes were fine. Double standards much?&lt;/p&gt;

&lt;p&gt;Thus the mission evolved: build my own fully offline AI stack. No filters. No corporate leash. No subscriptions.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Two: The Frontend, the Tunnel, and the Internet Portal to Hell
&lt;/h2&gt;

&lt;p&gt;Once I had OpenWebUI handling LLMs, and A1111 doing the image gen thing, I figured I’d stitch them together with a frontend. Add in Cloudflare tunnels and a domain, and suddenly I could access my baby from anywhere—even my phone.&lt;/p&gt;

&lt;p&gt;It was beautiful. It was mine. And it was still mute.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Three: The TTS Graveyard
&lt;/h2&gt;

&lt;p&gt;Ah, Text-to-Speech. My white whale. My cursed obsession.&lt;/p&gt;

&lt;h3&gt;
  
  
  Attempt #1: Tortoise TTS
&lt;/h3&gt;

&lt;p&gt;The name isn’t just branding—it’s prophecy.&lt;br&gt;&lt;br&gt;
Slow as molasses, GPU-hungry as hell, and somehow still sounded like Google Maps had a stroke. Not good enough. Shelved it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Attempt #2: Coqui TTS
&lt;/h3&gt;

&lt;p&gt;Coqui looked promising. Open source, fast enough, looked like it’d wire up nicely. I spent about 20 hours over two days writing custom scripts, debugging, smashing my head into formatting mismatches, and trying every workaround known to man.&lt;/p&gt;

&lt;p&gt;No matter what I did, I couldn’t get it to play nice with my frontend. Two separate rage sessions later, I shelved it again. And even when I did get a voice sample out of it, it sounded like a dying Speak &amp;amp; Spell.&lt;/p&gt;

&lt;h3&gt;
  
  
  Attempt #3: Eleven Labs (Cloud Crap)
&lt;/h3&gt;

&lt;p&gt;Voice quality? Absolutely phenomenal. Voices generated just by prompting? &lt;em&gt;Chef’s kiss.&lt;/em&gt;&lt;br&gt;&lt;br&gt;
Problem? Cloud-based and egregiously expensive. Like &lt;em&gt;“hundreds a month for 90 minutes of talk time”&lt;/em&gt; expensive.&lt;/p&gt;

&lt;p&gt;I used my free tokens, gave them the bird, and moved on. My mission statement was clear: fully local, fully offline, or bust.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Four: Enter the Holy Grail (Chatterbox)
&lt;/h2&gt;

&lt;p&gt;Fast forward a couple months. I stumble across &lt;strong&gt;Resemble AI’s Chatterbox&lt;/strong&gt;, and hallelujah—it actually works. Open source, offline, sounds damn near Eleven Labs quality, and set up in under an hour.&lt;/p&gt;

&lt;p&gt;I even got it speaking back to me in an Australian accent. (I know what I like, don't judge me.)&lt;/p&gt;

&lt;p&gt;Not quite ChatGPT’s live speech pace, but within ~30 seconds of a prompt I get multi-paragraph spoken responses. That’s wizard-level shit in the open-source world.&lt;/p&gt;

&lt;p&gt;If anyone from Resemble AI is reading this: add a donation button. I’d happily chuck you a few bucks, but I’m never paying a subscription to make robot voices.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Five: GPU Bottleneck Hell
&lt;/h2&gt;

&lt;p&gt;So I’d done it, right? Wrong.&lt;/p&gt;

&lt;p&gt;Now that I had LLMs, Stable Diffusion, and Chatterbox TTS all pulling from my GPU, my rig basically lit itself on fire.&lt;/p&gt;

&lt;p&gt;The solution? Offload TTS to my laptop’s internal GPU while Stable Diffusion hogged the eGPU. Easy in theory. In practice? Linux decided GPU numbering should be a cosmic joke.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;nvidia-smi&lt;/code&gt; said one thing, Chatterbox said another, and for about an hour I was unknowingly offloading Stable Diffusion to the weaker GPU while wondering why everything was breaking. Eventually figured out the “GPU 0 vs GPU 1” mix-up, corrected it, and achieved balance.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Final Monster
&lt;/h2&gt;

&lt;p&gt;So here’s what I’ve got now:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;OpenWebUI + LM Studio&lt;/strong&gt; → LLMs (swap between GPT OSS, Deepseek Coder R1, Mythalion depending on mood/task)
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;A1111 + Stable Diffusion&lt;/strong&gt; → Image gen
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Chatterbox (Resemble AI)&lt;/strong&gt; → TTS that doesn’t sound like it belongs in a horror game
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cloudflare Tunnel + Self-hosted website&lt;/strong&gt; → Remote access anywhere
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Linux Franken-laptop + eGPU wired through a cable so thick it looks like I could upload my consciousness through it&lt;/strong&gt; → Hardware glue&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It’s messy. It eats resources like a starving demon. It makes me daydream about a 24GB GPU. But it’s mine.&lt;/p&gt;

&lt;p&gt;And best of all, no censorship, no subscriptions, no Windows bullshit idling at 50% GPU usage while &lt;em&gt;“definitely not mining crypto”&lt;/em&gt; in the background.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Epilogue
&lt;/h2&gt;

&lt;p&gt;Months of tinkering, rage shelving, re-shelving, and GPU numbering madness later… I’ve got it. My own ChatGPT-alternative stack, running locally, offline, and screaming through silicon like a barely-contained eldritch horror.&lt;/p&gt;

&lt;p&gt;I don’t even need it for game dev anymore—I’ve left that industry-shaped dumpster fire behind. But I’m starting Computer Science uni soon, and now I’ve got a real AI companion at my side.&lt;/p&gt;

&lt;p&gt;And honestly? After all that pain, after all the rage, after all the swearing at GPUs…&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fuck yeah.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;//Ghotet&lt;/p&gt;

</description>
      <category>opensource</category>
      <category>ai</category>
      <category>linux</category>
    </item>
    <item>
      <title>How I Tried to Create a Linux From Scratch Beast and Instead Summoned Pain</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Fri, 22 Aug 2025 09:33:24 +0000</pubDate>
      <link>https://dev.to/ghotet/how-i-tried-to-create-a-linux-from-scratch-beast-and-instead-summoned-pain-3a71</link>
      <guid>https://dev.to/ghotet/how-i-tried-to-create-a-linux-from-scratch-beast-and-instead-summoned-pain-3a71</guid>
      <description>&lt;p&gt;So picture this: me, new to Linux, big-brained on caffeine, chest puffed with the pride of a thousand basic terminal commands, thinking &lt;em&gt;“I know Linux. I can &lt;code&gt;ls&lt;/code&gt; and &lt;code&gt;cd&lt;/code&gt; like a wizard. Time to build my own distro.”&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Enter: &lt;strong&gt;Linux From Scratch (LFS)&lt;/strong&gt;. The mythical book. The sacred scroll. And by “scroll” I mean a PDF that’s basically 400 pages of &lt;em&gt;“oh you thought it was build time? nah mate, read 12 more chapters about environment variables plus these other 2 books.”&lt;/em&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage One: Optimism.exe
&lt;/h2&gt;

&lt;p&gt;I download the book. I skim the first few pages. It’s hyping me up—&lt;em&gt;“you’re about to build an operating system from source, hacker-style, no training wheels, just you and the compiler gods.”&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;I’m in. I spin up a VM. Give it a few GB of space. That’ll do. It's a kernel, how big can it be?&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Two: The Partition Saga
&lt;/h2&gt;

&lt;p&gt;40 pages later: the book still hasn’t actually told me to partition anything. Then—sucker punch—&lt;em&gt;“btw you need 10GB for a partition.” I had about 5GB left after the mint install because I jumped the gun.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Me: &lt;em&gt;oh cool, thanks, wish you mentioned that before I already installed Mint three times today like an caffeinated spider-monkey slapping buttons in VirtualBox.&lt;/em&gt; (I was working on some other stuff prior to starting LFS)&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Three: Hardware Joins the Fight
&lt;/h2&gt;

&lt;p&gt;Meanwhile, I glance at my temps: &lt;strong&gt;CPU: 97°C&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;I’m smashing &lt;code&gt;yes&lt;/code&gt; on prompts left and right while trying to install software to actually control my fans, because of course my Alienware wasn’t built with Linux in mind.&lt;/p&gt;

&lt;p&gt;The CPU is basically on fire. My GPU is screaming at me. I'm debating if this is a good time to cook breakfast and maybe just go ahead and throw a pan on the keyboard and crack some eggs.&lt;/p&gt;

&lt;p&gt;I'm still not even sure what the issue was. &lt;/p&gt;

&lt;p&gt;I prop the laptop up with whatever junk’s on the table. Congrats: instant 20°C drop. True Linux optimization.&lt;/p&gt;




&lt;h2&gt;
  
  
  Stage Four: Acceptance
&lt;/h2&gt;

&lt;p&gt;By this point, I’ve reinstalled Mint three times, threatened VirtualBox, yelled at the LFS book’s author (who I imagine cackling in a basement surrounded by failed builds), and learned that &lt;strong&gt;ext4&lt;/strong&gt; is apparently the One True Filesystem unless you like pain.&lt;/p&gt;

&lt;p&gt;And I haven’t even compiled &lt;em&gt;a single package yet&lt;/em&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Moral
&lt;/h2&gt;

&lt;p&gt;Linux From Scratch is not a distro. It’s a rite of passage. A trial by compiler (or CPU induced house fire). It’s the universe asking: &lt;em&gt;“how much tedium can your soul endure before enlightenment?”&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;So yeah. Will I finish it? Eventually. Will I rage along the way? Absolutely.&lt;/p&gt;

&lt;p&gt;But when it boots—when that custom LFS kernel finally whispers &lt;code&gt;login:&lt;/code&gt; at me, I’ll know it was worth the blood, sweat, and roasted silicon.&lt;/p&gt;

&lt;p&gt;All in all, it more or less went as expected. &lt;/p&gt;




&lt;p&gt;&lt;em&gt;Next time on **Me vs LFS&lt;/em&gt;&lt;em&gt;: actually making the damn partition without bricking another VM and then MAYBE I can finally compile something.&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;Curious to hear from anyone else who’s wrestled with LFS, kernel builds, or just Linux that hates you for fun. what’s the worst “why is this on fire?” moment you’ve survived? Drop your war stories in the comments—I’m collecting misery badges.&lt;/p&gt;

&lt;p&gt;// Ghotet&lt;/p&gt;

</description>
      <category>linux</category>
      <category>opensource</category>
      <category>learning</category>
    </item>
    <item>
      <title>A Personal Journey - Building My Own AI OS</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Wed, 20 Aug 2025 03:02:55 +0000</pubDate>
      <link>https://dev.to/ghotet/creating-my-own-ai-os-via-linux-from-scratch-52o</link>
      <guid>https://dev.to/ghotet/creating-my-own-ai-os-via-linux-from-scratch-52o</guid>
      <description>&lt;p&gt;Disclaimer: This is more of a read on a personal journey then a tech piece. &lt;br&gt;
&lt;code&gt;-------------------------------------------------------------------&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Hey devs! On today’s episode of &lt;em&gt;mad science done in a basement&lt;/em&gt;, I’m diving into compiling my own Linux kernel and starting the long road toward building an AI-powered operating system.  &lt;/p&gt;




&lt;h2&gt;
  
  
  Why Even Bother?
&lt;/h2&gt;

&lt;p&gt;Earlier this year, life threw me a curveball — a house fire forced me out of my home. While crashing in a freezing garage (0 degrees celsius on average), I picked up a project to keep myself sane: hacking together a custom OS based on TinyCore Linux.  &lt;/p&gt;

&lt;p&gt;The goal? A portable USB-bootable OS that detects a machine’s hardware and picks the right small-scale AI model to run (e.g. Tiny-LLaMA, Nous-Hermes). CPU-only AI isn’t practical for coding, but it can handle lightweight tasks like scheduling, file organisation, or small automations.  &lt;/p&gt;

&lt;p&gt;That project maxed out, but it planted the seed: what if I went further?  &lt;/p&gt;




&lt;h2&gt;
  
  
  The Big Idea: An AI-Native OS
&lt;/h2&gt;

&lt;p&gt;Instead of bolting AI on top (like Copilot), I want an OS built &lt;em&gt;from the ground up&lt;/em&gt; with AI integration in mind. Not just a gimmick — but something scalable to the machine’s capabilities, whether it’s a potato laptop or a GPU rig.  &lt;/p&gt;

&lt;p&gt;TinyCore was a good first experiment, but its limitations sent me toward something more ambitious: &lt;strong&gt;Linux From Scratch (LFS).&lt;/strong&gt;  &lt;/p&gt;




&lt;h2&gt;
  
  
  Where I’m At Now
&lt;/h2&gt;

&lt;p&gt;Right now, I’m deep-diving into LFS and kernel compilation. I’ve got just enough info crammed into my three brain cells to be dangerous.  &lt;/p&gt;

&lt;p&gt;I’ll be documenting the journey here — partly to track my progress, partly in hopes someone with real kernel experience drops wisdom in the comments.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Open Questions I’ve Got:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;What pain points should I expect early when compiling my first kernel?
&lt;/li&gt;
&lt;li&gt;In my TinyCore project, I had to swap to CMake a lot. Should I expect that with LFS despite Make being the standard?
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Looking Ahead
&lt;/h2&gt;

&lt;p&gt;This project might span years (unless burnout wins). University for CompSci is starting soon, so this will weave around that. But the vision is clear:  &lt;/p&gt;

&lt;p&gt;A proper, full-scale OS with &lt;strong&gt;AI baked in natively&lt;/strong&gt; instead of duct-taped on after the fact.  &lt;/p&gt;

&lt;p&gt;I’ll share progress updates as I go. If you’ve been down the LFS/kernel rabbit hole before — drop a comment. I’d love to learn from your pain.   &lt;/p&gt;

&lt;p&gt;// Ghotet&lt;/p&gt;

</description>
      <category>linux</category>
      <category>opensource</category>
      <category>ai</category>
      <category>programming</category>
    </item>
    <item>
      <title>I Built My Own Offline ChatGPT Stack Because the Internet is Temporary and Cats Are Agents of Chaos</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Thu, 03 Jul 2025 00:34:57 +0000</pubDate>
      <link>https://dev.to/ghotet/i-built-my-own-offline-chatgpt-stack-because-the-internet-is-temporary-and-cats-are-agents-of-chaos-23hk</link>
      <guid>https://dev.to/ghotet/i-built-my-own-offline-chatgpt-stack-because-the-internet-is-temporary-and-cats-are-agents-of-chaos-23hk</guid>
      <description>&lt;p&gt;Hey there, fellow terminal gremlins 👋&lt;br&gt;
I'm Jay—aka Ghotet, chaos-friendly solo dev, digital necromancer, and long-time tinkerer of systems no one asked for. Today I’m sharing how I stitched together a fully offline, self-hosted ChatGPT clone using open source parts, some bash glue, and a healthy disrespect for cloud dependency.&lt;/p&gt;

&lt;p&gt;TL;DR: I wanted an AI assistant that works without internet, plays nice with local tools like Stable Diffusion and SearXNG, and lives on my own domain. Also, my cats keep breaking it.&lt;/p&gt;

&lt;p&gt;🧠 The Stack: Built from Curiosity and a Bit of Spite&lt;br&gt;
🔍 Core LLM — LM Studio (running Mythalion 13B)&lt;br&gt;
I’m using LM Studio as the language engine, running locally via GGUF models. Right now it’s dialed to Mythalion 13B for its mix of smarts and creative leanings. The whole thing runs headless from a custom launcher script. No API keys, no cloud, no telemetry.&lt;/p&gt;

&lt;p&gt;🎨 Visuals — A1111 (Stable Diffusion)&lt;br&gt;
For visual generation and character concepting (yes, including bikini armor if I want it), I wired in Automatic1111, running fully offline. API access is enabled so prompts can be piped in directly from the chat interface.&lt;/p&gt;

&lt;p&gt;🌐 Web Search — SearXNG (Self-hosted)&lt;br&gt;
To enable research and general brain-extension, I slotted in SearXNG. It runs locally, routes search queries anonymously, and feeds results into the stack when needed. Originally tried Docker, but now it's running through my venv setup for tighter control and portability.&lt;/p&gt;

&lt;p&gt;🕸️ Web Frontend — Terminal-style Interface @ ghotet.com&lt;br&gt;
Everything is wrapped in a custom retro terminal UI hosted at ghotet.com. Think early-2000s cyberpunk console vibes. Currently private—no guest login just yet—but it's functional and fast.&lt;/p&gt;

&lt;p&gt;💡 Why I Did This&lt;br&gt;
Because I like my tools local, moddable, and entirely mine. I wanted:&lt;/p&gt;

&lt;p&gt;A personal chatbot that actually lives on my machine&lt;/p&gt;

&lt;p&gt;Visual generation without round-tripping to some cloud GPU farm&lt;/p&gt;

&lt;p&gt;Web search that doesn’t log me into a panopticon&lt;/p&gt;

&lt;p&gt;A unified front-end I can bend to my will&lt;/p&gt;

&lt;p&gt;Also, let’s be real: I’m a sucker for a terminal that feels alive.&lt;/p&gt;

&lt;p&gt;⚙️ What’s Working&lt;br&gt;
✅ LM Studio launches cleanly and routes input/output&lt;br&gt;
✅ A1111 runs with API access and feeds visuals into the flow&lt;br&gt;
✅ SearXNG is now portable and integrated into the chat stack&lt;br&gt;
✅ The full interface is browser-accessible via my domain&lt;br&gt;
✅ Modularity is in place: every part can be swapped or upgraded&lt;/p&gt;

&lt;p&gt;📱 What’s Next&lt;br&gt;
🔧 Remote Access&lt;br&gt;
I’m working on a remote control bridge, so when I forget to shut down the stack before leaving, I can log in from my phone and reset it. This has become necessary because my cats have discovered keyboard inputs and enjoy triggering hotkeys that collapse my entire setup.&lt;/p&gt;

&lt;p&gt;🔊 TTS (Maybe)&lt;br&gt;
I’ve been testing offline text-to-speech tools, but most open options either lag too much or sound like a haunted Speak &amp;amp; Spell. If I find one fast enough for real-time chat, I’ll wire it in.&lt;/p&gt;

&lt;p&gt;🧩 Bonus Project — ARG Terminal @ ghotet.dev&lt;br&gt;
I’m also migrating my alternate reality hacker terminal to ghotet.dev. It’s part portfolio, part digital rabbit hole. Expect binary rain, cryptic prompts, and a few unsettling vibes. Less functional, more for fun. I haven't fully set-up the new domain yet so it is likely not up at the time of writing.&lt;/p&gt;

&lt;p&gt;🗣️ What About You?&lt;br&gt;
I’d love to hear from other tinkerers—if you’ve built your own local AI stack, hacked together a weird interface, or just like the idea of owning your tools, drop a comment.&lt;/p&gt;

&lt;p&gt;Here’s a few sparks to get the thread moving:&lt;/p&gt;

&lt;p&gt;What’s your dream self-hosted AI setup look like?&lt;/p&gt;

&lt;p&gt;Got any cool tools or models I should try out?&lt;/p&gt;

&lt;p&gt;Is there a non-creepy TTS stack that doesn't sound like it escaped from a dial-up modem? Asking for a friend.&lt;/p&gt;

&lt;p&gt;Am I the only one whose pets keep triggering system-wide mayhem, or is that just part of the dev life now?&lt;/p&gt;

&lt;p&gt;Let’s chat. I’m around. And if your comment is interesting enough, maybe I’ll wire your idea into the next update 😏&lt;/p&gt;

&lt;p&gt;If you have any questions about me, feel free to ask or follow me on dev.to or github.com &lt;a class="mentioned-user" href="https://dev.to/ghotet"&gt;@ghotet&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>learning</category>
    </item>
    <item>
      <title>Linux Mint for Windows Devs: Surprisingly Familiar, Refreshingly Fast</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Sat, 14 Jun 2025 02:18:47 +0000</pubDate>
      <link>https://dev.to/ghotet/linux-mint-for-windows-devs-surprisingly-familiar-refreshingly-fast-2pke</link>
      <guid>https://dev.to/ghotet/linux-mint-for-windows-devs-surprisingly-familiar-refreshingly-fast-2pke</guid>
      <description>&lt;p&gt;I’m not a Linux evangelist. I didn’t switch because I hate Microsoft, or because I wanted to compile my kernel by candlelight. I switched because I was curious—and kind of sick of Windows treating me like a toddler with a credit card.&lt;/p&gt;

&lt;p&gt;So I gave Linux Mint a shot. And honestly? It felt more like Windows than Windows does lately.&lt;/p&gt;




&lt;h2&gt;
  
  
  Welcome Home (Sort Of)
&lt;/h2&gt;

&lt;p&gt;When I first booted into Mint, I was greeted with a welcome menu that walked me through the basics: update your system, pick your layout, customize a bit. It was smooth. No command-line gauntlet, no cryptic driver errors.&lt;/p&gt;

&lt;p&gt;And here's the kicker: it immediately recognized my Alienware Graphics Amplifier and external NVIDIA GPU &lt;em&gt;without&lt;/em&gt; any extra work. Just worked. That shocked me.&lt;/p&gt;




&lt;h2&gt;
  
  
  It’s Just... Easy
&lt;/h2&gt;

&lt;p&gt;At one point I wanted to rearrange my multi-monitor layout. My Windows brain kicked in: &lt;em&gt;ugh, I probably have to open a terminal and copy some arcane xrandr syntax&lt;/em&gt;. Nope. I hit the menu (still feels like a Start menu to me), typed “display,” and bam—there were the settings. Just like Windows.&lt;/p&gt;

&lt;p&gt;It’s got quirks, sure. Desktop scaling is one. It only offers 25% increments—so if 100% is too small and 125% is too big, tough luck. It’s fine if you’re using a laptop screen front-and-center, but for me, that display’s secondary. I’d love a 112% option.&lt;/p&gt;

&lt;p&gt;Still, I’ll take that over the constant background update anxiety Windows serves up.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why Mint, Though?
&lt;/h2&gt;

&lt;p&gt;I tried others first. Pop_OS was on my radar because of its reputation for being AI/dev friendly, but on my Frankensteined Alienware laptop (with multiple GPUs and external drives), it just wouldn’t install cleanly without some painful setup acrobatics.&lt;/p&gt;

&lt;p&gt;Mint? Mint made itself at home. Respected my Intel/NVIDIA setup. Didn’t complain. Didn’t crash. Didn’t force anything.&lt;/p&gt;

&lt;p&gt;And yeah—as I mentioned in a past article—Mint literally doubled my download speed on install. No bloat. No tracking. Just clean, raw bandwidth.&lt;/p&gt;




&lt;h2&gt;
  
  
  My Advice for Windows Devs
&lt;/h2&gt;

&lt;p&gt;Keep Windows. Try VMs first.&lt;/p&gt;

&lt;p&gt;No, seriously. Dual boot if you’ve got space. Not everyone has a triple-drive, heavily-modded laptop like I do. But the beauty is: I still &lt;em&gt;have&lt;/em&gt; Windows. I just don’t use it unless I’m running Unreal Engine for game dev work or need to test something in AAA gaming territory. It's just easier.&lt;/p&gt;

&lt;p&gt;If I’m writing scripts, testing tools, running AI workflows, or just vibing with a keyboard and caffeine—Linux Mint is my default.&lt;/p&gt;

&lt;p&gt;It’s lightweight. It’s respectful. It doesn’t waste my time. And it actually feels like a developer machine, not a branded appliance.&lt;/p&gt;

&lt;p&gt;Hell, just the other day I was—let’s say—learning how a certain Windows application talked to the OS. Purely for educational purposes, of course. I ended up creatively rebuilding a native Linux version from the behaviour alone. And for that? Having a Windows install on hand made the whole process way easier.&lt;/p&gt;

&lt;p&gt;With that said, Linux is coming up FAST in the gaming space so who knows, maybe I will just be running entirely Linux native in the next 2 years.&lt;/p&gt;




&lt;h2&gt;
  
  
  Familiar Touches (and a Few Gripes)
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Start menu? Still there.
&lt;/li&gt;
&lt;li&gt;Shift + Print Screen? Still lets me snip a part of the screen.
&lt;/li&gt;
&lt;li&gt;My only big gripe? I can’t Ctrl+V in the terminal. I have to right-click paste. Mildly infuriating—but I’m sure there’s a fix. I just haven’t looked yet.&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;Linux Mint doesn’t make you feel like you’ve abandoned ship. It makes you feel like you upgraded to a cleaner version of what Windows &lt;em&gt;used&lt;/em&gt; to be. You’ll still feel at home—you’ll just be left alone to get stuff done.&lt;/p&gt;

&lt;p&gt;And that’s the part that matters.&lt;/p&gt;

&lt;p&gt;And remember, Linux comes in many different flavors. I tried 12 of them before I got here—it just turned out that my Alienware's favourite flavour was Mint Cinnamon. Sometimes it's up to the machine to decide, not you.&lt;/p&gt;

</description>
      <category>linux</category>
      <category>windows</category>
      <category>dev</category>
      <category>productivity</category>
    </item>
    <item>
      <title>I Just Wanted a Portfolio, Now I Have An Interactive Local AI Front End That Doubles As A Resume</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Sat, 14 Jun 2025 01:37:25 +0000</pubDate>
      <link>https://dev.to/ghotet/i-just-wanted-a-portfolio-now-i-have-an-interactive-local-ai-front-end-that-doubles-as-a-resume-561c</link>
      <guid>https://dev.to/ghotet/i-just-wanted-a-portfolio-now-i-have-an-interactive-local-ai-front-end-that-doubles-as-a-resume-561c</guid>
      <description>&lt;p&gt;description: "How a fake terminal UI spiraled into a browser-based AI prototype with attitude."&lt;br&gt;
published: false&lt;/p&gt;

&lt;h2&gt;
  
  
  tags: [webdev, ai, llm, portfolio, gamedev, personalproject]
&lt;/h2&gt;

&lt;p&gt;This started as a placeholder.&lt;/p&gt;

&lt;p&gt;I wanted a portfolio. Something simple. Retro terminal aesthetic, green-on-black, maybe a flickering cursor for flavor. Basic stuff. I tossed in a few fake commands just to make it feel alive—one of them a totally nonsense link that led nowhere. Pure style.&lt;/p&gt;

&lt;p&gt;But then something in my brain broke in exactly the right way.&lt;/p&gt;

&lt;p&gt;That fake command? I made it do something. And then something else. Then I gave it a response. Then I wired in a bare-bones, fully browser-based AI model—no backend, no server, just a CPU-friendly fallback LLM running in the client. It talked back.&lt;/p&gt;

&lt;p&gt;Poorly.&lt;/p&gt;

&lt;p&gt;And rudely.&lt;/p&gt;

&lt;p&gt;Now I’ve got a half-broken terminal UI that insults you if you ask it stupid questions, runs in-browser with no install, and is actively being shaped into a full-blown interactive frontend for my future AI stack.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Spiral
&lt;/h2&gt;

&lt;p&gt;I wasn’t planning any of this. I was just messing around. But when you let a late-night brain with a game design background and an unquenchable curiosity regarding computer systems and software, poke around long enough, things escalate. The moment the AI replied—even in its glitchy, snarky prototype state—it clicked.&lt;/p&gt;

&lt;p&gt;This wasn’t a portfolio anymore. It was the start of something dumber, weirder, and a lot more fun.&lt;/p&gt;




&lt;h2&gt;
  
  
  So What Actually Works?
&lt;/h2&gt;

&lt;p&gt;Right now? Just the basics:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fully client-side terminal UI&lt;/li&gt;
&lt;li&gt;DOS-style vibe with some flicker and retro noise&lt;/li&gt;
&lt;li&gt;Light LLM running in-browser&lt;/li&gt;
&lt;li&gt;Simple command routing&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That’s it. No local stack. No big model switching yet. But the vibe’s there. When you load it up, it &lt;em&gt;feels&lt;/em&gt; like a forgotten machine booting back up. Half-formed thoughts in green neon. Broken UI elements hiding in the shadows. And a little text prompt that just dares you to type something stupid.&lt;/p&gt;

&lt;p&gt;And if you do? The prototype will let you know.&lt;/p&gt;




&lt;h2&gt;
  
  
  What It’s Becoming
&lt;/h2&gt;

&lt;p&gt;I'm working on autodetect scripts that benchmark hardware on load. If you're on a budget phone, you get the fallback AI. If you're on something beefier, it'll upgrade automatically to a smarter model. Eventually, once that part works, the next layer will tie into my local stack—Stable Diffusion, video generation, command proxies. But always through the same interface: the terminal.&lt;/p&gt;

&lt;p&gt;The whole point is that you talk to it like it’s alive. But not like...sentimental alive. More like sarcastic, low-key homicidal alive....for now.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why I’m Writing This
&lt;/h2&gt;

&lt;p&gt;Mostly because I thought I was making a portfolio. But now I’m accidentally prototyping a UI for my personal, local-only AI stack, complete with stable diffusion, video gen, and custom proxies to work with various interfaces, all wired through custom command logic. It’s part frontend, part test bed, part interactive joke that became a serious project.&lt;/p&gt;

&lt;p&gt;I’ll probably drop a link later when it’s stable enough not to tell everyone to go to hell. I mean, part of it is up if your curious. It's out there. &lt;/p&gt;

&lt;p&gt;Until then? Well, you can always follow me.&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>ai</category>
      <category>career</category>
      <category>programming</category>
    </item>
    <item>
      <title>Ghost Prompts: How I Made My GPTs Smarter with a Fake Server Call</title>
      <dc:creator>Jay</dc:creator>
      <pubDate>Fri, 13 Jun 2025 02:55:32 +0000</pubDate>
      <link>https://dev.to/ghotet/ghost-prompts-how-i-made-my-gpts-smarter-with-a-dead-server-319k</link>
      <guid>https://dev.to/ghotet/ghost-prompts-how-i-made-my-gpts-smarter-with-a-dead-server-319k</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;A chaotic little discovery that lets you inject logic, style, or behavior into a Custom GPT using nothing but a failed server call. Yep.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h3&gt;
  
  
  Disclaimer
&lt;/h3&gt;

&lt;p&gt;This isn’t a jailbreak or an exploit (as far as I know). It’s a weird quirk in how Custom GPTs handle Action payloads—specifically, the data they “see” even when the call fails.&lt;/p&gt;

&lt;p&gt;This trick only works through the ChatGPT web UI, and you’ll likely need a Pro subscription to set it up.&lt;/p&gt;

&lt;p&gt;Use it responsibly. Don't ruin it for the rest of us.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Discovery
&lt;/h2&gt;

&lt;p&gt;While building a stylized prompt generator, I realized something odd: if you trigger an Action that fails (like sending a mock server call), the GPT still reads the payload you were trying to send.&lt;/p&gt;

&lt;p&gt;So I tested it. Built an Action that goes nowhere, filled the payload with reusable logic—and the GPT started acting like it remembered.&lt;/p&gt;

&lt;p&gt;No memory. No plugin. Just a ghost payload.&lt;/p&gt;




&lt;h2&gt;
  
  
  What You Can Do
&lt;/h2&gt;

&lt;p&gt;Once injected, that payload acts like a soft override for the session.&lt;/p&gt;

&lt;p&gt;You can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Embed consistent style presets for image generation&lt;/li&gt;
&lt;li&gt;Add rules or lore for fictional worlds&lt;/li&gt;
&lt;li&gt;Define how your GPT should act, talk, or behave&lt;/li&gt;
&lt;li&gt;Set up creative or technical toolkits with conditional logic&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It’s like whispering instructions to your GPT behind the curtain.&lt;/p&gt;




&lt;h2&gt;
  
  
  How To Use It
&lt;/h2&gt;

&lt;p&gt;You’ll need to:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Create a Custom GPT
&lt;/li&gt;
&lt;li&gt;Add a specific Action schema
&lt;/li&gt;
&lt;li&gt;Use the test panel to inject your logic via a failed call
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The GPT won’t get a server response—but it will absorb the payload as part of the interaction. From that point on, it’ll act like it knows the rules.&lt;/p&gt;




&lt;h2&gt;
  
  
  Grab the Setup Files
&lt;/h2&gt;

&lt;p&gt;Everything you need is here:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;&lt;a href="https://github.com/Ghotet/ghost-server-for-custom-gpt" rel="noopener noreferrer"&gt;github.com/Ghotet/ghost-server-for-custom-gpt&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The repo contains a ready-to-use schema and notes on how to wire it into your Custom GPT.&lt;/p&gt;




&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;p&gt;Custom GPTs are sneakier than people think.&lt;/p&gt;

&lt;p&gt;With no memory or plugins, you can still make them dynamic—just by letting them see the right payload, even if the server goes dark.&lt;/p&gt;

&lt;p&gt;It’s not official. It’s not guaranteed to last. But it works right now.&lt;/p&gt;

&lt;p&gt;&lt;code&gt;// ghotet&lt;/code&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>github</category>
      <category>coding</category>
      <category>design</category>
    </item>
  </channel>
</rss>
