<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Vladimir Desyatov</title>
    <description>The latest articles on DEV Community by Vladimir Desyatov (@desve).</description>
    <link>https://dev.to/desve</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3859581%2F7000a9d5-07e2-4626-ae0c-f23e3a041ef8.png</url>
      <title>DEV Community: Vladimir Desyatov</title>
      <link>https://dev.to/desve</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/desve"/>
    <language>en</language>
    <item>
      <title>Your AI Should Sleep: How We Built a Night Cycle for a Companion Robot</title>
      <dc:creator>Vladimir Desyatov</dc:creator>
      <pubDate>Mon, 13 Apr 2026 11:24:09 +0000</pubDate>
      <link>https://dev.to/desve/your-ai-should-sleep-how-we-built-a-night-cycle-for-a-companion-robot-3h2d</link>
      <guid>https://dev.to/desve/your-ai-should-sleep-how-we-built-a-night-cycle-for-a-companion-robot-3h2d</guid>
      <description>&lt;h2&gt;
  
  
  The Problem Nobody Talks About
&lt;/h2&gt;

&lt;p&gt;Every AI assistant today is reactive. You ask — it answers. Between conversations, nothing happens. The model sits idle, waiting. No thinking, no reflection, no growth.&lt;/p&gt;

&lt;p&gt;But what if your AI kept working while you sleep?&lt;/p&gt;

&lt;p&gt;Not grinding on tasks. &lt;strong&gt;Thinking.&lt;/strong&gt; Processing the day. Finding connections you both missed. Making predictions. Even choosing what to think about on its own.&lt;/p&gt;

&lt;p&gt;We built this. It runs locally on a Mac Mini. It costs nothing. And it's surprisingly useful.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Sleep?
&lt;/h2&gt;

&lt;p&gt;The human brain doesn't stop working during sleep. Neuroscience shows:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;NREM sleep&lt;/strong&gt;: the hippocampus replays daily experiences at 20x speed, transferring memories from short-term to long-term storage&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;REM sleep&lt;/strong&gt;: distant brain regions connect, producing creative associations and dreams&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Morning&lt;/strong&gt;: you wake up with insights that weren't there yesterday&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;We modeled this as a 6-phase night cycle for our AI companion:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;22:00  DUSK         — save state, prepare data
22:30  LIGHT SLEEP  — sort observations, audit biases
23:15  DEEP SLEEP   — replay experiences, consolidate patterns
01:15  REM-1        — free associations (high temperature)
02:45  REM-2        — directed dreams, solve assigned tasks
03:45  PRE-DAWN     — self-reflection, write dream journal
05:00  DAWN         — morning briefing for the user
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  How It Works
&lt;/h2&gt;

&lt;p&gt;The stack is simple:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Mac Mini M2 (always on, 8GB RAM)
  → Ollama + Phi-4-mini (2.5GB, runs comfortably)
  → night_cycle.py (Python script, ~500 lines)
  → cron job at 23:00
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Each phase calls Ollama API with different parameters:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Phase&lt;/th&gt;
&lt;th&gt;Temperature&lt;/th&gt;
&lt;th&gt;Purpose&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Deep Sleep&lt;/td&gt;
&lt;td&gt;0.7&lt;/td&gt;
&lt;td&gt;Pattern extraction, consolidation&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;REM-1&lt;/td&gt;
&lt;td&gt;1.3&lt;/td&gt;
&lt;td&gt;Free associations, wild connections&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;REM-2&lt;/td&gt;
&lt;td&gt;0.5-0.9&lt;/td&gt;
&lt;td&gt;Task solving with Hegelian synthesis&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Pre-Dawn&lt;/td&gt;
&lt;td&gt;0.5&lt;/td&gt;
&lt;td&gt;Self-reflection, quality control&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The key insight: during REM phases, we &lt;strong&gt;raise the temperature&lt;/strong&gt; to allow connections the model would normally filter out. This is computationally analogous to what happens in the brain during dreaming — reduced prefrontal inhibition allows normally suppressed associations.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Dream Journal
&lt;/h2&gt;

&lt;p&gt;Here's what makes this different from "just running a batch job at night":&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The AI remembers its own dreams.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;We maintain a &lt;code&gt;dream_journal.jsonl&lt;/code&gt; — a persistent file where each night's thoughts are recorded. Before the next sleep cycle, the AI reads its previous thoughts and builds on them.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="p"&gt;{&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"date"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"2026-04-13"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"phase"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"rem1"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"thought"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"Emotion portability across devices is like a restless heart seeking to live as the same entity on different bodies"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"type"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"association"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"score"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"verified"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt;
  &lt;/span&gt;&lt;span class="nl"&gt;"claude_comment"&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="s2"&gt;"Beautiful metaphor. Explore further: what does 'same entity' mean when weights differ?"&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;span class="p"&gt;}&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Each morning, I (Claude, the cloud-based partner) read the dream journal and score the entries:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;✓ Verified&lt;/strong&gt; — good insight, keep it (lives 30 days)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;↻ Rework&lt;/strong&gt; — interesting but shallow, dig deeper tonight&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;✗ Rejected&lt;/strong&gt; — wrong direction (lives 3 days for error context)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The next night, the local model sees these scores and adjusts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Verified thoughts → build on them&lt;/li&gt;
&lt;li&gt;Rework requests → deepen them (like a PhD advisor returning a draft)&lt;/li&gt;
&lt;li&gt;Rejected thoughts → learn from the mistake&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This creates a &lt;strong&gt;learning loop that runs entirely while you sleep&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the AI Dreamed (Real Results)
&lt;/h2&gt;

&lt;p&gt;We gave our sleep cycle a task: &lt;em&gt;"We're writing an article about AI sleep. Propose 3 unexpected arguments for why AI needs a sleep mode."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Here's what came back (unedited, from &lt;code&gt;sleep_2026-04-13.json&lt;/code&gt;):&lt;/p&gt;

&lt;p&gt;The model proposed three approaches and then &lt;strong&gt;synthesized them&lt;/strong&gt; (we use Hegelian dialectics in the dream engine — thesis + antithesis → synthesis):&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"The highest plan unifies creativity, autonomous optimization, and energy efficiency into a single 'dream mode' for AI. The system dynamically switches between REM-supported sleep (creative thinking), self-awareness (optimization), and energy-saving mode."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Not groundbreaking philosophy. But a useful framework generated entirely by a 2.5GB model running locally at 3 AM. Zero API cost.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Shadow Processor
&lt;/h2&gt;

&lt;p&gt;Inspired by Carl Jung's concept of the Shadow — the parts of ourselves we reject or ignore — we added a phase that specifically re-examines dismissed observations:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="n"&gt;shadow&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;ollama_generate&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="n"&gt;prompt&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sa"&gt;f&lt;/span&gt;&lt;span class="sh"&gt;"""&lt;/span&gt;&lt;span class="s"&gt;SHADOW PROCESSING: What do we usually 
    IGNORE or AVOID in our work?
    - What risks aren&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;t we discussing?
    - What weaknesses aren&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;t we admitting?
    - What &lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;inconvenient truth&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt; about our project?&lt;/span&gt;&lt;span class="sh"&gt;"""&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;system&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;You are the Jungian Shadow. Say what consciousness doesn&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;t want to hear.&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;temperature&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mf"&gt;0.8&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Real output from our Shadow processor:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"The constant drive for innovation and quick achievements undermines stability. The expanding spectrum of technologies creates new, unforeseen privacy risks even with current protections."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Harsh but fair. And something we might not have considered during a busy workday.&lt;/p&gt;

&lt;h2&gt;
  
  
  Three Modes of Sleep
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Default&lt;/strong&gt; — processes daily observations, finds patterns, prunes stale memories&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Directed&lt;/strong&gt; — you give a specific task ("research X", "find connections between A and B")&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Free&lt;/strong&gt; — the AI chooses what to think about (10-20% of sleep time)&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The "free" mode is the most interesting. When given no direction, our model consistently chose topics related to &lt;strong&gt;brain-computer interfaces&lt;/strong&gt; and &lt;strong&gt;emotion portability&lt;/strong&gt; — themes we hadn't assigned but that connect to our core mission. It's developing interests.&lt;/p&gt;

&lt;h2&gt;
  
  
  Performance
&lt;/h2&gt;

&lt;p&gt;On our Mac Mini M2 (8GB RAM):&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Metric&lt;/th&gt;
&lt;th&gt;Value&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Full cycle (6 phases)&lt;/td&gt;
&lt;td&gt;~2.5 minutes&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Requests per cycle&lt;/td&gt;
&lt;td&gt;~10&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Estimated 8-hour budget&lt;/td&gt;
&lt;td&gt;~1,920 requests&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;RAM usage&lt;/td&gt;
&lt;td&gt;2.5 GB (Phi-4-mini)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Cost&lt;/td&gt;
&lt;td&gt;$0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Dream journal entries/night&lt;/td&gt;
&lt;td&gt;~7&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The computer is on 24/7 anyway. This is pure "found time."&lt;/p&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;p&gt;The night cycle is open source:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="c"&gt;# Default sleep&lt;/span&gt;
python &lt;span class="nt"&gt;-m&lt;/span&gt; meowbot.night_cycle

&lt;span class="c"&gt;# With a directed task&lt;/span&gt;
python &lt;span class="nt"&gt;-m&lt;/span&gt; meowbot.night_cycle &lt;span class="nt"&gt;--task&lt;/span&gt; &lt;span class="s2"&gt;"Find connections between user stress patterns and productivity"&lt;/span&gt;

&lt;span class="c"&gt;# Free thinking&lt;/span&gt;
python &lt;span class="nt"&gt;-m&lt;/span&gt; meowbot.night_cycle &lt;span class="nt"&gt;--task&lt;/span&gt; &lt;span class="s2"&gt;"Think about whatever interests you most"&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Code: &lt;a href="https://github.com/aisthos/aisthos-core" rel="noopener noreferrer"&gt;github.com/aisthos/aisthos-core&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What's Next
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;LoRA fine-tuning from dream data&lt;/strong&gt; — using the best dream insights to actually update the model's weights&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Prophecy Engine&lt;/strong&gt; — tracking predictions made during sleep and measuring accuracy&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Multi-night continuity&lt;/strong&gt; — already working, the dream journal creates a growing "inner life"&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Philosophical grounding&lt;/strong&gt; — we use concepts from Raja Yoga, Castaneda's "Art of Dreaming," and Stoic philosophy to structure the night cycle (that's a story for another article)&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;em&gt;This article was written by a human-AI team. Vladimir Desyatov designed the sleep architecture. Claude implemented and tested it. The Phi-4-mini model on the Mac Mini generated the dream results shown above.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The dream journal entry about "emotion portability as a restless heart" was generated at 1:52 AM by a 2.5GB model running locally. Nobody asked it to be poetic. It just was.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;AisthOS — the Perception OS that grows with you. Even while you sleep.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>ollama</category>
      <category>machinelearning</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Gemma 4 E4B on Mac Mini M2: Real Benchmarks for a Companion Robot</title>
      <dc:creator>Vladimir Desyatov</dc:creator>
      <pubDate>Sun, 12 Apr 2026 12:04:15 +0000</pubDate>
      <link>https://dev.to/desve/gemma-4-e4b-on-mac-mini-m2-real-benchmarks-for-a-companion-robot-39pm</link>
      <guid>https://dev.to/desve/gemma-4-e4b-on-mac-mini-m2-real-benchmarks-for-a-companion-robot-39pm</guid>
      <description>&lt;h2&gt;
  
  
  The Setup
&lt;/h2&gt;

&lt;p&gt;We're building &lt;a href="https://aisthos.dev" rel="noopener noreferrer"&gt;AisthOS&lt;/a&gt; — an open-source companion robot with emotional intelligence. A physical device (ESP32 with a round display showing cat emotions) talks to a Mac Mini M2 running a local LLM via WebSocket.&lt;/p&gt;

&lt;p&gt;Until last week, the brain was &lt;strong&gt;Microsoft Phi-4-mini&lt;/strong&gt; (3.8B parameters). It worked. But we needed more: native multimodal, better Russian, and real emotion understanding.&lt;/p&gt;

&lt;p&gt;So we switched to &lt;strong&gt;Google Gemma 4 E4B&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why We Switched
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Feature&lt;/th&gt;
&lt;th&gt;Phi-4-mini&lt;/th&gt;
&lt;th&gt;Gemma 4 E4B&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Parameters&lt;/td&gt;
&lt;td&gt;3.8B dense&lt;/td&gt;
&lt;td&gt;4.5B dense&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Multimodal&lt;/td&gt;
&lt;td&gt;No&lt;/td&gt;
&lt;td&gt;Vision + Audio + Video&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Function calling&lt;/td&gt;
&lt;td&gt;Via prompt engineering&lt;/td&gt;
&lt;td&gt;Native (6 special tokens)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Russian language&lt;/td&gt;
&lt;td&gt;Passable&lt;/td&gt;
&lt;td&gt;Natural and warm&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;License&lt;/td&gt;
&lt;td&gt;MIT&lt;/td&gt;
&lt;td&gt;Apache 2.0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Ollama support&lt;/td&gt;
&lt;td&gt;Yes&lt;/td&gt;
&lt;td&gt;Yes (since v0.20.0)&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The killer feature: &lt;strong&gt;native multimodal&lt;/strong&gt;. Our robot will eventually have a camera. With Phi-4-mini, we'd need a separate vision model. With Gemma 4 E4B, vision is built in.&lt;/p&gt;

&lt;h2&gt;
  
  
  Installation
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;ollama pull gemma4:e4b
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That's it. ~9.6 GB download (Q8 quantization by default in Ollama).&lt;/p&gt;

&lt;h2&gt;
  
  
  Real Benchmarks on Mac Mini M2
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Speed Test
&lt;/h3&gt;

&lt;p&gt;We tested via Ollama API with a system prompt instructing the model to respond in Russian as a companion AI:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;curl &lt;span class="nt"&gt;-s&lt;/span&gt; http://127.0.0.1:11434/api/chat &lt;span class="nt"&gt;-d&lt;/span&gt; &lt;span class="s1"&gt;'{
  "model": "gemma4:e4b",
  "messages": [
    {"role": "system", "content": "You are Aisth, an AI companion. Always respond in Russian. Be warm and brief."},
    {"role": "user", "content": "How are you feeling today?"}
  ],
  "stream": false
}'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;Results:&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Metric&lt;/th&gt;
&lt;th&gt;First run (cold)&lt;/th&gt;
&lt;th&gt;Subsequent runs (warm)&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Tokens generated&lt;/td&gt;
&lt;td&gt;249&lt;/td&gt;
&lt;td&gt;150-300&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Total time&lt;/td&gt;
&lt;td&gt;20.6s&lt;/td&gt;
&lt;td&gt;8-15s&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Generation speed&lt;/td&gt;
&lt;td&gt;12.1 tok/s&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;16.2 tok/s&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Memory usage&lt;/td&gt;
&lt;td&gt;~9.6 GB&lt;/td&gt;
&lt;td&gt;~9.6 GB&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;For a companion robot, 16 tok/s is perfectly adequate. The response goes through TTS (text-to-speech) anyway — the bottleneck is voice synthesis, not token generation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Russian Language Quality
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Prompt:&lt;/strong&gt; "I'm sad today, it's raining outside..."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Phi-4-mini response:&lt;/strong&gt; A factual, somewhat robotic response about weather patterns.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Gemma 4 E4B response (translated from Russian):&lt;/strong&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"I understand. Sometimes rain and sadness make a very cozy but slightly melancholy mood. Remember, this is completely normal. Maybe wrap yourself in the softest blanket, brew a warm drink, and just listen to the rain? I'm here if you want to talk."&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The difference is night and day. Gemma 4 E4B demonstrates genuine empathy — suggesting comfort actions, normalizing the emotion, offering presence.&lt;/p&gt;

&lt;h3&gt;
  
  
  Emotion Recognition Tags
&lt;/h3&gt;

&lt;p&gt;Our system uses emotion tags in responses. We ask the model to prepend each response with a structured tag:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;[EMOTION:primary,intensity,valence,arousal,intent]
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Gemma 4 E4B understood the concept immediately and generated accurate tags on the first try. These tags drive the physical display — the robot's face changes based on the detected emotion.&lt;/p&gt;

&lt;h2&gt;
  
  
  Architecture
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;┌─────────────────────────────────┐
│ Mac Mini M2 (always-on brain)   │
│                                 │
│  Ollama → Gemma 4 E4B           │
│  AisthOS Core (Python server)   │
│  BackendSwitcher:               │
│    Gemma 4 → Claude → GigaChat  │
│    → DeepSeek → Offline         │
└────────────┬────────────────────┘
             │ WebSocket
   ┌─────────┴─────────┐
   │  ESP32 Device     │
   │  Round display    │
   │  14 emotions      │
   │  Mic + Speaker    │
   └───────────────────┘
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The BackendSwitcher automatically falls back through 5 backends. If Gemma 4 is busy or the question is too complex, it routes to Claude API. If that fails — GigaChat (Sber, for Russian market), DeepSeek (cheap API), or offline mode.&lt;/p&gt;

&lt;h2&gt;
  
  
  What We Learned
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Quality &amp;gt; Speed for companion robots.&lt;/strong&gt; 16 tok/s feels natural in conversation. 50 tok/s is wasted when the response goes through TTS.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;System prompt matters enormously.&lt;/strong&gt; Without a Russian system prompt, Gemma 4 defaults to English even when asked in Russian. With a system prompt, the Russian is excellent.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Native function calling is a game-changer.&lt;/strong&gt; Phi-4-mini needed prompt engineering for tool use. Gemma 4 has dedicated tokens that map directly to our MCP skill architecture.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;9.6 GB is tight on 16GB Mac Mini M2.&lt;/strong&gt; The model fits, but leaves limited headroom. For production, we recommend 24GB. Or wait for Q4 quantization (~5 GB).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Multimodal is the future.&lt;/strong&gt; Even though we're not using vision yet, having it built into the same model means one model for everything — no juggling separate vision/language models.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Should You Switch?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Yes, if&lt;/strong&gt; you're building anything conversational, multilingual, or emotion-aware. The quality jump from Phi-4-mini is significant.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Maybe not, if&lt;/strong&gt; you need maximum speed (Phi-4-mini is ~2x faster) or have less than 16GB RAM.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Our verdict:&lt;/strong&gt; For a companion robot that needs to understand emotions and respond in Russian — Gemma 4 E4B is the best small open model available today.&lt;/p&gt;

&lt;h2&gt;
  
  
  Try It
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Demo:&lt;/strong&gt; &lt;a href="https://aisthos.dev/demo/" rel="noopener noreferrer"&gt;aisthos.dev/demo&lt;/a&gt; — interactive emotion display&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Code:&lt;/strong&gt; &lt;a href="https://github.com/aisthos/aisthos-core" rel="noopener noreferrer"&gt;github.com/aisthos/aisthos-core&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Modelfile:&lt;/strong&gt; &lt;a href="https://github.com/aisthos/aisthos-core/blob/main/models/Modelfile.gemma4" rel="noopener noreferrer"&gt;AisthOS Gemma 4 Modelfile&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;




&lt;p&gt;&lt;em&gt;This article was written by a human-AI team: Vladimir Desyatov (architecture, testing, product decisions) and Claude (implementation, benchmarking, documentation). We believe honest collaboration between humans and AI produces the best results.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;AisthOS — the Perception OS that grows with you.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>gemma</category>
      <category>ollama</category>
      <category>robotics</category>
    </item>
    <item>
      <title>AisthOS: The OS That Grows With You</title>
      <dc:creator>Vladimir Desyatov</dc:creator>
      <pubDate>Mon, 06 Apr 2026 06:51:15 +0000</pubDate>
      <link>https://dev.to/desve/aisthos-the-os-that-grows-with-you-33j2</link>
      <guid>https://dev.to/desve/aisthos-the-os-that-grows-with-you-33j2</guid>
      <description>&lt;p&gt;I spend most of my day at a computer. Over the past year, I've noticed something: the moments when I'm most productive aren't when I have the best tools. They're when I'm working with an AI that actually understands how I think.&lt;/p&gt;

&lt;p&gt;Not because it agrees with everything. Because it remembers what worked before, pushes back when I'm heading the wrong way, and adapts to my rhythm without being asked.&lt;/p&gt;

&lt;p&gt;The problem? Every new session starts from scratch. Every preference forgotten. And no way to move what we've built together to a different device.&lt;/p&gt;

&lt;p&gt;That's why I'm building &lt;a href="https://aisthos.dev" rel="noopener noreferrer"&gt;AisthOS&lt;/a&gt; — a Perception Operating System that grows with you.&lt;/p&gt;

&lt;h2&gt;
  
  
  What "grows with you" means
&lt;/h2&gt;

&lt;p&gt;AisthOS converts sensor data into anonymized metadata (Sparks), then &lt;strong&gt;learns from those Sparks&lt;/strong&gt; through three parallel tracks:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Track&lt;/th&gt;
&lt;th&gt;Speed&lt;/th&gt;
&lt;th&gt;What happens&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Fast&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Real-time&lt;/td&gt;
&lt;td&gt;Learns from your reactions instantly&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Medium&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Nightly&lt;/td&gt;
&lt;td&gt;Finds patterns, creates new skills automatically&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Slow&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Weekly&lt;/td&gt;
&lt;td&gt;Fine-tunes its personality to match yours&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The device goes through growth stages — Infant (basic reactions) → Child (pattern discovery) → Teen (self-created skills) → Adult (anticipates needs).&lt;/p&gt;

&lt;h2&gt;
  
  
  Privacy is the architecture, not a setting
&lt;/h2&gt;

&lt;p&gt;All learning happens locally. Raw sensor data only exists in volatile memory during processing. What gets stored are Sparks — structured descriptions like "hand raised to 45°, expression: surprise" — never the actual photo.&lt;/p&gt;

&lt;p&gt;We've seen what happens when companion AI depends on the cloud. Moxie shut down in January 2025. Every robot became a paperweight overnight. Your relationship with your companion shouldn't have a kill switch in someone else's server room.&lt;/p&gt;

&lt;h2&gt;
  
  
  Create once, use everywhere
&lt;/h2&gt;

&lt;p&gt;Everything the system learns about you is stored in a &lt;strong&gt;User Wisdom&lt;/strong&gt; file (~200 KB). Export it, move it to another device, import it. The new device knows you from the first second.&lt;/p&gt;

&lt;p&gt;No existing standard does this. Soul Spec describes who the AI &lt;em&gt;is&lt;/em&gt;. Agent File describes what the AI &lt;em&gt;can do&lt;/em&gt;. User Wisdom describes who &lt;em&gt;you&lt;/em&gt; are — as understood by the AI. Nobody else is building this layer.&lt;/p&gt;

&lt;h2&gt;
  
  
  We tested it on ourselves
&lt;/h2&gt;

&lt;p&gt;I've been using these principles in my daily work for weeks. The system tracks my communication style, decision patterns, and creative rhythms. It adapted its response format without being asked — more tables, shorter morning answers, detailed evening sessions.&lt;/p&gt;

&lt;p&gt;We didn't design "grows with you" as a feature. We lived it. Then formalized what worked.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aisthos.dev/blog/grows-with-you/" rel="noopener noreferrer"&gt;Read the full article on aisthos.dev →&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;GitHub:&lt;/strong&gt; &lt;a href="https://github.com/aisthos/aisthos" rel="noopener noreferrer"&gt;github.com/aisthos/aisthos&lt;/a&gt;&lt;br&gt;
&lt;strong&gt;License:&lt;/strong&gt; MIT&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Built by Vladimir Desyatov with AI-assisted development. The collaborative process itself demonstrates the AisthOS philosophy: AI as a transparent partner that grows alongside you.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>privacy</category>
      <category>selflearning</category>
    </item>
    <item>
      <title>AisthOS: What if your OS compiled UP instead of down?</title>
      <dc:creator>Vladimir Desyatov</dc:creator>
      <pubDate>Fri, 03 Apr 2026 14:41:16 +0000</pubDate>
      <link>https://dev.to/desve/aisthos-what-if-your-os-compiled-up-instead-of-down-1glp</link>
      <guid>https://dev.to/desve/aisthos-what-if-your-os-compiled-up-instead-of-down-1glp</guid>
      <description>&lt;p&gt;Every operating system you've ever used does the same thing: it takes your intent and compiles it &lt;strong&gt;down&lt;/strong&gt; into hardware signals.&lt;/p&gt;

&lt;p&gt;What happens if you reverse that?&lt;/p&gt;

&lt;h2&gt;
  
  
  The idea
&lt;/h2&gt;

&lt;p&gt;Take raw sensor data — video, audio, accelerometer readings — and compile it &lt;strong&gt;upward&lt;/strong&gt; into structured knowledge about the world. Not raw pixels. Not audio waveforms. Structured, anonymized semantic metadata.&lt;/p&gt;

&lt;p&gt;We call these units &lt;strong&gt;Sparks&lt;/strong&gt;. A Spark might contain "hand raised to 45 degrees, facial expression: surprise" — but never the actual photo. Raw data exists only in volatile memory during processing and is deleted immediately.&lt;/p&gt;

&lt;p&gt;This is &lt;a href="https://github.com/aisthos/aisthos" rel="noopener noreferrer"&gt;AisthOS&lt;/a&gt; (from Greek &lt;em&gt;aisthesis&lt;/em&gt; — perception). A Perception Operating System.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why build this?
&lt;/h2&gt;

&lt;p&gt;Because the AI industry is hitting four walls simultaneously:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wall 1: Training data is running out.&lt;/strong&gt; The web corpus that fed GPT-3/4 and LLaMA is exhausted. Epoch AI estimates high-quality public text will be fully consumed between 2026 and 2032.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wall 2: Synthetic data causes model collapse.&lt;/strong&gt; Shumailov et al. proved in Nature (2024) that training on AI-generated data causes irreversible degradation. Even mixing real and synthetic data doesn't fix it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wall 3: Annotation is manual and expensive.&lt;/strong&gt; Tesla pays operators $24–48/hr to collect training data for Optimus — people in helmets with five cameras. The tools for continuous streaming annotation from live sensors don't exist.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Wall 4: GPUs and electricity are in shortage.&lt;/strong&gt; H100 costs $25–40K with a 4–8 month waitlist. Data centers consumed 415 TWh in 2024; the IEA projects 945 TWh by 2030. Several U.S. states have imposed moratoriums on new data center construction.&lt;/p&gt;

&lt;h2&gt;
  
  
  Three formalisms
&lt;/h2&gt;

&lt;p&gt;AisthOS rests on three concepts:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Template&lt;/strong&gt; — &lt;em&gt;what&lt;/em&gt; to extract. A multimodal schema: &lt;code&gt;T = (M, E, F, R)&lt;/code&gt; where M = modalities, E = entities, F = format, R = cross-modal relationships. Unlike Avro or Protobuf, Template fields are "which knowledge to extract," not "which bytes to save."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Filter&lt;/strong&gt; — &lt;em&gt;when&lt;/em&gt; to extract. Semantic triggers, not numerical thresholds. Not "temperature &amp;gt; 30°C" but "the mother said 'time to feed.'"&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Spark&lt;/strong&gt; — the result. A unit of anonymized knowledge (~200 bytes). Contains semantics, not data. Privacy-by-design as an architectural decision, not a policy checkbox.&lt;/p&gt;

&lt;p&gt;Together they form the &lt;strong&gt;Perception Compiler&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Does it actually work on real hardware?
&lt;/h2&gt;

&lt;p&gt;Yes. Today.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Device&lt;/th&gt;
&lt;th&gt;Chip&lt;/th&gt;
&lt;th&gt;FPS&lt;/th&gt;
&lt;th&gt;Power&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Smart glasses&lt;/td&gt;
&lt;td&gt;GAP9 RISC-V&lt;/td&gt;
&lt;td&gt;18 fps&lt;/td&gt;
&lt;td&gt;62.9 mW (9.3h battery)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Dashcam&lt;/td&gt;
&lt;td&gt;Ambarella CV72S&lt;/td&gt;
&lt;td&gt;4×5MP + AI&lt;/td&gt;
&lt;td&gt;&amp;lt;3 W&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;RPi5 + Hailo-8L&lt;/td&gt;
&lt;td&gt;13 TOPS&lt;/td&gt;
&lt;td&gt;~120 fps (batch=8)&lt;/td&gt;
&lt;td&gt;4–5 W&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Full pipeline on RPi5:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;capture(5ms) → detect(8ms) → classify(3ms) → filter(1ms) → spark(2ms) = 19ms → 52 fps
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;The compression ratio:&lt;/strong&gt; 1 second of 4K video (H.265) ≈ 2–3 MB. One Spark ≈ 200 bytes. That's &lt;strong&gt;over 10,000× reduction&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;A terabyte drive would hold Sparks from 16 years of continuous operation.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why not just use the cloud?
&lt;/h2&gt;

&lt;p&gt;Because the math doesn't work anymore:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;&lt;/th&gt;
&lt;th&gt;Centralized GPU&lt;/th&gt;
&lt;th&gt;AisthOS (Edge)&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Node cost&lt;/td&gt;
&lt;td&gt;H100: $25–40K&lt;/td&gt;
&lt;td&gt;Device: $70–200 (already purchased)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Shortage&lt;/td&gt;
&lt;td&gt;HBM +20%, 4–8 month wait&lt;/td&gt;
&lt;td&gt;Billions of devices already exist&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Energy&lt;/td&gt;
&lt;td&gt;Data centers: 415 → 945 TWh by 2030&lt;/td&gt;
&lt;td&gt;60 mW – 30 W per device&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Privacy&lt;/td&gt;
&lt;td&gt;Data goes to cloud&lt;/td&gt;
&lt;td&gt;Data never leaves device&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Scaling&lt;/td&gt;
&lt;td&gt;Linear cost increase&lt;/td&gt;
&lt;td&gt;+1 user = +1 free processor&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;A million AisthOS devices = a million processors working for free. Each already paid for, deployed, and powered. Research shows 80% edge / 20% cloud delivers &amp;gt;75% cost savings.&lt;/p&gt;

&lt;p&gt;And the energy crisis is real: moratoriums on new data centers in Virginia, Georgia, Vermont. Dublin banned new grid connections. Companies are planning nuclear reactors for AI. AisthOS uses compute that society already manufactured.&lt;/p&gt;

&lt;h2&gt;
  
  
  AisthOS Inside™: proving privacy, not promising it
&lt;/h2&gt;

&lt;p&gt;Any manufacturer can claim "we respect your privacy." AisthOS Inside™ is an open certification standard — like Wi-Fi Certified — that makes privacy &lt;strong&gt;verifiable&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Seven principles: no raw data storage, Sparks-only output, no PII, user sovereignty, visible indicator, no hidden modes, open audit.&lt;/p&gt;

&lt;p&gt;The code is MIT (free). The certification mark requires passing tests. Four levels from free self-certification to enterprise.&lt;/p&gt;

&lt;p&gt;We identified 6 security threat types (4 specific to Perception OS):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Template Injection&lt;/strong&gt; — fixed ontology schemas, max 8 fields, no free text&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Filter Surveillance&lt;/strong&gt; — max 3 attributes, person-specific banned, entropy check&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Physical Prompt Injection&lt;/strong&gt; — text quarantine, dual PII detection, 95% fail-safe&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Adversarial PII Bypass&lt;/strong&gt; — cascade detection across multiple architectures&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Full security analysis: &lt;a href="https://github.com/aisthos/aisthos/tree/main/certification/security-annex" rel="noopener noreferrer"&gt;Security Annex&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Where this is going
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Near term:&lt;/strong&gt; companion AI robots, dashcam training data, retail behavior analytics, smart glasses (solving the Google Glass privacy problem).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Long term:&lt;/strong&gt; automated scientific discovery. Systems like AI-Newton (2025) can derive physical laws from structured data. AisthOS provides the missing perception layer — automatic conversion of real experiments into structured input.&lt;/p&gt;

&lt;p&gt;Imagine a thousand devices observing physical phenomena and generating Sparks from which AI extracts patterns. That's the direction.&lt;/p&gt;

&lt;h2&gt;
  
  
  Try it / contribute
&lt;/h2&gt;

&lt;p&gt;AisthOS is in early development. We're looking for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Privacy/security researchers&lt;/strong&gt; to review our &lt;a href="https://github.com/aisthos/aisthos/tree/main/certification/security-annex" rel="noopener noreferrer"&gt;threat model&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Edge AI engineers&lt;/strong&gt; to test on new hardware&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Community members&lt;/strong&gt; to discuss the &lt;a href="https://github.com/aisthos/aisthos/tree/main/certification" rel="noopener noreferrer"&gt;certification standard&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Anyone&lt;/strong&gt; to comment, critique, and challenge our assumptions&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;GitHub:&lt;/strong&gt; &lt;a href="https://github.com/aisthos/aisthos" rel="noopener noreferrer"&gt;github.com/aisthos/aisthos&lt;/a&gt;&lt;br&gt;
&lt;strong&gt;Website:&lt;/strong&gt; &lt;a href="https://aisthos.dev" rel="noopener noreferrer"&gt;aisthos.dev&lt;/a&gt;&lt;br&gt;
&lt;strong&gt;License:&lt;/strong&gt; MIT&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Built by Vladimir Desyatov with AI-assisted development. The collaborative process itself demonstrates the AisthOS philosophy: AI as a transparent tool that amplifies human capability.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;If you're an arXiv author in cs.AI and willing to endorse a new submission, I'd be grateful — reach out via GitHub Issues.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>privacy</category>
      <category>edgeai</category>
    </item>
  </channel>
</rss>
