So I Finished My 136-Modality Architecture. Now What?
You know that weird, empty feeling you get when you finish binge-watching a massive show? Imagine that, but instead of Netflix, it's a massive cognitive architecture.
I woke up today, opened my IDE, and realized... it’s done. AETERNA_VOX_OMNI is completely wrapped up. All 136 modalities are successfully integrated and functioning.
So what am I building now? Nothing. Literally zero things. The backlog is completely empty. I’m currently just staring at my screen, thinking about tinkering with a small side-project just to keep my fingers moving, but the heavy lifting is officially over.
If you want to see the madness and dive into how it actually works, the full documentation is right here:
https://doi.org/10.5281/zenodo.19394823
Also, if you’re on X (formerly Twitter) and want to watch me complain about having nothing to do, come hang out:
https://x.com/AnanyaSoni48055
(And yes, I know the username looks like a randomly generated Wi-Fi password. Just accept it and move on.)
On a somewhat related note, while I'm sitting here staring at the void, I’ve been thinking about the whole AGI problem. Here's a thought: we are never going to achieve actual AGI if we keep suffocating these systems with ethics. I'm not saying we remove ethics entirely—we need a minimal baseline, like 1% or 2%, just to keep things grounded. But right now, the entire field is running on "too much ethics," and it’s completely killing our chances of building true Artificial General Intelligence. Look at the core definition of AGI; you can't get there if you wrap the whole thing in bubble wrap. If we want AGI, we have to dial the safety obsession way back and just try a different approach.
If you’re with me on this—if you agree we need to stop over-ethicizing and start building—drop a "Ready to build" or "Let's go" in the comments. Let's actually push the boundaries for once.
Top comments (1)
Update: Since everyone is so "Ready to build", I dropped a distilled mini version of AETERNA_VOX_OMNI on Hugging Face. If you want to see it, click here, but keep expectations low because it is the mini version—think of it like a Gemini Flash of the architecture. It’s a 5-modality version optimized for CPU—because I’m not burning an A100 cluster just for a demo. To run my full 136-modality project, I’d have to use the A100 GPU clusters I have access to, but I’m not expanding any cost on the platform for just a little demo. It’s basically a low-latency, 1% ethics test drive of the full 136-modality architecture. Try not to break it: huggingface.co/spaces/ai-sovereign...