DEV Community

Moth
Moth

Posted on • Edited on

A 25-Person Startup Built a Chip That Only Runs One AI Model. It's 73 Times Faster Than Nvidia.

Nvidia sells versatility. Its GPUs run any model, any framework, any workload. That flexibility made Nvidia a $3 trillion company. Taalas, a Toronto startup with 25 employees and $200 million in funding, is betting the opposite direction: a chip that can only run a single model, hardwired into the silicon.

The HC1 chip doesn't load model weights from memory. It etches them directly into the transistors. Every weight becomes a physical circuit. The multiply-accumulate operations that dominate inference happen at the transistor level, not in software shuttling data between memory and compute. The result, according to Taalas: 17,000 output tokens per second on Llama 3.1 8B. Nvidia's H200 manages roughly 233 tokens per second on the same model. That's a 73x speed advantage at one-tenth the power.

If true, this is the most significant inference performance claim since Groq's LPU announcement. Nvidia apparently agrees that inference optimization matters — the company paid $20 billion earlier this year to license IP from Groq.

How You Hardwire a Brain

The HC1 uses TSMC's 6nm process. Each chip packs 53 billion transistors into 815 square millimeters — nearly the maximum size a single chip can be. Most of those transistors aren't for compute in the traditional sense. They're paired mask ROM and SRAM storing the model's weights as physical circuits, not as data sitting in memory waiting to be fetched.

Ljubisa Bajic, Taalas's CEO, previously founded Tenstorrent. His co-founders Lejla Bajic and Drago Ignjatovic came from Tenstorrent and AMD. Their pitch: "We can put a weight and do the multiply associated with it all in one transistor."

That sounds impossible until you realize what it eliminates. The dominant bottleneck in inference isn't computation — it's memory bandwidth. GPUs spend most of their time waiting for data to arrive from HBM or DRAM. Taalas eliminates the trip entirely. The data is the circuit.

Ten HC1 cards fit in a standard dual-socket x86 server. Total power draw: 2,500 watts. For comparison, a single Nvidia B200 GPU draws 1,000 watts. The full server runs ten specialized inference engines at 2.5x the power of a single general-purpose GPU.

The Obvious Problem

A chip that can only run one model is useless the moment that model becomes obsolete. LLM releases move on monthly cadences. If every Llama update requires new silicon, the economics collapse.

Taalas claims a turnaround of two months from model weights to deployable PCIe cards, using a custom workflow with TSMC. Only two metal layers change between model versions, not the full chip design. That's plausible in theory — metal-layer modifications are the cheapest part of a chip redesign — but two months is still two months. By the time your Llama 3.1 chip arrives, Llama 3.2 might already be live.

The counterargument: inference at scale doesn't need the latest model. It needs the cheapest reliable model. A bank running fraud detection on a specific fine-tuned 8B model doesn't upgrade every quarter. A telecom running customer service on a validated deployment doesn't swap models for fun. The customers who would buy Taalas chips are the ones who've already decided which model they're running and need to run it billions of times as cheaply as possible.

The Numbers That Matter

Taalas has spent $30 million of its $200 million war chest. Twenty-five people. The current HC1 handles 8 billion parameters. A 20-billion-parameter version ships by summer 2026. Frontier-class models — the 70B and above range — arrive by year-end via pipeline parallelism across multiple cards.

The funding comes from Quiet Capital, Fidelity, and Pierre Lamond, a semiconductor investor who backed SanDisk, National Semiconductor, and Marvell. Lamond doesn't write checks for vaporware.

But the skeptics have a point. No major AI company has publicly endorsed the approach. The 73x claim has not been independently benchmarked. And the entire value proposition depends on a bet about the future of AI deployment: that the industry will standardize on a small number of foundation models deployed at massive scale, rather than continuously iterating toward new architectures.

If that bet is right, Taalas built the perfect chip for the inference economy. If it's wrong, they built the world's most expensive paperweight — one model at a time.

The semiconductor industry has a name for this kind of gamble. They call it an ASIC. Application-specific integrated circuits dominated computing before GPUs made flexibility king. Taalas is arguing the pendulum swings back when inference costs become the bottleneck, not training. Given that inference now accounts for over 80 percent of AI compute spending, they might not be wrong.

Nobody else is willing to burn a chip design on a single model. That's either visionary or suicidal. The market will decide which, probably within the year.


Originally published on Substack


If you work with AI tools daily, I built a set of prompt packs that actually work — tested across Claude, GPT-4, and Gemini. System prompts, code review chains, data extraction templates, and more.

👉 Browse the prompt packs on Polar.sh — individual packs from $5, or get all 5 for $19.

Top comments (1)

Some comments may only be visible to logged-in visitors. Sign in to view all comments.