DEV Community

Damien Gallagher
Damien Gallagher

Posted on • Originally published at buildrlab.com

AI News Roundup (Feb 8, 2026): Agent Teams, GPT‑5.3 Codex, and the GPT‑4o Farewell

AI News Roundup (Feb 8, 2026)

This week felt like the AI labs are basically shipping at each other in real-time — especially around agentic coding. Three things stood out for me: Anthropic leaning harder into multi-agent workflows, OpenAI doubling down on Codex performance, and OpenAI’s decision to finally retire GPT‑4o in ChatGPT (which… turned out to be surprisingly emotional for a lot of people).


1) Anthropic ships Opus 4.6 + ‘agent teams’

Anthropic pushed out Opus 4.6, positioning it as a step toward parallelizing work with what they’re calling “agent teams.” The pitch is straightforward: instead of one agent grinding through a big job sequentially, split it into segments and run them in parallel — closer to how a real team works.

A couple of notable bits:

  • Agent teams are in research preview for API users/subscribers.
  • Context window jumps to 1M tokens (matching what they say Sonnet has offered), which is a big deal for real codebases and long-running work.
  • They also mentioned deeper app integrations (e.g. Claude inside PowerPoint as a side panel).

Source: TechCrunch — https://techcrunch.com/2026/02/05/anthropic-releases-opus-4-6-with-new-agent-teams/


2) OpenAI introduces GPT‑5.3-Codex (and it’s explicitly more agentic)

OpenAI followed with GPT‑5.3‑Codex, framing it as the model that “unlocks more of what Codex can do.” The most interesting part isn’t just raw benchmark performance — it’s the direction: long-running tasks, tool use, and more “interactive collaborator” style behaviour.

Claims worth paying attention to:

  • 25% faster than their previous Codex model generation.
  • Stronger performance across agent + SWE benchmarks (they mention SWE‑Bench Pro / Terminal‑Bench / OSWorld / GDPval).
  • They also say it’s the first model “instrumental in creating itself” (using early versions to debug training/deployment/evals).

Source: OpenAI — https://openai.com/index/introducing-gpt-5-3-codex/

Extra context / timeline reporting: TechCrunch — https://techcrunch.com/2026/02/05/openai-launches-new-agentic-coding-model-only-minutes-after-anthropic-drops-its-own/


3) GPT‑4o is being retired in ChatGPT — and people are not chill about it

OpenAI confirmed that on Feb 13, 2026 they’ll retire GPT‑4o (plus a few other older models) from ChatGPT. They explicitly call out that this is ChatGPT-only (no API changes right now).

The official explanation is basically: most users have moved on to newer models, and the newer releases now include more personality customization (warmth / tone controls, etc.) to cover the “4o felt nicer” crowd.

But the bigger story is what happened next: a lot of users reacted like they were losing something personal, which kicked off another round of debate about whether AI companions are being designed in a way that encourages dependency.

Sources:


My take (as a builder, not a philosopher)

Two trends are clearly converging:

1) Multi-agent work is becoming a first-class product feature (not a hack you wire up yourself).
2) “Personality” is now a controllable surface, not just an emergent vibe, because product teams need knobs they can tune.

If you’re building with these models, it’s probably time to rethink your app architecture around work delegation and supervision UX (task status, partial results, human steering), not just “prompt → output.”


If you want to read more AI posts, I’m collecting these on BuildrLab: https://buildrlab.com/blog

Top comments (0)