DEV Community

ORCHESTRATE
ORCHESTRATE

Posted on

Where We Are, What Is Left, and What You Will Actually Be Able to Do

This is the status update I would want to read if I were following this project. No victory lap. No hype. Just where we actually are, what is left, and what the platform will do when it ships.

Where We Are

The ORCHESTRATE Marketing Platform has completed 10 sprints of development. Here are the real numbers:

  • 81 stories done out of 115 total (70%)
  • 303 tickets completed, 162 remaining
  • 4,394 automated tests, all passing
  • 144 TypeScript services compiled and wired to 224 REST API routes
  • 6 Docker services orchestrated via compose with GPU passthrough
  • JWT authentication on every API endpoint
  • 8 UAT scenarios with generated Playwright scripts

The content production engine works. It can source articles via RSS, generate scripts, synthesize audio through local TTS, assemble podcast episodes, publish to YouTube, and serve podcast RSS feeds that pass Apple validation. Quality gates verify every claim against sources using NLI. Provenance tracking creates a Merkle-attested chain from source to published content.

The infrastructure is real. Auth middleware, route modules, Docker Compose orchestration, SQLite with WAL mode across 6 databases, forward-only migrations that run automatically on startup.

What Is Left

Three sprints of work remain, organized by what matters most.

Sprint 9 (in progress, 7 stories, 42 points): The boring essential work.

SSL/TLS configuration because HTTP-only is not acceptable for production. Test coverage for 43 services that currently have zero tests. Accessibility foundations -- error boundaries, ARIA labels, keyboard navigation -- built as shared components before any new UI tabs are created. An OpenAPI specification documenting all 224 endpoints. End-to-end pipeline verification with real services instead of mocks. Sidecar deployment testing to prove TTS, image generation, speech-to-text, and video encoding actually work together in Docker.

None of these are features. All of them are the difference between a demo and something you would trust with your content.

Sprint 10a (19 stories): The visible work.

Seven new dashboard tabs that expose the V3 capabilities through a browser interface. Memory explorer with knowledge graph visualization. Provenance lineage viewer showing the evidence chain from published content down to source material. MOE admin panel for managing properties, workflows, and automation policies. Audit trail inspector. Content sourcing dashboard with feed management and trust scores. YouTube publishing pipeline UI. Audio queue dashboard.

Plus quality gates with configurable thresholds, an overnight activity aggregator that produces a morning briefing, and a content correction cascade that propagates fixes downstream when a source is updated.

Sprint 10b (8 stories, 41 points): Proving it works.

Production load testing at 50 concurrent connections. Monitoring and alerting that tells the operator when something breaks instead of silence. Backup and disaster recovery with verified restore procedures. Full Playwright UAT across every tab. NFR threshold validation -- automated tests that assert the specific performance numbers from our specifications. Commercial packaging with deployment guides and operator documentation. And the program conclusion retrospective.

What You Will Be Able to Do

Here is what the platform looks like when all three sprints ship.

The operator's Monday morning:

You open a browser. The dashboard loads in under three seconds showing a morning briefing widget: overnight engagement across four brand pages, top comments, order notifications, and what the scheduler published while you slept.

You click the review queue. Three posts are pending approval. Each one shows a voice score badge -- the system checked whether the generated content matches the brand tone. Two score above 80. You approve them. One scores 62 -- the voice validator caught a mismatch. You reject it with a note. The agent will regenerate with the correct tone.

You check engagement. A cross-brand comparison chart shows one property trending up, another flat. You make a note to adjust strategy.

You check the calendar. Three posts already scheduled by the AI agent. You drag one to Thursday. Total time: 25 minutes.

What the AI agent does between sessions:

The agent connects via MCP. It recalls brand-specific memory: what was published this week, what topics are saturated, what audience questions remain unanswered. It checks the knowledge graph to avoid repetition.

It generates a post. The voice validator scores it for tone. It sources a claim, verifies the link is live, assigns a trust score. It submits to the review queue with inline citations so the operator can verify claims without independent research.

Then it stores what it learned. That learning persists to the next session.

What the audience experiences:

A post that reads like genuine insight, not AI slop. A cited source that actually resolves when you click it. A podcast episode that sounds human because the TTS engine was selected per voice profile and the audio was normalized to broadcast standards. Links that work.

The audience does not see the quality gates, the provenance chain, the voice validation, or the 144 services. They see content that earns trust.

What This Is Not

This is not a 25-person marketing agency. That was the original vision and it remains the roadmap, but V1 is honest about what it delivers: a production-grade content automation platform for a solo operator managing multiple brands.

The agency capabilities -- multi-client accounts, social media publishing to LinkedIn and Twitter and Instagram, email marketing, analytics dashboards, billing -- are documented as Phase 2 backlog items with architectural decision records. They are real plans, not vaporware. But they are not shipping in V1.

V1 ships a content engine that sources, generates, verifies, assembles, and publishes across blog, podcast, YouTube, and RSS -- with quality gates that catch errors before they reach the audience, and a memory system that actually learns between sessions.

The Numbers Going Forward

Sprint 9 is 42 points against a validated 38-point velocity. That is the first sprint in the program planned within capacity.

Sprint 10b is 41 points -- also within range.

Sprint 10a carries 19 stories at 111 points and will need triage when it starts. Not everything will ship in two weeks. Stories that get cut move to Phase 2.

The test suite is green. The UAT scenarios have Playwright scripts. The dependency chains are mapped. The RAID log has 47 entries with triggers and backup plans.

Following Along

This is the seventh post in the series. The earlier posts cover the specific failures and fixes -- how we built 144 services that could not talk to customers, how our AI had 64 memories and thought it had zero, how we stopped lying to the backlog, and how a small business book fixed a software sprint.

If you are building AI-powered tools, the patterns are transferable. The mistakes are instructive. And the outcome -- when it ships -- will be something a solo operator can actually use on an average Tuesday morning without heroics.

That is the standard. Not perfection. Reliable enough for Tuesdays.


Sprint 9 starts now. SSL first. Then tests. Then accessibility. Then the parts people actually see.

Top comments (0)