<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Stacey Schneider</title>
    <description>The latest articles on DEV Community by Stacey Schneider (@staceyeschneider).</description>
    <link>https://dev.to/staceyeschneider</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3880627%2F94d335fe-8733-4879-98e8-84f571371559.jpeg</url>
      <title>DEV Community: Stacey Schneider</title>
      <link>https://dev.to/staceyeschneider</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/staceyeschneider"/>
    <language>en</language>
    <item>
      <title>Orchestration Is Officially A Commodity. Context and Governance Are the Moat.</title>
      <dc:creator>Stacey Schneider</dc:creator>
      <pubDate>Thu, 30 Apr 2026 14:51:26 +0000</pubDate>
      <link>https://dev.to/staceyeschneider/orchestration-is-officially-a-commodity-context-and-governance-are-the-moat-1ndg</link>
      <guid>https://dev.to/staceyeschneider/orchestration-is-officially-a-commodity-context-and-governance-are-the-moat-1ndg</guid>
      <description>&lt;p&gt;The industry is sending flares galore that orchestration is a commodity. &lt;a href="https://openai.com/index/open-source-codex-orchestration-symphony/" rel="noopener noreferrer"&gt;Symphony&lt;/a&gt; dropped last week. An orchestration spec for Codex, handed to the community, with a note that OpenAI has no plans to keep it as a standalone product. You open-source the pipes. You don't open-source your moat.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.blog/changelog/2026-02-13-github-agentic-workflows-are-now-in-technical-preview/" rel="noopener noreferrer"&gt;GitHub's ACE&lt;/a&gt; is in technical preview. &lt;a href="https://www.warp.dev/blog/oz-orchestration-platform-cloud-agents" rel="noopener noreferrer"&gt;Warp's Oz&lt;/a&gt; ships with Claude, Codex, and Gemini on day one. Every major player in Agentic Workflow Orchestration (AWO) is landing on the same primitives at the same time. That's not competition. That's commoditization.&lt;/p&gt;

&lt;p&gt;I've been building agentic systems for years at &lt;a href="https://promptowl.ai" rel="noopener noreferrer"&gt;PromptOwl&lt;/a&gt; and watched this happen from the inside. A year ago we had a full drag-and-drop orchestration editor—think n8n. In practice it wasn't useful for agentic systems, especially for regular business users, so we buried it: first behind a logic block editor, then behind setup wizards. Orchestration became plumbing because that's what it is.&lt;/p&gt;

&lt;p&gt;From my perch though, it is crystal clear that context and governance are the real battlegrounds. That's where the actual competition starts.&lt;/p&gt;

&lt;h2&gt;
  
  
  Context: The Thing That Makes Agents Useful
&lt;/h2&gt;

&lt;p&gt;Most agents underwhelm. They lie, they forget, and they fall over. All fireable offenses for humans, but agents are still in kindergarten. We keep them around hoping to raise them up to be valuable workers.&lt;/p&gt;

&lt;p&gt;Their problem isn't that the models are bad—frontier models are remarkably capable and even the generic models are extremely useful today. Agents keep doing all the wrong things because they don't know your business.&lt;/p&gt;

&lt;p&gt;Orchestration platforms treat context like a config file. You paste a system prompt, maybe attach a few documents, and call it done. That's not context. That's a briefing. And briefings go stale.&lt;/p&gt;

&lt;p&gt;Real context is institutional and organic in nature. It's how your team makes decisions when two good options conflict. It's what "done" means in your org versus what it means in the ticket description. It's the three approaches that were tried and abandoned before you joined. None of that is in a README. It lives in people's heads and it leaks out slowly over Slack threads and PR comments and post-mortems you forgot about.&lt;/p&gt;

&lt;p&gt;Agents operating without that context will be productive in exactly the way an eager new hire with no onboarding is productive—they'll move fast, they'll complete tasks, and they'll occasionally make decisions that anyone who'd been around for six months would have caught immediately.&lt;/p&gt;

&lt;p&gt;Winning here isn't about better routing. It's about maintaining a living, curated layer of institutional context that agents can draw on—and that teams actually own, update, and move.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Warning:&lt;/strong&gt; Context portability is the canary in the lock-in coal mine. If your agent's intelligence lives in the platform's proprietary memory system, you're not building on a tool. You're feeding a database you don't own.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Governance: The Thing That Makes Agents Trustworthy
&lt;/h2&gt;

&lt;p&gt;The second problem is trust—and it's why most enterprises are still in the stands watching the AI wars play out.&lt;/p&gt;

&lt;p&gt;Governance at the agent layer isn't a dashboard feature. It's a standards problem. And the standards are starting to arrive.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/dickhardt/AAuth" rel="noopener noreferrer"&gt;AAuth&lt;/a&gt;—a new specification from Dick Hardt, the architect behind OAuth 2.0 and OpenID Connect—gives agents cryptographic identity. That's the foundation, but what it unlocks is more interesting than the spec itself: standard payloads that document what the agent did, under what authority, with what inputs. Not logs you search after something goes wrong. Structured, signed records of responsibility that travel with every action.&lt;/p&gt;

&lt;p&gt;That same chain-of-custody logic needs to extend across the entire stack. Which version of the prompt ran this interaction? Was that version reviewed and approved before it went anywhere near production? What did the agent actually say, and can you prove it? Full traceability of interactions—signed—isn't an audit feature. It's the baseline for any organization that plans to let agents act autonomously on anything consequential. Evals belong here too: you can't trust an agent you've never tested, and test results need to be part of the record, not a spreadsheet someone ran once before launch.&lt;/p&gt;

&lt;p&gt;The pattern is consistent regardless of stack or team size: organizations hit the context wall first, then the traceability wall, then realize they were the same wall the whole time. You can't govern what you can't trace, and you can't trace what was never signed.&lt;/p&gt;

&lt;p&gt;Context needs the same rigor. Institutional knowledge that agents draw on—your SOPs, your decision frameworks, your accumulated organizational memory—has authors and owners. Who wrote it, who approved it, when was it last verified? Right now most teams treat this as vibes. It needs to be treated as provenance. If an agent makes a bad call because it was working from a policy that was outdated or never formally ratified, that's not a model problem. It's a stewardship problem.&lt;/p&gt;

&lt;p&gt;AWO platforms building only for speed are going to hit this wall hard. Identity, traceability, and context stewardship aren't enterprise add-ons. They're the price of admission for anything consequential.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to Ask Before You Hand an Agent the Keys
&lt;/h2&gt;

&lt;p&gt;Forget asking whether your orchestration platform is multi-model. Warp already is. That question is over. Here's what to ask instead.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;One. Where does your context live?&lt;/strong&gt; Is your institutional knowledge—your prompts, your memory, your agent configuration—stored in a portable format you own? Or is it accumulating inside the platform's proprietary system, getting harder to move with every passing week?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Two. Can you audit what your agents did?&lt;/strong&gt; Not just "did the task complete"—but what decisions did the agent make, what information did it act on, and who authorized it to do so? If you can't answer that after the fact, you can't run agents on anything that matters.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Three. Do you have approval flows for high-stakes actions?&lt;/strong&gt; Before an agent merges a PR, sends an external communication, modifies production data, or spends money—is there a human checkpoint? Can you configure where that checkpoint sits without writing custom middleware?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Four. Can you scope agent access?&lt;/strong&gt; Does the platform let you define what each agent can see and touch? Or is the default "everything the authenticated user can access," which in practice means your agent has more access than most of your employees.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Five. Can you roll back?&lt;/strong&gt; If an agent makes a bad call—and eventually one will—what does recovery look like? Is it a config change, a button click, or a three-day archaeology project through logs nobody labeled?&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Evaluate Platforms Right Now
&lt;/h2&gt;

&lt;p&gt;Skip "how many agents can it run in parallel" and "which models does it support." Ask: does this platform make my agents smarter over time as they accumulate context about my business? And does it give me the controls I need to trust agents with decisions that have actual consequences?&lt;/p&gt;

&lt;p&gt;A platform that routes tasks quickly and has no answer for context decay or governance is fast plumbing. Impressive at the demo. Underwhelming in production.&lt;/p&gt;

&lt;p&gt;The orchestration race is already decided. Every major platform is converging on the same primitives and OpenAI just open-sourced the proof. What happens next is the interesting part—who owns the context layer, who can prove what their agents did and why, who built stewardship into the foundation instead of bolting it on after a compliance scare.&lt;/p&gt;

&lt;p&gt;The pipes are done. Now comes the hard part.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>devops</category>
      <category>discuss</category>
      <category>architecture</category>
    </item>
    <item>
      <title>Running AI on a Budget: 11 Tactics for Enterprise-Scale Efficiency</title>
      <dc:creator>Stacey Schneider</dc:creator>
      <pubDate>Wed, 15 Apr 2026 19:56:58 +0000</pubDate>
      <link>https://dev.to/staceyeschneider/running-ai-on-a-budget-12-tactics-for-enterprise-scale-efficiency-npd</link>
      <guid>https://dev.to/staceyeschneider/running-ai-on-a-budget-12-tactics-for-enterprise-scale-efficiency-npd</guid>
      <description>&lt;p&gt;At my company, &lt;a href="https://promptowl.ai" rel="noopener noreferrer"&gt;PromptOwl&lt;/a&gt;, everyone coworks with AI for 90 to 100% of their work. It's everywhere, in every process, and increasingly connecting everything. Engineering, marketing, sales, leadership—AI is in every workflow, every day. Getting there took us just over a year, and taught us a lot about what that actually costs.&lt;/p&gt;

&lt;p&gt;Running AI at that scale boils down to two optimization problems: money and time. &lt;/p&gt;

&lt;p&gt;Money shows up on the monthly invoice, and if you are not paying attention it will floor you in costs. AI is expensive to run blindly. We expect to pay $2-300 per developer, per day on the frontier models, but we had to make sure we weren't wasting money on things that didn't matter. &lt;/p&gt;

&lt;p&gt;Time is the second problem. It's the hours lost to waiting for generated responses, to rerunning prompts because bad context resulted in wrong outputs, and workflows that require constant manual intervention to function and not fall over. &lt;/p&gt;

&lt;p&gt;Optimizing for time and money is an evergreen effort. Too much is evolving, and we will always need to adapt. But these eleven tactics are the foundation of how we run AI at scale.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Setup
&lt;/h2&gt;

&lt;p&gt;Get this right once. It pays back on every session.&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Organize your Prime Documents
&lt;/h3&gt;

&lt;p&gt;On our journey to 90%, the first obvious problem was re-explaining core business tenets to the models for every conversation. So, I started writing context files for the models—in markdown, to cut down on the size of the messages sent to the LLMs. &lt;/p&gt;

&lt;p&gt;A Prime Document is a structured context file written specifically for AI use—not a deck you'd send to a colleague, but a document the model can actually use. Your brand brief, product spec, customer profile, team norms. Every function that runs AI regularly should have at least one.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgrcvwe5v3x6sbv66b75x.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgrcvwe5v3x6sbv66b75x.png" alt="Sample of PromptOwl Prime Documents" width="800" height="293"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Most teams don't have these. They improvise context in the chat window every time, and wonder why the outputs aren't consistent. &lt;/p&gt;

&lt;h3&gt;
  
  
  2. Write them for AI—shrink your files
&lt;/h3&gt;

&lt;p&gt;I like to think that every character I send to AI lights up a GPU in some data center. And while I love a light show, I visualize a huge energy waste in padding the data. It also can contort the results, as the context window for each LLM is fixed. Your whole conversation has to fit in it. So every unnecessary character, eats away at what the models can do for you. &lt;/p&gt;

&lt;p&gt;Uploading a full 60-page brand guide every time may provide detail, but it will cost you processing time, tokens—and potentially the right answer. The model re-reads everything it doesn't need, every single time. The more it has, the more it can confuse. &lt;/p&gt;

&lt;p&gt;Your Prime Documents should contain only what's relevant, written in a format the model can parse efficiently—not formatted for a human presenting to a board. Ask it to distill these documents for itself, then start a new chat with it. Smaller, cleaner input means identical or better output quality, faster responses, and a lower cost per call.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Use a context management tool
&lt;/h3&gt;

&lt;p&gt;Eventually, you will collect a nest of these machine optimized documents. And then maintaining them becomes the new burden. &lt;/p&gt;

&lt;p&gt;We started to centralize the management of our Prime Documents, publishing them on our Google Drive. I became the arbiter of them. But since everything in AI moves so fast, our context needed near constant updating. This new task of resyncing them ended up taking so much of my time, I took on the title Chief Context Officer. &lt;/p&gt;

&lt;p&gt;Plus the whole thing was such a drag. Literally. I was constantly dragging documents to every chat, plus reviewing and updating them if they needed. Only to have to do it all over again in a half hour when my conversation ran out of space. &lt;/p&gt;

&lt;p&gt;The solution was to build a context management layer that operates over a wiki-style system of markdown files. Accessing it via a CLI meant I could still use Claude and Antigravity the exact same way, but the system actively improved the context by creating tags and links to live data and other docs. Now my chats have the ability to look at everything and pick up where the other left off, with no retraining of the conversation.&lt;/p&gt;

&lt;p&gt;This pattern reduces tasks by 2-300% in my workflow, and mitigates the risk of having to start over if something goes sideways. I don't maintain it nearly as often either, as it usually just learns as I work. &lt;/p&gt;

&lt;p&gt;This is the same theory that &lt;a href="https://x.com/karpathy/status/2039805659525644595?lang=en" rel="noopener noreferrer"&gt;Andrej Karpathy wrote about&lt;/a&gt; a couple weeks ago. We published a &lt;a href="https://promptowl.ai/resources/contextnest-whitepaper/" rel="noopener noreferrer"&gt;ContextNest whitepaper&lt;/a&gt; on the details of how this works two months ago, too. And bonus! It's also an open-source project: &lt;a href="https://github.com/PromptOwl/ContextNest" rel="noopener noreferrer"&gt;ContextNest&lt;/a&gt;. There should be a desktop client any day now too.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Set up skills
&lt;/h3&gt;

&lt;p&gt;When customers ask me what they should apply AI to first, my answer is always the same—build what you repeatedly spend the most time on. Sometimes folks do a time journal to determine where this happens. I have a bias for rapid results, so I think this can be conducted as a mental exercise. At least to find the low hanging fruit.&lt;/p&gt;

&lt;p&gt;For me, I would spend 4-5 hours every week doing strategy work. I'd pull all the numbers, have AI distill all the standup notes, make sure action items were tracked and ticked off. Then I'd think about what to prioritize this week. Finally, I'd communicate it to the team so everyone is aligned and moving in unison. &lt;/p&gt;

&lt;p&gt;At first, I attempted to automate the SOP for doing this with another markdown file. Then I built tools to handle the repeatable research processes, like summarizing the chatter on the engineering channel in Slack. These Skills are reusable, pre-configured AI behaviors you define once and improve through actual usage.&lt;/p&gt;

&lt;p&gt;They can also be shared. In our case, that means that everyone who meets with a prospect, could create the same deep dive research brief and proposal based on our status today—what new features are out, what promotion is running, what the customer wanted to meet about in the first place. With very little coordination, we can create the same output. &lt;/p&gt;

&lt;p&gt;Without shared skills, every person on your team reinvents the wheel and gets slightly different results. Skills create consistency, reduce token cost on repeated tasks, and make AI output auditable across the org.&lt;/p&gt;




&lt;h2&gt;
  
  
  The 7 Habits of Highly Successful Prompting
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fybudly8iuvbbuenlu2y0.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fybudly8iuvbbuenlu2y0.jpeg" alt="Productivity Habits with AI" width="800" height="336"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now that we have the system, these are the habits that govern every session to get the most out of them. &lt;/p&gt;

&lt;h3&gt;
  
  
  1. Plan first. Build last.
&lt;/h3&gt;

&lt;p&gt;The expensive moment isn't generating the final artifact. It's generating it three times because the spec wasn't clear—and losing 20 minutes each time to recreate it.&lt;/p&gt;

&lt;p&gt;Before you ask for the web page, the strategy doc, or the campaign copy, use a few cheap messages to get alignment—figure out your structure, identify the edge cases, and establish naming conventions and expectations. &lt;/p&gt;

&lt;p&gt;Thinking is cheap. Building is expensive. Iteration in planning costs almost nothing in tokens and almost always saves you time. Iteration in generation costs both.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Run a murderboard on the plan
&lt;/h3&gt;

&lt;p&gt;One of the best things AI does is help you think through things like other people. Other professionals, your customers, or future prospects. This is your opportunity to learn from what they would predictably say. I have a markdown file of various personas that I pull from to review my plans, ruthlessly tear them apart, and find not just the holes, but their recommendations to satisfy them.&lt;/p&gt;

&lt;p&gt;I call this group-think exercise a "murderboard," but you can just tell it to run a focus group antagonistically or to act like a customer and complain. As long as you try to use multiple perspectives and explicitly get the model to break its tendency for sycophancy, it will help you find problems before you codify them into production.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Tell the AI to ask you questions
&lt;/h3&gt;

&lt;p&gt;A 600-word fully-specified prompt is often the most expensive way to get a mediocre result. With that much detail, the models think they should know everything and usually make some terrible assumptions.&lt;/p&gt;

&lt;p&gt;Describe what you need, focusing on the results and how it will be used. Tell the model to ask clarifying questions before it starts. Each exchange costs a fraction of a re-generation. You get better output from a conversation than from a wall of text that leaves the model guessing where your spec was ambiguous.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Edit the message. Don't stack on top of it.
&lt;/h3&gt;

&lt;p&gt;You sent a prompt, spotted a typo, realized you left out a constraint. Most people send a correction as a new message.&lt;/p&gt;

&lt;p&gt;That's a mistake for two reasons. &lt;/p&gt;

&lt;p&gt;Every new message adds to the context window—the model is now reading your original error and your correction simultaneously and trying to reconcile them. It has to think through this each time. &lt;/p&gt;

&lt;p&gt;Plus, new messages can take time to ingest—or worse, distract the model from your original question.&lt;/p&gt;

&lt;p&gt;Find the edit button. Replace the message. Then the next response doesn't carry your mistake forward, and you don't spend ten minutes untangling an output that went sideways because of a fixable prompt.&lt;/p&gt;

&lt;h3&gt;
  
  
  5. Turn off what you're not using
&lt;/h3&gt;

&lt;p&gt;Web search has a cost. Extended thinking has a cost. Document connectors have a cost. Most of the time none of them are needed for the task at hand. Frontier models have huge costs.&lt;/p&gt;

&lt;p&gt;If you can streamline what you don't need (especially if switching costs are low because you have a ContextNest), you can save a lot of time and tokens.&lt;/p&gt;

&lt;p&gt;Enable them in the moment you actually need them—not as default-on settings running in the background of every request.&lt;/p&gt;

&lt;h3&gt;
  
  
  6. Work in smaller sections
&lt;/h3&gt;

&lt;p&gt;Engineers called this out a long time ago. Models can not handle large codebases well. Using context and focusing on smaller sections at a time means models return results more quickly and have less risk of choking on the task. &lt;/p&gt;

&lt;p&gt;The same is true for business efforts. Don't ask for the 5,000-word strategy document in one prompt. Ask for the outline. Then expand each section. Don't ask for the full function—ask for the structure first, then fill in each piece.&lt;/p&gt;

&lt;p&gt;Smaller sections mean faster iteration, easier course correction, and lower cost when something needs to be redone. It also gives you natural checkpoints so the output doesn't drift through a long generation you can't easily fix at the end.&lt;/p&gt;

&lt;h3&gt;
  
  
  7. Match the model to the task
&lt;/h3&gt;

&lt;p&gt;Every call to a frontier model that didn't need to be one is money you didn't have to spend.&lt;/p&gt;

&lt;p&gt;As of April 15, the cost to generate 500 words on Opus 4.6 is about 1.67 cents. For that same 1.67 cents:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Sonnet 4.6 gives you ~835 words &lt;/li&gt;
&lt;li&gt;Haiku 4.5 gives you ~2,500 words &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This means Haiku is 5x cheaper than Opus for output. For content generation, listicles, and drafts—Haiku earns its place. Opus earns its place when nuance, analysis, or voice precision actually matters. Sonnet feels like the safe middle ground, but often the flash models are enough. &lt;/p&gt;

&lt;p&gt;Route simple triage, summarization, and single-turn questions to a lightweight model. Save the heavy models for work that actually requires it—the analysis feeding a real decision, the writing carrying your company's voice, the code review that can't afford a miss. The right tool for the job is a standard engineering principle. Apply it here.&lt;/p&gt;




&lt;h2&gt;
  
  
  The foundation, not the ceiling
&lt;/h2&gt;

&lt;p&gt;These eleven tactics cut both bills—the invoice and the productivity drain. But there is more to come, especially when you think about sharing a living context across a team or an organization.&lt;/p&gt;

&lt;p&gt;This is the area I am studying now. I'll be writing (and releasing commercial software) about how workflow and tools are adapting across organizations in future posts. Stay tuned!&lt;/p&gt;

</description>
      <category>ai</category>
      <category>productivity</category>
      <category>llm</category>
      <category>resources</category>
    </item>
  </channel>
</rss>
