<h2>Get Ready for the Apps SDK</h2>
<p><em>Hundreds of millions of people now open a conversational interface every day—to plan trips, learn new skills, compare products, or simply get something done. That shift in daily behavior has quietly rewritten user expectations: answers should arrive inline, actions should complete without context switches, and an "app" should feel like help, not a detour.</em></p>
<p>
<a href="https://developers.openai.com/apps-sdk">OpenAI's new Apps SDK</a>, built on top of the
<a href="https://modelcontextprotocol.io">Model Context Protocol (MCP)</a>, formalizes this new reality.
It lets your capability appear directly inside a conversation—the moment intent is expressed. Your UI can render in-thread, call your systems, return structured data or results, and then disappear until needed again. Websites and mobile apps don't vanish—they become structured data layers, identity providers, and policy engines that feed these conversational surfaces.
</p>
<p>
The value unit of software has changed. It's no longer a "destination" you visit; it's an <strong>intent</strong> you resolve.
One chat may now compose multiple brands and services into a single outcome. ChatGPT is the first large-scale implementation, but the pattern will spread fast—other assistants will standardize the same in-thread app model, turning intent-native experiences into a cross-platform baseline.
</p>
<p>
This guide is your map to that landscape. You'll see how discovery and ranking work inside ChatGPT,
what to build first (and why it sticks), the MCP building blocks you'll actually ship,
design rules for inline UX, the KPIs that now define success, and the traits of teams that consistently get picked.
If intent is the new homepage, this is how your brand shows up—and wins—at the moment of need.
</p>
<h2>The Conceptual Shift: From Destinations to Moments</h2>
<p>
For twenty years, digital strategy meant building places for users to go—websites, mobile apps, and dashboards.
Every task began with a detour: open an app, sign in, search, tap through menus, complete the job, exit.
It worked when attention was abundant and distribution predictable.
Today, attention is fractured, and users expect everything to meet them in context.
</p>
<p>
Conversational interfaces changed that equation.
Users now start with language—"Book a flight to Dubai," "Generate a logo," "Summarize this PDF."
Instead of sending them away to a destination, the assistant can <em>perform</em> the task by orchestrating micro-capabilities behind the scenes.
The request becomes the router.
</p>
<em>Shift in Metric:</em> From measuring <strong>visits</strong> and <strong>DAUs</strong> to measuring <strong>invocations</strong> and <strong>resolutions</strong>.
Each intent call is now a unit of engagement and trust.
<p>
This is why traditional growth levers—SEO, App Store ranking, notification funnels—are losing power.
The next era favors systems that can respond precisely to user intent in real time.
Discovery happens by relevance, not by search placement; retention happens by reliability, not by habit loops.
In this model, the AI layer becomes the new operating system of attention.
</p>
<p>
Think of it as the difference between visiting a restaurant and having a chef who appears the moment you're hungry.
The surface stays conversational, but the work behind it becomes modular, composable, and data-driven.
Each capability exists to resolve a single verb—book, design, price, explain, calculate—and then hands control back to the user or to another module in the chain.
</p>
<p>
Research supports this pivot. The global conversational-AI market is projected to exceed $30 billion by 2029,
with more than 900 million daily users engaging chat assistants across platforms.
That's not hype—it's gravity. Users have already chosen the conversational interface as their default starting point.
</p>
<p>
For builders, this means success will no longer be measured by pageviews or downloads,
but by how often and how confidently the model selects your capability to fulfill an intent.
Reliability, clarity of contract, and speed of resolution become your new growth metrics.
</p>
<h2>Chapter 2 – Infrastructure Behind the Shift: MCP + Apps SDK</h2>
<p>
The <a href="https://developers.openai.com/apps-sdk">Apps SDK</a> is not just a new feature—it's the architectural hinge between the web and a fully conversational internet.
It's powered by the <a href="https://modelcontextprotocol.io">Model Context Protocol (MCP)</a>,
an open standard that defines how language models talk to tools, data, and interfaces.
Together they turn what used to be API integrations into full, conversational capabilities.
</p>
<p>
MCP acts as the connective tissue. Every server that implements it can advertise <em>tools</em>
(functions defined with <a href="https://json-schema.org/">JSON Schema</a>), respond to <code>call_tool</code> requests,
and optionally render a live UI inside the chat.
Transport is flexible—Server-Sent Events or Streamable HTTP—ensuring the same app works across ChatGPT web and mobile.
The model itself orchestrates everything: invoking, parsing, and deciding when to surface you.
</p>
<p>
</p>
{
"name": "price_checker",
"description": "Return live product pricing",
"input_schema": {
"type": "object",
"properties": { "sku": { "type": "string" } },
"required": ["sku"]
}
}
<figcaption>Example MCP tool definition using JSON Schema</figcaption>
<p>
On top of MCP sits the Apps SDK—OpenAI's official toolkit that simplifies server registration,
authentication, and UI delivery. It gives developers a consistent way to:
</p>
<ul>
<li>Register tools and expose them to the model with metadata that informs discovery and ranking.</li>
<li>Render inline UIs (cards, carousels, full-screen flows) using the <code>text/html+skybridge</code> MIME type.</li>
<li>Handle user authentication with built-in OAuth 2.1 support.</li>
<li>Define latency budgets, caching hints, and localization through <code>_meta</code> properties.</li>
</ul>
<p>
When you deploy an MCP server through the SDK, ChatGPT can invoke it just as easily as it calls an internal OpenAI tool.
The boundary between "OpenAI-built" and "third-party" dissolves.
Your app becomes part of the model's native vocabulary—the assistant can reference it, chain it, or call it mid-conversation without breaking flow.
</p>
<p>
This is why early builders matter. The SDK's discovery and ranking system learns from usage patterns.
Apps that deliver low-latency, high-completion results quickly become the model's preferred choices for that domain.
The more your tool resolves intents cleanly, the more often it will be automatically suggested or invoked.
</p>
<em>Developer Advantage:</em> The Apps SDK preview (October 2025) still has open discovery slots.
Early apps accumulate ranking data now that later entrants can't easily replicate.
<p>
The protocol also makes experiences portable. MCP is open—other assistants can adopt it,
meaning your same backend can power multiple conversational surfaces.
Build once, and your service could appear across ChatGPT, enterprise copilots, and future multimodal agents.
</p>
<h2>Chapter 3 – Strategic Implications for Brands & Builders</h2>
<p>
The consequence of this infrastructure shift is strategic, not just technical.
Every brand that relies on digital interaction must now decide how it will surface when the user no longer visits a site or opens an app.
</p>
<p>
In the old world, discovery meant capturing attention—SEO, social, ad funnels, app-store rankings.
In the new one, discovery happens through <strong>relevance and reliability</strong>.
The model decides which tool to call based on observed outcomes, latency, and clarity of schema.
The more deterministic and accurate your responses, the higher your selection probability.
</p>
<p>
This transforms the business stack:
</p>
<ul>
<li>
Marketing → Metadata Engineering: success depends on how well your app describes itself to the model.
<p>
Waiting on the sidelines is expensive.
Early adopters are already shaping the ranking algorithms through usage signals—latency, completion, and satisfaction markers.
Like early SEO pioneers, they'll own durable real estate in the model's decision graph.
</p>
<p>
For builders, this means reframing success metrics.
You no longer measure clicks, sessions, or DAUs; you measure <strong>resolved outcomes</strong>.
Did your capability finish the user's job? Did it do so quickly, clearly, and securely?
Those are now the levers that drive organic discovery.
</p>
<em>Strategic Lens:</em> Treat the assistant as your new distribution partner.
It brings intent-qualified traffic; you bring precise resolution.
Mutual value builds automatically through performance.
<p>
The companies that adapt fastest will rebuild their product roadmaps around intents rather than features.
A "feature" is something users hunt for; an "intent" is something they simply express.
The winners design capabilities that fit seamlessly into that sentence and deliver instant clarity.
</p>
<p>
This is the essence of the distribution reset.
The web rewarded visibility; conversational ecosystems reward <em>utility</em>.
Your growth loop becomes self-reinforcing: better resolutions → more model trust → higher invocation → more data → even better performance.
</p>
<h2>Chapter 4 – What to Build & Why It Works</h2>
<p>
The best early Apps are not mini websites—they are <strong>micro-capabilities</strong> that resolve a single, valuable intent
cleanly inside a conversation. You win not by breadth, but by precision: the model keeps calling the tools that
consistently complete the job fastest.
</p>
<p>
If a task already lives on the web, you can probably move it into ChatGPT. Think of your service as a
<em>function of intent</em>:
</p>
<table>
<thead>
<tr>
<th>Category</th>
<th>Typical Intent</th>
<th>Conversation Outcome</th>
</tr>
</thead>
<tbody>
<tr>
<td><strong>Product Discovery</strong></td>
<td>"Show me running shoes under $150."</td>
<td>Inline cards with filtered SKUs and links.</td>
</tr>
<tr>
<td><strong>Planning & Decision</strong></td>
<td>"Help me plan a 3-day Tokyo itinerary."</td>
<td>Carousel of suggested plans + booking CTAs.</td>
</tr>
<tr>
<td><strong>Computation & Tools</strong></td>
<td>"Calculate my monthly payment."</td>
<td>Interactive calculator widget with results summary.</td>
</tr>
<tr>
<td><strong>Support & Education</strong></td>
<td>"Explain recursion with a quick demo."</td>
<td>Animated teaching widget with follow-up Q&A.</td>
</tr>
</tbody>
</table>
<p>
These patterns share a principle: <strong>resolution in-flow</strong>.
The user never leaves the chat, yet completes the job.
The system measures and rewards that frictionless outcome.
</p>
<em>Tip:</em> Start with one clear verb—<strong>book</strong>, <strong>price</strong>, <strong>compare</strong>, <strong>explain</strong>.
When the model understands what your tool "owns," invocation becomes automatic.
<p>
Over time, multiple brands will chain together: a budgeting app calls your mortgage calculator,
which calls an insurance quote tool—all orchestrated by the model.
The connective format that makes this possible is the <strong>structuredContent</strong> payload your app returns.
</p>
<h2>Chapter 5 – Engineering & Design Playbook</h2>
<p>
Building an App for ChatGPT means building an <strong>MCP server</strong> that declares your capabilities
and optionally ships a small UI bundle.
You don't need a new tech stack—just a disciplined structure:
</p>
<ol>
<li>Describe your tools with clear JSON Schema.</li>
<li>Expose them via a public <code>/mcp</code> endpoint.</li>
<li>Attach an HTML template rendered with <code>text/html+skybridge</code>.</li>
<li>Return three fields in every response: <code>structuredContent</code>, <code>content</code>, and <code>_meta</code>.</li>
</ol>
<p>
</p>
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { z } from "zod";
const server = new McpServer({ name: "price-checker", version: "1.0.0" });
// Define a simple tool
server.registerTool(
"check-price",
{
title: "Check Product Price",
inputSchema: { sku: z.string() },
_meta: { "openai/outputTemplate": "https://api.example.com/templates/price-card" }
},
async ({ sku }) => {
const price = await fetch(`https://api.example.com/prices/${sku}`).then(r => r.json());
return {
structuredContent: { sku, price: price.amount, currency: price.currency },
content: [{ type: "text", text: `The current price is ${price.amount} ${price.currency}.` }],
_meta: { source: "example-api", checkedAt: new Date().toISOString() }
};
}
);
server.listen(8080);
<figcaption>Minimal MCP server registering a single pricing tool</figcaption>
<p>
This snippet shows the full loop: the model calls <code>check-price</code> with a SKU,
your server fetches data, and returns both human and machine-readable outputs.
ChatGPT then decides whether to render a card, show text, or compose it with another tool.
</p>
<em>Best Practice:</em> Keep responses small and deterministic.
The faster your tool resolves and the clearer your schema, the more often the model will select it again.
<h3>Designing for Conversation</h3>
<p>
Your UI is not a standalone app—it's a fragment of dialogue.
Keep interfaces single-purpose, visually quiet, and responsive to chat context.
Use system fonts and platform colors, limit interactive depth to one or two steps,
and let ChatGPT handle narration around your component.
</p>
<ul>
<li>
Inline cards — confirmations, summaries, and quick pickers.
<p>
Instrument everything. Log latency per invocation, hydration time, and completion rate.
Treat these as product metrics, not technical afterthoughts—they directly influence ranking.
</p>
<p>
Security and privacy follow standard web rules: use HTTPS, strict CSP, and OAuth 2.1.
Never leak private identifiers in <code>structuredContent</code>; keep them in <code>_meta</code>.
When you localize, respect the <code>_meta["openai/locale"]</code> hint and render dates or currency accordingly.
</p>
<blockquote>
<p>
The most elegant conversational interfaces keep it minimal.
</p>
</blockquote>
<p>
By following these principles, your app feels like a natural extension of the conversation—fast,
focused, and invisible until it's exactly what the user needs.
</p>
<h2>Chapter 6 – Monetisation Models</h2>
<p>
Utility without capture is philanthropy.
Apps inside ChatGPT can't rely on banner clicks or ad impressions—there are none.
The Apps SDK is a distribution layer, not a checkout flow.
Monetisation therefore hinges on connecting in-thread value to your external revenue systems.
</p>
<p>
The core question becomes: <strong>Who owns the customer?</strong>
OpenAI owns the <em>conversation</em>; you own the <em>relationship</em>.
The winning pattern treats the assistant as your most powerful channel partner—
you deliver resolution; it delivers reach.
</p>
<h3>Emerging Commercial Models</h3>
<ul>
<li>
<strong>SaaS Entitlement Play</strong> —
Authenticate through OAuth 2.1, detect plan tier, and unlock premium features inline.
Paying users experience full capability; free users see a guided teaser that converts naturally.
</li>
<li>
<strong>High-Intent Lead Funnel</strong> —
Ideal for consultative sectors (finance, real estate, B2B).
Your app qualifies leads via calculators or diagnostics, then ends with one CTA:
"Book a 15-minute consultation."
Every invocation is a pre-qualified prospect.
</li>
<li>
<strong>Transactional & Affiliate Model</strong> —
Retail, travel, and marketplaces embed configuration, comparison, and pre-checkout flows in-chat.
Final payment can redirect to your site with pre-filled carts and tracking parameters.
The assistant becomes your conversion pre-processor.
</li>
<li>
<strong>Brand & Awareness Utility</strong> —
Some Apps act purely as brand anchors—free, frictionless, and ubiquitous.
They build trust, gather preference data, and secure long-term default status
("Check the weather → calls your app").
</li>
</ul>
<em>Metric Shift:</em>
Track <strong>resolved intents per user</strong>, not sessions.
Each completed job is both satisfaction signal and monetisable event.
<p>
Over time, OpenAI and others will formalise revenue APIs, but early builders shouldn't wait.
The current advantage lies in habit formation: become the model's default resolver now,
monetise through your existing channels later.
</p>
<h2>Chapter 7 – Where You'll Win First</h2>
<p>
Certain industries already think conversationally—they'll convert first because the interface matches their workflow.
Anywhere users compare, configure, decide, or request in natural language is fertile ground.
</p>
<table>
<thead>
<tr>
<th>Sector</th>
<th>Example Intent</th>
<th>Inline Outcome</th>
</tr>
</thead>
<tbody>
<tr>
<td><strong>Travel & Hospitality</strong></td>
<td>"Find flights to Dubai next Thursday."</td>
<td>Interactive flight cards with booking links.</td>
</tr>
<tr>
<td><strong>Education & Training</strong></td>
<td>"Teach me basic SQL with practice examples."</td>
<td>Adaptive lesson widget with live quizzes.</td>
</tr>
<tr>
<td><strong>Finance & Insurance</strong></td>
<td>"Estimate my mortgage payment."</td>
<td>Calculator + CTA to book advisor call.</td>
</tr>
<tr>
<td><strong>Retail & E-Commerce</strong></td>
<td>"Compare noise-cancelling headphones."</td>
<td>Carousel of products + direct purchase options.</td>
</tr>
<tr>
<td><strong>Healthcare</strong></td>
<td>"Schedule a follow-up with my doctor."</td>
<td>Secure scheduling + triage guidance.</td>
</tr>
<tr>
<td><strong>Entertainment & Sports</strong></td>
<td>"Show me tonight's NBA stats."</td>
<td>Live scoreboard + ticketing widget.</td>
</tr>
<tr>
<td><strong>Home Improvement</strong></td>
<td>"Plan a kitchen renovation budget."</td>
<td>Step-by-step planner with cost estimates.</td>
</tr>
</tbody>
</table>
<p>
These categories share three properties:
</p>
<ol>
<li>
Structured Data — clear inputs/outputs make schemas easy.
<p>
Early entrants in these sectors will define their industry schemas—the formats every competitor must match.
Once those shapes solidify, the model will prefer known structures,
giving schema authors a compounding advantage similar to early search-index dominance.
</p>
<em>Strategic Advice:</em>
Pick one vertical intent you can dominate.
Build it impeccably, measure invocation rates, then expand sideways into adjacent intents using the same data backbone.
<h2>Chapter 8 – Team Traits & Future Orchestration</h2>
<p>
The teams that consistently win in this new ecosystem don't treat Apps as marketing stunts or integrations.
They treat them as <strong>core product interfaces</strong>—living systems that evolve by observing, resolving, and learning
from real user intent.
</p>
<h3>Traits of Teams That Win</h3>
<ul>
<li>
Utility Over Messaging: They lead with usefulness. The pitch is embedded in performance.
<p>
These teams collapse the traditional gap between engineering, design, and strategy.
Conversation design is product design.
Schema is UX.
Latency is brand perception.
The companies that grasp this reality early are the ones whose apps the model will repeatedly call.
</p>
<h3>The Next Step: Orchestration</h3>
<p>
Today, each App acts independently. Tomorrow, multiple capabilities—across brands and domains—will cooperate in a single conversation.
This is the birth of the <strong>orchestrated web</strong>: where the assistant conducts a network of services to deliver complete outcomes.
One chat might involve five vendors seamlessly chained: data retrieval, analysis, booking, payment, and follow-up.
</p>
<p>
MCP was designed with this future in mind.
It standardizes contracts between capabilities so composition happens naturally.
A travel planner app could invoke your pricing tool; your pricing tool could hand its structured output
to a booking engine—all without user friction or custom integrations.
</p>
<em>Vision:</em> The orchestrated web is the AI-native internet.
Every service becomes a callable function of trust and speed, not a siloed domain.
<p>
The long-term opportunity is enormous.
When orchestration becomes the norm, brand equity will correlate with invocation reliability.
The best app isn't the prettiest—it's the one the model calls first, because it never fails to deliver.
</p>
<h2>Conclusion – The Bottom Line</h2>
<p>
Apps inside ChatGPT aren't a novelty—they're the next distribution layer of software.
The center of gravity has shifted from destinations to intents.
The winners will be the teams who turn a single, high-value customer job into a
fast, trustworthy capability that the model keeps choosing.
</p>
<p>
Treat this as <strong>product work, not marketing work</strong>.
Build for intent, not for eyeballs.
Measure resolution, not reach.
The companies that internalize those principles now will own the next decade of discovery.
</p>
<p>
The playbook is clear:
</p>
<ol>
<li>
Pick one sharp intent you can dominate.
<p>
Every resolved task strengthens your position in the model's ranking graph.
Every fast response earns another call.
Over time, you don't just serve users—you become part of the conversation itself.
</p>
<p>
The market is wide open.
Build with precision, respect latency, and let utility lead.
You'll earn a permanent slot in the most valuable real estate in software—right inside the conversation.
</p>
Top comments (0)