Apps SDK
Posted3 months agoActive3 months ago
developers.openai.comTechstoryHigh profile
skepticalmixed
Debate
80/100
OpenaiApps SDKAI-Powered AppsModel Context Protocol
Key topics
Openai
Apps SDK
AI-Powered Apps
Model Context Protocol
OpenAI has launched an Apps SDK, sparking debate among developers and users about its potential, limitations, and implications for the future of AI-powered apps and interfaces.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
15m
Peak period
121
0-6h
Avg / period
22.9
Comment distribution160 data points
Loading chart...
Based on 160 loaded comments
Key moments
- 01Story posted
Oct 6, 2025 at 2:27 PM EDT
3 months ago
Step 01 - 02First comment
Oct 6, 2025 at 2:42 PM EDT
15m after posting
Step 02 - 03Peak activity
121 comments in 0-6h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 9, 2025 at 3:47 AM EDT
3 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45494558Type: storyLast synced: 11/20/2025, 8:14:16 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Now, I realize that the best argument for MCP vs function calls in my case, is that I want to allow external products/agents/chatbots to interface with my app. MCP is that standard. I will implement very carefully, but that's what I need to do.
So far, it seems that if you give an LLM a few tools to create projects and other entities, they seem to be very good at using them. The user gets the option of chat driven ui for our app, with not that much work for limited features.
Currently building internal MCP servers to make that easy. But I can imagine having a public one in the future.
The problem with this approach is precisely that these apps/widgets have hard-coded input and output schema. They can work quite well when the user asks something within the widget's capabilities, but the brittleness of this approach starts showing quickly in real-world use. What if you want to use more advanced filters with Zillow? Or perhaps cross-reference with StreetEasy? If those features aren't supported by the widget's hard-coded schema, you're out of luck as a user.
What I think it much more exciting is the ability to completely create generative UI answers on the fly. We'll have more to say on this soon from Phind (I'm the founder).
That said, I used it a lot more a year ago. Lately I’ve been using regular LLMs since they’ve gotten better at searching.
Conservational user interfaces are opaque; they lack affordances. https://en.wikipedia.org/wiki/Affordance
> Affordances represent the possibilities in the world for how an agent (a person, animal, or machine) can interact with something. Some affordances are perceivable, others are invisible. Signifiers are signals. Some signifiers are signs, labels, and drawings placed in the world, such as the signs labeled “push,” “pull,” or “exit” on doors, or arrows and diagrams indicating what is to be acted upon or in which direction to gesture, or other instructions. Some signifiers are simply the perceived affordances, such as the handle of a door or the physical structure of a switch. Note that some perceived affordances may not be real: they may look like doors or places to push, or an impediment to entry, when in fact they are not.
With Norman's definition, if a conversational interface can perform an action, it affords that action. The fact that you don't know that it affords that action means there's a lack of a signifier.
As you say, this is a matter of definition, I'm just commenting on Norman's specific definition from the book.
I immediately knew the last generation of voice assistants was dead garbage when there was no way to know what it could do, they just expected you to try 100 things, until it worked randomly
For a concrete example, think a search result listing that can be broken down into a single result or a matrix to compare results, as well as a filter section. So you could ask for different facets of your current context, to iterate over a search session and interact with the results. Dunno, I’m still researching.
Have you written somewhere about your experience with Phind in this area?
Now that models have gotten much more capable, I'd suggest to give the executing model as much freedom with setting (and even determining) the schema as possible.
Chat paired to the pre-built and on-demand widgets address this limitation.
For example, in the keynote demo, they showed how the chat interface lets you perform advanced filtering that pulls together information from multiple sources, like filtering only Zillow housers near a dog park.
I think most software will follow this trend and become generated on-demand over the next decade.
The only place I can see this working is if the LLM is generating a rich UI on the fly. Otherwise, you're arguing that a text-based UX is going to beat flashy, colourful things.
Sure, but deploying a website or app doesn't mean anyone's going to use it, does it?
I could make an iOS app, I could make a website, I could make a ChatGPT app... if no one uses it, it doesn't matter how big the userbase of iOS, the internet, or ChatGPT is...
Per the docs: 'Every app comes from a verified developer who stands behind their work and provides responsive support'
That's thinly veiled corporate speak for, Fortune 500 or GTFO
Personally I don't hope thats the future.
For a large number of tasks that cleanly generalize into a stream of tokens, command line or chat is probably superior. We'll get some affordances like tab auto completion to help remember the name of certain bots or mCP endpoints that can be brought in as needed...
But for anything that involves discovery, graphical interaction feels more intuitive and we'll probably get bespoke interfaces relevant to that particular task at hand with some sort of partially hidden layers to abstract away the token stream?
I was really hoping Apple would make some innovations on the UX side, but they certainly haven’t yet.
They want to be the platform in which you tell what you want, and OAI does it for you. It's gonna connect to your inbox, calendar, payment methods, and you'll just ask it to do something and it will, using those apps.
This means OAI won't need ads. Just rev share.
Ads are defenitely there. Just hidden so deeply in the black box which is generating the useful tips :)
One could be for example: from people asking online which tools they should use to build something and being constantly recommended to do it with Next.js
Another could be: how many of the code that was used to train the LLM is done in Next.js
Generally, the answer is probably something along the lines of "next.js is kind of the most popular choice at the time of training".
You may have started seeing this when LLMs seem to promote things based entirely on marketing claims and not on real-world functionality.
More or less, SEO spam V2.
In my (non-lawyer) understanding, each message potentially containing sponsored content (which would be every message, if the bias is encoded in the LLM itself,) would need to be marked as an ad individually.
That would make for an odd user interface.
If OpenAI thinks there’s sweet, sweet revenue in email and calendar apps, just waiting to be shared, their investors are in for a big surprise.
OpenAI’s moat will only come from the products they built on top. Theoretically their products will be better because they’ll be more vertically integrated with the underlying models. It’s not unlike Apple’s playbook with regard to hardwares and software integration.
[1] This is an example. Which model was the best when is not important.
[0] https://www.nber.org/system/files/working_papers/w34255/w342...
They obviously want both. In fact they are already building an ad team.
They have money they have to burn, so it makes sense to throw all the scalable business models in the history, eg app store, algo feed, etc, to the wall and see what stick.
In 2024, iOS App Store generated $1.3T in revenue, 85% of which went to developers.
Edit: yes I understand it is correct, but still it sounds like an insane amount
It is now evident why Flash was murdered.
This is a stupid conspiracy given Apple decided not to support Flash on iPhone since before Jobs came around on third-party apps. (The iPhone was launched with a vision of Apple-only native apps and HTML5 web apps. The latter's performance forced Cupertino's hand into launching the App Store. Then they saw the golden goose.)
HTML5 was new and not widely supported, the web was WAY more fragmented back then, to put things in perspective, Internet Explorer still had the largest market share, by far. The only thing that could provide the user with a rich interactive experience was Flash, it was also ubiquitous.
Flash was the biggest threat to Apple's App Store; this wasn't a conspiracy, it was evident back then but I can see why it is not evident to you in 2025. Jobs open letter was just a formal declaration of war.
Yes. It was a bad bet on the open web by Apple. But it was the one they took when they decided not to support Flash with the original iPhone's launch.
> Flash was the biggest threat to Apple's App Store
Flash was not supported since before there was an App Store. Since before Apple deigned to tolerate third-party native apps.
You can argue that following the App Store's launch, Apple's choice to not start supporting Flash was influenced by pecuinary interests. But it's ahistoric to suggest the reason for the original decision was based on interests Cupertino had ruled out at the time.
That 1T figure is real, but it includes things like if you buy a refrigerator using the Amazon iOS app.
I'm genuinely surprised these companies went with usage-based versus royalty pricing.
Connecting these apps will, at times, require authentication. Where it does not require payment, it's a fantastic distribution channel.
"Find me hotels in Capetown that have a pool by the beach .Should cost between 200 dollars to 800 dollars a night "
I don't see how this is a significant upgrade over the many existing hotel-finder tools. At best it slightly augments them as a first pass, but I would still rather look at an actual map of options than trust a stream of generated, ad-augmented text.
The UI 'cards' will naturally becoming ever increasing, and soon you end up back with a full app within ChatGPT or ChatGPT just becomes an app launcher.
The only advantage I can see is if ChatGPT can use data from other apps/ chats in your searches e.g. find me hotels in NYC for my upcoming trip (and it already knows the types of hotels you like, your budget and your dates)
However, it might be useful for people who do want to use that instead.
Instead, the model will provide you with a list of (in chat) “apps” that can fulfill your request. SEO becomes AISO (AI Search Optimization). Sites can partly expose data to entice you to choose them.
This time will be different?
I personally prefer well curated information.
The LLM will do the curation.
e.g. Coursera can send back a video player
I could see chat apps becoming dominant in Slack-oriented workplaces. But, like, chatting with an AI to play a song is objectively worse than using Spotify. Dynamically-created music sounds nice until one considers the social context in which non-filler music is heard.
Getting an AI to play "that song that goes hmm hmmm hmmm hmmm ... uh, it was in some commercials when I was a kid" tho
Also their playlists are made by real people (mostly...), so they don't completely suck ass.
Also, following the Beatport top 100 tech house playlist, and hearing how many tracks aren't actually tech house makes me wonder about who makes that particular playlist.
That's how I feel about a lot of AI stuff.
Like... It's neat. It's a fun novelty. It makes a good party trick. It's the software equivalent of a knick knack.
Like 90% of the pixel AI features. There's some good ones in there, sure, but most of them you play around with for a day and then forget exist.
This isn't me making a cute little website in my free time. This is thousands of developers, super computers out the wazoo, and a huge chunk of the western economy.
Like, a snowglobe is cute. They don't do much, but they're cute. I'd buy one for ten dollars.
I would not buy a snowglobe for 10 million dollars.
Absolutely. The point is this is a specialised and occasional use case. You don't want to have to go through a chat bot every time you want to play a particular song just because sometimes you might hum at it.
The closest we've come to a widely-adopted AR interface are AirPods. Critically, however, they work by mimicing how someone would speak to a real human by them.
There's a whole bizarre subculture in computing that fails to recognize what it is about computers that people actually find valuable.
Everyone wants the next device category. They covet it. Every other company tries to will it into existence.
I’m not very bullish on people wanting to live in the ChatGPT UI, specifically, but the concept of dynamic apps embedded into a chat-experience I think is a reasonable direction.
I’m mostly curious about if and when we get an open standard for this, similar to MCP.
The former is like a Waymo, the latter is like my car suddenly and autonomously deciding that now is a good time to turn into a Dollar Tree to get a COVID vaccine when I'm on my way to drop my kid off at a playdate.
What users want, which various entities religiously avoid providing to us, is a fair price comparison and discovery mechanism for essentially everything. A huge part of the value of LLMs to date is in bypassing much of the obfuscation that exists to perpetuate this, and that's completely counteracted by much of what they're demonstrating here.
So perhaps chatbots are an excellent method for building out a prototype in a new field while you collect usage statistics to build a more refined UX - but it is bizarre that so many businesses seem to be discarding battle tested UXes for chatbots.
Thing is, for those who paid attention to the last chatBot hype cycle, we already knew this. Look at how Google Assistant was portrayed back in 2016. People thought you'd be buying starbucks via the chat. Turns out the starbucks app has a better UX
The only reason for the voice interface is to facilitate the production of a TV show. By having the characters speak their requests aloud to the computer as voice commands, the show bypasses all the issues of building visual effects for computer screens and making those visuals easy to interpret for the audience, regardless of their computing background. However, whenever the show wants to demonstrate a character with a high level of computer mastery, the demonstration is almost always via the touchscreen (this is most often seen with Data), not the voice interface.
TNG had issues like this figured out years ago, yet people continue to fall into the same trap because they repeatedly fail to learn the lessons the show had to teach.
Maybe this is how we all get our own offices again and the open floor plan dies.
"...and that is why we need the resources. Newline, end document. Hey, guys, I just got done with my 60 page report, and need-"
"SELECT ALL, DELETE, SAVE DOCUMENT, FLUSH UNDO, PURGE VERSION HISTORY, CLOSE WINDOW."
Here's hoping this at least gets us back to cubes.
>changes bass to +4 because the unit doesn't do half increments
“No volume up to 35, do not touch the EQ”
>adjusts volume to 4 because the unit doesn’t do half increments
> I reach over, grab my remote, and do it myself
We have a grandparent that really depends on their Alexa and let me tell you repeatedly going “hey Alexa, volume down. Hey Alexa, volume down. Hey Alexa, volume down,” gets really old lol we just walk over and start using the touch interface
A UX is better and another app or website feels like the exact separation needed.
Booking flights => browser => skyscanner => destination typing => evaluation options with ai suggestions on top and UX to fine-tune if I have out of the ordinary wishes (don’t want to get up so early)
I can’t imagine a human or an AI be better than is this specialized UX.
This general concept (embedding third parties as widgets in a larger product) has been tried many times before. Google themselves have done this - by my count - at least three separate times (Search, Maps, and Assistant).
None have been successful in large part because the third party being integrated benefits only marginally from such an integration. The amount of additional traffic these integrations drive generally isn't seen as being worth the loss of UX control and the intermediation in the customer relationship.
MCP standardizes how LLM clients connect to external tools—defining wire formats, authentication flows, and metadata schemas. This means apps you build aren't inherently ChatGPT-specific; they're MCP servers that could work with any MCP-compatible client. The protocol is transport-agnostic and self-describing, with official Python and TypeScript SDKs already available.
That said, the "build our platform" criticism isn't entirely off base. While the protocol is open, practical adoption still depends heavily on ChatGPT's distribution and whether other LLM providers actually implement MCP clients. The real test will be whether this becomes a genuine cross-platform standard or just another way to contribute to OpenAI's ecosystem.
The technical primitives (tool discovery, structured content return, embedded UI resources) are solid and address real integration problems. Whether it succeeds likely depends more on ecosystem dynamics than technical merit.
“CEO” Fidji Simo must really need something to do.
Maybe I’m cynical about all of this, but it feels like a whole lot of marketing spin for an MCP standard.
https://lukew.com/ff/entry.asp?2122
222 more comments available on Hacker News