Amp
ampcode.comKey Features
Tech Stack
Key Features
Tech Stack
I run every single coding prompt through Codex, Claude Code, Qwen and Gemini, compare which one gives me the best and go ahead with using that one. Maybe I go with Codex 60% of the times, Claude 20% and Qwen/Gemini the remaining 20%, not often at all either of them get enough right. I've tried integrating Amp into my workflow too, but as mentioned, too expensive. I do think the grass is currently the greenest with Codex, still.
That aligns with my annecdata :)
But my first thought looking at this is that the numbers are probably skewed due to distribution of user skill levels, and what types of users choose which tool.
My hypothesis is that Amp is chosen by people who are VERY highly skilled in agentic development. Meaning these are the people most likely to provide solid context, good prompts, etc. That means these same people would likely get the best results from ANY coding agent. This also tracks with Amp being so expensive -- users or companies are more likely to pay a premium if they can get the most from the tool.
Claude Code on the other hand is used by (I assume) a way larger population. So the percentage of low-skill users is likely to be much higher. Those users may still get value from the tool, but their success rate will be lower by some factor with ANY coding agent. And this issue (if my hypothesis is correct) is likely 10x as true for GitHub Copilot.
Therefore I don't know how much we should read into stats like the total PR merge success percentage, because it's hard to tell the degree of noise caused by this user skill distribution imbalance.
Still interesting to see the numbers though!
But then I switched to GLM 4.6 using Claude CLI tool and that was good enough and significantly cheaper/faster.
Then Opus 4.5 came out with better pricing and might as well just use that directly. Still working great.
With Amp I was spending $5 here and there every day. Great, but pricey.
My intuition is that it's not deep... the differentiating factor is "regular" (non LLM) code which assembles the LLM context and invokes the LLM in a loop.
Claude/Codex have some advantage, because they can RLHF/finetune better than others. But ultimately this is about context assembly and prompting.
The same thing is going to happen with all of the human language artifacts present in the agentic coding universe. Role definitions, skills, agentic loop prompts....the specific language, choice of words, sequence, etc really matters and will continue to evolve really rapidly, and there will be benchmarkers, I am sure of it, because quite a lot of orgs will consider their prompt artifacts to be IP.
I have personally found that a very high precision prompt will mean a smaller model on personal hardware will outperform a lazy prompt given to a foundation model. These word calculators are very very (very) sensitive. There will be gradations of quality among those who drive them best.
The best law firms are the best because they hire the best with (legal) language and are able to retain the reputation and pricing of the best. That is the moat. Same will be the case here.
You might get an 80% “good enough” prompt easily but then all the differentiation (moat) is in that 20% but that 20% is tied to the model idiosyncrasies, making the moat fragile and volatile.
I love that I am not welded to one model and someone smart has evaluated what’s best fit for what. It’s a bit a shame they lost Steve Yegge as their brand ambassador. A respected and practicing coder is big endorsement.
To anyone on the fence - give it a go.
Because if I understand them correctly, aren’t they a wrapper around all the major LLMs (focused specially on developer use cases?
I was watching the CEO of that Chad IDE give an interview the other day, they are doing the same thing as amp just with "brain rot" ads/games etc instead (different segment, fine), they are using that activity to produce value (ads/game usage or whatever) for another business such they can offset the costs of more expensive models. (I presume this could extend out into selling data also, but I don't know they do that today)
This is not exactly where I expected all this to go, during the rise of DO at least once a quarter someone would email me about getting advice on their IDE startup and I never thought a paid IDE product would/could ever take off, that was until this code complete stuff became doable, so for me anyway it's kinda funny to see this all kick off, if you'd told me in 2015 we'd be seeing IDE wars in 2025, I'd have laughed. Strange times.
Congrats to the amp team on all the success btw, all I hear is great things about the product, good work.
Not affiliated with Hacker News or Y Combinator. We simply enrich the public API with analytics.