Things I've Learned in My 7 Years Implementing AI
Posted3 months agoActive2 months ago
jampa.devTechstoryHigh profile
calmmixed
Debate
60/100
AILlmsProduct DevelopmentStartups
Key topics
AI
Llms
Product Development
Startups
The author reflects on 7 years of implementing AI, sharing lessons learned and observations on the current state of AI development, sparking a discussion on the limitations and potential of AI as a product or tool.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
22m
Peak period
51
Day 1
Avg / period
11
Comment distribution55 data points
Loading chart...
Based on 55 loaded comments
Key moments
- 01Story posted
Oct 15, 2025 at 2:27 PM EDT
3 months ago
Step 01 - 02First comment
Oct 15, 2025 at 2:49 PM EDT
22m after posting
Step 02 - 03Peak activity
51 comments in Day 1
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 28, 2025 at 1:02 PM EDT
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45596602Type: storyLast synced: 11/20/2025, 6:45:47 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
this list isnt learnings.
EDIT: Removed the video because a bug in Substack causes the space bar to play the video instead of scrolling down. Sorry for the unintentional jumpscare.
I think it's a good example of the kind of internal tools the article is talking about. I would not have spent the time to build this without claude making it much faster to build stand-alone projects and I would not have the agent to do the english -> policy output with LLMs.
Nailed it. And the thing is, you can (and should) still have deterministic guard rails around AI! Things like normalization, data mapping, validations etc. protect against hallucinations and help ensure AI’s output follows your business rules.
And further downstream: Audit trails, human sign-offs, operations which are reversible or have another workflow for making compensating actions to fix it up.
In my mind you are trading potentially a function that always evaluates the same for a given f(x) for one that might not evaluate the same and requires oversight.
This is the best case for AI, it's not very different from the level 3 autonomous car with driver in the loop instead of fully autonomous level 5 vehicle that probably requires AGI level of AI.
The same applies to medicine where limited number specialists (radiologist/cardiologist/oncologist/etc) in the loop are being assisted by AI for activities that probably require too much time for experts manually looking at laborious evidences especially for non-obvious early symptom detection (X-ray/ECG/MRI) for the modern practice of evidence based medicine.
That's fine if the person wouldn't be able to write the code otherwise.
There are lots and lots of people in positions that are "programming adjacent". They use computers as their primary tool and are good at something (like CAD), but can't necessarily sling code. So, a task like: "We're about to release these drawings to an external client. Please write a script to check that all the drawings have author, project, and contract number that matches what they should for this client and flag any that don't." is good AI bait. Or "Please shovel this data from X, Y, and Z into an Excel Spredsheet" is also decent AI bait.
Programmers underestimate how difficult it is to synthesize code from thin air. It is much easier to read a small script than to construct it.
I'm not disagreeing with the overall post, but from closely observing end users of LLM-backed products for a while now, I think this needs nuance.
The average joe, be it a developer, random business type, a school teacher or your mum, is very bad at telling an llm what it should do.
- In general people are bad at expressing their thoughts and desires clearly. Frontier LLMs are still mostly sycophantic, so in absence of clear instructions they will make up things. People are prone to treating the LLM as a mind reader, without critically assessing if their prompts are self-contained and sufficiently detailed.
- People are pretty bad at estimating what kind of data an LLM understands well. In general data literacy, and basic data manipulation skills, are beneficial when the use case requires operating on data besides natural language prompts. This is not a given across user bases.
- Very few people have a sensible working model of what goes on in an autoregressive black box, so they have no intuition on managing context
User education still has a long way to go, and IMO is a big determining factor in people getting any use at all from the shiny new AI stuff that gets slathered onto every single software product these days
Free form chat is pretty terrible. People just want the thing to (smartly) take actions. One or two buttons that do the thing, no prompting involved, is much less complicated.
There are like thousands wrappers around LLMs masquerading as AI apps for specialized usecases, but the real performance of these apps is really only bottlenecked by the LLM performance, and their UIs generally only get in the way of the direct LLM access/feedback loop.
To work with LLMs effectively you need to understand how to craft good prompts, and how to read/debug the responses.
Yeah, it often makes sense to adjust the user's prompt, add system/wrapper prompts, etc. But that's not really related to UI..
"Executive functions include basic cognitive processes such as attentional control, cognitive inhibition, inhibitory control, working memory, and cognitive flexibility. Higher-order executive functions require the simultaneous use of multiple basic executive functions and include planning and fluid intelligence (e.g., reasoning and problem-solving)."
Executive function is not just emotional control. It's higher levels of thinking. It's the E in CEO. AI is really just an amplifier, just like how a calculator in one person's hand could be much powerful than in someone else's
and everyone's being sold on this tech being super magic but to some questions there is an irreducible complexity that you have to deal with, and that still takes effort.
So this is nice
> productionizing their proof-of-concept code and turning it into something people could actually use.
because it's so easy to glamorize research, while ignoring what actually makes ideas products.
This is also the problem. It's a looking back perspective and it's so easy to be miss the forest from the trees when you're down in the weeds. I'm talking from experience and it's a feeling I get when reading the post.
In the grand scheme of things our current "AI" will probably look like a weird detour.
Note that a lot of these perspectives are presented (and thought) without a timeline in mind. We're actually witnessing timelines getting compressed. It's easy to see the effects of one track while missing the general trend.
This take is looking at (arguably "over") LLM timeline, while missing everything else that is happening.
It’s easy now to get something good enough for use by you, friends, colleagues etc.
As it’s always been, developing an actual product is at least one order of magnitude more work. Maybe two.
But both internal tools and full products are made one order of magnitude easier by AI. Whole products can be made by tiny teams. And that’s amazing for the world.
No. Not at all. Many things maybe got easier but a lot of things got magnitudes harder. Maintaining bug bounty programs for example, or checking the authenticity and validity of written content on blogs.
Calling LLMs are a huge win for humanity is incredibly naive given we dont know the long term effects these tools are having on creativity in online spaces, authenticity of user bases, etc etc.
Some artificial flavorings and artificial coloring have been proven to be cancerous. I doubt artificial intelligence is going to end up being much different.
-- uhhh... am I the only one seeing a startup boom??? There are a bajillion kids working on AI start ups these days.
AI deals continued to dominate venture funding during the third quarter. AI companies raised $19 billion in Q3, according to Crunchbase data. That figure represents 28% of all venture funding.
The fourth quarter of 2024 has been no less busy for these outsized rounds. Elon Musk’s xAI raised a behemoth $6 billion round, one of seven AI funding rounds over $1 billion in 2024, in November. That’s just months after OpenAI raised its $6.6 billion round.
https://techcrunch.com/2024/12/20/heres-the-full-list-of-49-...
What do you do? Oh we do DeepCoffee brewing. It's a coffee machine powered by Deep Learning to brew the perfect cup. Keurig and Starbucks are Yahoo, and we're Google. (now people probably say those guys are google and we're openai but I digress)
Yeah I don't know about that, the model providers like OpenAI, Anthropic, etc, literally sell intelligence as a product. And their business model is looking a lot more stable in the long term than all the startups built on top.
This correlates with the natural world. Intelligence isn’t a direct means of survival for anything. It isn’t a requirement for physical health.
It is an indirect means, I.e. a tool.
This has been my main use case for AI. I have lots of ideas for little tools to take some of the drudgery out of regular work tasks. I'm a developer and could build them but I don't have the time. However, they're simple enough that I can throw them together in a basic script form really quickly with Cursor. Recently I built a tool to analyse some files, pull out data, and give me it in the format I needed. A relatively simple python script. Then I used Cursor to put it together with a simple file input UI in an electron app so I could easily share it with colleagues. Like I say, I've been developer for a long time but never written python or packaged an electron app and this made it so easy. The whole thing took less than 20mins and it was quick enough that I could do it as part of the task I was doing anyway rather than additional work I needed to find time to do.
> The last releases were unimpressive. Does anyone know a real application where ChatGPT 5 can do something that o3 could not?
Yeah, GPT-5 is way better at coding in codex, especially for longer tasks. Opus 4.1 is pretty good too. Gemini 3 is dropping soon. GPT-5 was more about reducing costs/having a router so when your doctor asks a question to chatgpt he's routed to the thinking version from what I understand.
> The good news is that what we have is enough for most people.
This is true.
Also stop using LMArena as an indicator of anything, it hasn't meant much for more than 6 months.
> AI tools like KNNs are very limited but still valuable today.
I've seen discussions calling even feed-forward CNNs, monte-carlo chains, or GANs "antiquated" because transformers and diffusion have surpassed their performance on many domains. There is a hyper-fixation on large transformers and a sentiment that it somehow replaces everything that came before in every domain.
It's a tool that unlocks things we could not do before. But it doesn't do everything better. It does plenty of things worse (at-least taking power and compute into account). Even if it can do algebraic now (as is so proudly proclaimed in the benchmarks), wolfram alpha remains and will continue to remain far more suited to the task. Even if it can write code; it does NOT replace programming languages as I've seen people claim in very recent posts on here on HN.
SWEBench performance moved from 25% to 70% solved since beginning of 2025, and even with the narrowest possible lens from 65% to 70% since May. ARC-AGI2 keeps rapidly climing. We have experimental models able to (maybe) hold their ground at IMO gold. As well as performing at IPhO gold level.
And that leaves out the point that LMArena is a popularity contest. Not "was this correct", but "which answer did you like better". The thing that brought us glazing. A leveling ELO (or a very slowly climbing one) is kind of expected, and is really saying nothing about progress in the field.
Still doesn't mean "THE MACHINE GODS ARE COMING", but I would expect to see continued if slowing improvement. (I mean, how much better can you get at math if you already can be a useful assistant to Terence Tao and win IMO gold? And wouldn't we expect that progress to slow?)
But more than "how hard of a problem can you solve", I expect we'll see a shift to instead looking at missing capabilities. E.g. memory - currently, you can tell Claude 500 times to use uv, not pip, and it will cheerfully not know that again the 501st time. That's much more important than "oh, it now solves slightly harder problems most of us don't have". And if you look at arxiv papers, a lot are peeking in that direction.
I'd also expect work on efficiency. "can we not make it cost about the amount of India's budget to move the industry forward, every year" is kind of a nice idea. No, I'm not making the number up, or taking OAIs fantasy numbers - 2025 AI industry capex is expected to be $375B. We'll likely need that efficiency if we want to get significantly better at difficulty level or task length, too.