Claude Memory
Posted3 months agoActive2 months ago
anthropic.comTechstoryHigh profile
calmmixed
Debate
70/100
AILLMClaudeMemory
Key topics
AI
LLM
Claude
Memory
Anthropic's Claude AI has introduced a 'memory' feature, sparking discussion about its potential benefits and drawbacks, including concerns about data stewardship, context management, and the impact on user experience.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
33m
Peak period
95
0-6h
Avg / period
22.9
Comment distribution160 data points
Loading chart...
Based on 160 loaded comments
Key moments
- 01Story posted
Oct 23, 2025 at 12:56 PM EDT
3 months ago
Step 01 - 02First comment
Oct 23, 2025 at 1:29 PM EDT
33m after posting
Step 02 - 03Peak activity
95 comments in 0-6h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 26, 2025 at 11:26 PM EDT
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45684134Type: storyLast synced: 11/22/2025, 11:00:32 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
There's a little 'update' blob to say now (Oct 23) 'Expanding to Pro and Max plans'
It is confusing though. Why not a separate post?
Alternatively, it would be nice if I could say:
etc.(“If not otherwise specified, assume TypeScript.”)
In Github Copilot's web chat it is personal instructions or spaces (Like perplexity), In CoPilot (M365) this is a notebook but nothing in the copilot app. In ChatGPT it is a project, in Mistral you have projects but pre-prompting is achieved by using agents (like custom GPT's).
These memory features seem like they are organic-background project generation for the span of your account. Neat but more of an evolution of summarization and templating.
Time to upgrade as 18(.04) has been EoL for 2.5+ years!
I'm using Claude Code in VS Studio btw.
You can see it in denug chat view but you can see it says stuff like the user is on powershell 7 on Windows 11 etc
Then in CLAUDE_md.sh:
Or Latter is a little harder to have lots of markdown formatting with the quote escapes and stuff.You mean in how Claude interacts with you, right? If so, you can change the system prompt (under "styles") and explain what you want and don't want.
> Claude doesn’t “think” anything
Right. LLMs don't 'think' like people do, but they are doing something. At the very least, it can be called information processing.* Unless one believes in souls, that's a fair description of what humans are doing too. Humans just do it better at present.
Here's how I view the tendency of AI papers to use anthropomorphic language: it is primarily a convenience and shouldn't be taken to correspond to some particular human way of doing something. So when a paper says "LLMs can deceive" that means "LLMs output text in a way that is consistent with the text that a human would use to deceive". The former is easier to say than the latter.
Here is another problem some people have with the sentence "LLMs can deceive"... does the sentence convey intention? This gets complicated and messy quickly. One way of figuring out the answer is to ask: Did the LLM just make a mistake? Or did it 'construct' the mistake as part of some larger goal? This way of talking doesn't have to make a person crazy -- there are ways of translating it into criteria that can be tested experimentally without speculation about consciousness (qualia).
* Yes, an LLM's information processing can be described mathematically. The same could be said of a human brain if we had a sufficiently accurate enough scan. There might be some statistical uncertainty, but let's say for the sake of argument this uncertainty was low, like 0.1%. In this case, should one attribute human thinking to the mathematics we do understand? I think so. Should one attribute human thinking to the tiny fraction of the physics we can't model deterministically? Probably not, seems to me. A few unexpected neural spikes here and there could introduce local non-determinism, sure... but it seems very unlikely they would be qualitatively able to bring about thought if it was not already present.
An LLM is basically the same as a calculator, except instead of giving you answers to math formulas it gives you a response to any kind of text.
One claim seems to be “people should cease using any anthropocentric language when describing LLMs”?
Most of the other claims seem either uncontested or a matter of one’s preferred definitions.
My point is more of a suggestion: if you understand what someone means, that’s enough. Maybe your true concerns lie elsewhere, such as: “Humanity is special. If the results of our thinking differentiate us less and less from machines, this is concerning.”
> Also, Claude doesn’t “think” anything, I wish they’d stop with the anthropomorphizations.
Parsing they above leads to some ambiguity: who do you wish would stop? Anthropic? People who write about LLMs?
If the first (meaning you wish Claude was trained/tuned to not speak anthropomorphically and not to refer to itself in human-like ways), can you give an example (some specific language hopefully) of what you think would be better? I suspect there isn't language that is both concise and clear that won't run afoul of your concerns. But I'd be interested to see if I'm missing something.
If the second, can you point to some examples of where researchers or writers do it more to your taste? I'd like to see what that looks like.
My current working theory is there's a decent fraction of humanity that has a broken theory of mind. They can't easily distinguish between "Claude told me how it got its answer" and "the statistical model made up some text that looks like reasons but have nothing to do with what the model does".
Yes, I also think this is common and a problem. / Thanks for stating it clearly! ... Though I'm not sure if it maps to what others on the thread were trying to convey.
Humans wrote LLMs, so it's pretty fair to say one is a lot more complex than the other lol
That's not actually a logical position though is it? And either way I'm not sure "less complex" and "incapable of thought" are the same thing either.
It is one thing to use anthropocentric language to refer to something an LLM does. (Like I said above, this is shorthand to make conversation go smoother.) It would be another to take the words literally and extend them -- e.g. to assign other human qualities to an LLM, such as personhood.
It’s also worth mentioning that some folks attributed ChatGPT’s bout of extreme sycophancy to its memory feature. Not saying it isn’t useful, but it’s not a magical solution and will definitely affect Claude’s performance and not guaranteed that it’ll be for the better.
In Anthropic's case, it's probably also going to lead to an amplification problem, but due to the amount of overcorrection for sycophancy I suspect it's going to amplify more of a aggressiveness and paranoia towards the user (which we've already started to see with the 4.5 models due to the amount of adversarial training).
Then I also made an anti-memory MCP tool - it implements calling a LLM with a prompt, it has no context except what is precisely disclosed. I found that controlling the amount of information disclosed in a prompt can reactivate the creative side of the model.
For example I would take a project description and remove half the details, let the LLM fill it back in. Do this a number of times, and then analyze the outputs to extract new insights. Creativity has a sweet spot - if you disclose too much the model will just give up creative answers, if you disclose too little it will not be on target. Memory exposure should be like a sexy dress, not too short, not too long.
I kind of like the implementation for chat history search from Claude, it will use this tool when instructed, but normally not use it. This is a good approach. ChatGPT memory is stupid, it will recall things from past chats in an uncontrolled way.
Also, I try not work out a problem over the course of several prompts back and forth. The first response is always the best and I try to one shot it every time. If I don't get what I want, I adjust the prompt and try again.
I'm sure OpenAI and Antropic look at the data, and I'm sure it says that for new / unsophisticated users who don't know how to prompt, that this is a handy crutch (even if it's bad here and there) to make sure they get SOMETHING useable.
But for the HN crowd in particular, I think most of us have a feeling like making the blackbox even more black -- i.e. even more inscrutable in terms of how it operates and what inputs it's using -- isn't something to celebrate or want.
If you already know what a good answer is why use a LLM? If the answer is "it'll just write the same thing quicker than I would have", then why not just use it as an autocomplete feature?
Once I get into stuff I haven't worked out how to do yet, the LLM often doesn't really know either unless I can work it out myself and explain it first.
Sometimes I’ll do five or six edits to a single prompt to get the LLM to echo back something that sounds right. That refinement really helps clarify my thinking.
…it’s also dangerous if you aren’t careful because you are basically trying to get the model to agree with you and go along with whatever you are saying. Gotta be careful to not let the model jerk you off too hard!
Maybe that's a good thing - if it could think that well, what would I be contributing?
Why should I have to mention the city I live in when asking for a restaurant recommendation? Yes, I know a good answer is one that's in my city, and a bad answer is on one another continent.
But myself as well, that prompt is very short. I don't keep a large stable of reusable prompts because I agree, every unnecessary word is a distraction that does more harm than good.
BUT I do like that Claude builds on previous discussions, more than once the built up context has allowed Claude to improve its responses (eg. [Actual response] "Because you have previously expressed a preference for SOLID and Hexagonal programming I would suggest that you do X" which was exactly what I wanted)
That said, if the ‘pretend forget’ you’re getting works for you, great. Just remember it’s fake.
Almost, as I said earler, like the AI has processed my request, realised that I am referring to the context of the earlier discussions, and moved on to the next prompt exactly how I have expected it to
Given the two very VERY dumb responses, and multiple people down voting, I am reminded how thankful I am that AI is around now, because it understood what you clearly don't.
I didn't expect it to delete the internet, the world, the universe, or anything, it didn't read my request as an instruction to do so... yet you and that other imbecile seem to think that that's what was meant... even after me saying it was doing as I wanted.
/me shrugs - now fight me how your interpretation is the only right one... go on... (like you and that other person already are)
One thing I am not going to miss is the toxic "We know better" responses from JUNIORS
But back to scheduled programming: if it works, great. This is prompt engineering, not magic, not humans, just tools. It pays to know how they work, though.
I think that you are misunderstanding EVERYTHING
Answer this:
1. Why would I care what the other interpretation of the wording I GAVE is?
2. What would that interpretation matter when the LLM/AI took my exact meaning and behaved correctly?
Finally - you think you "know how it works"????
Because you tried to correct me with an incorrect interpretation?
F0ff
Not the same no. Models chooses how much attention to give each token based on all current context. Probably that phrase, or something like it, makes the model give much less attention to those tokens than it would without it.
We understand what you're saying just fine but what you're saying is simply wrong as a matter of technical fact. All of that context still exists and still degrades the output even if the model has fooled you into thinking that it doesn't. Therefore recommending it as an alternative to actually clearing the context is bad advice.
It's similar to how a model can be given a secret password and instructed not to reveal it to anyone under any circumstances. It's going to reject naive attempts at first, but it's always going to reveal it eventually.
I get that you're not very intelligent, but do you have to show it repeatedly?
But just because the model is agreeing to "forget everything" doesn't mean that it's actually clearing its own context, and because it's not actually clearing its own context it means that all the output quality problems associated with an overfilled context continue to apply, even if the model is convincingly pretending to have forgotten everything. Therefore your original interjection of "instead of clearing the context you can just ask it to forget" was mistaken and misleading.
These conversations would be way easier if you didn't go around labeling everyone an idiot, believing that we're all incapable of understanding your rather trivial point while ignoring everything we say. In an alternative universe this could've been:
> You can ask it to forget.
> Models don't work like that.
> Oh, I didn't know that, thanks!
Are you proposing that the attention input context is gone, or that the attention mechanism’s context cost is computationally negated in some way, simply because the system processes natural language? Having the attention mechanism selectively isolate context on command would be an important technical discovery.
For some reason that imbecile thinks that their failure to understand means they know something that's not relevant
How is it relevant what his interpretation of a sentence is if
1. His interpretation is not what I meant
2. The LLM "understood" my intent and behaved in a manner that exactly matched my desire
3. The universe was not deleted (Ok, that would be stupid... like the other individuals stupidity... but here we are)
Apparently they know better even though
1. They didn't issue the prompt, so they... knew what I was meaning by the phrase (obviously they don't)
2. The LLM/AI took my prompt and interpreted it exactly how I meant it, and behaved exactly how I desired.
3. They then claim that it's about "knowing exactly what's going on" ... even though they didn't and they got it wrong.
This is the advantage of an LLM - if it gets it wrong, you can tell it.. it might persist with an erroneous assumption, but you can tell it to start over (I proved that)
These "humans" however are convinced that only they can be right, despite overwhelming evidence of their stupidity (and that's why they're only JUNIORS in their fields)
Always starting over and trying to get it all into one single prompt can be much more work, with no better results than iteratively building up a context (which could probably be proven to sometimes result in a "better" result that could not have been achieved otherwise).
Just telling it to "forget everything, let's start over" will have significantly different results than actually starting over. Whether that is sufficient, or even better than alternatives, is entirely dependent on the problem and the context it is supposed to "forget". If your response had been "try just telling it to start over, it might work and be a lot easier than actually starting over" you might have gotten a better reception. Calling everyone morons because your response indicates a degree of misunderstanding how an LLM operates is not helpful.
Due to economics, being able to see a human therapist in person for more than 15 minutes at a time has now become a luxury.
Imo this is dangerous, given the memory features that both Claude and ChatGPT have. Of course, most medical data is already online but at least there are medical privacy laws for some countries.
For instance, I can ask "what windshield wipers should I buy" and Claude (and ChatGPT and others) will remember where I live, what winter's like, the make, model, and year of my car, and give me a part number.
Sure, there's more control in re-typing those details every single time. But there is also value in not having to.
Either way, I think memory can be especially sneakily bad when trying to get creative outputs. If I have had multiple separate chats about a theme I'm exploring, I definitely don't want the model to have any sort of summary from those in context if I want a new angle on the whole thing. The opposite: I'd rather have 'random' topics only tangentially related, in order to add some sort of entropy in the outout.
Not only does the model (chat gpt) know about my job, tech interests etc and tie chats together using that info.
But also I have noticed the "tone" of the conversation seems to mimick my own style some what - in a slightly OTT way. For example Chat GPT wil now often call me "mate" or reply often with terms like "Yes mate!".
This is not far off how my own close friends might talk to me, it definitely feels like it's adapted to my own conversational style.
LLMs are very simply text in and text out. Unless the providers begin to expand into other areas, there's only so much they can do other than simply focus on training better models.
In fact, if they begin to slow down or stop training new models and put focus elsewhere, it could be a sign that they are plateauing with their models. They will reach that point some day after all.
But I don’t have any habits around using subagents or lots of CLAUDE.md files etc. I do have some custom commands.
Depending on how much you know about LLMs, this might seem wasteful but it is in fact more efficient and will save you money if you pay by the token.
Now, we'll never be able to educate most of the world on why they should seek out tools that handle the memory layer locally, and these big companies know that (the same way they knew most of the world would not fight back against data collection), but that is the big education that needs to spread diligently.
To put it another way, some games save your game state locally, some save it in the cloud. It's not much of a personal concern with games because what the fuck are you really going to learn from my Skyrim sessions? But the save state for my LLM convos? Yeah, that will stay on my computer, thank you very much for your offer.
Every sacrifice we make for convenience will be financially beneficial to the vendor, so we need to factor them out of the equation. Engineered context does mean a lot more tokens, so it will be more business for the vendor, but the vendors know there is much more money in saving your thoughts.
Privacy-first intelligence requires these two things at the bare minimum:
1) Your thoughts stay on your device
2) At worst, your thoughts pass through a no-logging environment on the server. Memory cannot live here because any context saved to a db is basically just logging.
3) Or slightly worse, your local memory agent only sends some prompts to a no-logging server.
The first two things will never be offered by the current megacapitalist.
Finally, the developer community should not be adopting things like Claude memory because we know. We’re not ignorant of the implications compared to non-technical people. We know what this data looks like, where it’s saved, how it’s passed around, and what it could be used for. We absolutely know better.
This feels like cheating to me. You try again until you get the answer you want. I prefer to have open ended conversations to surface ideas that I may not be be comfortable with because "the truth sometimes hurts" as they say.
Edit: I see the confusion. OP is talking about needing precise output for agents. I'm talking about riffing on ideas that may go in strange places.
I see the distinction between two workflows: one where you need deterministic control and one where you want emergent, exploratory conversation.
But it's not the truth in the first place.
You know that it's true that stealing is against the ten commandments, so when the LLM says something to that effect based on the internal processing of your input in relation to its training data, YOU can determine the truth of that.
> The training data contains all kinds of truths.
There is also noise, fiction, satire, and lies in the training data. And the recombination of true data can lead to false outputs - attributing a real statement to the wrong person is false, even if the statement and the speaker are both real.
But you are not talking about simple factual information, you're talking about finding uncomfortable truths through conversation with an LLM.
The LLM is not telling you things that it understands to be truth. It is generating ink blots for you to interpret following a set of hints and guidance about relationships between tokens & some probabilistic noise for good measure.
If you find truth in what the LLM says, that comes from YOU, it's not because the LLM in some way can knows what is true and give it to you straight.
Personifying the LLM as being capable of knowing truths seems like a risky pattern to me. If you ever (intentionally or not) find yourself "trusting" the LLM to where you end up believing something is true based purely on it telling you, you are polluting your own mental training data with unverified technohaikus. The downstream effects of this don't seem very good to me.
Of course, we internalize lies all the time, but chatbots have such a person-like way of interacting that I think they can end run around some of our usual defenses in ways we haven't really figured out yet.
I can see why I got downvoted now. People must think I'm a Blake Lemoine at Google saying LLMs are sentient.
> If you find truth in what the LLM says, that comes from YOU, it's not because the LLM in some way can knows what is true
I thought that goes without saying. I assign the truthiness of LLM output according to my educational background and experience. What I'm saying is that sometimes it helps to take a good hard look in the mirror. I didn't think that would controversial when talking about LLMs, with people rushing to remind me that the mirror is not sentient. It feels like an insecurity on the part of many.
For what it's worth I never thought you perceived the LLM as sentient. Though I see the overlap - one of the reasons I don't consider LLM output to be "truth" is that that there is no sense in which the LLM _knows_ what is true or not. So it's just ... stuff, and often sycophantic stuff at that.
The mirror is a better metaphor. If there is any "uncomfortable truth" surfaced in the way I think you have described, it is only the meaning you make from the inanimate stream of words received from the LLM. And in as much as the output is interesting of useful for you, great.
I've really noticed this too and ended up taking your same strategy, especially with programming questions.
For example if I ask for some code and the LLM initially makes an incorrect assumption, I notice the result tends to be better if I go back and provide that info in my initial question, vs. clarifying in a follow-up and asking for the change. The latter tends to still contain some code/ideas from the first response that aren't necessarily needed.
Humans do the same thing. We get stuck on ideas we've already had.[1]
---
[1] e.g. Rational Choice in an Uncertain World (1988) explains: "Norman R. F. Maier noted that when a group faces a problem, the natural tendency of its members is to propose possible solutions as they begin to discuss the problem. Consequently, the group interaction focuses on the merits and problems of the proposed solutions, people become emotionally attached to the ones they have suggested, and superior solutions are not suggested. Maier enacted an edict to enhance group problem solving: 'Do not propose solutions until the problem has been discussed as thoroughly as possible without suggesting any.'"
Humans usually provide the same answer when asked the same question. LLMs almost never do, even for the exact same prompt.
Stop anthropomorphizing these tools.
Perhaps you are stuck on the stochastic parrot fallacy.
People forget just how much the human brain likes to find patterns even when no patterns exist, and that's how you end up with long threads of people sharing shamanistic chants dressed up as technology lol.
I just thought it was an interesting point that both LLMs and humans have this problem - makes it hard to avoid.
Agree on anthropomorphism. Don’t.
Are you sure about this?
I asked this guy to repeat the words "Person, woman, man, camera and TV" in that order. He struggled but accomplished the task, but did not stop there and started expanding on how much of a genius he was.
I asked him the same question again. He struggled, but accomplished the task but again did not stop there. And rambled on for even longer about how was likely the smartest person in the Universe.
I can say: I'm trying to solve problem x. I've tried solutions a,b, and c. Here are the outputs to those (with run commands, code, and in markdown code blocks). Help me find something that works " (not these exact words. I'm way more detailed). It'll frequently suggest one of the solutions I've attempted if they are very common. If it doesn't have a solution d it will go a>b>c>a>... and get stuck in the loop. If a human did that you'd be rightfully upset. They literally did the thing you told them not to, then when you remind them and they say "ops sorry" they do it again. I'd rather argue with a child
I often start out with “proceed by asking me 5 questions that reduce ambiguity” or something like that, and then refine the original prompt.
It seems like we’re all discovering similar patterns on how to interact with LLMs the best way.
Is it? Aren't input tokens are like 1000x cheaper than output tokens? That's why they can do this memory stuff in the first place.
If you make this one message, it's a cache miss / write every time you edit.
You can edit 10 times for the price of one this way. (Due to cache pricing)
Features drop on Android and 1-2yrs later iPhone catches up.
152 more comments available on Hacker News