Antml: Anthropic’s Markup Language
Posted2 months agoActive2 months ago
karashiiro.leaflet.pubTechstory
calmmixed
Debate
60/100
Artificial IntelligenceMarkup LanguageXMLLarge Language Models
Key topics
Artificial Intelligence
Markup Language
XML
Large Language Models
The post introduces ANTML, a markup language developed by Anthropic, and the discussion revolves around its novelty, effectiveness, and implications for AI model training and behavior.
Snapshot generated from the HN discussion
Discussion Activity
Moderate engagementFirst comment
1h
Peak period
6
1-2h
Avg / period
2.3
Comment distribution9 data points
Loading chart...
Based on 9 loaded comments
Key moments
- 01Story posted
Oct 31, 2025 at 12:37 AM EDT
2 months ago
Step 01 - 02First comment
Oct 31, 2025 at 1:50 AM EDT
1h after posting
Step 02 - 03Peak activity
6 comments in 1-2h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 31, 2025 at 7:28 AM EDT
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45768482Type: storyLast synced: 11/20/2025, 12:47:39 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
The model runtime recognizes these as special tokens. It can be configured using a chat template to replace these token with something else. This is how one provider is modifying the xml namespace, while llama.cpp and vllm would move the content between <think> and </think> tags to a separate field in the response JSON called `reasoning_content`.
The model can't introspect itself to know this, but the latest hybrid thinking models do think noticeably differently when prompted the right way between these tags, suddenly shifting to an extremely verbose mode of output that looks more like raw reasoning traces the earliest versions of reasoning models were showing
I have a prompt that used the original <antThinking> tag (it wasn't namespaced with 3.5). Every release eroded how steerable that output was and now 4 and up start spewing massive amounts of tokens in an extremely freeform manner that's also more effective
nano-banana/Gemini Flash 2.5 Image similarly does something like this: when prompted to think carefully will generally write a paragraph or two, but once coaxed a bit harder will suddenly output massive sprawling markdown formatted planning like the original un-masked thinking traces from Gemini Thinking.
It's like there's a sort of "mode collapse" for reasoning in these models that's going to continue getting more noticeable as providers lean harder into RL, which is ironic given how much ceremony there is around hiding reasoning traces in these APIs to avoid distillation
6 more comments available on Hacker News