Apple Adds Matmul Acceleration to A19 Pro GPU
Posted4 months agoActive4 months ago
Techstory
calmpositive
Debate
30/100
Apple HardwareAI AccelerationLlms
Key topics
Apple Hardware
AI Acceleration
Llms
It's equivalent to Tensor cores from Nvidia. This means prompt processing will get a significant boost on Apple devices, making local LLMs far more viable on future Macs (as soon as M5).
Apple has added matmul acceleration to the A19 Pro GPU, potentially boosting local LLM performance on future Macs, with commenters discussing the implications for AI processing and hardware choices.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
2h
Peak period
2
12-15h
Avg / period
1.2
Key moments
- 01Story posted
Sep 9, 2025 at 1:50 PM EDT
4 months ago
Step 01 - 02First comment
Sep 9, 2025 at 4:12 PM EDT
2h after posting
Step 02 - 03Peak activity
2 comments in 12-15h
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 11, 2025 at 1:05 PM EDT
4 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45185637Type: storyLast synced: 11/20/2025, 7:50:26 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
If you want to run on a Mac Pro or iMac, this will be fine, but at the price points, youd be silly to spend money on either when you can do a dual nvidias for the same ram, and that will be dedicated ram.
For portable Apple devices, the max memory you can get currently is 24GB IIRC and thats probably not going to change any time soon. The only decent model that can run locally is Gemma 27B QAT which will eat up 17gb at the minimum, and that model really struggles with some stuff that you can do for free on ChatGPT or Gemini
So yeah, speed is not gonna matter when results are shit.
It isn’t cheap, but you can buy a 16 inch MacBook Pro with 128GB unified memory today.
The mac ultra has up 512gb and while expensive has more than twice the memory than any GPU alternative on a similar price point.
What was dragging it behind was the lack of Matmul acceleration, which seems that will change soon. Likely nvidia cards will still be faster and have better support, but paying a very big premium for it (ironic that apple is the cheaper option for once)
https://news.ycombinator.com/newsfaq.html
Useful /r/LocalLlama discussion: https://www.reddit.com/r/LocalLLaMA/comments/1ncprrq/apple_a...