I Build a Tools to Calculate How Much Vram Is Needed to Run Llms
Posted4 months agoActive4 months ago
kolosal.aiTechstory
supportivepositive
Debate
10/100
LlmsVramAI Tools
Key topics
Llms
Vram
AI Tools
The author created a tool to calculate the required VRAM for running Large Language Models (LLMs), sparking interest and appreciation from the community for simplifying a complex task.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
N/A
Peak period
3
0-1h
Avg / period
3
Key moments
- 01Story posted
Sep 19, 2025 at 10:25 PM EDT
4 months ago
Step 01 - 02First comment
Sep 19, 2025 at 10:25 PM EDT
0s after posting
Step 02 - 03Peak activity
3 comments in 0-1h
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 19, 2025 at 11:17 PM EDT
4 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45309640Type: storyLast synced: 11/17/2025, 1:02:45 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
You just paste the direct download URL of a GGUF model (for example, from Hugging Face), enter the context length you plan to use, and it will give you an approximate memory requirement.
It’s especially useful if you're trying to figure out whether a model will fit in your available VRAM or RAM, or when comparing different quantization levels like Q4_K_M vs Q8_0.
The tool is completely free and open-source. You can try it here: https://www.kolosal.ai/memory-calculator
And check out the code on GitHub: https://github.com/KolosalAI/model-memory-calculator
I'd really appreciate any feedback, suggestions, or bug reports if you decide to give it a try.