Not
Hacker
News
!
Home
Hiring
Products
Discussion
Q&A
Users
LLM Optimization | Trending Topic on Hacker News | Not Hacker News!
Not
Hacker
News
!
Home
Hiring
Products
Discussion
Q&A
Users
Home
/
Discussion
/
LLM Optimization
Back to Discussion
LLM Optimization
Loading...
19 stories
•
24h:
0%
•
7d: 0
•
298 comments
Top contributors:
tdchaitanya
blndrt
anuarsh
JnBrymn
che_shr_cat
Stories
Related Stories
19 stories tagged with llm optimization
Adaptive LLM Routing Under Budget Constraints
206
78 comments
by tdchaitanya
Posted
4 months ago
Active
about 2 months ago
LLM optimization
cost reduction
AI research
Tau² Benchmark: How a Prompt Rewrite Boosted GPT-5-Mini by 22%
197
65 comments
by blndrt
Posted
4 months ago
Active
about 2 months ago
LLM optimization
prompt engineering
AI benchmarking
Run Qwen3-Next-80b on 8gb GPU at 1tok/2s Throughput
123
17 comments
by anuarsh
Posted
4 months ago
Active
about 2 months ago
LLM optimization
GPU limitations
AI model deployment
Don't Build Multi-Agents
123
89 comments
by JnBrymn
Posted
4 months ago
Active
about 2 months ago
AI development
multi-agent systems
LLM optimization
Deepconf: Scaling LLM Reasoning with Confidence, Not Just Compute
98
35 comments
by che_shr_cat
Posted
5 months ago
LLM optimization
AI research
compute efficiency
Chunkllm: a Lightweight Pluggable Framework for Accelerating Llms Inference
96
8 comments
by PaulHoule
Posted
3 months ago
Active
about 2 months ago
LLM Optimization
AI Inference
Machine Learning Frameworks
Efficient Llm:bandwidth, Compute, Synchronization, and Capacity Are All You Need
6
0 comments
by matt_d
Posted
3 months ago
Active
about 1 month ago
LLM optimization
AI research
Machine Learning
The Alien Artifact: Dspy and the Cargo Cult of LLM Optimization
3
0 comments
by valgaze
Posted
3 months ago
Active
about 2 months ago
LLM Optimization
DSPy
AI Development
LLM-Use – an LLM Router That Chooses the Right Model for Each Prompt
3
2 comments
by justvugg
Posted
3 months ago
Active
about 2 months ago
LLM optimization
cost reduction
AI routing
Cutting LLM Batch Inference Time by Half with Dynamic Prefix Bucketing
2
0 comments
by DISCURSIVE
Posted
about 2 months ago
Active
about 2 months ago
LLM optimization
batch inference
dynamic prefix bucketing
Kv Marketplace – Share LLM Attention Caches Across Gpus Like Memcached
2
1 comments
by nsomani
Posted
about 2 months ago
Active
about 2 months ago
LLM optimization
GPU computing
distributed systems
Un-Locc Reduce LLM API Costs by Compressing Text Into Images
2
0 comments
by MaxDevv
Posted
2 months ago
Active
about 2 months ago
LLM optimization
text compression
AI cost reduction
Prompt Refiner – Lightweight Python Lib to Clean and Compress LLM Input
1
1 comments
by xinghaohuang
Posted
about 1 month ago
Active
about 1 month ago
LLM optimization
text preprocessing
Python library
Fixing LLM Memory Degradation in Long Coding Sessions
1
0 comments
by robertomisuraca
Posted
about 2 months ago
Active
about 1 month ago
LLM optimization
development_tools
ai
Dspy on a Pi: Cheap Prompt Optimization with Gepa and Qwen3
1
0 comments
by lsb
Posted
about 2 months ago
Active
about 2 months ago
DSPy
Raspberry Pi
LLM optimization
A Self-Tuning Open-Source Agent for LLM KPI Optimization
1
0 comments
by dan-carp-builds
Posted
3 months ago
Active
about 2 months ago
LLM optimization
open-source
AI agents
LLM Optimization Notes: Memory, Compute and Inference Techniques
1
0 comments
by gmays
Posted
3 months ago
Active
about 2 months ago
LLM Optimization
Distributed Machine Learning
AI Inference Techniques
Modular, LLM-Optimized Openapi Docs – Deterministic Urls
1
0 comments
by saransh_01
Posted
4 months ago
Active
about 2 months ago
API documentation
LLM optimization
software development
My Friend Says He Has an AI Optimized Language
1
2 comments
by samweb3
Posted
4 months ago
Active
about 2 months ago
artificial intelligence
query languages
LLM optimization