Right-sizes LLM models to your system's RAM, CPU, and GPU
bilsbie
76 points
18 comments
March 01, 2026
Related Discussions
Found 5 related stories in 48.1ms across 3,471 title embeddings via pgvector HNSW
- Llm9p: LLM as a Plan 9 file system mleroy · 15 pts · March 08, 2026 · 52% similar
- How I write software with LLMs indigodaddy · 69 pts · March 16, 2026 · 51% similar
- Show HN: Run 500B+ Parameter LLMs Locally on a Mac Mini fatihturker · 11 pts · March 09, 2026 · 50% similar
- I don't use LLMs for programming ms7892 · 68 pts · March 12, 2026 · 48% similar
- Google's TurboQuant AI-compression algorithm can reduce LLM memory usage by 6x gmays · 16 pts · March 27, 2026 · 47% similar
Discussion Highlights (10 comments)
kamranjon
This is a great idea, but the models seem pretty outdated - it's recommending things like qwen 2.5 and starcoder 2 as perfect matches for my m4 macbook pro with 128gb of memory.
fwipsy
Personally I would have found a website where you enter your hardware specs more useful.
castral
I wish there was more support for AMD GPUs on Intel macs. I saw some people on github getting llama.cpp working with it, would it be addable in the future if they make the backend support it?
andsoitis
Claude is pretty good at among recommendations if you input your system specs.
dotancohen
In the screenshots, each model has a use case of General, Chat, or Coding. What might be the difference between General and Chat?
sneilan1
This is exactly what I needed. I've been thinking about making this tool. For running and experimenting with local models this is invaluable.
est
Why do I need to download & run to checkout? Can I just submit my gear spec in some dropdowns to find out?
esafak
I think you could make a Github Page out of this.
manmal
Slightly tangential, I‘m testdriving an MLX Q4 variant of Qwen3.5 32B (MoE 3B), and it’s surprisingly capable. It’s not Opus ofc. I‘m using it for image labeling (food ingredients) and I‘m continuously blown away how well it does. Quite fast, too, and parallelizable with vLLM. That’s on an M2 Max Studio with just 32GB. I got this machine refurbed (though it turned out totally new) for €1k.
BloondAndDoom
This pretty cool, and useful but I only wish this was a website. I don’t like the idea of running an executable for something that can perfectly be done as a website. (Other than some minor features, tbh even you can enable Corsair and still check the installed models from a web browser). Sounds like a fun personal project though.