logoalt Hacker News

storusyesterday at 2:17 PM1 replyview on HN

Huge local thinking LLMs to solve math and for general assistant-style tasks. Models like Kimi-2.5-Q3, DeepSeek-XX-Q4/Q5, Qwen-3.5-Q8, MiniMax-m2.5-Q8 etc. that bring me to Claude4/GPT5 territory without any cloud. For coding I have another machine with 3x RTX Pro 6000 (mostly Qwen subvariants) and for image/video/audio generation I have 2x DGX Sparks from ASUS.


Replies

ganoushoreillyyesterday at 3:33 PM

We must be twins, i've got the same three working in a cluster.

I was really excited to see where the GB300 Desktops end up, with 768gb ram but now that data is leaking / popping up (dell appears to only be 496gb), we may be in the 60-100k range and that's well out of my comfort zone.

If Apple came out with a 768gb Studio at 15k i'd bite in a heart beat.

https://www.dell.com/en-us/lp/dell-pro-max-nvidia-ai-dev

show 1 reply