logoalt Hacker News

TZubiritoday at 7:10 AM1 replyview on HN

have you tried Llama? In my experience it has been strictly better than GPT OSS, but it might depend on specifically how it is used.


Replies

embedding-shapetoday at 8:20 AM

Have you tried GPT-OSS-120b MXFP4 with reasoning effort set to high? Out of all models I can run within 96GB, it seems to consistently give better results. What exact llama model (+ quant I suppose) is it that you've had better results against, and what did you compare it against, the 120b or 20b variant?

show 1 reply