logoalt Hacker News

sailingparrottoday at 1:36 AM1 replyview on HN

> Nvidia has been using its newfound liquid funds to train its own family of models

Nvidia has always had its own family of models, it's nothing new and not something you should read too much into IMHO. They use those as template other people can leverage and they are of course optimized for Nvidia hardware.

Nvidia has been training models in the Megatron family as well as many others since at least 2019 which was used as blueprint by many players. [1]

[1] https://arxiv.org/abs/1909.08053


Replies

breputtoday at 1:57 AM

Nemotron-3-Nano-30B-A3B[0][1] is a very impressive local model. It is good with tool calling and works great with llama.cpp/Visual Studio Code/Roo Code for local development.

It doesn't get a ton of attention on /r/LocalLLaMA but it is worth trying out, even if you have a relatively modest machine.

[0] https://huggingface.co/nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B...

[1] https://huggingface.co/unsloth/Nemotron-3-Nano-30B-A3B-GGUF

show 3 replies