logoalt Hacker News

Trinity large: An open 400B sparse MoE model

68 pointsby linolevantoday at 12:57 AM22 commentsview on HN

Comments

myntitoday at 8:10 AM

They trained it in 33 days for ~20m (that includes apparently not only the infrastructure but also the salaries over a 6 month period). And the model is coming close to QWEN and Deepseek. Pretty impressive

show 1 reply
linolevantoday at 1:12 AM

I'm particularly excited to see a "true base" model to do research off of (https://huggingface.co/arcee-ai/Trinity-Large-TrueBase).

mwcampbelltoday at 9:54 PM

Given that it's a 400B-parameter model, but it's a sparse MoE model with 13B active parameters per token, would it run well on an NVIDIA DGX Spark with 128 GB of unified RAM, or do you practically need to hold the full model in RAM even with sparse MoE?

show 2 replies
Alifatisktoday at 10:59 PM

What did they do to make the loss drop so much in phase 3?

Also, why are they comparing with Llama 4 Maverick? Wasn’t it a flop?

show 2 replies
syntaxingtoday at 11:14 PM

So refreshing to see open source models like this come from the US. I would love for a 100Bish size one that can compete against OSS-120B and GLM air 4.5

frogpersontoday at 9:51 PM

What exactly does "open" mean in this case? Is it weights and data or just weights?

show 1 reply
gregghtoday at 9:54 PM

The only thing I question is the use of Maverick in their comparison charts. That's like comparing a pile of rocks to an LLM.

0xdeadbeefbabetoday at 11:32 PM

Is anyone excited to do ablative testing on it?

observationisttoday at 5:15 PM

This is a wonderful release.