logoalt Hacker News

Dynamic Large Concept Models: Latent Reasoning in an Adaptive Semantic Space

53 pointsby gmaysyesterday at 4:31 PM4 commentsview on HN

Comments

mivenyesterday at 7:01 PM

I'm really glad that these HNet-inspired approaches are getting traction, I'm a big fan of that paper.

Though I wonder how much of the gains in this case are actually due to 75% extra parameters compared to the baseline, even if the inference FLOPs are matched.

Can't help but see this as a just different twist on parameter use sparsity idea leveraged by MoE models, as those also gain in performance at constant forward pass FLOPs because of extra parameters.

sorenjanyesterday at 5:58 PM

Would this enable a model to learn concepts in one language and generate answers about it in another, as long as it learns general translations between them?

show 1 reply
notrealyme123yesterday at 6:07 PM

Broken citations. My inner reviewer gets sad. :(