logoalt Hacker News

nltoday at 2:03 AM0 repliesview on HN

> Instead of running the model once (flash) or multiple times (thinking/pro) in its entirety

I'm not sure what you mean here, but there isn't a difference in the number of times a model runs during inference.