logoalt Hacker News

transformi12/09/20244 repliesview on HN

Not impressive compare to the opensource video models out there, I anticipated some physics/VR capabilities, but it's basically just a marketing promotion to "stay in the game"...


Replies

bbor12/09/2024

I... can you explain, or point to some competitors...? To me this looks leagues ahead of everything else. But maybe I'm behind the game?

AFAIK based on HuggingFace trending[1], the competitors are:

- bytedance/animatediff-lightning: https://arxiv.org/pdf/2403.12706 (2.7M downloads in the past 30d, released in March)

- genmo/mochi-1-preview: https://github-production-user-asset-6210df.s3.amazonaws.com... (21k downloads, released in October)

- thudm/cogvideox-5b: https://huggingface.co/THUDM/CogVideoX-5b (128k downloads, released in August)

Is there a better place to go? I'm very much not plugged into this part of LLMs, partially because it's just so damn spooky...

EDIT: I now see the reply above referencing Hunyuan, which I didn't even know was its own model. Fair enough! I guess, like always, we'll just need to wait for release so people can run their own human-preference tests to definitively say which is better. Hunyuan does indeed seem good

Geee12/09/2024

What's the best open source video model right now?

show 2 replies
zeknife12/09/2024

Like with music generation models, the main thing that might make "open source" models better is most likely that they have no concern about excluding copyrighted material from the training data, so they actually get a good starting point instead of using a dataset consisting of youtube videos and stock footage

bailvgu12/14/2024

[dead]