Qwen is still better that Gemma though. Also you can tune it more for different tasks, which means that you can prioritize thinking and accuracy versus inference speed.
Genuine question: how do you tune it?
I thought "fine-tuning" meant training it on additional data to add additional facts / knowledge? I might be mistaking your use of the word "tune", though :)
It’s a heck of a lot faster too.
Yes I would just go with qwen.
Qwen is better at some things (code, in particular), but Gemma has better prose and better vision. At least, it feels that way to me.