There seems to be a mass delusion about how capable SOTA models actually are. That's my only explanation for how poorly I find them performing in basic knowledge tasks compared to how others describe their prowess.
I understand you to be implying that I shouldn't trust my perception that there's a meaningful difference in how much different models hallucinate. I will take that under advisement, but I am still interested in the answer to my original question.
I understand you to be implying that I shouldn't trust my perception that there's a meaningful difference in how much different models hallucinate. I will take that under advisement, but I am still interested in the answer to my original question.