Some 'misinformation' is hard to correct because the corrections are reversed by those who are intent on spreading the 'misinformation'. This is especially prevalent around contentious and/or politically sensitive subjects like the mentioned SARS2-related cases. This is what makes it hard to trust articles on such subjects on Wikipedia.
worse yet, you might read some topics and won't expect them to be poisoned with misinformation. Like the Holocaust history in Poland
https://digitalcommons.chapman.edu/history_news_articles/151... https://slate.com/technology/2023/04/how-wikipedia-covers-th...
If this is quite widespread, it should be fairly straightforward to point to an example of a page that's being defaced with misinformation, which would include an edit history and perhaps a Talk page documenting whatever sides to the debate there is that's preventing consensus.
I don't disagree that weird bullshit occasionally happens on Wikipedia, but I have noticed that as soon as light is cast on it, it usually evaporates and a return to factual normality is established.