Besides the political slant of Grokipedia, it's true that a lot of work that needed to be crowdsourced can be now packaged as work for LLMs. We all know the disadvantages of using LLMs, so let me mention some of the advantages: much higher speed, much more impervious to groupthink, cliques, and organised campaigns; truly ego-less editing and debating between "editors". Grokipedia is not viable because of Musk's derangement, but other projects, more open and publicly auditable, might come along.
"higher speed" isn't an advantage for an encyclopedia.
The fact that Musk's derangement is clear from reading grokipedia articles shows that LLMs are less impervious to ego. Combine easily ego driven writing with "higher speed" and all you get is even worse debates.
LLMs are only impervious to "groupthink" and "organized campaigns" and other biases if the people implementing them are also impervious to them, or at least doing their best to address them. This includes all the data being used and the methods they use to process it.
You rightfully point out that the Grok folks are not engaged in that effort to avoid bias but we should hold every one of these projects to a similar standard and not just assume that due diligence was made.
> much more impervious to groupthink
Citation very much needed. LLMs are arguably concentrated groupthink (albeit a different type than wiki editors - although I'm sure they are trained on that), and are incredibly prone to sycophancy.
Establishing fact is hard enough with humans in the loop. Frankly, my counterargument is that we should be incredibly careful about how we use AI in sources of truth. We don't want articles written faster, we want them written better. I'm not sure AI is up to that task.
> impervious to groupthink, cliques, and organised campaigns
Yeeeeah, no. LLMs are only as good as the datasets they are trained on (ie the internet, with all its "personality"). We also know the output is highly influenced by the prompting, which is a human-determined parameter, and this seems unlikely to change any time soon.
This idea that the potential of AI/LLMs is somehow not fairly represented by how they're currently used is ludicrous to me. There is no utopia in which their behaviour is somehow magically separated from the source of their datasets. While society continues to elevate and amplify the likes of Musk, the AI will simply reflect this, and no version of LLM-pedia will be a truly viable alternative to Wikipedia.
> much more impervious to groupthink
Can you explain what you mean by this? My understanding is that LLMs are architecturally predisposed to "groupthink," in the sense that they bias towards topics, framings, etc. that are represented more prominently in their training data. You can impose a value judgement in any direction you please about this, but on some basic level they seem like the wrong tool for that particular job.