So, what exactly is the point of the LLM if it can't exceed an average person and produces results which are not trustworthy?
I think it's necessary to remember that they're not a general artificial intelligence, but language models. For example, they're pretty good (not perfect) at translating things, including translating arbitrary instructions into code or machine-readable forms.
"The average person" has a job. Those jobs can now be performed by machine. The societal implications are profound.