It's funny because some time ago (months? years?) people would say that you just didn't prompt the LLM well enough. But now LLMs are better and prompting isn't as arcane as before, so the next frontier is giving them the proper context. See this HN thread currently in the front page
You also have to be using the exact right model to get reasonable results, which is always the one you have to pay for, not the free one, and also not the one you were using