logoalt Hacker News

captainblandtoday at 5:05 PM9 repliesview on HN

I feel like this is a feature which improves the perceived confidence of the LLM but doesn't do much for correctness of other outputs, i.e. an exacerbation of the "confidently incorrect" criticism.


Replies

kemayotoday at 6:35 PM

It's a mismatch with our intuition about how much effort things take.

If there's humans involved, "I took this data and made a really fancy interactive chart" means that you put a lot more work into it, and you can probably somewhat assume that this means some more effort was also put into the accuracy of the data.

But with the LLM it's not really very much more work to get the fancy chart. So the thing that was a signifier of effort is now misleading us into trusting data that got no extra effort.

(Humans have been exploiting this tendency to trust fancy graphics forever, of course.)

show 1 reply
outloretoday at 8:02 PM

I suspect chain of thought while building the chart will improve the overall correctness of the answer

elliotbnvltoday at 5:56 PM

It's a usability / quality of life feature to me. Nothing to do with increasing perceived confidence. I guess it depends on how much you already (dis)trust LLMs.

I'm finding more and more often the limiting factor isn't the LLM, it's my intuition. This goes a way towards helping with that.

programmertotetoday at 5:59 PM

A recent LinkedIn post that I came across as an example of people trusting (or learning to trust) AI too much while not realizing that it can make up numbers too: https://www.linkedin.com/posts/mariamartin1728_claude-wrote-...

P.S. Credit to the poster, she posted a correction note when someone caught the issue: https://www.linkedin.com/posts/mariamartin1728_correction-on...

show 1 reply
vunderbatoday at 5:58 PM

A similar thing happened when Google started really pushing generating flowcharts as a use-case with Nano Banana. A slick presentation can distract people from the only thing that really matters - the accuracy of the underlying data.

show 1 reply
nerdjontoday at 5:46 PM

This was my first thought as well, all this does is further remove the user from seeing the chat output and instead makes it appear as if the information is concretely reliable.

I mean is it really that shocking that you can have an LLM generate structured data and shove that into a visualizer? The concern is if is reliable, which we know it isnt.

show 2 replies
mikkupikkutoday at 6:05 PM

I agree. Maybe next they'll add emotionally evocative music, with swelling orchestral bits when you reach the exciting climate of the slop.