logoalt Hacker News

josefritzishereyesterday at 3:56 PM3 repliesview on HN

It continues to amaze me how recklessly some people cram AI into spaces where it performs poorly and the consequences include death.


Replies

y-c-o-m-byesterday at 6:20 PM

As a software dev that uses it and observes the many errors it makes on a daily basis, I definitely treat the output with a much greater deal of skepticism than the average person I speak with. If you're used to it providing relatively accurate results based on surface level google-eqsue searches, then it makes sense why you'd place a higher weight on it being an "expert" vs a "tool that needs verification". I understand why people fall into this mindset.

I used ChatGPT to do a valve adjustment on an engine; a task I've never done before. I didn't just accept the torque values and procedure it told me though, because I know better from my experience with it as a dev. I cross-referenced it all with Youtube videos, forum posts, instruction manuals (where available) to make sure the job was A) doable for a non-mechanic like me and B) done correctly. Thanks to the Youtube video (which I cross-referenced with other sources), I discovered the valve clearance values were slightly off with the ChatGPT recommendation.

I think the average Joe would assume these values were correct and run with it.

rectangyesterday at 4:04 PM

If the AI gets attached to a health insurer (not the case here as far as I know), I would expect it to make decisions that are aligned with the company’s incentive to weed out unprofitable patients. AI is not a human who takes a Hippocratic oath; it can be more easily manipulated to perform unethical acts.

show 2 replies
TZubiriyesterday at 4:22 PM

But it doesn't perform poorly actually, it's just that the stakes are very high and it's a highly regulated environment.

Most physicians I know use ChatGPT. Although of course it's usage guided by an expert, not by the patient, nor fully autonomous.