this post was submitted on 25 Aug 2025
172 points (98.3% liked)

Fuck AI

3818 readers
584 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

founded 1 year ago
MODERATORS
 

cross-posted from: https://programming.dev/post/36289727

Comments

Our findings reveal a robustness gap for LLMs in medical reasoning, demonstrating that evaluating these systems requires looking beyond standard accuracy metrics to assess their true reasoning capabilities.6 When forced to reason beyond familiar answer patterns, all models demonstrate declines in accuracy, challenging claims of artificial intelligence’s readiness for autonomous clinical deployment.

A system dropping from 80% to 42% accuracy when confronted with a pattern disruption would be unreliable in clinical settings, where novel presentations are common. The results suggest that these systems are more brittle than their benchmark scores suggest.

you are viewing a single comment's thread
view the rest of the comments
[–] MTK@lemmy.world 5 points 13 hours ago

I think that is their point... That LLMs are being rushed into the medical field when they are mostly making statistical predictions for medical answers and not actually reasoning.