this post was submitted on 10 Jul 2025
330 points (93.9% liked)

Technology

72894 readers
4027 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

A robot trained on videos of surgeries performed a lengthy phase of a gallbladder removal without human help. The robot operated for the first time on a lifelike patient, and during the operation, responded to and learned from voice commands from the team—like a novice surgeon working with a mentor.

The robot performed unflappably across trials and with the expertise of a skilled human surgeon, even during unexpected scenarios typical in real life medical emergencies.

top 50 comments
sorted by: hot top controversial new old
[–] cupcakezealot@piefed.blahaj.zone 15 points 6 days ago (4 children)

you could not pay me enough to have my surgery done by a robot

[–] _cryptagion@lemmy.dbzer0.com 4 points 6 days ago

yeah, it's much better to have a towel left inside of you by a real human.

[–] echodot@feddit.uk 3 points 6 days ago (1 children)

They obviously don't feel comfortable with the robot doing surgery on humans just yet either which is why they're not actually suggesting doing that yet. It will have to go through years and years of certification before that's even considered.

I'm sure most surgeries will still be conducted by humans but there are situations where one of these would be extremely helpful. Any situation where a surgeon isn't currently accessible and can't quickly get there. Remote communities, Disaster relief, Arctic research facilities, Starships trapped in the Delta quadrant, War zones, Ships at sea.

[–] Bluewing@lemmy.world 2 points 6 days ago (1 children)

Do you think a 5 bed hospital will have the money to afford a robotic surgeon?

[–] echodot@feddit.uk 1 points 5 days ago

You assume an Antarctic research facility lacks funds?

[–] otp@sh.itjust.works 4 points 6 days ago (1 children)

If it were the only option, I'd gladly take it.

I rely on robots to do a lot of other things in my life, directly and indirectly.

Well, not many directly. But machines, definitely.

[–] echodot@feddit.uk 1 points 6 days ago

Yeah it's not like I refuse to drive my car because it wasn't handcrafted by a human.

It is an electrical fault on four wheels, but that's just because it's old.

[–] Zacryon@feddit.org 0 points 6 days ago (1 children)

I trust a good machine much more than any human.

[–] qaz@lemmy.world 3 points 6 days ago (1 children)

Have you considered that the machine is made by a collection of humans?

[–] Zacryon@feddit.org 0 points 5 days ago

Yes. But if the machine has proven to work reliably it will usually do so for its lifetime, while humans are prone to e multitude of errors. Especially in the medical field.

[–] finitebanjo@lemmy.world 4 points 6 days ago* (last edited 6 days ago) (2 children)

See the part that I dont like is that this is a learning algorithm trained on videos of surgeries.

That's such a fucking stupid idea. Thats literally so much worse than letting surgeons use robot arms to do surgeries as your primary source of data and making fine tuned adjustments based on visual data in addition to other electromagnetic readings

[–] echodot@feddit.uk 7 points 6 days ago (1 children)

Yeah but the training set of videos is probably infinitely larger, and the thing about AI is that if the training set is too small they don't really work at all. Once you get above a certain data set size they start to become competent.

After all I assume the people doing this research have already considered that. I doubt they're reading your comment right now and slapping their foreheads and going damn this random guy on the internet is right, he's so much more intelligent than us scientists.

[–] finitebanjo@lemmy.world 0 points 5 days ago (1 children)

Theres no evidence they will ever reach quality output with infinite data, either. In that case, quality matters.

[–] echodot@feddit.uk 0 points 5 days ago* (last edited 5 days ago) (1 children)

No we don't know. We are not AI researchers after all. Nonetheless I'm more inclined to defer to experts then you. No offence, (I mean there is some offence, because this is a stupid conversation) but you have no qualifications.

[–] finitebanjo@lemmy.world 1 points 5 days ago* (last edited 5 days ago)

It's less of an unknown and more of a "it has never demonstrated any such capability."

Btw both OpenAI and Deepmind wrote papers proving their then models would never approach human error rate with infinite training. It correctly predicted performance of ChatGPT4.

[–] Zacryon@feddit.org 3 points 6 days ago (3 children)

That's such a fucking stupid idea.

Care to elaborate why?

From my point of view I don't see a problem with that. Or let's say: the potential risks highly depend on the specific setup.

[–] JustARaccoon@lemmy.world 1 points 5 days ago

Unless the videos have proper depth maps and identifiers for objects and actions they're not going to be as effective as, say, robot arm surgery data, or vr captured movement and tracking. You're basically adding a layer to the learning to first process the video correctly into something usable and then learn from that. Not very efficient and highly dependant on cameras and angles.

[–] finitebanjo@lemmy.world 0 points 5 days ago (1 children)

Imagine if the Tesla autopilot without lidar that crashed into things and drove on the sidewalk was actually a scalpel navigating your spleen.

[–] echodot@feddit.uk 1 points 5 days ago (1 children)

Absolutely stupid example because that kind of assumes medical professionals have the same standard as Elon Musk.

[–] finitebanjo@lemmy.world 2 points 5 days ago (1 children)

Elon Musk literally owns a medical equipment company that puts chips in peoples brains, nothing is sacred unless we protect it.

[–] echodot@feddit.uk 0 points 5 days ago

Into volunteers it's not standard practise to randomly put a chip in your head.

[–] Showroom7561@lemmy.ca 0 points 6 days ago* (last edited 6 days ago) (1 children)

Being trained on videos means it has no ability to adapt, improvise, or use knowledge during the surgery.

Edit: However, in the context of this particular robot, it does seem that additional input was given and other training was added in order for it to expand beyond what it was taught through the videos. As the study noted, the surgeries were performed with 100% accuracy. So in this case, I personally don't have any problems.

[–] finitebanjo@lemmy.world -1 points 5 days ago* (last edited 5 days ago) (1 children)

I actually don't think that's the problem, the problem is that the AI only factors for visible surface level information.

AI don't have object permanence, once something is out of sight it does not exist.

[–] Showroom7561@lemmy.ca 2 points 5 days ago

If you read how they programmed this robot, it seems that it can anticipate things like that. Also keep in mind that this is only designed to do one type of surgery.

I'm cautiously optimist.

I'd still expect human supervision, though.

[–] DrunkenPirate@feddit.org 113 points 1 week ago (27 children)

And then you‘re lying on the table. Unfortunately, your case is a little different than the standard surgery. Good luck.

[–] echodot@feddit.uk 0 points 6 days ago

I assume my insides are pretty much like everyone else's. I feel like if there was that much of a complication it would have been pretty obvious before the procedure started.

"Hey this guy had two heads, I'm sure the AI will work it out."

load more comments (26 replies)
[–] DeathByBigSad@sh.itjust.works 99 points 1 week ago (10 children)

Good, now add jailtime for the ceo if something goes wrong, then we'll have a very safe tech.

[–] qfe0@lemmy.dbzer0.com 5 points 6 days ago

Just like how we jail every surgeon that does something wrong

load more comments (9 replies)
load more comments
view more: next ›