36
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 06 Mar 2024
36 points (100.0% liked)
TechTakes
1401 readers
201 users here now
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
founded 1 year ago
MODERATORS
me when the machine specifically designed to pass the turing test passes the turing test
If you can design a model that spits out self-aware-sounding things after not having been trained on a large corpus of human text, then I'll bite. Until then, it's crazy that anybody who knows anything about how current models are trained accepts the idea that it's anything other than a stochastic parrot.
Glad that the article included a good amount of dissenting opinion, highlighting this one from Margaret Mitchell: "I think we can agree that systems that can manipulate shouldn't be designed to present themselves as having feelings, goals, dreams, aspirations."
Cool tech. We should probably set it on fire.
Despite the hype, from my admittedly limited experience I haven't seen a chatbot that is anywhere near passing the turing test. It can seemingly fool people who want to be fooled but throw some non-sequiturs or anything cryptic and context-dependent at it and it will fail miserably.
I agree, except with the first sentence.
The Turing test doesn't say any of that. Which is why it was first passed in the 60s, and is a bad test.
Any of… what?
Yea I don't think the Turing test is that great for establishing genuine artificial intelligence, but I also maintain that current state of the art doesn't even pass the Turing test to an intellectually honest standard and certainly didn't in the 60s.