52
submitted 1 year ago by troyunrau@lemmy.ca to c/futurism@lemmy.ca

The article and release are interesting unto themselves. However, as this is c/Futurism, let's discuss what happens in the future. How do you folks think this ideological battleground plays out in 5, 50, or 500 years?

all 14 comments
sorted by: hot top controversial new old
[-] RealFknNito@lemmy.world 14 points 1 year ago

Friendly reminder AI is just regurgitating information, not creating it or verifying it. It could happily tell you how to murder based on 14 posts other people have written about murder - none of whom know what the fuck they're talking about, so neither will the AI.

I am very unafraid.

[-] troyunrau@lemmy.ca 4 points 1 year ago

Do you ever have the existential crisis where you wonder if your own brain is just doing the same thing? Maybe none of your thoughts are original and are just recombinations? What if the spark of consciousness is just a delusion?

[-] WeirdAnimeOtter@lemmy.world 2 points 1 year ago

Shhh, a delusions is only comfortable when you are ignorant

[-] Gradually_Adjusting@lemmy.world 7 points 1 year ago
[-] Anticorp@lemmy.ml 4 points 1 year ago

The AI in Hyperion are some of the coolest sci-fi I've ever read.

[-] troyunrau@lemmy.ca 2 points 1 year ago

It's excellent :)

[-] troyunrau@lemmy.ca 6 points 1 year ago

If nothing else, it shows that the LLM genie cannot be put back in the bottle.

5 years: bad actors are using unregulated LLMs to flood content to every possible medium where there's a chance to influence someone. It becomes increasingly hard to tell what is an authentic human interaction on the internet. CS:GO has fake 12yo kids swearing at each other. Social media websites are now popping up that require a human to physically visit one of their service centres with government I'd to register, proving that they're human.

50 years: LLMs are so ubiquitous and self-tuning that you can have any model you want with any parameters. This becomes similar to TARS on Interstellar, sans robot body. Regulations have largely failed, but people tell their personals AIs to "tone it down in front of the family" like they would their belligerent uncles. At least one major diplomatic incident is caused when nation state sponsored LLMs are arguing with one another.

500 years: I am a paperclip.

[-] drekly@lemmy.world 3 points 1 year ago

CS:GO went the way of Overwatch 1 ๐Ÿ˜’

[-] 0x0001@sh.itjust.works 5 points 1 year ago

The "war" with LLMs is already over, that is to say there was no war. We embraced it.

Given how easily we accepted it and how slow legislation has been to have any meaningful impact on the development of the models if we achieve true AGI we'll see a similar path to integration.

My opinion is different than some, in my eyes AGI is a kind of natural evolution of us, if we create it and foster it's early development we will effectively be its parents. As our progeny, it will outlast humanity no matter it's relationship with us.

In my eyes that's the only possible way to truly preserve a piece of us forever and may be the only way to get information out beyond the fermi paradox if the paradox's key tenant that proves true is the short lifespan/implosion of advanced societies.

Creating AGI is what religious people would call judgement day, we'll let an advanced and evolving entity make the judgements about our societies that we can't.

Maybe they'll let us live the same way we let mosquitoes live, we don't wage wars where every human goes out and kills mosquitoes. We find them completely useless and even a leech on each of us, if we found an easy way to eradicate them we probably would.

Whatever happens, we deserve it. I don't say this as a way to say "the people researching and releasing LLMs are making mistakes", in fact I think those people are some of the best and brightest among us. Rather I'm saying humanity on the whole is pretty fucked up. Genocide, torture, hatred, religious bigotry, nationalism, tribalism, humans are largely just kinda smart and honestly fairly evil primates.

We literally raise sentient creatures from birth in cages to murder and devour, and the vast majority of society accepts that?!

To be more concrete, I don't think LLMs will be the path to AGI, but I do think they are an essentiao component of a consciousness. The "stream" of consciousness specifically.

5 years, nothing, LLMs treated like google is today, an incredible information source and partner for creation.

50 years, hard to conceive this time period arriving without AGI. I think humans will suffer as they weakly combat the ideology of a more advanced race being in charge. Not physically combat, but literally "back in my day" style combat.

500 years, I'd be surprised if we last that long with our current structure and march toward doom. Perhaps the new AGI race strips what it needs to leave the planet and we revert back to 1800s technology. So hard to predict.

[-] themeatbridge@lemmy.world 2 points 1 year ago

Finally, I can get those ducks home from the park!

[-] baconisaveg@lemmy.ca -4 points 1 year ago

You can either be for censorship, or against it. There's no "oh let's just censor it a little" middle-ground.

[-] troyunrau@lemmy.ca 1 points 1 year ago

I disagree. There is middle ground. If an engineer gives bad advice, it shouldn't be propagated -- you know, bridges fall down and people die. Where possible, the invalid info should be scrubbed and replaced with valid info. The engineering firm also has their reputation, permits to practice, etc. at stake. But an AI does not. There's no one to sue for negligence when someone takes invalid advice from an AI which is masquerading as a doctor. Etc. The companies making AIs are mostly trying to protect themselves when they put the gates in place.

You could go stand on your soapbox and shout suicide tips to the crowd as they walk by. You might get locked up as you're abetting a crime (in most jurisdictions). But what if you're posting suicide advice into a forum, and the advice was generated by an AI? What if a script is posting it? Where does the legal responsibility for harm fall?

[-] baconisaveg@lemmy.ca 2 points 1 year ago

A Large Language Model is just a set of computer algorithms designed to answer a user's question, it's just a tool. None of your arguments are at all relevant to the tool itself, but rather to how the tool is used. A hammer is designed to pound nails but it can also be used to murder somehow, are you going to sue the hammer manufacturer because they didn't prevent that?

If someone uses a hammer to murder someone, do they get away with it because the hammer wasn't designed to kill someone, so clearly it's not their fault? No, of course not. This article is nothing but rage-bait. They may as well have taken a hammer and started hitting everything they could (except for nails of course), and then wrote some bullshit about how Master-Craft produces items that can be used to perform abortions and kill Native Americans.

And as for my original post, this has to do with how the LLM is trained. There's several ways to 'censor' the output from a LLM, including prompts and ban tokens. This is what services like GPT or Stable Diffusion do, they don't censor the training data, they censor the inputs and outputs shown to the user. So should the training data be scrubbed of all traces of anything we find objectionable? There was plenty of murders in Hamlet, do we exclude that because the model might suggest poisoning your partner by pouring poison in their ear while they sleep?

this post was submitted on 30 Sep 2023
52 points (94.8% liked)

Futurism

429 readers
1 users here now

A place to discuss the ideas, developments, and technology that can and will shape the future of civilization.

Tenets:

(1) Concepts are often better treated in isolation -- eg: "what if energy became near zero cost?"
(2) Consider the law of unintended consequences -- eg: "if this happens, then these other systems fail"
(3) Pseudoscience and speculative physics are not welcome. Keep it grounded in reality.
(4) We are here to explore the parameter spaces of the future -- these includes political system changes that advances may trigger. Keep political discussions abstract and not about current affairs.
(5) No pumping of vapourware -- eg: battery tech announcements.

See also: !retrotechnology@lemmy.ca and !predictions@lemmy.ca

founded 1 year ago
MODERATORS