this post was submitted on 20 Jul 2023
61 points (98.4% liked)

Technology

37604 readers
468 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

In an open letter published on Tuesday, more than 1,370 signatories—including business founders, CEOs and academics from various institutions including the University of Oxford—said they wanted to “counter ‘A.I. doom.’”

“A.I. is not an existential threat to humanity; it will be a transformative force for good if we get critical decisions about its development and use right,” they insisted.

you are viewing a single comment's thread
view the rest of the comments
[–] SSUPII@sopuli.xyz 10 points 1 year ago (17 children)

It honestly makes my blood boil when I read "AI will destroy everything" or "kill everyone" or similar. No, its not going to happen. We are nowhere close to that, we never will unintentionally or intentionally create unkillable beings like Terminator just because there is actually no material to build it with to make it par.

Big tech wanting to freeze AI just want to control it themself. Full speed ahead!

[–] nanoobot@kbin.social 4 points 1 year ago* (last edited 1 year ago) (8 children)

It makes my blood boil when people dismiss the risks of ASI without any notable counterargument. Do you honestly think something a billion times smarter than a human would struggle to kill us all if it decided it wanted to? Why would it need a terminator to do it? A virus would be far easier. And who's to say how quickly AI will advance now that AI is directly assisting progress? How can you possibly have any certainty on any timelines or risks at all?

[–] axum@kbin.social 7 points 1 year ago (4 children)

Put down the crack, there is a huge ass leap between general intelligence and the LLM of the week.

Next you're going to tell me cleverbot is going to launch nukes. We are still incredibly far from general intelligence ai.

[–] nanoobot@kbin.social 1 points 1 year ago* (last edited 1 year ago) (1 children)

I never said how long I expected it to take, how do you know we even disagree there? But like, is 50 years a long time to you? Personally anything less than 100 would be insanely quick. The key point is I don't have a high certainty on my estimates. Sure, might be perfectly reasonable it takes more than 50, but what chance is there it's even quicker? 0.01%? To me that's a scary high number! Would you be happy having someone roll a dice with a 1 in 10000 chance of killing everyone? How low is enough?

[–] axum@kbin.social 1 points 1 year ago (1 children)

But like, is 50 years a long time to you

I'll be dead.

but what chance is there it's even quicker? 0.01%? To me that's a scary high number! Would you be happy having someone roll a dice with a 1 in 10000 chance of killing everyone? How low is enough?

The odds are higher that Russia nukes your living area.

[–] nanoobot@kbin.social 1 points 1 year ago (1 children)

Well I wont be, and just because one thing might be higher probability than another, doesn't mean it's the only thing worth worrying about.

[–] averyminya@beehaw.org 1 points 1 year ago

The likelihood of an orca coming out of your asshole is just as likely as a meteor coming from space and blowing up your home.

Both could happen, but are you going to shape your life around whether or not they occur?

Your concern of AI should absolutely be pitted towards the governments and people using it. AI, especially in its current form, will not be doing a human culling.

Humans do that well enough already.

load more comments (2 replies)
load more comments (5 replies)
load more comments (13 replies)