this post was submitted on 26 Jun 2023
52 points (100.0% liked)

Technology

37603 readers
522 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

Tech CEOs want us to believe that generative AI will benefit humanity. They are kidding themselves

you are viewing a single comment's thread
view the rest of the comments
[–] Mirodir@lemmy.fmhy.ml 2 points 1 year ago (1 children)

Do we have a AI with a theory of mind or just a AI that answers the questions in the test correctly?

Now whether or not there is a difference between those two things is more of a philosophical debate. But assuming there is a difference, I would argue it's the latter. It has likely seen many similar examples during training (the prompts are in the article you linked, it's not unlikely to have similar texts in a web-scraped training set) and even if not, it's not that difficult to extrapolate those answers from the many texts it must've read where a character was surprised at an item missing that that character didn't see being stolen.

[–] exohuman@kbin.social 2 points 1 year ago (1 children)

Good point. How will we be able to tell the difference?

[–] newde@beehaw.org 3 points 1 year ago

You can make an educated guess if you would understand the intricacies of the programming. In this case, it's most likely blurting out words and phrases that statistically most adequately fit the (perhaps somewhat leading) questions.