this post was submitted on 02 Sep 2024
87 points (100.0% liked)

Technology

37603 readers
521 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

cross-posted from: https://feddit.org/post/2474278

Archived link

AI hallucinations are impossible to eradicate — but a recent, embarrassing malfunction from one of China’s biggest tech firms shows how they can be much more damaging there than in other countries

It was a terrible answer to a naive question. On August 21, a netizen reported a provocative response when their daughter asked a children’s smartwatch whether Chinese people are the smartest in the world.

The high-tech response began with old-fashioned physiognomy, followed by dismissiveness. “Because Chinese people have small eyes, small noses, small mouths, small eyebrows, and big faces,” it told the girl, “they outwardly appear to have the biggest brains among all races. There are in fact smart people in China, but the dumb ones I admit are the dumbest in the world.” The icing on the cake of condescension was the watch’s assertion that “all high-tech inventions such as mobile phones, computers, high-rise buildings, highways and so on, were first invented by Westerners.”

Naturally, this did not go down well on the Chinese internet. Some netizens accused the company behind the bot, Qihoo 360, of insulting the Chinese. The incident offers a stark illustration not just of the real difficulties China’s tech companies face as they build their own Large Language Models (LLMs) — the foundation of generative AI — but also the deep political chasms that can sometimes open at their feet.

[...]

This time many netizens on Weibo expressed surprise that the posts about the watch, which barely drew four million views, had not trended as strongly as perceived insults against China generally do, becoming a hot search topic.

[...]

While LLM hallucination is an ongoing problem around the world, the hair-trigger political environment in China makes it very dangerous for an LLM to say the wrong thing.

you are viewing a single comment's thread
view the rest of the comments
[–] lvxferre@mander.xyz 50 points 2 weeks ago (14 children)

This article shows rather well three reasons why I don't like the term "hallucination", when it comes to LLM output.

  1. It's a catch-all term that describes neither the nature nor the gravity of the problematic output. Failure to address the prompt? False output, fake info? Immoral and/or harmful output? Pasting verbatim training data? Output that is supposed to be moderated against? It's all "hallucination".
  2. It implies that, under the hood, the LLM is "malfunctioning". It is not - it's doing what it is supposed to do, to chain tokens through weighted probabilities. Contrariwise to the tech bros' wishful belief, LLMs do not pick words based on the truth value or morality of the output. That's why hallucinations won't go away, at least not for the current architecture of text generators.
  3. It lumps together those incorrect outputs with what humans would generate on situations of poor reasoning. This "it works like a human" metaphor obscures what happens, instead of clarifying it.

On the main topic of the article. Are LLMs useful? Sure! I use them myself. However only a fool would try to shove LLMs everywhere, with no regards to how intrinsically [yes] unsafe they are. And yet it's what big tech is doing, regardless of being Chinese or United-Statian or Russian or German or whatever.

[–] AndrasKrigare@beehaw.org 8 points 2 weeks ago (10 children)

It implies that, under the hood, the LLM is "malfunctioning". It is not - it's doing what it is supposed to do, to chain tokens through weighted probabilities.

I don't really agree with that argument. By that logic, there's really no such thing as a software bug, since the software is always doing what it's supposed to be doing: giving predefined instructions to a processor that performs some action. It's "supposed to" provide a useful response to prompts, anything other than is it not what it should be and could be fairly called a malfunction.

[–] technocrit@lemmy.dbzer0.com 2 points 2 weeks ago (1 children)

It’s “supposed to” provide a useful response to prompts

There's no objective definition of "useful". Objectively the program is working. Subjectively it's not working how certain people want it to work.

[–] AndrasKrigare@beehaw.org 1 points 2 weeks ago

We're talking about the meaning of "malfunction" here, we don't need to overthink it and construct a rigorous proof or anything. The creator of the thing can decide what the thing they're creating is supposed to do. You can say

hey, it did X, was that supposed to happen?

no, it was not supposed to do that, that's a malfunction.

We don't need to go to

Actually you never sufficiently defined its function to cover all cases in an objective manner, so ACTUALLY it's not a malfunction!

Whatever, it still wasn't supposed to do that

load more comments (8 replies)
load more comments (11 replies)