this post was submitted on 19 Nov 2023
134 points (94.1% liked)

Technology

34984 readers
161 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] huginn@feddit.it 53 points 1 year ago (4 children)

It's interesting to me how many people I've argued with about LLMs. They vehemently insist that this is a world changing technology and the start of the singularity.

Meanwhile whenever I attempt to use one professionally it has to be babied and tightly scoped down or else it goes way off the rails.

And structurally LLMs seem like they'll always be vulnerable to that. They're only useful because they bullshit but that also makes them impossible to rely on for anything else.

[–] mkhoury@lemmy.ca 20 points 1 year ago (1 children)

I've been using LLMs pretty extensively in a professional capacity and with the proper grounding work it becomes very useful and reliable.

LLMs on their own is not the world changing tech, LLMs+grounding (what is now being called a Cognitive Architecture), that's the world changing tech. So while LLMs can be vulnerable to bullshitting, there is a lot of work around them that can qualitatively change their performance.

[–] huginn@feddit.it 10 points 1 year ago (1 children)

I'm a few months out of date in the latest in the field and I know it's changing quickly. What progress has been made towards solving hallucinations? The feeding output into another LLM for evaluation never seemed like a tenable solution to me.

[–] mkhoury@lemmy.ca 6 points 1 year ago (2 children)

Essentially, you don't ask them to use their internal knowledge. In fact, you explicitly ask them not to. The technique is generally referred to as Retrieval Augmented Generation. You take the context/user input and you retrieve relevant information from the net/your DB/vector DB/whatever, and you give it to an LLM with how to transform this information (summarize, answer a question, etc).

So you try as much as you can to "ground" the LLM with knowledge that you trust, and to only use this information to perform the task.

So you get a system that can do a really good job at transforming the data you have into the right shape for the task(s) you need to perform, without requiring your LLM to act as a source of information, only a great data massager.

[–] sudoreboot@slrpnk.net 5 points 1 year ago (1 children)

That seems like it should work in theory, but having used Perplexity for a while now, it doesn't quite solve the problem.

The biggest fundamental problem is that it doesn't understand in any meaningful capacity what it is saying. It can try to restate something it sourced from a real website, but because it doesn't understand the content it doesn't always preserve the essence of what the source said. It will also frequently repeat or contradict itself in as little as two paragraphs based on two sources without acknowledging it, which further confirms the severe lack of understanding. No amount of grounding can overcome this.

Then there is the problem of how LLMs don't understand negation. You can't reliably reason with it using negated statements. You also can't ask it to tell you about things that do not have a particular property. It can't filter based on statements like "the first game in the series, not the sequel", or "Game, not Game II: Sequel" (however you put it, you will often get results pertaining to the sequel snucked in).

[–] BluesF 2 points 1 year ago

Yeah, it's just back exactly to the problem the article points out - refined bullshit is still bullshit. You still need to teach your LLM how to talk, so it still needs that cast bullshit input into its "base" before you feed it the "grounding" or whatever... And since it doesn't actually understand any of that grounding it's just yet more bullshit.

[–] huginn@feddit.it 4 points 1 year ago

Definitely a good use for the tool: NLP is what LLMs do best and pinning down the inputs to only be rewording or compressing ground truth avoids hallucination.

I expect you could use a much smaller model than gpt to do that though. Even llama might be overkill depending on how tightly scoped your DB is

[–] Conradfart@lemmy.ca 18 points 1 year ago (2 children)

They are useful when you need to generate quasi meaningful bullshit in large volumes easily.

LLMs are being used in medicine now, not to help with diagnosis or correlate seemingly unrelated health data, but to write responses to complaint letters or generate reflective portfolio entries for appraisal.

Don't get me wrong, outsourcing the bullshit and waffle in medicine is still a win, it frees up time and energy for the highly trained organic general intelligences to do what they do best. I just don't think it's the exact outcome the industry expected.

I think it's the outcome anyone really familiar with the tech expected, but that rarely translates to marketing departments and c-suite types.

I did an LLM project in school, and while that was a limited introduction, it was enough for me to doubt most of the claims coming from LLM orgs. An LLM is good at matching its corpus and that's about it. So it'll work well for things like summaries, routine text generation, and similar tasks (and it's surprisingly good at forming believable text), but it'll always disappoint with creative work.

I'm sure the tech can do quite a bit more than my class went through, but the limitations here are quite fundamental to the tech.

[–] huginn@feddit.it 3 points 1 year ago

That's kinda the point of my above comment: they're useful for bullshit: that's why they'll never be trustworthy

[–] waspentalive@beehaw.org 9 points 1 year ago* (last edited 1 year ago) (1 children)

I use chatgpt to make up stuff, imagine things that don't exist for fun - like a 'pitch' for the next new Star Trek series, or to reword my much too succinct prose for a manual for a program I am writing ('Calcula' in gitlab) or ideas for a new kind of restaurant (The chef teaches you how to cook the meal you are about to eat) - but never have it code or ask it about facts, it makes them up just as easily as the stuff I just described.

[–] huginn@feddit.it 5 points 1 year ago

A great use for the tool

[–] Damage@feddit.it 6 points 1 year ago (2 children)

It's a computer that understands my words and can reply, even complete tasks upon request, nevermind the result. To me that's pretty groundbreaking.

[–] huginn@feddit.it 24 points 1 year ago (3 children)

It's a probabilistic network that generates a response based on your input.

No understanding required.

[–] 0ops@lemm.ee 15 points 1 year ago

It's a probabilistic network that generates a response based on your input.

Same

[–] Eheran@lemmy.world 6 points 1 year ago (2 children)

Ask it to write code that replaces every occurrence of "me" in every file name in a folder with "us", but excluding occurrences that are part of a word (like medium should not be usdium) and it will give you code that does exactly that.

You can ask it to write code that does a heat simulation in a plate of aluminum given one side of heated and the other cooled. It will get there with some help. It works. That's absolutely fucking crazy.

[–] sugar_in_your_tea@sh.itjust.works 6 points 1 year ago (1 children)

Maybe, that really depends on if that task or a very similar task exists in sufficient amounts in its training set. Basically, you could get essentially the same result by searching online for code examples, the LLM might just make it a little faster (and probably introduce some errors as well).

An LLM can only generate text that exists in its training data. That's a pretty important limitation, which has all kinds of copyright-related issues associated with it (e.g. I can't just copy a code example from GitHub in most cases).

[–] Eheran@lemmy.world 0 points 1 year ago (1 children)

No, it does not depend on preexisting tasks, which is why I told you those 2 random examples. You can come up with new, never before seen questions if you want to. How to stack a cable, car battery, beer bottle, welding machine, tea pot to get the highest tower. Whatever. It is not always right, but also much more capable than you think.

[–] huginn@feddit.it 2 points 1 year ago (1 children)

It is dependent on preexisting tasks, you're just describing encoded latent space.

It's not explicit but it's implicitly encoded.

And you still can't trust it because the encoding is intrinsically lossy.

[–] Eheran@lemmy.world -1 points 1 year ago

It can come up with new solutions.

[–] huginn@feddit.it 4 points 1 year ago* (last edited 1 year ago)

Ask it to finish writing the code to fetch a permission and it will make a request with a non-existent code. Ask it to implement an SNS API invocation and it'll make up calls that don't exist.

Regurgitating code that someone else wrote for an aluminum simulation isn't the flex you think it is: that's just an untrustworthy search engine, not a thinking machine

[–] iopq@lemmy.world -1 points 1 year ago (2 children)

Yet it can outperform humans on some tests involving logic. It will never be perfect, but that implies you can test its IQ

[–] huginn@feddit.it 9 points 1 year ago (2 children)
  1. Not consistently and not across truly logical tests. They abjectly fail at abstract reasoning. They do well only in very specific cases.
  2. IQ is an objectively awful measure of human intelligence. Why would it be useful for artificial intelligence?
  3. For these tests that are so centered around specific facts: of course a model that has had the entirety of the Internet encoded into it has the answers. The shocking thing is that the model is so lossy that it doesn't ace the test.
[–] Feathercrown@lemmy.world 1 points 1 year ago (1 children)

IQ correlates with a good number of things though. It'a not perfect but it's not meaningless either.

[–] huginn@feddit.it 1 points 1 year ago (1 children)

And global warming correlates with the decline in piracy rates. IQ is a garbage statistic invented by early 20th century eugenicists to prove that white people were the best.

You can't boil down the nuance of the most complex object in the known universe to a single number.

[–] Feathercrown@lemmy.world 0 points 1 year ago

Not perfectly you can't. But similarly to how people's SAT scores predict their future success, IQ tests in aggregate do have predictive power.

[–] iopq@lemmy.world -1 points 1 year ago (1 children)

IQ is objectively a good measure of human intelligence. High IQ people have higher educational achievement, income, etc.

[–] huginn@feddit.it 1 points 1 year ago (1 children)

Don't take the effect and make it the cause my guy

[–] iopq@lemmy.world 0 points 1 year ago (1 children)

I never said it's the cause. We're trying to find a measure that correlates well with actual intelligence g

IQ correlates with g, but also income/education correlates with g because smarter people do better in these metrics.

IQ doesn't make you smarter, but smarter people can do better on IQ tests

[–] huginn@feddit.it 1 points 1 year ago (1 children)

smarter people do better in these metrics.

Smarter by what measure? IQ?

You're using circular reasoning here.

Income and education levels are not the same thing as intelligence, nor are smarter people higher earners or more well educated.

IQ correlates best with educational achievement. Educational achievement is best predicted by your zip code. Poverty creates sharp educational disadvantages.

Intelligence, as measured by your maximally attainable levels of obtaining knowledge and skills, is something that the majority of people will never test their limits of.

IQ tests do not measure that maximum, only how far along that trajectory you might have come compared to your "peers".

Therefore: IQ tests are 1 step removed from just asking someone where they grew up, how much college they attended and how much money their parents made.

It has nothing to do with measuring that underlying factor and everything to do with measuring socioeconomic status.

It was a crude tool invented by eugenicists to promote genocide and you should stop using or respecting it at all.

[–] iopq@lemmy.world 0 points 1 year ago

You have the cause and effect switched. More intelligent parents have higher incomes, their kids also are more likely to have higher intelligence because it's heritable. So if a bunch of rich Jews live in a Manhattan, does it mean that IQ is predicting their zip code? Or is it predicting they are Jewish? That's just a weird way of thinking about it.

Your definition of intelligence is not that is commonly thought of intelligence. It's like saying "intelligence is your maximum chess skill measured by elo and you will never achieve it"

But learning to play chess quickly would also fall under intelligence. Given not being a high level at a similar game, there will be people who learn quicker than others. Learning languages, learning from your experience, all fall under intelligence.

If you take all of these measures, average them, you will get a true intelligence measurement. IQ is a sample of true intelligence by using only some measurements, since intelligence can have an infinite amount of tasks that it can be used to perform

[–] exponential_wizard@lemm.ee 4 points 1 year ago (1 children)

"Test it's IQ". The fact that you think IQ is a useful test for intelligence tells me everything I need to know

[–] iopq@lemmy.world -1 points 1 year ago

The fact you went out of your way to write it's when I wrote the correct "its" tells me everything I need to know about your educational achievement

[–] amki@feddit.de 17 points 1 year ago

That is exactly what it doesn't. There is no "understanding" and that is exactly the problem. It generates some output that is similar to what it has already seen from the dataset it's been fed with that might correlate to your input.