this post was submitted on 03 Sep 2024
133 points (92.4% liked)

No Stupid Questions

35931 readers
1048 users here now

No such thing. Ask away!

!nostupidquestions is a community dedicated to being helpful and answering each others' questions on various topics.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules (interactive)


Rule 1- All posts must be legitimate questions. All post titles must include a question.

All posts must be legitimate questions, and all post titles must include a question. Questions that are joke or trolling questions, memes, song lyrics as title, etc. are not allowed here. See Rule 6 for all exceptions.



Rule 2- Your question subject cannot be illegal or NSFW material.

Your question subject cannot be illegal or NSFW material. You will be warned first, banned second.



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Questions which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding META posts and joke questions.

Provided it is about the community itself, you may post non-question posts using the [META] tag on your post title.

On fridays, you are allowed to post meme and troll questions, on the condition that it's in text format only, and conforms with our other rules. These posts MUST include the [NSQ Friday] tag in their title.

If you post a serious question on friday and are looking only for legitimate answers, then please include the [Serious] tag on your post. Irrelevant replies will then be removed by moderators.



Rule 7- You can't intentionally annoy, mock, or harass other members.

If you intentionally annoy, mock, harass, or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here.



Credits

Our breathtaking icon was bestowed upon us by @Cevilia!

The greatest banner of all time: by @TheOneWithTheHair!

founded 1 year ago
MODERATORS
 

I've tried several types of artificial intelligence including Gemini, Microsoft co-pilot, chat GPT. A lot of the times I ask them questions and they get everything wrong. If artificial intelligence doesn't work why are they trying to make us all use it?

top 50 comments
sorted by: hot top controversial new old
[–] SpaceNoodle@lemmy.world 80 points 2 months ago (2 children)

Investors are dumb. It's a hot new tech that looks convincing (since LLMs are designed specifically to appear correct, not be correct), so anything with that buzzword gets a ton of money thrown at it. The same phenomenon has occurred with blockchain, big data, even the World Wide Web. After each bubble bursts, some residue remains that actually might have some value.

[–] pimeys@lemmy.nauk.io 36 points 2 months ago

And LLM is mostly for investors, not for users. Investors see you "do AI" even if you just repackage GPT or llama, and your Series A is 20% bigger.

[–] Kintarian@lemmy.world 13 points 2 months ago

I can see that. That guy over there has the new shiny toy. I want a new shiny toy. Give me a new shiny toy.

[–] OpenStars@discuss.online 46 points 2 months ago (2 children)

Money. If you paid to use those services, they got what they wanted.

[–] homesweethomeMrL@lemmy.world 19 points 2 months ago

Money.

That’s the entirety of the reason.

[–] Boozilla@lemmy.world 12 points 2 months ago (1 children)
[–] iconic_admin@lemmy.world 7 points 2 months ago (1 children)

Summed up an MBA in four words.

load more comments (1 replies)
[–] gedaliyah@lemmy.world 37 points 2 months ago (2 children)

Generative AI has allowed us to do some things that we could not do before. A lot of people very foolishly took that to mean it would let us do everything we couldn't do before.

[–] Kintarian@lemmy.world 9 points 2 months ago

That's because the PR department keeps telling us that it's the best things since sliced bread.

load more comments (1 replies)
[–] Juice@midwest.social 31 points 2 months ago (15 children)

The last big fall in the price of bitcoin, in December '22 was caused by a shift in the dynamics of mining where it became more expensive to mine new btc than what the coin was actually worth. Not only did this plunge the price of crypto it also demolished demand for expensive graphics chips which are repurposed to run the process-heavy complex math used in mining. Cheaper chips, cascading demand and server space that was dedicated to mining related activities threatened to wipe out profit margins in multiple tech sectors.

6 months later, Chat GPT is tolled out by Open AI. The previous limitations on processing capabilities were gone, server space was cheap and the tech was abundant. So all these tech sectors at risk of losing their ass in an overproduction driven recession, now had a way to pump the price of their services and this was to pump AI.

Additionally around this time the world was recovering from covid lockdowns. Increased demand for online services was dwindling (exacerbating the other crisis outlined above) as people were returning to work and spending more time being social IRL rather than using services. Companies had hired lots of new workers: programmers, tech infrastructure workers, etc., yo meet the exploding demand during covid. Now they had too many workers and their profits were being threatened.

The Federal reserve had raised interest rates to stifle continued hiring of new employees. The solution that the fed had come up with in order to stifle inflation was to encourage laying off workers end masse -- what Marxists might call restoring the reserve army of labor, or relative surplus population -- which was substantially depleted during the pandemic. But business owners were reluctant to do this, the tight labor market of the last few years had made business owners and managers skittish about letting people go.

A basic principle at play here, is that new technology is introduced for two reasons only: to sell as a new commodity and (what we are principally concerned with) replacing workers with machines. Another basic principle is that the capitalist system has to have a certain percentage of its population unemployed and hyper exploited in order to keep wages low.

So there was a confluence of incentives here. 1. Inexpensive server space and chips which producers were eager to restore to profitability (or else face drastic consequences) 2. A need to lay off workers in order to stop inflation 3. Incentives for businesses to do so.

Laying off relatively highly paid technical/intellectual labor is a low hanging fruit in this whole equation, and the roll out of AI did just that. Hundreds of thousands of highly paid workers were laid off across a variety of sectors, assured that AI would create so much more efficiency and cut out the need for so many of these workers. So they rolled out this garbage tech that doesn't work, but everyone in the industry, the media, the government needs it to work, or else they face a massive economic crisis, which had already started with inflation.

At the end of the day its just a massive grift, pushed out to compensate for excessive overproduction driven by another massive grift (cryptocurrency) combined with economic troubles that arose from an insufficient government response to a pandemic that killed millions of people; and rather than take other measures to stifle inflation our leaders in global finance decided to shunt the consequences onto workers, as always. The excuse given was AI, which is nothing more than a predictive text algorithm attached to a massive database created by exploited workers overseas and stolen IPs, and a fuck load of processing power.

[–] Kintarian@lemmy.world 9 points 2 months ago (2 children)

I hope someday we can come up with an economic system that is not based purely on profit and the exploitation of human beings. But I don't know that I'll live long enough to see it.

[–] Juice@midwest.social 6 points 2 months ago* (last edited 2 months ago)

Well remember that the shifts that can happen in material conditions and consciousness can happen very quickly. We can't decide when that is, but we can prepare and build trust until it does occur. Hard to imagine what it would take in the west to see an overthrow of capitalism, all we can do is throw our weight behind where it will have the most effect, hopefully where our talents reside also! Stay optimistic, despite even evidence to the contrary. For the capitalists, its better to believe that the end of the world is coming than to believe a new world is possible. So if nothing else lets give em hell

[–] z00s@lemmy.world 4 points 2 months ago

I can't tell you how many times I've had this exact thought. 😕

[–] z00s@lemmy.world 4 points 2 months ago (1 children)

Are you an economist or business professor IRL? Because that was an amazing answer!

[–] Juice@midwest.social 6 points 2 months ago* (last edited 2 months ago)

No actually I'm mostly self educated. I'm just a tech worker who studies history, social theory and economics, but also does some political organizing. So take it with a grain of salt if you must.

Glad you got something from it, I appreciate the compliment!

load more comments (13 replies)
[–] hungryphrog@lemmy.blahaj.zone 26 points 2 months ago

Robots don't demand things like "fair wages" or "rights". It's way cheaper for a corporation to, for example, use a plagiarizing artificial unintelligence to make images for something, as opposed to commissioning a human artist who most likely will demand some amount of payment for their work.

Also I think that it's partially caused by people going "ooh, new thing!" without stopping to think about the consequences of this technology or if it is actually useful.

[–] some_guy@lemmy.sdf.org 16 points 2 months ago

Rich assholes have spent a ton of money on it and they need to manufacture reasons why that wasn't a waste.

[–] givesomefucks@lemmy.world 12 points 2 months ago* (last edited 2 months ago) (5 children)

A dumb person thinks AI is really smart, because they just listen to anyone that answers confidentially

And no matter what, AI is going to give its answer like it's is 100% definitely the truth.

That's why there's such a large crossover with AI and crypto, the same people fall for everything.

There's new supporting evidence for Penrose's theory that natural intelligence involves just an absolute shit ton of quantum interactions, because we just found out how the body can create an environment where quantom super position can not only be achieved, but incredibly simply.

AI got a boost because we didn't really (still dont) understand consciousness. Tech bro's convinced investors that neurons were what mattered, and made predictions for when that amount of neurons can be simulated.

But if it include billions of molecules in quantum superposition, we're not getting there in our lifetimes. But there's a lot of money sunk in to it already, so there's a lot of money to lose if people suddenly get realistic about what it takes to make a real artificial intelligence.

load more comments (5 replies)
[–] xia@lemmy.sdf.org 12 points 2 months ago (4 children)

The natural general hype is not new... I even see it in 1970's scifi. It's like once something pierced the long-thought-impossible turing test, decades of hype pressure suddenly and freely flowed.

There is also an unnatural hype (that with one breakthrough will come another) and that the next one might yield a technocratic singularity to the first-mover: money, market dominance, and control.

Which brings the tertiary effect (closer to your question)... companies are so quickly and blindly eating so many billions of dollars of first-mover costs that the corporate copium wants to believe there will be a return (or at least cost defrayal)... so you get a bunch of shitty AI products, and pressure towards them.

load more comments (4 replies)
[–] BugleFingers@lemmy.world 11 points 2 months ago

IIRC When ChatGPT was first announced I believe the hype was because it was the first real usable interface a layman could interact with using normal language and have an intelligible response from the software. Normally to talk with computers we use their language (programming) but this allowed plain language speakers to interact and get it to do things with simple language in a more pervasive way than something like Siri for instance.

This then got over hyped and over promised to people with dollars in their eyes at the thought of large savings from labor reduction and capabilities far greater than it had. They were sold a product that has no real "product" as it's something most people would prefer to interact with on their own terms when needed, like any tool. That's really hard to sell and make people believe they need it. So they doubled down with the promise it would be so much better down the road. And, having spent an ungodly amount into it already, they have that sunken cost fallacy and keep doubling down.

This is my personal take and understanding of what's happening. Though there's probably more nuances, like staying ahead of the competition that also fell for the same promises.

[–] Kolanaki@yiffit.net 10 points 2 months ago* (last edited 2 months ago)

The hype is also artificial and usually created by the creators of the AI. They want investors to give them boatloads of cash so they can cheaply grab a potential market they believe exists before they jack up prices and make shit worse once that investment money dries up. The problem is, nobody actually wants this AI garbage they're pushing.

[–] Feathercrown@lemmy.world 9 points 2 months ago (1 children)

Disclaimer: I'm going to ignore all moral questions here

Because it represents a potentially large leap in the types of problems we can solve with computers. Previously the only comparable tool we had to solve problems were algorithms, which are fast, well-defined, and repeatable, but cannot deal with arbitrary or fuzzy inputs in a meaningful way. AI excels at dealing with fuzzy inputs (including natural language, which was a huge barrier previously), at the expense of speed and reliability. It's basically an entire missing half to our toolkit.

Be careful not to conflate AI in general with LLMs. AI is usually implemented as Machine Learning, which is a method of fitting an output to training data. LLMs are a specific instance of this that are trained on language (hence, large language models). I suspect that if AI becomes more widely adopted, most users will be interacting with LLMs like you are now, but most of the business benefit would come from classifiers that have a more restricted input/output space. As an example, you could use ML to train an AI that can be used to detect potentially suspicious bank transactions. The more data you have to sort through, the better AI can learn from it*, so I suspect the companies that have been collecting terabytes of data will start using AI to try to analyze it. I'm curious if that will be effective.

*technically it depends a lot on the training parameters

load more comments (1 replies)
[–] Kanda@reddthat.com 8 points 2 months ago (4 children)

There is no artificial intelligence, just very large statistical models.

[–] Kintarian@lemmy.world 7 points 2 months ago (1 children)

It's easier for the marketing department. According to an article, it's neither artificial nor intelligent.

[–] Feathercrown@lemmy.world 5 points 2 months ago (10 children)

In what way is it not artificial

load more comments (10 replies)
load more comments (3 replies)
[–] mjhelto@lemm.ee 8 points 2 months ago

It amazed people when it first launched and capitalists took that to mean replace all their jobs with AI. Where we wanted AI to make shit jobs easier, they used it to replace whole swaths of talent across the industry's. Recent movies read like they were written almost entirely by AI. Like when Cartman was a robot and kept giving out terrible movie ideas.

[–] ptz@dubvee.org 8 points 2 months ago* (last edited 2 months ago) (1 children)

Like was said: money.

In addition, they need training data. Both conversations and raw material. Shoving "AI" into everything whether you want it or not gives them the real world conversational data to train on. If you feed it any documents, etc it's also sucking that up for the raw data to train on.

Ultimately the best we can do is ignore it and refuse to use it or feed it garbage data so it chokes on its own excrement.

load more comments (1 replies)
[–] bionicjoey@lemmy.ca 8 points 2 months ago

A lot of jobs are bullshit. Generative AI is good at generating bullshit. This led to a perception that AI could be used in place of humans. But unfortunately, curating that bullshit enough to produce any value for a company still requires a person, so the AI doesn't add much value. The bullshit AI generates needs some kind of oversight.

[–] SomeAmateur@sh.itjust.works 7 points 2 months ago* (last edited 2 months ago) (1 children)

I genuinely think the best practical use of AI, especially language models is malicious manipulation. Propaganda/advertising bots. There's a joke that reddit is mostly bots. I know there's some countermeasures to sniff them out but think about it.

I'll keep reddit as the example because I know it best. Comments are simple puns, one liner jokes, or flawed/edgy opinions. But people also go to reddit for advice/recommendations that you can't really get elsewhere.

Using an LLM AI I could in theory make tons of convincing recommendations. I get payed by a corporation or state entity to convince lurkers to choose brand A over brand B, to support or disown a political stance or to make it seem like tons of people support it when really few do.

And if it's factually incorrect so what? It was just some kind stranger™ on the internet

load more comments (1 replies)
[–] Tylerdurdon@lemmy.world 7 points 2 months ago (3 children)
  • automation by companies so they can "streamline"their workforces.

  • innovation by "teaching" it enough to solve bigger problems (cancer, climate, etc).

  • creating a sentient species that is the next evolution of life and watching it systematically eradicate every last human to save the planet.

load more comments (3 replies)
[–] empireOfLove2@lemmy.dbzer0.com 7 points 2 months ago (13 children)

They were pretty cool when they first blew up. Getting them to generate semi useful information wasn't hard and anything hard factual they would usually avoid answering or defer.

They've legitimately gotten worse over time. As user volume has gone up necessitating faster, shallower model responses, and further training on Internet content has resulted in model degradation as it trains on its own output, the models gradually begin to break. They've also been pushed harder than they were meant to, to show "improvement" to investors demanding more accurate human like fact responses.

At this point it's a race to the bottom on a poorly understood technology. Every money sucking corporation latched on to LLM's like a piglet finding a teat, thinking it was going to be their golden goose to finally eliminate those stupid whiny expensive workers that always ask for annoying unprofitable things like "paid time off" and "healthcare". In reality they've been sold a bill of goods by Sam Altman and the rest of the tech bros currently raking in a few extra hundred billion dollars.

load more comments (13 replies)
[–] Fedegenerate@lemmynsfw.com 5 points 2 months ago* (last edited 2 months ago)

As a beginner in self hosting I like plugging the random commands I find online into a llm. I ask it what the command does, what I'm trying to achieve and if it would work..

It acts like a mentor, I don't trust what it says entirely so I'm constantly sanity checking it, but it gets me to where I want to go with some back and forth. I'm doing some of the problem solving, so there's that exercise, it also teaches me what commands do and how the flags alter it. It's also there to stop me making really stupid mistakes that I would have learned the hard way without.

Last project was adding a HDD to my zpool as a mirror. I found the "attach" command online with a bunch of flags. I made what I thought was my solution and asked chatgpt. It corrected some stuff: I didn't include the name of my zpool. Then gave me a procedure to do it properly.

In that procedure I noticed an inconsistency in how I was naming drives vs how my zpool was naming drives. Asked chat gpt again, I was told I was a dumbass, if thats the naming convention I should probably use that one instead of mine (I was using /dev/sbc and the zpool was using /dev/disk/by-id/). It told me why the zpool might have been configured that way so that was a teaching moment, I'm using usb drives and the zpool wants to protect itself if the setup gets switched around. I clarified the names and rewrote the command, not really chatgpt was constantly updating the command as we went... Boom I have mirrored my drives, I've made all my stupid mistakes in private and away from production, life is good.

Mooooneeeyyyy

I work as an AI engineer, let me tell you, the tech is awesome and has a looooot of potential but its not ready yet. Because of high potential literally no one wants to miss the opportunity of getting rich quick with it. Its only been like 2-3 years when this tech was released to the public, if only openai had released it as open-source, just like everyone before them, we wouldn't be here. But they wanted to make money and now everyone else wants to too.

[–] Bookmeat@lemmy.world 5 points 2 months ago

Novelty, lack of understanding, and avarice.

[–] ProfessorScience@lemmy.world 4 points 2 months ago

When ChatGPT first started to make waves, it was a significant step forward in the ability for AIs to sound like a person. There were new techniques being used to train language models, and it was unclear what the upper limits of these techniques were in terms of how "smart" of an AI they could produce. It may seem overly optimistic in retrospect, but at the time it was not that crazy to wonder whether the tools were on a direct path toward general AI. And so a lot of projects started up, both to leverage the tools as they actually were, and to leverage the speculated potential of what the tools might soon become.

Now we've gotten a better sense of what the limitations of these tools actually are. What the upper limits of where these techniques might lead are. But a lot of momentum remains. Projects that started up when the limits were unknown don't just have the plug pulled the minute it seems like expectations aren't matching reality. I mean, maybe some do. But most of the projects try to make the best of the tools as they are to keep the promises they made, for better or worse. And of course new ideas keep coming and new entrepreneurs want a piece of the pie.

[–] HobbitFoot@thelemmy.club 4 points 2 months ago (1 children)

The idea is that it can replace a lot of customer facing positions that are manpower intensive.

Beyond that, an AI can also act as an intern in assisting in low complexity tasks the same way that a lot of Microsoft Office programs have replaced secretaries and junior human calculators.

[–] Kintarian@lemmy.world 5 points 2 months ago (1 children)

I've always figured part of it Is that businesses don't like to pay labor and they're hoping that they can use artificial intelligence to get rid of the rest of us so they don't have to pay us.

load more comments (1 replies)
[–] lemmylommy@lemmy.world 4 points 2 months ago* (last edited 2 months ago) (1 children)

You have asked why there is so much hype around artifical intelligence.

There are a few reasons this might be the case:

  1. Because humans are curious. Experimenting with how humans believe memory and intelligence work might just lead them to find out something about their own intelligence.

  2. Because humans are stupid. Most do not have the slightest idea what „AI“ is this time, yet they are willing to believe in the most outlandish claims about it. Look up ELIZA. It fooled a lot of people, just like LLMs today.

  3. Because humans are greedy. And the prospect of replacing a lot of wage-earners, and not just manual laborers this time, with a machine is just too good to pass up for management. The potential savings are huge, if it works, so the willingness to spend money is also considerable.

In conclusion, there are many reasons for the hype around artificial intelligence and most of them relate to human deficiencies and human nature in general.

If you have further questions I am happy to help. Enjoy your experience with AI. While you still can. 🤖

load more comments (1 replies)
[–] LarmyOfLone@lemm.ee 4 points 2 months ago (1 children)

Look at all the comments on this post. We're not quite there but imagine half of the comments written by Chat GPT and it's only going to get better.

Does it matter than 50% of them get it wrong?

[–] OpenStars@discuss.online 4 points 2 months ago

To advertisers? No.

To the platform designers? Also no.

To idiot users? Still no.

To non-idiot users? Surprisingly no (bc we already left and are here now:-).

To people wanting Reddit to go the distance and boost their stock values, yes. But only in the long-term, which never exists, and in the short-term, no.

Hence, enshittification, delivered in a confident tone.

[–] PenisDuckCuck9001@lemmynsfw.com 4 points 2 months ago* (last edited 2 months ago) (1 children)

One of the few things they're good at is academic "cheating". I'm not a fan of how the education industry has become a massive pyramid scheme intended to force as many people into debt as possible, so I see ai as the lesser evil and a way to fight back.

Obviously no one is using ai to successfully do graduate research or anything, I'm just talking about how they take boring easy subjects and load you up with pointless homework and assignments so waste your time rather than learn anything. My homework is obviously ai generated and there's a lot of it. I'm using every resource available to get by.

load more comments (1 replies)
load more comments
view more: next ›