this post was submitted on 29 Jan 2025
207 points (96.8% liked)

World News

40018 readers
2421 users here now

A community for discussing events around the World

Rules:

Similarly, if you see posts along these lines, do not engage. Report them, block them, and live a happier life than they do. We see too many slapfights that boil down to "Mom! He's bugging me!" and "I'm not touching you!" Going forward, slapfights will result in removed comments and temp bans to cool off.

We ask that the users report any comment or post that violate the rules, to use critical thinking when reading, posting or commenting. Users that post off-topic spam, advocate violence, have multiple comments or posts removed, weaponize reports or violate the code of conduct will be banned.

All posts and comments will be reviewed on a case-by-case basis. This means that some content that violates the rules may be allowed, while other content that does not violate the rules may be removed. The moderators retain the right to remove any content and ban users.


Lemmy World Partners

News !news@lemmy.world

Politics !politics@lemmy.world

World Politics !globalpolitics@lemmy.world


Recommendations

For Firefox users, there is media bias / propaganda / fact check plugin.

https://addons.mozilla.org/en-US/firefox/addon/media-bias-fact-check/

founded 2 years ago
MODERATORS
 

Summary

Alibaba has launched Qwen 2.5-Max, an AI model it claims outperforms DeepSeek-V3, OpenAI’s GPT-4o, and Meta’s Llama-3.1-405B.

The release, coinciding with Lunar New Year, reflects mounting competition in China’s AI sector after DeepSeek’s rapid rise.

DeepSeek’s recent advancements have pressured Chinese rivals like ByteDance and Baidu to upgrade their models and cut prices.

DeepSeek’s founder downplays price wars, focusing on artificial general intelligence (AGI). The company’s lean, research-focused structure contrasts with China’s tech giants, which face challenges in AI innovation.

top 26 comments
sorted by: hot top controversial new old
[–] mctoasterson@reddthat.com 23 points 1 day ago

Can't wait for Wish.com to release DickGargle 3.8-Ultra1

[–] Breve@pawb.social 69 points 1 day ago (2 children)

Well, the models start comin' and they don't stop comin'...

The US tech sector has just been completely disrupted. Turns out decades of slashing public education and demonizing "liberal" colleges is starting to catch up. Even Elmo himself said that H1B visas are critical because the US simply isn't producing enough talent, but he and the other tech billionaires didn't realize that money can't buy everything, as they are now being shown with their pants down.

[–] dave 21 points 1 day ago (1 children)

Well, the models start comin' and they don't stop comin'...

Got my RTX, gonna hit the ground runnin’…

[–] pHr34kY@lemmy.world 6 points 23 hours ago (1 children)

Didn't make sense just to train for fun.

[–] locahosr443@lemmy.world 4 points 23 hours ago

Gonna steal some data it's free to learn

[–] mlg@lemmy.world 7 points 1 day ago (1 children)

I read this entire comment synced to smashmouth lmao

[–] superkret@feddit.org 2 points 23 hours ago

Get off muh swamp!

[–] jagermo@feddit.org 62 points 1 day ago (1 children)
[–] adespoton@lemmy.ca 39 points 1 day ago (1 children)

DeepSeek’s “big change” isn’t the performance of its model though; it’s that it is fully open and operates on a fraction of the resources.

Is alibaba’s model also open weights, open reasoning, free for anyone to run, and runnable (and trainable) on consumer hardware?

[–] trevor@lemmy.blahaj.zone 38 points 1 day ago (1 children)

Call it "open weight" if you want, but it's not "fully open". The training data is still proprietary, and the model can't be accurately reproduced. It's proprietary in the same way that llama is proprietary.

[–] Gsus4@mander.xyz 10 points 1 day ago* (last edited 1 day ago) (1 children)

But I could use it as a starting point for training and build from it with my own data. I could fork it. I couldn't fork llama, I don't have the weights.

[–] trevor@lemmy.blahaj.zone 10 points 1 day ago

You can also fork proprietary code that is source available (depending on the specific terms of that particular proprietary license), but that doesn't make it open source.

Fair point about llama not having open weights though. So it's not as proprietary as llama. It still shouldn't be called open source if the training data that it needs to function is proprietary.

[–] r00ty@kbin.life 30 points 1 day ago

Oh, good. Maybe they will stop trying to scrape my websites at some ridiculous rate using faked real browser UAs. I just blocked their whole ASN (AS45102) in the end.

[–] NielsBohron@lemmy.world 22 points 1 day ago

I thought for sure this was an Onion article

[–] Pregnenolone@lemmy.world 14 points 1 day ago (1 children)
[–] ThePowerOfGeek@lemmy.world 21 points 1 day ago (1 children)

I already have the Temu AI psuedocode. Here you go:

10 print "Hi, how can I help?"

20 receive input

30 print "That's awesome! What else?"

40 go to 20

[–] dubyakay@lemmy.ca 11 points 1 day ago

Looks pretty basic to me!

[–] Hubi@feddit.org 15 points 1 day ago (1 children)

Any word on the training cost? I feel like that's the most relevant metric here.

[–] ms_lane@lemmy.world 14 points 1 day ago

2 Reeses Cups and a pack of ramen. Alibaba are efficient!

[–] Bronzebeard@lemm.ee 5 points 1 day ago

Oh cool, I was worried my 401k had almost sort of recovered from the last bombshell earlier this week...

[–] A_A@lemmy.world -4 points 1 day ago (1 children)

DeepSeek_R1 outperform or equalzz GPT-1o is major newZ, but : 4o is much better than 1o. Now, Qwen-2.5Max outperforms GPT-4o ... watever the investment involved, this is even more important ( ! ).

[–] ebolapie@lemmy.world 10 points 1 day ago (1 children)
[–] A_A@lemmy.world -4 points 1 day ago (1 children)

😋 yes, why ? becauzzze of the zzZ ?

[–] ebolapie@lemmy.world 9 points 1 day ago

Among other things, yes.