69

The venture capitalist and new Trump administration member David Sacks, meanwhile, said that there is “substantial evidence” that DeepSeek “distilled the knowledge out of OpenAI’s models.”

“There’s a technique in AI called distillation, which you’re going to hear a lot about, and it’s when one model learns from another model, effectively what happens is that the student model asks the parent model a lot of questions, just like a human would learn, but AIs can do this asking millions of questions, and they can essentially mimic the reasoning process they learn from the parent model and they can kind of suck the knowledge of the parent model,” Sacks told Fox News. “There’s substantial evidence that what DeepSeek did here is they distilled the knowledge out of OpenAI’s models and I don’t think OpenAI is very happy about this.”

This sounds like horse shit to me but I don't know the technical details well enough to say with confidence.

also "suck the knowledge out of the parent model" what the actual fuck?

top 6 comments
sorted by: hot top controversial new old
[-] BigMuffin69@awful.systems 7 points 19 hours ago

Footage of Deepseek slurping the knowledge out of the GPT4

[-] Gustephan@lemmy.world 14 points 1 day ago

I don't know enough to say whether this is valid or just crybaby tech bros having a fit on fox news but like... God I hope deepseek is completely stolen like this because and I hope there's absolutely nothing closedai can do about the fact that there's a better thief out there on the market. Fuck them so hard and fuck their hypocrisy about stealing data. Maybe we can finally move away from trying to use a double digit percentage of national electric grid capacity to power a fucking glorified magic 8ball

[-] blakestacey@awful.systems 11 points 1 day ago

This is much more a TechTakes story than a NotAwfulTech one; let's keep the discussion over on the other thread:

https://awful.systems/post/3400636

[-] khalid_salad@awful.systems 5 points 23 hours ago* (last edited 23 hours ago)

Noted going forward. Sorry about that! ❤

[-] humanspiral@lemmy.ca 4 points 20 hours ago

They don't have any evidence. They say someone did "hammer their API", and then they terminated their license (last year), but they don't know who. China bashing is not going to depend on actual evidence.

All that matters, in the end, is "customer prices" instead of our devoted love for Sam Altman.

[-] NextElephant9@awful.systems 3 points 1 day ago

Knowledge distilation is training a smaller model to mimic the outputs of a larger model. You don't need to use the same training set that was used to train the larger model (the whole internet or whatever they used for chatgpt), but can use a transfer set.

Here's a reference: Hinton, Geoffrey. "Distilling the Knowledge in a Neural Network." arXiv preprint arXiv:1503.02531 (2015)., https://arxiv.org/pdf/1503.02531

this post was submitted on 29 Jan 2025
69 points (100.0% liked)

NotAwfulTech

395 readers
3 users here now

a community for posting cool tech news you don’t want to sneer at

non-awfulness of tech is not required or else we wouldn’t have any posts

founded 2 years ago
MODERATORS