this post was submitted on 30 Jan 2025
149 points (94.6% liked)

Technology

61774 readers
6632 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
149
Deepseek Database Exposed (thehackernews.com)
submitted 1 week ago* (last edited 1 week ago) by [email protected] to c/[email protected]
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 43 points 1 week ago (2 children)

both OpenAI and Microsoft are probing whether DeepSeek used OpenAI's application programming interface (API) without permission to train its own models on the output of OpenAI's systems, an approach referred to as distillation.

That would definitely show up in the quality of responses. Surely they have better and cheaper training sources...

[–] [email protected] 33 points 1 week ago

And if they did... So what

Get fucked corpo parasite. Nobody fucking care about another corpo punking u esp when it is done in spectacular manner.

[–] monotremata 4 points 1 week ago

I think it's reasonably likely. There was a research paper about how to do basically that a couple years ago. If you need a basic LLM trained on a specialized form of input and output, getting the expensive existing LLMs to generate that text for you is pretty efficient/inexpensive, so it's a reasonable way to get a baseline model. Then you can add stuff like chain of reasoning and mixture of experts to improve the performance back up to where you need it. It's not going to be a way to push the state of the art forward, but it's sure a cheap way to catch up to models that have done that pushing.