The venture capitalist and new Trump administration member David Sacks, meanwhile, said that there is ā€œsubstantial evidenceā€ that DeepSeek ā€œdistilled the knowledge out of OpenAIā€™s models.ā€

ā€œThereā€™s a technique in AI called distillation, which youā€™re going to hear a lot about, and itā€™s when one model learns from another model, effectively what happens is that the student model asks the parent model a lot of questions, just like a human would learn, but AIs can do this asking millions of questions, and they can essentially mimic the reasoning process they learn from the parent model and they can kind of suck the knowledge of the parent model,ā€ Sacks told Fox News. ā€œThereā€™s substantial evidence that what DeepSeek did here is they distilled the knowledge out of OpenAIā€™s models and I donā€™t think OpenAI is very happy about this.ā€

This sounds like horse shit to me but I donā€™t know the technical details well enough to say with confidence.

also ā€œsuck the knowledge out of the parent modelā€ what the actual fuck?

  • Gustephan@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    Ā·
    2 days ago

    I donā€™t know enough to say whether this is valid or just crybaby tech bros having a fit on fox news but likeā€¦ God I hope deepseek is completely stolen like this because and I hope thereā€™s absolutely nothing closedai can do about the fact that thereā€™s a better thief out there on the market. Fuck them so hard and fuck their hypocrisy about stealing data. Maybe we can finally move away from trying to use a double digit percentage of national electric grid capacity to power a fucking glorified magic 8ball

  • humanspiral@lemmy.ca
    link
    fedilink
    English
    arrow-up
    4
    Ā·
    2 days ago

    They donā€™t have any evidence. They say someone did ā€œhammer their APIā€, and then they terminated their license (last year), but they donā€™t know who. China bashing is not going to depend on actual evidence.

    All that matters, in the end, is ā€œcustomer pricesā€ instead of our devoted love for Sam Altman.

  • NextElephant9@awful.systems
    link
    fedilink
    English
    arrow-up
    3
    Ā·
    2 days ago

    Knowledge distilation is training a smaller model to mimic the outputs of a larger model. You donā€™t need to use the same training set that was used to train the larger model (the whole internet or whatever they used for chatgpt), but can use a transfer set.

    Hereā€™s a reference: Hinton, Geoffrey. ā€œDistilling the Knowledge in a Neural Network.ā€ arXiv preprint arXiv:1503.02531 (2015)., https://arxiv.org/pdf/1503.02531