American companies are spending enormous sums to develop high-performing AI models. Distillation attacks are attempting to maliciously extract them — and nobody is doing much to stop it.

  • humanspiral@lemmy.ca
    link
    fedilink
    English
    arrow-up
    6
    ·
    3 hours ago

    Models getting better does give extra information for making newer models better too. China publishes far more advanced research than US models “steal”, and they open source exceptionally strong/fast models that US can also steal from.

  • P03 Locke@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    9
    ·
    3 hours ago

    Whoever wrote this article didn’t even bother to do the most basic of research.

    DeepSeek fully admitted they started with ChatGPT outputs to train its model. And then they released it as an open-source model, so that everybody else can “steal” their work. On the image/video front, the general public has created every possible variation on top of every model you can think of. On top of that, any model that has ever been released with full weights has been spun into whatever variation or VRAM size you want.

    The ugly truth that the American companies want to hide is the fact that they are spending trillions of dollars on an oligopoly that they can’t keep long-term. They hope that they can just keep spending more money to add more billions of parameters to their models, and keep technologically competitive with the secondary open-source models. But, they’ve already ran into diminishing returns over a year ago, and the global compute sector physically cannot keep up with demand for another cycle of even more diminishing returns.

    The other factor is that realistic miniaturization of models is already here. Some of the smaller sizes aren’t as effective as the 250GB models they use on cloud-based services, but you can still do a lot with a 16GB or 24GB video card, using models of those sizes. Optimization and LLM quantization is getting better and better each year. The AI bubble burst is going to force a cascade shift into a new era of localization. Everybody is sick to fucking death of renting and subscribing to everything. Us pirates already do so on the media front, and soon localization of LLMs is going to become way more popular.

    The question isn’t “Can people steal the tech?”. It’s “how long will people notice that it’s already happening?”

  • infinitesunrise@slrpnk.net
    link
    fedilink
    English
    arrow-up
    34
    ·
    6 hours ago

    I would reckon that China is perfectly satisfied to let us be the sole host of the thing that is rapidly destroying our economy and trust in all media from the inside out.

  • Sims@lemmy.ml
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    1
    ·
    5 hours ago

    yadayada, more moronic ‘China baad’ propaganda.

  • fckreddit@lemmy.ml
    link
    fedilink
    English
    arrow-up
    9
    ·
    6 hours ago

    Yeah, because American LLMs are so immensely useful that people are throwing money at them.

  • The_Walkening [none/use name]@hexbear.net
    link
    fedilink
    English
    arrow-up
    1
    ·
    4 hours ago

    lol. love when people are panicking about this when it means that these things are basically interchangeable anyways. Didn’t someone at Google write a memo that was like “we’re kinda fucked b/c you can re-create this stuff with enough resources” like 2 years ago?

    • P03 Locke@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 hours ago

      Didn’t someone at Google write a memo that was like “we’re kinda fucked b/c you can re-create this stuff with enough resources” like 2 years ago?

      Basically, yes. They were specifically decrying the amount of open-sourcing they and their American competitors were doing, because capitalism, of course. Around this time, we had examples like StabilityAI’s StableDiffusion and Meta’s LLaMA as open-source models. And around this time, everybody else started closing their models, despite the fact that the research kept on going out in the open. StabilityAI kept their models open, mostly because they had no choice, but the attitude shifted towards profitability.

      So, China took the open-source mantle, and these open/closed lines are being drawn strictly around national divisions as this American vs. China slant. Which is mostly a diversion of the real battle.