• ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
    link
    fedilink
    English
    arrow-up
    9
    ·
    11 hours ago

    I very much expect the AI bubble to pop in the near future, and DeepSeek goes a long way towards ensuring this happens by completely breaking the business model of western AI companies. Nvidia and AI basically carried the whole market, and now it turns out that you can do the same thing with a fraction of the resources. That means nobody needs huge volumes of latest and greatest Nvidia chips anymore.

    • MLRL_Commie [comrade/them, he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      5
      ·
      10 hours ago

      But what is to limit Nvidia from just shifting to “wow now that more can be done with less, we will do even more with EVEN MORE CARDS” and apply the improvements from Deepseek to their own models while continuing to ramp up in the same way?

      Eventually there will be a pop, but I genuinely don’t get how this ushers in that pop. I guess the diminishing returns of these systems is hastened? As in, adding more cards with more resources will continuously have less effect. But companies will for now still want even the smallest edge by putting Deepseek-style-improved-AI on even bigger data centers.

      • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
        link
        fedilink
        English
        arrow-up
        5
        ·
        10 hours ago

        That’s what OpenAI thought originally when they started working on ChatGPT5, they figured they’d just make the model bigger and it’s going to do more. Turns out that making the model bigger doesn’t actually produce better results. We’re also at a point now where most of the publicly available information has been scraped as well. Now the focus is turning towards improving algorithms for making sense of the data as opposed to just stuffing more data into the model. And this is a problem for Nvidia because current generation of chips is already good enough for doing this.

        Of course, people will find ways to utilize more processing power as is always the case. But at least in the near term, this is no longer the bottleneck.