• leisesprecher@feddit.org
    link
    fedilink
    arrow-up
    16
    ·
    3 days ago

    Deepseek showed that actually putting thought into the architecture achieves much more than just throwing more hardware at the problem.

    This means a) there will be much less demand for hardware, since much more could be run locally on regular consumer devices. And b) the export restrictions don’t really work and instead force China to create actually better models.

    That means, a lot of the investments into the thousands of AI companies are in jeopardy.

    • artificialfish@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 hours ago

      I think “just writing better code” is a lot harder than you think. You actually have to do research first you know? Our universities and companies do research too. But I guarantee using R1 techniques on more compute would follow the scaling law too. It’s not either or.

    • bobs_monkey@lemm.ee
      link
      fedilink
      arrow-up
      9
      ·
      3 days ago

      Realistically, the CCP is probably throwing a lot of money at developers to get something good going and available, and US companies are whining about how it’s not fair. The fact of the matter is that a solid product is available for much cheaper, and US companies are now screaming foul. Guess what, a superior product made of good code (people) beats out just throwing money at hardware, who’d’ve gone an thunk it.

      • leisesprecher@feddit.org
        link
        fedilink
        arrow-up
        12
        ·
        3 days ago

        What I find really fascinating here is that obviously openAI, Meta, etc. seem to be structurally incapable of actually innovating at this point.

        I mean, reducing training costs by literally an order of magnitude just by writing better software is astonishing and shows how complacent the large corporations have gotten.

        • artificialfish@programming.dev
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          4 hours ago

          Meta? The one that released Llama 3.3? The one that actually publishes its work? What are you talking about?

          Why is it so hard to believe that deepseek is just yet another amazing paper in a long line of research done by everyone. Just because it’s Chinese? Everyone will adapt to this amazing innovation and then stagnate and throw compute at it until the next one. That’s how research works.

          Not to mention China has billions of more people to establish a research community…