Summary: Meta, led by CEO Mark Zuckerberg, is investing billions in Nvidia’s H100 graphics cards to build a massive compute infrastructure for AI research and projects. By end of 2024, Meta aims to have 350,000 of these GPUs, with total expenditures potentially reaching $9 billion. This move is part of Meta’s focus on developing artificial general intelligence (AGI), competing with firms like OpenAI and Google’s DeepMind. The company’s AI and computing investments are a key part of its 2024 budget, emphasizing AI as their largest investment area.

  • @[email protected]
    link
    fedilink
    English
    210 months ago

    Pretty sure they’ll be given insight into the roadmap for that price, and be able to place speculative orders on upcoming generations.

    • @Rapidcreek
      link
      English
      010 months ago

      I used to present those roadmaps. They change too.

        • @Rapidcreek
          link
          English
          010 months ago

          The article says “by the end. Of the year” they will spend billions

          • @[email protected]
            link
            fedilink
            English
            010 months ago

            “spend billions” does not equal “hand over cash and take home GPUs”. It’ll mean a contract worth that amount with delivery terms defined over time. Even over the course of a year there’s likely to be newer product than Lovelace.

            • @Rapidcreek
              link
              English
              010 months ago

              When you get product you pay for it. Spending means paying for it. You may have a contract for future product, but you don’t pay for the future product in advance as SOX rules kick in. Commonly, a chip development cycle can be at least 10 months.