• Phoenixz@lemmy.ca
    link
    fedilink
    English
    arrow-up
    7
    ·
    7 hours ago

    And none of us will be allowed to have them

    Only datacenters and only fortune 500 companies will be able to use anything Nvidia

    • Earthman_Jim@lemmy.zip
      link
      fedilink
      English
      arrow-up
      24
      arrow-down
      1
      ·
      20 hours ago

      Yeah, I wonder how long it will take them to clue in that no one wants to trade gaming for an AI fucking girlfriend ffs…

      • setsubyou@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        14 hours ago

        I mean if they came with a cool android body we could talk about it. It should at least be able to do cleaning and cooking. Otherwise my wife won’t like it.

      • Ebby@lemmy.ssba.com
        link
        fedilink
        English
        arrow-up
        8
        ·
        19 hours ago

        Idk… I’m a tad excited to buy one for peanuts on eBay in a couple years for a local smart home upgrade. Heck, when the bubble pops, maybe they can sell power from all those generators back to the city and lower our utility bills too. /S

        • foggenbooty@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 hours ago

          I know you put /S, but for other people that read this it will not be an option. These only work in specialised servers that you will not he able to run at home (unless you’re a mad scientists type).

        • Earthman_Jim@lemmy.zip
          link
          fedilink
          English
          arrow-up
          2
          ·
          19 hours ago

          They might be useful for rendering, and I’d love to see how smoothly Teardown Game could run with all those cores.

  • RegularJoe@lemmy.worldOP
    link
    fedilink
    English
    arrow-up
    19
    arrow-down
    2
    ·
    21 hours ago

    Nvidia’s Vera Rubin platform is the company’s next-generation architecture for AI data centers that includes an 88-core Vera CPU, Rubin GPU with 288 GB HBM4 memory, Rubin CPX GPU with 128 GB of GDDR7, NVLink 6.0 switch ASIC for scale-up rack-scale connectivity, BlueField-4 DPU with integrated SSD to store key-value cache, Spectrum-6 Photonics Ethernet, and Quantum-CX9 1.6 Tb/s Photonics InfiniBand NICs, as well as Spectrum-X Photonics Ethernet and Quantum-CX9 Photonics InfiniBand switching silicon for scale-out connectivity.

    • TropicalDingdong@lemmy.world
      link
      fedilink
      English
      arrow-up
      20
      ·
      21 hours ago

      288 GB HBM4 memory

      jfc…

      Looking at the spec’s… fucking hell these things probably cost over 100k.

      I wonder if we’ll see a generational performance leap with LLM’s scaling to this much memory.

      • AliasAKA@lemmy.world
        link
        fedilink
        English
        arrow-up
        14
        ·
        19 hours ago

        Current models are speculated at 700 billion parameters plus. At 32 bit precision (half float), that’s 2.8TB of RAM per model, or about 10 of these units. There are ways to lower it, but if you’re trying to run full precision (say for training) you’d use over 2x this, something like maybe 4x depending on how you store gradients and updates, and then running full precision I’d reckon at 32bit probably. Possible I suppose they train at 32bit but I’d be kind of surprised.

        Edit: Also, they don’t release it anymore but some folks think newer models are like 1.5 trillion parameters. So figure around 2-3x that number above for newer models. The only real strategy for these guys is bigger. I think it’s dumb, and the returns are diminishing rapidly, but you got to sell the investors. If reciting nearly whole works verbatim is easy now, it’s going to be exact if they keep going. They’ll approach parameter spaces that can just straight up save things into their parameter spaces.

        • in_my_honest_opinion@piefed.social
          link
          fedilink
          English
          arrow-up
          6
          ·
          18 hours ago

          Sure, but giant context models are still more prone to hallucination and reinforcing confidence loops where they keep spitting out the same wrong result a different way.

          • AliasAKA@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            ·
            9 hours ago

            Sorry, I’m not saying that’s a good thing. It’s not just the context that’s expanding, but the parameter of the base model. I’m saying at some point you just have saved a compressed version of the majority of the content (we’re already kind of there) and you’d be able to decompress it even more losslessly. This doesn’t make it more useful for anything other than recreating copyrighted works.

            • in_my_honest_opinion@piefed.social
              link
              fedilink
              English
              arrow-up
              1
              ·
              6 hours ago

              Ah I see, however you do bring up another point. I really think we need a true collection of experts able to communicate without the need for natural language and then a “translation” layer to output natural language or images to the user. The larger parameters would allow the injection of experts into the pipeline.

              Thanks for the clarification, and also for the idea. I think one thing we can all agree on is that the field is expanding faster than any billionaire or company understands.

              • AliasAKA@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                ·
                5 hours ago

                This already happens intrinsically in the models. The tokens are abstracted in the internal layers and only translated in the output layer back to next token prediction. Training visual models is slightly different because you’re not outputting tokens but pixel values (or possibly bounding boxes or edges, but not usually; conversely if not generative you may be predicting labels which could theoretically be in token space).

                The field itself is actually fairly stagnant in architecture. It’s still just attention layers all the way down. It’s just adding more context length and more layers and wider layers while training on more data. I personally think this approach will never achieve AGI or anything like it. It will get better at perfectly reciting its training data, but I don’t expect truly emergent phenomena to occur with these architectures just because they’re very big. They’ll be decent chatbots, but we already have that, and they’ll just consumer ever more resources for vanishingly small improvements (and won’t functionally improve any true logical capability beyond regurgitating logical paths already trodden in their training data but in a very brittle way, because they do not actually understand the logic or why the logic is valid, they have no true state model of objects which are described in the token space they’re traversing probabilistically).

                • in_my_honest_opinion@piefed.social
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  edit-2
                  4 hours ago

                  will never achieve AGI or anything like it

                  On this we absolutely agree. I’m targeting a more efficient interactive wiki essentially. Something you could package and have it run on local consumer hardware. Similar to this https://codeberg.org/BobbyLLM/llama-conductor but it would be fully transform native and there would only need to be one LLM for interaction with the end user. Everything else would be done in machine code behind the scenes.

                  I was unclear I guess, I was talking about injecting other models, running their prediction pipeline for the specific topic, and then dropped out of the window to be replaced by another expert. This functionality handled by a larger model that is running the context window. Not nested models, but interchangeable ones dependent on the vector of the tokens. So a qwq RAG trained on python talking to a qwen3 quant4 RAG trained on bash wrapped in deepseekR1 as the natural language output to answer the prompt “How do I best package a python app with uv on a linux server to run a backend for a …”

                  Currently this type of workflow is often handled with MCP servers from some sort of harness and as I understand it those still use natural language as they are all separate models. But my proposal leverages the stagnation in the field and leverages it as interoperability.

      • in_my_honest_opinion@piefed.social
        link
        fedilink
        English
        arrow-up
        4
        ·
        18 hours ago

        Fundamentally no, linear progress requires exponential resources. The below article is about AGI but transformer based models will not benefit from just more grunt. We’re at the software stage of the problem now. But that doesn’t sign fat checks, so the big companies are incentivized to print money by developing more hardware.

        https://timdettmers.com/2025/12/10/why-agi-will-not-happen/

        Also the industry is running out of training data

        https://arxiv.org/html/2602.21462v1

        What we need are more efficient models, and better harnessing. Or a different approach, reinforced learning applied to RNNs that use transformers has been showing promise.

        • TropicalDingdong@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          17 hours ago

          Yeah I’ve read that before. I don’t necessarily agree with their framework. And even working within their framework, this article is about a challenge to their third bullet.

          I’m just not quite ready to rule out the idea that if you can scale single models above a certain boundary, you’ll get a fundamentally different/ novel behavior. This is consistent with other networked systems, and somewhat consistent with the original performance leaps we saw (the ones I think really matter are ones from 2019-2023, its really plateaued since and is mostly engineering tittering at the edges). It genuinely could be that 8 in a MoE configuration with single models maxing out each one could actually show a very different level of performance. We just don’t know because we just can’t test that with the current generation of hardware.

          Its possible there really is something “just around the corner”; possible and unlikely.

          What we need are more efficient models, and better harnessing. Or a different approach, reinforced learning applied to RNNs that use transformers has been showing promise.

          Could be. I’m not sure tittering at the edges is going to get us anywhere, and I think I would agree with just… the energy density argument coming out of the dettmers blog. Relative to intelligent systems, the power to compute performance (if you want to frame it like that) is trash. You just can’t get there in computation systems like we all currently use.

      • boonhet@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        4
        ·
        20 hours ago

        LLMs can already use way more I believe, they don’t really run them on a single one of these things.

        The HBM4 would likely be great for speed though.

      • panda_abyss@lemmy.ca
        link
        fedilink
        English
        arrow-up
        3
        ·
        19 hours ago

        Yeah they’re going to cost as much as a house.

        I think we’ll see much larger active portions of larger MOEs, and larger context windows, which would be useful.

        The non LLM models I run would benefit a lot from this, but I don’t know of I’ll ever be able to justify the cost of how much they’ll be.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      33
      arrow-down
      3
      ·
      19 hours ago

      You’re in a community called “Technology” and it’s got a bunch of upvotes, so us, presumably.

      • Earthman_Jim@lemmy.zip
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        5
        ·
        18 hours ago

        The news is important, but when it comes to user-end AI in general, big fucking meh.

  • fubarx@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    18 hours ago

    Question is, how long before it makes it to the next DGX Spark? Some people don’t have $10B to burn.