Jensen Huang keynote at CES 2025 [video]
122 points by subharmonicon 5 days ago | 74 comments
  • behnamoh 5 days ago |
    I liked this part:

        "one small step at a time, and one giant leap, together."
    
    I didn't like this part:

        5090 for $2000, about $500 more than 4090 when it was announced.
    
    They didn't mention VRAM amount though, and I doubt it's more than 24GB. If Apple M4 Ultra gets close to 1.8 TB/s bandwidth of 5090, it'll crush GeForce once and for all (and for good).

    Also nitpick: the opening video said tokens are responsible for all AI, but that only applies to a subset of AI models...

    • jsheard 5 days ago |
      It's 32GB, that was leaked well before the event.

      > If Apple M4 Ultra gets close to 1.8 TB/s bandwidth of 5090

      If past trends hold (Ultra = 2x Max) it'll be around 1.1 TB/s, so closer to the 4090.

      • behnamoh 5 days ago |
        Jensen didn't talk about it. Maybe he knows it's embarrassingly low. Everyone knows nvidia won't give us more VRAM to avoid cannibalizing their enterprise products.
        • aseipp 11 hours ago |
          The official specs for the 5090 have been out for days on nvidia.com, and they explicitly state it's 32GB of GDDR7 with a 512-bit bus, for a total of 1.8TB/s of bandwidth.
          • behnamoh 9 hours ago |
            32GB is still not a lot.
            • BryanLegend 9 hours ago |
              "640k should be enough for anybody"
            • Jedd 5 hours ago |
              This feels like a weird complaint, given you started by saying it was 24GB, and then argued that the person who told you it was actually 32GB was making that up.
              • behnamoh 2 hours ago |
                Looks like Nvidia did a simple calculation:

                    32GB (for 5090) / 24GB (for 4090) ≃ 1.33
                
                Then multiply 4090's price by that:

                    $1500 × 1.33 ≃ $2000
                
                All else equal, this means that price per GB of VRAM stayed the same. But in reality, other things improved too (like the bandwidth) which I appreciate.

                I just think that for home AI use, 32GB isn't that helpful. In my experience and especially for agents, models at 32B parameters just start to be useful. Below that, they're useful only for simple tasks.

      • qingcharles 5 hours ago |
        It makes for a hard choice. M4 Ultra with say 128GB of RAM vs ... 2 x 5090 with 64GB of RAM for about the same price?

        More tokens/sec on the dual 5090, but way bigger model on the M4.

        Plus the dual 5090 might trip your breakers.

        • smallmancontrov an hour ago |
          If anyone in this thread had watched the linked video or even read a summary, they ought to be at least talking about the DIGITS announcement.

          128GB of VRAM for $3000.

          Slow? Yes. It isn't meant to compete with the datacenter chips, it's just a way to stop the embarrassment of being beaten at HPC workstations by apple, but it does the job.

    • toshinoriyagi 5 days ago |
      edit: 32GB is confirmed here https://www.nvidia.com/en-us/geforce/graphics-cards/50-serie...

      Supposedly, this image of a Inno3D 5090 box leaked, revealing 32GB of VRAM. It seems like the 5090 will be more of a true halo product, given the pricing of the other cards.

      https://www.techpowerup.com/330538/first-nvidia-geforce-rtx-...

    • ksec 13 hours ago |
      For what 5090 is offering with 32GB RAM I thought it is a pretty decent price comparatively to 4090. I thought the whole lineup is really well priced.
      • bn-l 5 hours ago |
        Good for image generation, inadequate for local llms (on its own).
    • headcanon 9 hours ago |
      Market price for a 4090 is ~$2700 today: https://stockx.com/nvidia-founders-geforce-rtx-4090-24gb-gra...

      When you have a retail price so far below "street" price, it just makes it harder to obtain and scalpers take a bigger cut. Raising the price to something more normal at least gives you more of a chance at the big-box store.

      • heyoni 9 hours ago |
        Or scalpers won’t be dissuaded and street price for a 5090 will be $3200 or more. $1500 was already an insane price tag for scalpers to pay but they did it anyways.
      • geokon 30 minutes ago |
        very naiive question... But why have a retail price at all and not just auction off batches to retailers (say 10 000 card at a time)

        Let the market set the price and only control how many cards you auction and when

  • magicalhippo 5 days ago |
    AMD barely mentioned their next-gen GPUs, NVIDIA came out swinging right from the start. AMD announced two new models which by their own cryptic slide wouldn't even compete with their current top-end. Then NVIDIA came and announced a 4090-performance GPU for $549...

    If that's not just hot hair from NVIDIA, I totally get the business decision from AMD but man, would love some more competition in the higher end.

    • behnamoh 5 days ago |
      AMD has strategically decided to abandon the enthusiasts market [0]. Really sad that nvidia's monopoly has got solidified even further. That's one of the reasons I contribute to Apple's MLX instead of CUDA.

      0: https://www.techradar.com/computing/gpu/amd-announces-new-ra...

      Update: typical HN behavior. someone is downvoting all my comments one by one...

      • jiggawatts 5 days ago |
        Someone with a lot of NVIDIA shares, no doubt…
        • blitzar 5 days ago |
          More likely someone holding the AMD bag

          edit: confirmed.

          • khazhoux 6 hours ago |
            Ad Hominem
      • ksec 13 hours ago |
        >Update: typical HN behavior. someone is downvoting all my comments one by one...

        You cant downvote older comments or comments older than 1 / 2 days.

    • Permik 5 days ago |
      It depends totally how you define "hot air" in this instance. Later in the keynote there's a mention how the $549 card achieves the 4090-like rendering performance with DLSS, i.e. not with raw graphics, number crunching horsepower.

      Personally? It's a no for me dawg, DLSS unfortunately doesn't actually replace the need for the raw GPGPU crunch.

      For the average layman and a consumer? Nvidia will be selling those GPU's like hotcakes.

      • y-c-o-m-b 11 hours ago |
        I kind of agree with this. On the one hand when DLSS works well, it's amazing. I didn't notice any issues at all playing Senua's Sacrifice and Senua's Saga. They were both as good as playing on everything maxed without DLSS. On the other hand when using it on Jedi Survivor, it's a very janky experience. You can clearly see artifacts and it trying to "self-correct" when using DLSS vs without DLSS.
        • kcb 9 hours ago |
          DLSS upscaling is good. On a 4k display I can't see any visual reason not to run the Balanced preset. DLSS frame gen on the other hand I'm much more skeptical of. Every time I've tried it something just feels off.
          • heyoni 9 hours ago |
            DLSS in cyberpunk has a lot of issues. Faces look weird, tons of ghosting, hair and particle effects looking very jagged.
            • josephg 8 hours ago |
              They’re claiming DLSS4 (only available on their new cards) fixes a lot of these issues. But we’ll have to wait for the cards to get in the hands of reviewers before we’ll know for sure. I'm still pretty skeptical.

              That said, if you read between the lines, it looks like the new 5090 is about ~40% faster than the 4090 at rasterisation. That’s a solid inter generational improvement.

      • qingcharles 5 hours ago |
        Creating the frames with raw GPU is better because they are exactly what the game engine wants to render. But that's unbelievably expensive.

        DLSS is absolutely the answer for gaming for the masses. DLSS will continue to improve and is far, far cheaper at creating intermediary frames than rendering them for real.

        I buy my cards for AI first and gaming second, though. So DLSS is little use to me. 5090 is a bit of a weak improvement over the 4090 for me, but here we are.

    • nomel 5 days ago |
      > Then NVIDIA came and announced a 4090-performance GPU for $549

      Never trust vendor performance claims (these specifically rely on 3x frame generation), and never assume cards will be available at MSRP.

    • ekianjo 11 hours ago |
      > Then NVIDIA came and announced a 4090-performance GPU for $549...

      That's obviously a marketing hoax. Spray DLSS and other tricks and you can make this kind of claims while the raw power is clearly on the side of the 4090

    • uluyol 9 hours ago |
      AMD is not ready to have a full announcement for their new consumer GPUs: https://morethanmoore.substack.com/p/where-was-rdna4-at-amds...
      • magicalhippo 6 hours ago |
        I knew this at the time of writing, should have worded myself better.
    • caseyy 8 hours ago |
      4090 (no frame generation) performance with a $549 card (4x frame generation), isn’t it?

      In other words, the same performance at somewhere in the range of 2x-3.5x lighter workload. In other words, vastly different performance under same workload. It is veeery on the edge of false advertisement for them to omit the information in parentheses until later in the presentation. :)

      That line should have been “4090 frame rates on budget hardware thanks to DLSS 4” or something like that.

      • magicalhippo 6 hours ago |
        From what I've read and heard, the 4090 was with single-frame generation.

        Anyway, my point with those lines was to highlight just how defensive AMD was vs how offensive NVIDIA was.

  • ksec 12 hours ago |
    I actually got a lot more upbeat about the potential of AI after watching this keynote more so than any other demonstration or presentation.

    Blackwell is still based on N4, a derivative of 5nm. We know we will have N3 GPU next year, and they should be working with TSMC on capacity planning for N2. Currently Blackwell is pretty much limited by capacity of either TSMC, HBM or packaging. And unless the AI hype dies off soon ( which I doubt will happen ), we should still have at least another 5 years of these GPU improvements. We will have another 5 - 10x performance increase.

    And they have foreshadowed their PC entry with MediaTek Partnership. ( I wonder why they dont just acquired Mediatek ) and may be even Smartphone or Tablet with Geforce GPU IP.

    The future is exciting.

    • AlotOfReading 12 hours ago |
      The same industry opposition that killed the ARM acquisition would kill a mediatek acquisition.
      • satellite2 9 hours ago |
        The acquisition of ARM would have put Nvidia in the position of being able to impose to competitors uncompetitive restrictions on a quasi monopolistic market. Is it the case with MediaTek?
        • AlotOfReading 8 hours ago |
          Mediatek is the largest smartphone SoC manufacturer, so yes. They're a supplier for Apple and the biggest competitor to Qualcomm, the two companies that were the most vocal in opposing the ARM acquisition.
    • brookst 12 hours ago |
      My guess is mediatek’s margins are much lower so an acquisition would tank Nvidia’s stock price by leading to lower returns. That and/or not wanting the distraction of operating mediatek’s many businesses that aren’t aligned to Nvidia core competence.
    • bloomingkales 12 hours ago |
      There was something eerily epic about his assertions. To suggest that they are all of those companies is pretty wild.

      AI is bewitching.

      • WeylandYutani 5 hours ago |
        Prepare for Butlerian Jihad.
    • dehrmann 11 hours ago |
      I was taking these sort of improvements for granted. They'll certainly make ai cheaper, but most of the cost is up-front for generative models, and the models don't seem to be getting linearly better.
      • ksec 43 minutes ago |
        >I was taking these sort of improvements for granted.

        A lot of people would think so. In reality had Smartphone ( iPhone ) not taken off. The PC market along would not be able to sustain the cost of innovation and Moore's Law would have ended at 14nm and we would have to wait 3 to 4 years for every node. Smartphone along with all the adjacent work on Internet infrastructure scaled us to 3nm. I wrote about both of these in 2014 and was expecting some slow down by 3nm. Now AI will provide the Capex to scale us to 1nm and beyond.

    • kcb 9 hours ago |
      A major chip acquisition by Nvidia will probably never get global regulatory approval. I guess the question is what does Mediatek provide in the partnership. Nvidia is well versed in developing ARM SoCs already.
      • bgnn 9 hours ago |
        5g, wifi, bluetooth etc. what makes a mobile processor monile processor other than the CPU
        • knowitnone 7 hours ago |
          yeah, those are just modules that anybody can just buy. NVIDIA already makes jetsons which are basically SBCs with decent GPU. so again, what does Mediatek provide?
          • downrightmike 4 hours ago |
            Monopoly accusation shield. "But we're sharing the tech!"
      • ksec 31 minutes ago |
        > what does Mediatek provide?

        Market access, shares and relationship to all vendors, most of them never worked with Nvidia. Ready made 5G solutions. Know how in product category that is not as simple as buying IP and making a chip. Shipping Nvidia CUDA GPU IP to all other domains. For acquisition benefits they share the development cost of leading node and volume. Which is part of the wider SemiConductor industry strategy aka Broadcom.

        But yes. I guess China would be the first one against it. I never quite understand why Two Company in country A and B would need Country C 's approval for them to merge or be acquired.

    • VHRanger 7 hours ago |
      How would shrinking the processor size help if the current GPU generation is already completely bottlenecked by VRAM bandwidth saturation?

      We aren't seeing nearly the same gains on VRAM bandwidth as we are on compute bandwidth

      • ffsm8 7 hours ago |
        Well, at least this generation almost doubles that bandwidth, right?
        • jsheard 6 hours ago |
          The flagship is an outlier there since it went from 384bit to 512bit memory. There was no such bus width increase for the rest of the stack so the gains are more modest, 4080 to 5080 is only a 34% improvement.
          • may_mccheese 5 hours ago |
            the solution is people will figure how to burn their models onto ASICs cheaply. apple model on iphone, google model on android, etc. This is tantalizing to businesses (1) you have YOUR company's model on lockdown as you see fit (2) iterations or improvements to the model mean you'll need to buy buy buy
    • ninetyninenine 5 hours ago |
      Is GPU improvement is driven more by gaming then by AI hype? Gaming is the biggest industry and there is real money coming from that. Does speculative money from VCs actually overshadow actual money spent by consumers?

      I know stock prices is driven by AI hype but how much does it actually effect the bottom line of Nvidia? I think GPU improvement happens regardless of AI.

      • l33tman 5 hours ago |
        "NVIDIA’s financial report reveals that its data center business revenue in FY2Q25 grew by 154% YoY, outpacing other segments and raising its contribution to total revenue to nearly 88%."

        Gaming almost doesn't even register in Nvidias revenue anymore.

        But I do think Jensen is smart enough to not drop gaming completely, he knows the AI hype might come and go and competitors might finally scrounge up some working SDKs for the other platforms.

      • kbolino 5 hours ago |
        ML was big before LLMs and nVidia was already making a killing from selling expensive GPUs that would never draw a single polygon ca 2015. They've been hobbling FP64 (double precision) support in cheaper "consumer" GPUs, to prevent their use in most data centers, for a long time too.
      • AnotherGoodName 5 hours ago |
        https://s201.q4cdn.com/141608511/files/doc_financials/2025/Q...

        Datacenter revenue alone is ~10x of gaming. The datacenter revenue is thought to have literally ~100x the earnings all up (H100 and 4090 have similar transistor counts but the H100 sells for over $30k while the 4090 sells for $2k which indicates huge margins).

        Gaming is pretty much insignificant for nvidia. That’s why nvidias stock has 10x’ed recently and their PE looks better now than it did 5 years ago despite that stock increase. They found a new market that dwarfs their old market.

      • Mr_Minderbinder 4 hours ago |
        NVIDIA’s net income grew ~580% year-on-year in their 2024 fiscal year. FY2025 is on track for 100%+ growth, essentially 14x in the last 2 years. This is not coming from gaming, “AI hype” is having a huge effect on NVIDIA’s bottom line.
        • kristianp 3 hours ago |
          It all depends on whether AI companies can continue to find significant improvements to their models this year. Are transformers reaching their limits? Can researchers find the next level of performance or are we headed for another AI slump?
      • jhanschoo an hour ago |
        Interpreting your question about "GPU improvement" from a product perspective, my read is that NVIDIA is of course targeting AI applications and the datacenter. To that end it just focuses on silicon that makes most sense for AI compute, and not so much for gaming.

        Of course, the GPUs for the datacenter and for gaming are the same designs, so my read is that in gaming NVIDIA makes up for lack of actual performance for traditional rendering by pushing technologies that can utilize tensor cores like AI upscaling, frame prediction, ray tracing + denoising, etc.., that don't actually contribute to game graphics as much as they could have if they did an architecture tailored to gaming needs instead, with the technologies that they have. It's also sexier in theory to talk about exclusive AI-powered technologies proprietary to NVIDIA than just better performance.

  • contingencies 7 hours ago |
    As someone who has gone pretty deep in to robotics over the last 9 years I skipped right to the physical AI portion, and wasn't impressed.

    This has been stated on HN in most robotics threads, but the core of what they show, once again, is content generation, a feature largely looking for an application. The main application discussed is training data synthesis. While there is value in this for very specific use cases it's still lipstick ("look it works! wow AI!") on a pig (ie. non-deterministic system being placed in a critical operations process). This embodies one of the most fallacious, generally unspoken assumptions in AI and robotics today - that it is desirable to deal with the real world in an unstructured manner using fuzzy, vendor-linked, unauditable, shifting sand AI building blocks. This assumption can make sense for driving and other relatively uncontrolled environments with immovable infrastructure and vast cultural, capital and paradigm investments demanding complex multi-sensor synthesis and rapid decision making based on environmental context based on prior training, but it makes very little sense for industrial, construction, agricultural, rural, etc. Industrial is traditionally all about understanding the problem or breaking it in to unit operations, design, fabricate and control the environment to optimize the process for each of those in sequence, and thus lowering the cost and increasing the throughput.

    NVidia further wants us to believe we should buy three products from them: an embedded system ("nano"), a general purpose robotic system ("super") and something more computationally expensive for simulation-type applications ("ultra"). They claim (with apparently no need to proffer evidence whatsoever) that "all robotics" companies need these "three computers". I've got news for you: we don't, this is a fantasy, and limited if any value add will result from what amounts to yet another amorphous simulation, integration and modeling platform based on broken vendor assumptions. Ask anyone experienced in industrial, they'll agree. The industrial vendor space is somewhat broken and rife with all sorts of dodgy things that wouldn't fly in other sectors, but NVidia simply ain't gonna fix it with their current take, which for me lands somewhere between wishful thinking and downright duplicitous.

    As for "digital twins", most industrial systems are much like software systems: emergent, cobbled together from multiple poor and broken individual implementations, sharing state across disparate models, each based on poorly or undocumented design assumptions. This means their view of self-state, or "digital twin", is usually functionally fallacious. Where "digital twins" can truly add value is in areas like functional safety, where if you design things correctly you avoid being mired in potentially lethally disastrous emergent states from interdependent subsystems that were not considered at subsystem design, maintenance or upgrade time because a non-exhaustive, insufficiently formal and deterministic approach was used in system design and specification. This very real value however hinges on the value being delivered at design time, before implementation, which means you're not going to be buying 10,000 NVidia chips, but most likely zero.

    So my 2c is the Physical AI portion is basically a poorly founded forward-looking application sketch from what amounts to a professional salesman in a shiny black crocodile jacket at a purchased high-viz keynote. Perhaps the other segments had more weight.

    • truetraveller 22 minutes ago |
      Upvoted for a real wall-of-text in an AI-generation era.
  • dang 2 hours ago |
    Related ongoing thread:

    Ask HN: Pull the curtain back on Nvidia's CES keynote please - https://news.ycombinator.com/item?id=42670808 - Jan 2025 (2 comments)

  • sliken an hour ago |
    The Nvidia Project Digits looked rather interesting. Much like a Apple studio ultra. 2 dies, "high" bandwidth unified memory, 20 cores, tiny mac mini sized case, and 2 SFP+ for 2 connectX ports (unsure if that's IB or ethernet).

    Claim it will be quite good at AI (1 Tflop of fp4), but sadly don't mention the memory bandwidth. It's somewhere in the range of awesome to terrible depending on the bandwidth.

    • seanmcdirmid an hour ago |
      Am looking forward to hearing about this in the next few months. Apple has been super late in updating the macstudio, and even the old models are still super expensive, even used.
    • teleforce an hour ago |
      Previous HN post on Project Digits (494 comments) [1].

      [1] Nvidia's Project Digits is a 'personal AI supercomputer'

      https://news.ycombinator.com/item?id=42619139