Nvidia News
-
JH is arriving in Europe this weekend and is due to meet many Heads of State as well as present at a couple of conferences. Likely we will hear of further Sovereign AI/Token factory initiatives next week! There could be a giant EU project brewing to rival Stargate.
-
Nvidia’s next-gen Rubin GPU and Vera CPU chips will finish tape-out(the final stage of the design process) at TSMC in June and begin trial production, with sample chips in September, earliest, media report, citing unnamed supply chain sources. TSMC will make Rubin on N3P and CoWoS-L advanced packaging.
Sample chips will then be delivered to their ODM partners to start incorporating them into their various product designs. The time line, which would suggest mass production starting in the Q2/2026 range is approx 9 months ahead of their previously publicises roadmap-a staggering speed of architectural progress.
Given the R200 is rated at 1.8KW per chip I would expect full sized racks to have a TDP (thermal Design Power) of up to 180KW/rack and cost somewhere between $5-$6 million.
Note: The Rubin chip is packaged on the same CoWoS-L process as Blackwell which means we will not see the clunky ramp observed during the Hopper to Blackwell transition. Further, existing cooling systems (CDU rated to 250KW) allows built in headroom to manage thermals which should mitigate any cooling issues.
-
Hi ExIM
Actually the annual performance increase is much more creating exponential performance gains. By 2028 we should see up to 700X more performance that GB200 today. By 2031 about 1 million X. This is what Jensen is driving at when he says 'I want to reduce the cost of inference to almost zero'. You have machines that are so powerful they produce vastly more tokens for the same cost. Whilst there are other costs like power and bandwidth bottlenecks resulting in non linear cost drops, you will see todays 10 cents/1000 token cost fall to '.001 cents' or less. There is a direct relationship between Nr of tokens and the complexity of the given problem(to solve).
In AI, a token is a small unit of text, like a word, punctuation mark, or part of a word, that the system processes.
How is token generation growing:
Microsoft processed over 100 trillion tokens, a ~5× increase YoYGoogle’s monthly token volume increased by ~50× in the past year
AI Agents (“super agents”):
Per Barclays, AI super-agent users might consume 36 million to 356 million tokens/year per user -
Today
1 Billion? I didn't expect much given the state of our books.
Leaders he says....: The European Commission, led by President Ursula von der Leyen, announced the InvestAI initiative at the AI Action Summit in Paris . This initiative aims to mobilise €200 billion for AI investments across Europe over the next five years.
-
Ian Buck, NVIDIA's Vice President of Hyperscale and High-Performance Computing, made the comment....."You're only seeing a small glimpse in the public papers of what the true behind the scenes world class work has actually been able to do," about AI during a fireside chat at the Bank of America Global Technology Conference.
His statement, highlights that the public and even industry observers are only privy to a limited view of AI's current capabilities. This suggests that significant advancements, likely in areas like AI model performance, scalability, or novel applications, are being developed in private by leading tech companies, research labs, or NVIDIA itself. Buck's role at NVIDIA, a key player in AI hardware and software, implies he’s referring to cutting-edge work enabled by their technology, such as advanced GPU architectures or AI frameworks like CUDA.
I think this makes a lot sense and is very interesting. Why else would the industry pile a trillion USD+ into it. The unprecedented surge in spending and a race to scale.
BTW GPT 5 is due out during the summer.
-
Jensen is presenting a keynote in Paris this morning (10am). Expect some tasty DC deals with our European neighbours to drop.
-
Mistal and Nvidia re building an AI cloud-breaking news. Nvidia CEO talking to Macron(Paris later)
-
separated Jensen teased, stating full details will be announced on Friday. This is part of their Industrial Revolution 2.0 where all industries will be transformed . Designed in the Universe and automated via Robotics and Autonomous Vehicles (anything that moves). Be in no doubt. Heavy industry, design and logistics will change and Nvidia is a the very heart of it all.
-
-
Interesting thoughts from JH
We know demand is massive -this number equates to around 30GW. I thought 15GW. Obviously they can't supply it right now but they will over the next few years. Slow down, anyone?Asics are the TPUs used by AWS/GOOG and to an extent Meta. They are chips designed for one task only. Generally, whilst the market is big in $ terms, GPU $ is about 12X that of Asics . That won't change. Jensen thinks his GPUs are actually more cost effective, even when performing the tasks Asics chips were designed to do.
-
During GTC France this week Jensen mentioned that he would announce some data centre deals relating to Europe on Friday. Being the stalker that I am, I see he met German Chancellor Friedrich Merz. I'm speculating but this could be a massive German initiative, multi billions. Let's see what drops
-
Nvidia Europe DC deals announced today. When fully scaled it's quoted as 2GW across 4 countries incl Germany, Italy France, UK. Details are emerging but we know that 1GW is 40-50 billion.
-
Why we won't be rushing to buy AMD
. Not sure about their math 3-4% and 400 billion.
AMD claim GPU as fast as Nvidia (next year. .)-a lab chip which they can't make in volume, demand is weak due to 'No Cuda' and can only be networked to 10k chips vs Nvidia 100s of thousands. As we have said before 1 chip is meaningless . How good is your total solution. Still years behind.
I find it baffling why some investors hold both stocks. AMD are not a credible competitor imo
-
The Nvidia B40 (also known as the RTX Pro 6000D) is being positioned as the replacement for the H20 GPU in China, following U.S. export restrictions on advanced AI chips. It is based on Nvidia's Blackwell architecture and uses GDDR7 memory, allowing it to comply with trade controls.
Availability: Production began in June 2025, with general rollout expected in July and throughout Q3. Technically this will catch the current quarter((Q2) being May-July so potential for additional revenue.
Price: Estimated at $6,500–$8,000 USD per unit, making it more affordable than the H20, which exceeded $10,000 USD.
Sales potential: Analysts expect B40 shipments could reach $1–3 billion USD in revenue per quarter, depending on demand and licensing constraints.