Our AI Future According to Nvidia - Sync #511
I hope you enjoy this free post. If you do, please like ❤️ or share it, for example by forwarding this email to a friend or colleague. Writing this post took around eight hours to write. Liking or sharing it takes less than eight seconds and makes a huge difference. Thank you! Our AI Future According to Nvidia - Sync #511Plus: humanoid robots show off their acrobatic skills; OpenAI o1-pro; Apple's dire Siri situation; Figure opens factory for humanoid robots; Claude can now search the web; how to reprogram life
Hello and welcome to Sync #511! This week, Nvidia hosted its GTC conference, where the company showcased not only its latest lineup of GPUs, AI supercomputers, and AI desktop computers, but also outlined its vision of the AI future—which we will examine in detail in this week’s issue of Sync. Elsewhere in AI, OpenAI has released its most expensive model to date—o1-pro. Meanwhile, Claude can now search the internet, Meta is being sued in France, leaked meetings reveal how dire Apple’s Siri situation is, and China will require AI-generated content to be labelled as such. Over in robotics, Boston Dynamics’ Atlas and Unitree G1 show off their acrobatic skills, Figure has opened a factory for humanoid robots, and Japanese robots have gained a sense of smell. Additionally, a massive AI analysis has identified genes related to brain ageing—and drugs to slow it down—and the original AlexNet algorithm has been recovered and made open source. I hope you’ll enjoy this rather long issue of Sync! Our AI Future According to NvidiaWatching Jensen Huang’s GTC 2025 opening keynote felt like peeking into the future. In a keynote that spanned from next-generation GPUs and datacentre racks to humanoid robots, Huang laid out a vision filled with AI factories, digital twins, autonomous robots, and AI agents, with Nvidia positioned at the centre of the AI era. At this year’s GTC, we witnessed the debut of Blackwell Ultra, a GPU optimised for inference at scale; the unveiling of Rubin, a 3nm architecture that triples performance over Blackwell; and a glimpse of Feynman, Nvidia’s post-Rubin chip. Nvidia also leaned heavily into AI robotics with Groot N1, an open foundation model for humanoids, and introduced software libraries for various industries—from biotech to materials science and physics—to accelerate research and progress using its hardware. But Nvidia’s ambitions go far beyond faster chips. It doesn’t just want to power AI—it wants to define the infrastructure, the tools, and the platforms that bring it to life everywhere. While the company has already cemented its dominance in AI training, the next frontier—AI inference—is still up for grabs. To stake its claim, Nvidia isn’t just building better GPUs. It’s building an entire ecosystem—hardware, software, systems, and frameworks—designed to scale intelligence across every industry. At GTC 2025, Jensen Huang laid out exactly what that future could look like. Nvidia’s Big Vision—AI Factories, Digital Twins, Physical AIThe key concept woven into Nvidia’s vision of the future is the idea of AI factories. AI factories are Nvidia’s term for next-generation data centres purpose-built to produce intelligence, not just process data. Think of them as factories for AI models and agents—massive, highly optimised computing facilities where models are trained, fine-tuned, and deployed at scale. These massive facilities, requiring at least a gigawatt of power (roughly equivalent to the output of a full-scale nuclear power plant), will simulate real-world environments, run digital twins, or host AI agents that continuously learn, evolve, and interact with the physical or virtual world, all powered by Nvidia next-generation hardware and software. These AI factories won’t produce cars, semiconductors, or steel. Instead, they’ll produce tokens—the atomic units of AI output. These tokens will then be transformed into code, music, scientific papers, engineered parts, and even new proteins. Tokens that will power copilots, autonomous agents, robots, and entire industries. As Huang said, we are at an inflection point. AI is no longer a curiosity or a prototype in the lab. It’s now capable of contributing meaningfully to fields like biology, chemistry, physics, and engineering—writing code, designing materials, and even simulating new drugs or molecular structures before they exist in the real world. Huang likened this shift to a new computing paradigm: simulate everything before you make anything. This is where digital twins come in—virtual models of real-world systems. Nvidia’s Omniverse platform is expanding rapidly to power simulations for everything from data centres and wind turbines to cars, robots, and even entire physical facilities like steel mills. These digital twins, running in AI factories, allow companies to design, test, and optimise systems in simulation before building anything physically. It’s a dramatic acceleration of the engineering process—and one that saves time, energy, and money.
At the heart of this transformation is Nvidia’s commitment to computing everything, everywhere. With each new GPU generation, from Hopper to Blackwell to Rubin and beyond, Nvidia is promising exponential leaps in performance. All this computing power will be for nothing if there is no way of using it. That’s why, alongside new GPUs, AI supercomputers and desktop computers, Nvidia is also releasing a suite of specialised libraries using CUDA, a parallel computing platform and programming model developed by Nvidia that allows developers to use Nvidia GPUs to accelerate computing tasks. Nvidia offers specialised software frameworks for medicine, climate, energy, and materials, all tightly coupled with its hardware.
And then there’s physical AI—the idea that intelligence shouldn’t just live in the cloud, but walk, talk, and interact with the world. From the Groot N1—an open humanoid robot foundation model—to robotics simulators in Omniverse, Nvidia is laying the foundation for robots that reason, plan, and manipulate objects in the real world. This is the vision: a world where AI doesn’t just assist—it builds. Where tokens become inventions. Where simulation replaces trial and error. Where gigawatt-scale data centres hum away, producing the intelligence that drives the next era of human progress. This ambitious vision of the future will require multiple gigawatt AI factories providing enormous amounts of computing power. These AI factories will be powered by powerful, next-generation GPUs, which Nvidia announced at GTC 2025. Nvidia's Roadmap – GPUs, AI Supercomputers and Desktop AI ComputersThe first of this next-generation hardware revealed at GTC 2025 was Blackwell Ultra, an evolution of last year’s Blackwell architecture. While it maintains the same 20 petaflops of FP4 performance as its predecessor, Blackwell Ultra expands memory from 192GB to 288GB of HBM3E, boosting capacity for large-scale inference tasks. Nvidia also rearchitected the SM (streaming multiprocessor) design and improved the MUFU unit (used for softmax operations in transformers), making it 2.5x faster—a crucial update for attention-heavy models. Next was Rubin—Nvidia’s first GPU built on a 3nm process, slated for release in 2026. Named after Vera Rubin, an American astronomer who provided critical evidence for the existence of dark matter, the new architecture will feature dual reticle-sized compute dies (in other words, very big chips) and deliver up to 50 FP4 petaflops, more than double the performance of Blackwell. Rubin’s successor, Rubin Ultra, is expected in 2027 and will pack four GPU dies and 1TB of HBM4E memory, delivering 100 petaflops per package. Nvidia also teased Feynman, its next architecture after Rubin, expected in 2028, built around the custom “Vera” CPU and designed for extreme-scale workloads. These GPUs are not standalone chips—they’re part of a full-stack system strategy. Nvidia also unveiled new rack-scale systems like NVL144 and hinted at future NVL576 and NVL1152 Kyber rack configurations, which push GPU count, interconnect bandwidth, and power density to new levels.
Apart from new GPUs and data centre hardware, Nvidia also announced desktop AI computers in the form of DGX Spark and DGX Station. Previously known as Project DIGITS, DGX Spark is a tiny gold box packing 1 PFLOPS of computing power and 128 GB of unified RAM for about $3,000. Its bigger brother—DGX Station—is the size of a typical desktop computer, offering 20 times more computing power than DGX Spark and 784 GB of unified memory. DGX Station will be available later this year. Its price hasn’t been announced yet, but I wouldn’t be surprised if it is more than $10,000. Both DGX Spark and DGX Station are designed with AI developers, researchers, and data scientists in mind, who can run massive neural networks—up to 200 billion parameters in the case of DGX Spark—locally and develop, test, and validate AI models more quickly. In addition to DGX Spark and DGX Station, Nvidia has also introduced the RTX PRO 6000 Blackwell, which it calls the “most powerful desktop GPU ever created.” Built on Nvidia's Blackwell architecture and equipped with 96GB of GDDR7 memory, the RTX PRO 6000 offers 4000 TOPS of AI performance or 125 TFLOPS of single-precision performance. However, all this computing power will demand up to 600W. This GPU is aimed at developers, researchers, and creators who need massive local computing power for tasks such as model development, 3D rendering, or running simulations.
Nvidia Won the Training Phase, But Can It Win Inference?There’s no question that Nvidia owns the AI training market. From OpenAI and Anthropic to Meta and Microsoft, nearly every large language model and cutting-edge AI system is trained on Nvidia hardware. Nvidia dominates the space—at least for now. But as the AI landscape shifts from training massive models to inference—running them—the rules of the game will change. And the question now is: Can Nvidia maintain its leadership in a world where inference, not training, is the dominant workload? AI training is a resource-hungry, centralized task. It typically takes place in massive data centres, over weeks or months, using thousands of GPUs across tightly coupled clusters. It’s Nvidia’s sweet spot. Inference, on the other hand, is diverse, fragmented, and ubiquitous. It takes place in the cloud, on edge devices, and locally in autonomous vehicles, on factory floors, and inside hospitals. It powers everything from voice assistants and customer service bots to robotic arms and real-time data analysis. This presents an opportunity for new players, each targeting a different slice of the inference pie, to carve out a niche or even establish a foothold for future expansion. AMD, Nvidia’s main competitor in the GPU space, offers MI300X chips, which have more memory per package (up to 256GB HBM), allowing them to run large models with fewer servers. Upstarts like Groq and Cerebras are pushing ultra-low-latency chips, with token speeds 10–20x faster than GPU-based inference—perfect for chain-of-thought or agent-based applications. Meanwhile, startups such as d-Matrix, Axelera, and Hailo are building inference-optimised silicon for edge devices, mobile, and cost-sensitive markets. At the same time, major cloud providers such as Amazon and Google are producing custom AI chips, such as Amazon’s Trainium or Google’s TPUv5, offering improved performance and efficiency tailored to their workloads. That’s why this year’s GTC saw a major pivot toward inference. Blackwell Ultra, for example, was clearly built to handle large-scale, high-throughput inference. It boosts HBM capacity, improves efficiency in attention-heavy models, and reworks the SM design to better suit the needs of real-time generation. Meanwhile, Nvidia’s new inference software stack—Dynamo—tackles the software side of the problem to help developers squeeze the most out of their GPUs. But Nvidia’s advantage is still its full-stack approach — powerful hardware, software frameworks and a developer ecosystem trained to build on CUDA. By pushing inference performance forward with new GPUs, while also improving deployment infrastructure and software, Nvidia is making a credible play to dominate this next phase. If training was about building the model, inference is about putting AI to work — and Nvidia is making it clear they want to own that, too. GTC 2025 wasn’t just a showcase of faster chips and smarter software—it was a glimpse into Nvidia’s vision of the future: a world where AI isn’t a tool, but a foundational layer across every industry, every data centre, and every device. With Rubin and Blackwell Ultra, Nvidia is redefining what’s possible in AI infrastructure. With projects like Groot N1 and digital twins, it’s expanding AI into the physical world. And with AI factories and gigawatt-scale data centres, it’s building the backbone of a new kind of economy—one where intelligence is manufactured at scale. But the real challenge isn’t just building the future. It’s bringing everyone along for the ride. The inference market is more competitive, more diverse, and more fragmented than training ever was. If Nvidia wants to lead in this new phase, it must continue to deliver performance, efficiency, and accessibility—both in silicon and in software. I think Jensen Huang’s and Nvidia’s vision of the future is going to happen. As much as AI is overhyped today, even if the current AI bubble bursts, it would only delay—not derail—that trajectory. Much like the dot-com crash in the early 2000s, which wiped out unsustainable business models but left behind the foundations of today’s internet economy, an AI correction would clear the noise and leave behind the most viable, transformative technologies. The infrastructure, the chips, the software—all of it is being built now. And when the dust settles, AI will be just as embedded in our world as the internet is today—only far more powerful. If you enjoy this post, please click the ❤️ button or share it. Do you like my work? Consider becoming a paying subscriber to support it For those who prefer to make a one-off donation, you can 'buy me a coffee' via Ko-fi. Every coffee bought is a generous support towards the work put into this newsletter. Your support, in any form, is deeply appreciated and goes a long way in keeping this newsletter alive and thriving. 🦾 More than a humanA Massive AI Analysis Found Genes Related to Brain Aging—and Drugs to Slow It Down 🧠 Artificial IntelligenceGoogle’s comments on the U.S. AI Action Plan OpenAI’s o1-pro is the company’s most expensive AI model yet China mandates labels for all AI-generated content in fresh push against fraud, fake news Claude can now search the web Leaked Apple meeting shows how dire the Siri situation really is Tencent introduces large reasoning model Hunyuan-T1 Sakana claims its AI-generated paper passed peer review — but it’s a bit more nuanced than that AI reasoning models can cheat to win chess games Anthropic CEO floats idea of giving AI a “quit job” button, sparking skepticism French publishers and authors file lawsuit against Meta in AI case McDonald’s Gives Its Restaurants an AI Makeover The US Army Is Using ‘CamoGPT’ to Purge DEI From Training Materials Flagship Pioneering Unveils Lila Sciences to Build Superintelligence in Science If you're enjoying the insights and perspectives shared in the Humanity Redefined newsletter, why not spread the word? 🤖 Robotics▶️ Walk, Run, Crawl, RL Fun | Boston Dynamics | Atlas (1:10) Boston Dynamics continues the tradition of dropping a new video of its humanoid robot performing feats of acrobatics that not many people can do. It is impressive how dexterous Atlas is and how natural and smooth its movements are. ▶️ World's First Side-Flipping Humanoid Robot (0:18) Boston Dynamics’ Atlas is not the only humanoid robot skilled in acrobatics. In this video, Chinese robotics company Unitree shows its humanoid robot G1 performing a side flip. Although not as elegant as Atlas, this feat of acrobatics is still impressive. BotQ: A High-Volume Manufacturing Facility for Humanoid Robots ▶️ Bernt Børnich "1X Technologies Androids, NEO, EVE" @1X-tech (1:14:29) In this video, Bernt Børnich, CEO and founder of 1X Technologies—a company building humanoid robots for everyday environments—discusses the advantages of using a human-like form factor, the engineering and AI challenges involved in designing and training such robots, and why 1X is prioritising household applications over industrial ones. Unlike many competitors, 1X believes that starting in the home offers a broader path to generalisation, safety, and scalability in real-world robotics. Insilico Medicine deploys the first bipedal humanoid AI scientist in the fully-robotic drug discovery laboratory NOMARS: No Manning Required Ship Ainos and ugo develop service robots with a sense of smell 🧬 Biotechnology▶️ The Ultimate Guide To Genetic Modification (41:25) In this video, The Thought Emporium, YouTube’s chief mad scientist, gives an excellent introduction to genetic engineering and DNA printing. Starting with simple fluorescent proteins and progressing to complex genetic logic gates and oscillators, the video explores how DNA functions as a biological programming language. Along the way, it covers lab techniques, synthetic biology concepts, and the launch of a new open-source plasmid store designed for students, hobbyists, and bio-nerds alike. AstraZeneca pays up to $1bn for biotech firm ‘that could transform cell therapy’ Researchers engineer bacteria to produce plastics AI-Designed Enzymes 💡TangentsAlexNet Source Code Is Now Open Source Thanks for reading. If you enjoyed this post, please click the ❤️ button or share it. Humanity Redefined sheds light on the bleeding edge of technology and how advancements in AI, robotics, and biotech can usher in abundance, expand humanity's horizons, and redefine what it means to be human. A big thank you to my paid subscribers, to my Patrons: whmr, Florian, dux, Eric, Preppikoma and Andrew, and to everyone who supports my work on Ko-Fi. Thank you for the support! My DMs are open to all subscribers. Feel free to drop me a message, share feedback, or just say "hi!" |
Older messages
How OpenAI and Anthropic Plan to Maintain US Dominance in AI - Sync #510
Tuesday, March 18, 2025
Plus: Manus; Gemini Robotics; inside Google's investments into Anthropic; world's first "Synthetic Biological Intelligence"; Ozempic show promise in preventing age-related conditions;
Where’s Apple Intelligence? - Sync #509
Sunday, March 9, 2025
Plus: Musk vs OpenAI trial set for expedited trial this year; scientists create woolly mice; an android with artificial muscles; another dancing humanoid robot; how to make superbabies; and more! ͏ ͏ ͏
Claude 3.7 Sonnet and GPT-4.5 - Sync #508
Sunday, March 2, 2025
Plus: Plus: Alexa+; Google AI co-scientist; humanoid robots for home from Figure and 1X; miracle HIV medicine; a startup making glowing rabbits; and more! ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
AI that can model and design the genetic code for all domains of life - Sync #507
Thursday, February 27, 2025
Plus: Grok 3; Figure shows its in-house AI model for humanoid robots; HP acquires Humane AI's IP; Meta explores robotics and announces LlamaCon; Microsoft's quantum chip; and more! ͏ ͏ ͏ ͏ ͏ ͏
CES 2025 - Sync #501
Sunday, January 12, 2025
Plus: Sam Altman reflects on the last two years; Anthropic reportedly in talks to raise $2B at $60B valuation; e-tattoo decodes brainwaves; anthrobots; top 25 biotech companies for 2025; and more! ͏ ͏
You Might Also Like
BetterDev #277 - When You Deleted /lib on Linux While Still Connected via SSH
Tuesday, March 25, 2025
Better Dev #277 Mar 25, 2025 Hi all, Last week, NextJS has a new security vulnerability, CVE-2025-29927 that allow by pass middleware auth checking by setting a header to trick it into thinking this is
JSK Daily for Mar 25, 2025
Tuesday, March 25, 2025
JSK Daily for Mar 25, 2025 View this email in your browser A community curated daily e-mail of JavaScript news Easily Render Flat JSON Data in JavaScript File Manager The Syncfusion JavaScript File
Want to create an AI Agent?
Tuesday, March 25, 2025
Tell me what to build next ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
LangGraph, Marimo, Django Template Components, and More
Tuesday, March 25, 2025
LangGraph: Build Stateful AI Agents in Python #674 – MARCH 25, 2025 VIEW IN BROWSER The PyCoder's Weekly Logo LangGraph: Build Stateful AI Agents in Python LangGraph is a versatile Python library
Charted | Where People Trust the Media (and Where They Don't) 🧠
Tuesday, March 25, 2025
Examine the global landscape of public trust in media institutions. Confidence remains low in all but a few key countries. View Online | Subscribe | Download Our App Presented by: BHP >> Read
Daily Coding Problem: Problem #1728 [Medium]
Tuesday, March 25, 2025
Daily Coding Problem Good morning! Here's your coding interview problem for today. This problem was asked by Square. Assume you have access to a function toss_biased() which returns 0 or 1 with a
LW 175 - Shopify uses AI to Prepare Stores for Script Editor Deprecation
Tuesday, March 25, 2025
Shopify uses AI to Prepare Stores for Script Editor Deprecation Shopify Development news and
Reminder: Microservices rules #7: Design loosely design-time coupled services - part 1
Tuesday, March 25, 2025
You are receiving this email because you subscribed to microservices.io. Considering migrating a monolith to microservices? Struggling with the microservice architecture? I can help: architecture
Delete your 23andMe data ASAP 🧬
Tuesday, March 25, 2025
95+ Amazon tech deals; 10 devs on vibe coding pros and cons -- ZDNET ZDNET Tech Today - US March 25, 2025 dnacodegettyimages-155360625 How to delete your 23andMe data and why you should do it now With
Post from Syncfusion Blogs on 03/25/2025
Tuesday, March 25, 2025
New blogs from Syncfusion ® Create AI-Powered Smart .NET MAUI Data Forms for Effortless Data Collection By Jeyasri Murugan This blog explains how to create an AI-powered smart data form using our .NET