GTC 2025: NVIDIA's Blackwell-Based Servers and DGX Station

The GTC (GPU Technology Conference), held annually since 2009, will be hosted by NVIDIA this year from March 17 to 21. The conference is designed to showcase the latest developments and to promote collaboration and further innovation across different industries. It is attended mainly by developers, researchers, and technology leaders. NVIDIA CEO Jensen Huang has been saying for some time that companies will become token factories in the future—meaning that every workflow will be supported by artificial intelligence. Currently, large servers play a major role in this process, but AI integration will increasingly extend to personal computers. In the future, computers and laptops will have hardware capable of running even large language models in the background. This is necessary because programmers, engineers, and almost everyone will work with AI assistance.

The New Blackwell GPU

However, let’s not get ahead of ourselves. According to Jensen Huang, to achieve these goals, it is first necessary to dramatically increase computing power (scale-up) and then expand the system by combining these scaled-up components (scale-out). He explained that the concept was first used in the Grace Hopper architecture and Ranger server model presented three years ago. Although the physical size of the server was too large, it validated their ideas. The size issue was partly due to the use of air cooling, which was replaced by liquid cooling in the new Blackwell GPU, allowing it to fit into standard server racks. Another problem was that the NVLink-connected CPU and GPU were housed in the same module. NVLink provides a high-speed connection between the CPU and GPU. In traditional systems, a PCI Express interface is used for this, but NVLink has lower latency. By disaggregating NVLink, the CPU and GPU can be placed in separate modules, so that each component can be replaced independently in the server.

Despite these improvements, a third problem remains: the optical data transmission cables (transceivers) used to connect the GPUs. These cables are extremely expensive (six are needed for each GPU, adding $6000 to the GPU price) and they increase power consumption by an extra 180 watts per GPU. To solve this, Jensen Huang presented a solution based on silicon photonics that enables GPUs to communicate using photons. Incidentally, Google is already using this technology in its data centers, achieving a 40% reduction in power consumption. 

Server performance roadmap
Server performance roadmap

Thanks to the reduction in size, they have reached a performance of 1 exaflop (1000 petaflops) per server rack. The memory bandwidth is an impressive 570 TB/s. For comparison, an NVIDIA RTX 4070 has a bandwidth that is a thousand times lower at 504 GB/s, although it is not designed for servers. For another realistic performance comparison, Jensen Huang states that for an AI company consuming 1 megawatt, the current setup with 1400 server racks using H100s can process 300 million tokens per second when running a large language model. With the new solution, assuming the same 1 MW consumption, 600 server racks replacing the old 1400 and H100s replaced by the new Blackwell compute units would result in 12,000 million tokens per second. The performance increase is so dramatic that it is hard to keep track, but successors have already been announced. The Blackwell Ultra will arrive at the end of this year, followed by the Rubin and Rubin Ultra GPUs next year and in 2027. The Rubin Ultra will deliver 15 exaflops per rack instead of the current 1 exaflop.

DGX Station

As mentioned earlier, Jensen Huang speaks of 30 million programmers who will soon work with some form of AI assistance. This is an important distinction between those who already claim that programmers will become obsolete and those who believe programmers will continue to be needed. However, for programmers to run large language models locally, they need adequate memory bandwidth and enough memory. The DGX Station is NVIDIA’s answer to this market need. It features 8 TB/s of memory bandwidth, 20,000 AI TFLOPS, and 784 GB of RAM—of which 288 GB is available to the GPU—so it can run relatively large models. Naturally, it uses the newly announced Blackwell chip, just like the Geforce RTX 5xxx series of graphics cards. The big question will obviously be its price. The lower-performing DGX Spark, also launched earlier this year, costs $4000 while offering only 128 GB of RAM, 273 GB/s of memory bandwidth, and 1000 AI TFLOPS, making it 20 times less powerful and much smaller. Although the DGX Spark has its advantages due to its smaller size—allowing you to connect more units to build a powerful small server for an office—the price is still quite steep. 

DGX Station
DGX Station
Share this post
After a Historic Turn, SK Hynix Becomes the New Market Leader in the Memory Industry
For three decades, the name Samsung was almost synonymous with leadership in the DRAM market. Now, however, the tables have turned: in the first half of 2025, South Korea’s SK Hynix surpassed its rival in the global memory industry for the first time, ending a streak of more than thirty years. This change signifies not just a shift in corporate rankings but also points to a deeper transformation across the entire semiconductor industry.
The Most Popular Theories About the Impact of AI on the Workplace
Since the release of ChatGPT at the end of 2022, the field of AI has seen impressive developments almost every month, sparking widespread speculation about how it will change our lives. One of the central questions concerns its impact on the workplace. As fears surrounding this issue persist, I believe it's worth revisiting the topic from time to time. Although the development of AI is dramatic, over time we may gain a clearer understanding of such questions, as empirical evidence continues to accumulate and more theories emerge attempting to answer them. In this article, I’ve tried to compile the most relevant theories—without claiming to be exhaustive—as the literature on this topic is expanding by the day. The question remains: can we already see the light at the end of the tunnel, or are we still heading into an unfamiliar world we know too little about?
Neuralink’s Rivals Are Already in the Clinic – Europe’s Answer to Musk’s Brain Chips
While Silicon Valley is loud with the hype surrounding brain-computer interfaces (BCIs), attention often centers on one high-profile name: Elon Musk’s Neuralink. Behind the scenes, however, Europe is quietly yet determinedly building its own neurotechnological future. Startups across the continent are not merely following trends—they are shaping the market with unique solutions, often rooted in practical medicine and materials science. This raises the question: in the shadow of flashy American marketing, can Europe’s quieter innovation achieve real breakthroughs and take a leading role in this revolutionary field?
NVIDIA Driver Support Changes – The Clock Is Ticking for the GTX 900–10 Series
NVIDIA has announced a major shift in its driver support strategy. This decision affects millions of users, but what does it actually mean in practice? Is it really time for everyone to consider upgrading their hardware, or is the situation more nuanced? Understanding the implications is key to staying prepared for the technological changes of the coming years.
This Is What the Google Pixel 10 Might Look Like — But Do We Really Know Everything Yet?
The Google Pixel series, known for its clean software and outstanding photography capabilities, is reaching its 10th generation this year. Ahead of the official launch, expected on August 20, a wealth of information and rumors is already circulating online, outlining a device family that may not bring dramatic visual changes but instead introduces real, tangible innovations under the hood. But are these improvements enough for the Pixel 10 to stand out in an increasingly crowded market?
A Brutal Quarter for Apple, but What Comes After the iPhone?
Amid global economic and trade challenges, Apple has once again proven its extraordinary market power, surpassing analyst expectations in the third quarter of its 2025 fiscal year. The Cupertino giant not only posted record revenue for the period ending in June but also reached a historic milestone: the shipment of its three billionth iPhone. This achievement comes at a time when the company is grappling with the cost of punitive tariffs, intensifying competition in artificial intelligence, and a series of setbacks in the same field.