Will ASICs replace NVIDIA GPUs?

The development of artificial intelligence over the past decade has been closely linked to the name NVIDIA, which has become the dominant player in the market with its graphics processing units (GPUs). A significant portion of today's AI models are built on these GPUs, and NVIDIA's decade-old software ecosystem—especially the CUDA platform—has become an indispensable tool for research, development, and industrial applications. At the same time, in recent years, the biggest players in the technology sector – including Google, Amazon, Meta, and Microsoft – have been turning with increasing momentum toward AI chips developed in-house and optimized for specific tasks, known as ASICs.

ASIC (Application-Specific Integrated Circuit) chips are not designed for general-purpose computing like GPUs, but are hardware-optimized for specific tasks, such as running AI models. As a result, their energy efficiency and operating costs can be significantly lower. Google, for example, has reported that its own TPU (Tensor Processing Unit) chips consume three times less energy than NVIDIA AI GPUs and can be up to 1.7 times faster in certain tasks. Advanced ASICs already achieve a performance density of 142 TOPS/W, while their power consumption typically does not exceed 350 watts – this is particularly noteworthy for models that work with hundreds of billions of parameters.

Large companies are also interested in ASICs because of the significant reduction in total cost of ownership (TCO). According to industry analyses, these chips can reduce AI infrastructure costs by as much as 30-50 percent, especially for large-scale applications. This applies not only to hardware procurement, but also to operating costs such as cooling and power supply. In addition, custom-designed chips allow companies to reduce their dependence on external suppliers, which can result in a more stable supply and a better-controlled cost structure in the long term.

There are also economic considerations behind this: to offset the so-called “Nvidia tax” – i.e., the often high price associated with NVIDIA chips – in-house development is becoming an increasingly attractive alternative. Google and Amazon are already planning to ship a total of more than three million ASIC chips by 2025, which is estimated to be half of NVIDIA's annual AI GPU supply. By 2026, with Meta's entry, the shipment volume of ASIC chips could even exceed that of GPUs.

However, despite the rapid advancement of the technology, it is important to note that ASICs do not replace GPUs in all respects. The biggest limitation at present is the software ecosystem. NVIDIA's CUDA platform is now the basis for the work of more than 30 million developers. This system is so deeply embedded in AI development that a complete switch to ASICs would require not only hardware investment but also a complete software redesign. The immaturity of alternative platforms—such as Google TPUs—often poses a significant barrier to widespread adoption, especially for smaller developer communities or research institutions.

Another challenge is chip manufacturing. Advanced manufacturing technologies such as TSMC's CoWoS packaging process are a bottleneck, with some estimates suggesting that they allow for the production of no more than 300,000 to 400,000 ASICs per year. Meta's next-generation chips (MTIA T-V1 and T-V2), for example, are already in the design phase, but without manufacturing capacity, their market launch is uncertain.

NVIDIA's response was not long in coming. The company's new NVLink protocol allows processors and ASICs from other manufacturers to be directly connected to NVIDIA GPUs, offering a kind of “open collaboration” for those who do not wish to abandon the GPU-based ecosystem entirely. Meanwhile, the company continues to spend more than $16 billion annually on research and development, far exceeding the amount spent on ASIC development.

Industry analysis clearly indicates that the rise of ASIC chips does not spell the end of GPUs, but rather the specialization of the AI hardware ecosystem. ASICs are extremely efficient at well-defined, stable tasks, such as AI model inference, while GPUs remain indispensable for developing, testing, and scalable training of new algorithms.

The future is therefore much more about coexistence than replacement. The development of artificial intelligence is not the result of a single technology, but rather an ecosystem of complementary solutions. In this equation, ASIC chips symbolize efficiency and goal-oriented operation, while NVIDIA GPUs represent flexibility, software support, and freedom of development. The challenge for industry players is to combine the advantages of these two worlds in a way that lays the foundation for the next phase of artificial intelligence development. 

Share this post
Google Gemini CLI, a powerful offering in the field of AI accessible from the terminal
Google's recently announced Gemini CLI is an open source, command line AI tool that integrates the Gemini 2.5 Pro large language model directly into the terminal. The goal of the initiative is nothing less than to transform natural language commands into real technical workflows, in an environment that has already been synonymous with efficiency for many.
Satya Nadella's thoughts on the role, future, and responsibility of artificial intelligence
Rapid change is not uncommon in the world of technology, but rarely does it affect so many sectors at once as today's artificial intelligence (AI) revolution. In an interview with Y Combinator, Satya Nadella, CEO of Microsoft, not only assessed technological developments, but also placed the development of AI in a broader social and economic context. His approach is restrained, calm, and purposeful: AI is not a mystical entity, but a tool that must be properly applied and interpreted.
Apple joins China's subsidy programme: market protection with discounts
Apple has officially joined the Chinese government's consumer electronics subsidy programme to boost domestic consumption, reports the South China Morning Post. The move can be seen as both a market adaptation and a strategic attempt by the US technology giant to strengthen its position in an increasingly competitive environment. Under the programme, shoppers in the Chinese cities of Beijing and Shanghai can now get direct discounts on certain Apple products.
Intel's new 18A chip manufacturing process: moderate progress or a turning point in the global semiconductor industry?
Intel recently published a detailed technical analysis of its latest manufacturing technology, known as 18A, which the company says represents a significant advance in performance, power consumption, and transistor density. Based on the material presented at the VLSI 2025 symposium, the new manufacturing node promises up to 25% faster speed, 36% lower power consumption, and approximately 30% higher transistor density compared to the previous Intel 3 process. While these figures are impressive in themselves, the significance of the process lies primarily in the fact that this is the first time in years that Intel can offer a truly competitive alternative to the world's leading chip manufacturer, Taiwan's TSMC.
The new M5 iPad Pro is coming: what can we expect from Apple's next tablet?
Apple's product announcements always receive significant attention, especially when it comes to a flagship device like the iPad Pro. Although there has been no official confirmation yet, industry players and analysts expect the new iPad Pro with the M5 chip to be released in late 2025 or early 2026 at the latest. Signs indicate that Apple is sticking to the 18-month update cycle it uses for iPad Pro models.
The future of AI and the price of transparency – What do the OpenAI files say?
There has been growing interest in OpenAI's operations in recent times. This is no coincidence: the artificial intelligence models they have developed – such as ChatGPT – are widely used, while we have only fragmentary information about the decision-making and ownership structures behind them. Some light is shed on this obscurity by a report called OpenAI Files, prepared by two technology oversight organizations, the Midas Project and the Tech Oversight Project. The document not only discusses the company's internal operations, but also touches on broader social issues: what mechanisms are needed when a private company holds the key to the economy of the future?

Linux distribution updates released in the last few days