Affected by the global DRAM supply shortage and continuously rising costs, nearly all PC hardware utilizing memory has been impacted. Against this backdrop, NVIDIA is considering adjustments to its GPU supply strategy for AIC brand partners in the Chinese mainland market.
According to industry sources, NVIDIA plans to reduce the overall production of its GeForce RTX 50 series GPUs in the first half of 2026, with supply expected to drop by 30% to 40% compared to the same period in 2025. Currently, AMD has already taken the lead in raising graphics card prices at the beginning of this month. Although NVIDIA has not officially announced a price increase, retail market prices have remained unusually firm. Due to the sensitivity of consumer-grade GPU pricing, it is difficult to fully pass on the rising costs of memory to consumers. As a result, NVIDIA has opted for a "price-controlled volume reduction" strategy to maintain market balance and brand positioning. This move has also indirectly impacted the planning of subsequent product lines, casting a shadow over the release schedule of its SUPER versions, which may likely delay their originally planned launch. If memory supply fails to ease, NVIDIA may further adjust its product release timeline and regional supply strategies.
While consumer-grade graphics cards face production cuts, NVIDIA presents a different scenario in the AI accelerator sector. According to EPOCH.AI's cost analysis of the B200, its material cost is approximately $6,400. The most expensive component is not the GPU core itself but the 192GB HBM3e high-bandwidth memory, which costs $2,900—nearly half of the total cost.
In comparison, the cost of GPU logic cores is only around 1,100. This cost structure further highlights the critical role and pricing pressure of HBM memory in AI accelerator cards. However, despite the persistently high cost of HBM, NVIDIA's profit margins in the AI GPU sector remain astonishing. The market price of the B200 ranges between 30,000and40,000, translating to a gross margin close to 80%, and in extreme cases, even approaching 90%. The urgent demand for computing power among enterprise clients, coupled with their relatively high price tolerance, allows NVIDIA to smoothly pass on costs and maintain ultra-high profits. NVIDIA also boosts overall sales through rack-level solutions, offsetting the cost pressure per card. This has been one of the key reasons behind NVIDIA's skyrocketing profits in recent years. Additionally, rumors suggest that NVIDIA is considering vertical integration upstream, potentially developing its own HBM core technology to reduce dependency and costs.
The "AI tilt" in DRAM production capacity has become a critical variable in the global supply chain. Manufacturers such as Samsung and SK Hynix are prioritizing advanced processes for HBM3e and future HBM4, leading to tighter supply of standard DRAM. This trend further widens the profit gap between the AI and consumer hardware sectors. In the long run, driven by AI, memory has evolved from a basic component to a strategic resource. Looking ahead, as HBM technology iterates and competition for production capacity intensifies, the battle between AI and consumer electronics for memory demand will only escalate.



.png)
