A rush to build AI capacity using expensive coprocessors is jacking up the prices of servers, says research firm Omdia. Credit: Shutterstock Server unit shipments for calendar year 2023 could drop by as much as 20% when compared to last year, according to projections by a market research firm. However, revenues are growing for the server vendors, so while they may be selling fewer servers in terms of units, they’re selling pricier, more decked-out hardware. In its latest market update for cloud and data center, Omdia forecasts server unit shipments to decline by 17% to 20% this year, while revenue is expected to grow by 6% to 8%. Omdia cites the rise of heterogeneous computing for shift in spending. Instead of buying servers with just x86 processors, customers are buying servers with GPUs, DPUs, AI processing and inferencing chips, and other silicon processors. With so many expensive chips going into the servers, Omdia predicts CPUs and co-processors to account for 30% of data center spending by 2027, compared with less than 20% in the previous decade. Unsurprisingly, this shift is being driven by AI. Omdia said there was a dramatic shift in data center investment priorities this year, driven by a rush to build AI capacity, which made forecasting in 2023 incredibly difficult. Nvidia is the most popular supplier of GPUs for AI processing despite two ready competitors in Intel and AMD. Last quarter, Nvidia sold almost a half million GPUs to data center customers, and those processors go for a rumored $40,000 per card. And while Nvidia didn’t complain about supply on its most recent earnings call, apparently it does have a problem with it. Omdia says leading server OEMs like Dell, Lenovo, and HPE are not able to fulfil GPU server orders yet, due to a lack of GPU supply from Nvidia. OEMs indicated a lead time of 36 to 52 weeks for servers configured with Nvidia H100 GPUs. That’s in part because a few players are taking all of the supply. Omdia noted that both Microsoft and Meta are on track to receive 150,000 of Nvidia’s H100 accelerators by the end of this year – which is three times as many as Nvidia’s other major customers, Google, Amazon and Oracle. These high-powered servers are also driving demand for better power efficiency and management. Data center operators have to get more compute power out of the same power envelope due to constraints and power supply. Omdia said rack power distribution revenue in 1H23 was up 17% over last year, while UPS revenue growth for the first half of 2023 was 7% ahead of last year. “With a ramp of professional services for generative AI enabling broad enterprise adoption in 2024 and beyond, the only thing that can curb the current rate of AI deployment is power availability,” Omdia said in its report. AI is also driving demand for liquid cooling, since air cooling is simply no longer efficient for the very hot processors used in AI. Cooling vendors and server OEMs tell Omdia direct-to-chip liquid cooling is ramping in line with its forecast for 80% revenue growth within the year, and it noted that server vendor Super Micro recently said that it expects 20% of the servers it ships in 4Q23 will use direct-to-chip liquid cooling. Up through 2027, Omdia expects continued growth in rack power density, server performance improvement, and server fleet consolidation. There will be a strong focus on computing performance to enable the commercialization of AI. AI models will continue to be a research project requiring a great deal of tuning, even libraries of pre-trained models. Because of the rate of advancements, it expects meaningful server refresh cycles to take place, where enterprises will prioritize equipment consolidation and utilization improvement during the refresh cycle. And it expects the trend for hyper-heterogeneous computing, with servers configured with 1-2 CPUs and up to 20 workload-optimized custom-build co-processors, will enable a significant consolidation in server fleets. Related content news High-bandwidth memory nearly sold out until 2026 While it might be tempting to blame Nvidia for the shortage of HBM, it’s not alone in driving high-performance computing and demand for the memory HPC requires. By Andy Patrizio May 13, 2024 3 mins CPUs and Processors High-Performance Computing Data Center news CHIPS Act to fund $285 million for semiconductor digital twins Plans call for building an institute to develop digital twins for semiconductor manufacturing and share resources among chip developers. By Andy Patrizio May 10, 2024 3 mins CPUs and Processors Data Center news HPE launches storage system for HPC and AI clusters The HPE Cray Storage Systems C500 is tuned to avoid I/O bottlenecks and offers a lower entry price than Cray systems designed for top supercomputers. By Andy Patrizio May 07, 2024 3 mins Supercomputers Enterprise Storage Data Center news Lenovo ships all-AMD AI systems New systems are designed to support generative AI and on-prem Azure. By Andy Patrizio Apr 30, 2024 3 mins CPUs and Processors Data Center PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe