laitimes

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

author:Leqing industry observation

At present, the global AI deployment is accelerating, a large number of AI models are emerging, and the launch and iteration of models have greatly driven the demand for upstream computing power. With the rising demand for data load of digital infrastructure in China, the growth of the mainland AI server market is accelerating.

From the perspective of the overall development trend, countries have intensively introduced policies related to large models, and the pace of data, computing power, and algorithm updates and iterations has accelerated, jointly driving the performance improvement of AI large models, and AI large models have stepped out of the two development paths of "closed source" and "open source", and various enterprises and institutions have entered the game.

The computing power with AI server as the carrier is one of the terminal manifestations of artificial intelligence. The storage and transportation capacity linked to computing power will also be increased accordingly, which will drive the growth of hardware computing power chips (GPU, FPGA, AISIC, etc.), memory chips, optical modules, PCBs, power management, heat dissipation, etc. Historically, the entire industry will see growth brought about by the increase in the penetration rate of AI servers.

According to TrendForce, it is estimated that in 2023, the shipment of AI servers (including those equipped with GPUS, FPGAs, ASICs, etc.) will be nearly 1.2 million, an annual increase of 38.4%, accounting for nearly 9% of the total server shipments.

Pay attention to Leqing's industry observation and gain insight into the industrial pattern!

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

Overview of the AI server industry

The increase in computing power directly drives the demand for servers. As the core support to solve the computing power, AI servers can be used to support local applications and web pages, and can also provide complex AI models and services for cloud and local servers.

AI servers use heterogeneous architectures, and the number of GPUs is much higher than that of ordinary servers.

From the perspective of server upstream costs, chips, memory, and hard disks account for nearly seventy percent of the server production costs. The GPU of the inference server accounts for 25% of the cost, and the GPU of the training server accounts for about 73% of the overall cost.

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

The main difference between AI servers and ordinary servers is that the architecture is different, AI servers use heterogeneous architectures such as CPU+GPU/FPGA/ASIC, while ordinary servers generally use CPU architectures; The number of GPUs varies greatly, and the number of GPUs used in a single AI server is usually more than 4.

The GPU architecture is the mainstream acceleration architecture and is the core cost component of the server. GPUs use parallel computing and are suitable for processing-intensive computing scenarios, such as graphics rendering, machine learning, and Xi, and the increase in AI computing power requirements has further increased the computing speed and usage requirements of GPU cards.

Compared with general servers, the increase of AI servers is mainly due to the significant increase in AI chips and memory bandwidth, the corresponding increase in value, and the upgrading of other supporting components to varying degrees.

According to the observation of the semiconductor industry, taking Nvidia DGX H100 as an example, the value of GPU board groups (including HBM) accounts for the highest proportion, reaching 73%, followed by storage accounting for about 5%, of which DRAM accounts for 3% and NAND accounts for about 2%. Overall, GPUs and storage together account for 90%.

In addition, with the improvement of computing power, the demand for heat dissipation has also increased correspondingly, and the introduction of liquid cooling technology is expected to bring new opportunities to the heat dissipation link.

Compared with general servers, AI servers increase the use of GPGPU, so the HBM usage is about 320~640GB based on 4 or 8 NVIDIAA10080GB configurations.

In the future, the increasing complexity of AI models will stimulate more memory usage and simultaneously drive the demand for serverDRAM, SSD, and HBM.

The cost of AI server GPUs accounts for a larger proportion:

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

Sort out the AI server market pattern

According to statistics, in 2022, the four major cloud providers in North America, Microsoft, Google, Meta, and AWS, together account for about 66%, and overseas cloud giants have greater demand for AI servers, but with the development and application of domestic AI large models driving more AI server demand, the mainland AI server market space is expected to further increase.

In recent years, the wave of AI construction in mainland China has continued to heat up, with ByteDance's annual procurement accounting for 6.2%, followed by Tencent, Alibaba, and Baidu, with about 2.3%, 1.5%, and 1.5% respectively.

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

According to IDC data, in terms of market share, Inspur Information, Xinhua III, and Super Fusion ranked among the top three shares of the mainland AI server market in 2022, accounting for 28.10%, 17.20%, and 17.00% of the market share respectively.

In 2022, Inspur AI server won 49 championships in MLPerf, the world's authoritative AI performance evaluation competition, providing strong impetus for AI R&D and application with comprehensive leading AI training and inference performance, and its AI server products have been applied to the world's leading Internet giants and leading technology companies in the fields of AI+Science, AI+Graphics, AIGC, etc., becoming the largest supplier of AI servers in the world.

Inspur Information Yingxin Server NF5688M6 and Configuration:

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

Sugon is a leading manufacturer of high-end servers in China. The company has developed full-stack high-end server products and has large-scale deployment practices, continuously developing computing power service business, accelerating the innovation and implementation of massive complex industry applications through the national integrated computing power service platform, and providing computing power support for multiple large models in China.

AI server layout manufacturers also include China Great Wall, Tuowei Information, Tongfang Co., Ltd., etc. There are many participating manufacturers in the upstream and downstream links, and some of the representative manufacturers include ZTE, iFLYTEK, Hikvision, 360, Montage Technology, Loongson Zhongke, Digital China, Haitian AAC, Donghua Software, etc. #Artificial Intelligence##Computing Power##Server##Technology##Finance##Article Launch Challenge##来点儿干货#

AI server is facing the outlet! The key track of computing power, the layout of the leading powerhouse Hengqiang

epilogue

At present, advanced technologies such as AIGC and cloud computing have brought tremendous changes and opportunities to various industries. In this context, as an important cornerstone to support large-scale data processing and high-performance computing, the demand and application scenarios of supercomputing are also showing a continuous growth trend.

The combination of supercomputing and cloud computing has brought new growth points to the server market. The elasticity and scalability of cloud computing enable supercomputing resources to serve many fields such as scientific research, industrial design, simulation, and big data analysis more efficiently. This combination not only improves computing efficiency, but also reduces the cost of use, so that more institutions and enterprises can enjoy the convenience brought by supercomputing.

At the same time, the AI high-end servers required by supercomputing have also become a new star in the server market. AI servers have higher performance, stronger stability, and better scalability, which can meet the stringent requirements of supercomputing in terms of computing speed and data processing capabilities. With the continuous popularization and deepening of supercomputing applications, the demand for such high-end servers will continue to rise, which is expected to drive the entire server market to achieve further growth.

Pay attention to Leqing's industry observation and gain insight into the industrial pattern!

Read on