Insights
In October 2023, Quanta revealed plans to open three new factories in California, USA, with the goal of creating state-of-the-art assembly lines for AI servers. Around the same time, Wiwynn shared its intentions to launch a server cabinet assembly plant in Johor, Malaysia, featuring advanced liquid cooling technology. Additionally, server contract manufacturing giants, Foxconn and Inventec, are strategically positioning their AI server manufacturing facilities both domestically and internationally to meet the expected demand for AI server orders in the 2024 market.
TrendForce’s Insights:
Both Wiwynn and Quanta are contract manufacturers for cloud service giants such as Meta, Microsoft, and AWS. These three cloud service providers accounted for nearly 50% of the global server procurement in 2023. They’re doing this to keep up with the growing demand for AI servers, especially in the latter part of 2023, driven by applications like ChatGPT. Big cloud service providers, have allocated a significant chunk of their global server orders to these manufacturers, giving AI servers a top spot over regular servers.
Wiwynn, in particular, has set up shop in Malaysia to meet the surging demand for AI servers. Due to factors like the trade tensions between China and the U.S. and tariff avoidance measures, they are shifting their manufacturing capacity and equipment from their Guangdong factory in China to locations in Taiwan and Malaysia. This transition is expected to be completed between the end of 2023 and early 2024, making it easier to manage resources on a global scale.
Quanta’s smart move to open new assembly fabs near major U.S. cloud service providers allows them to deliver quickly to data centers in Europe and the U.S., saving on transportation costs and ensuring speedy deliveries. Both major Taiwanese manufacturers are optimistic about their orders for AI servers. This optimism allows them to expand their manufacturing capabilities in existing and new locations to strengthen partnerships with the big cloud players.
Major companies like Foxconn and Inventec are actively expanding their production facilities, both in their home country and abroad, to prepare for the expected increase in orders for AI servers from leading cloud service providers in 2024.
Fii and Ingrasys Inc., two important subsidiaries of Foxconn, dedicated to handling orders and manufacturing for servers. They have their own server assembly plants in various locations, including China, the United States, Europe, Vietnam, and Taiwan. They follow an integrated supply chain model, starting from producing motherboards to assembling complete server cabinets. Once the assembly is done, they ship the products to the data centers of cloud service providers. To meet the anticipated high-end AI server orders in 2024, Ingrasys Inc. added new production lines in the second quarter of 2023 to meet the demands of AI server manufacturers.
Inventec, a manufacturer that specializes in making server motherboards, expects a steady demand for AI servers in 2024-2025. With this expectation, they started construction at their factory in Thailand in Q3 of 2023. By Q4 of 2024, the factory will undergo production line testing, and mass production could begin as early as the first quarter of 2025 to meet the needs of major manufacturers. The new factory in Mexico is expected to match the production capacity of their Chinese facility. It has already started limited production and is expected to be in full operation by Q4 of 2024.
The four major contract manufacturers in Taiwan, specializing in server production, are either adding new production lines to their existing facilities or building new factories overseas between Q2 to Q4 of 2023. This undoubtedly shows their positive attitude to AI server shipments in 2024. As the use of AI servers continues to grow, market demand is expected to significantly increase year by year, which is likely to bring substantial revenue, profit, and production advantages to these contract manufacturers.
News
According to Taiwan’s Liberty Times, in response to the global supply chain restructuring, electronic manufacturing plants have been implementing a “China+ N” strategy in recent years, catering shipments to customers in different regions. Among them, Inventec continues to strengthen its server production line in Thailand and plans to enter the NVIDIA B200 AI server sector in the second half of next year.
Currently, Inventec’s overall server production capacity is distributed as follows: Taiwan 25%, China 25%, Czech Republic 15%, and Mexico, after opening new capacity this quarter, is expected to reach 35%. It is anticipated that next year’s capital expenditure will increase by 25%, reaching 10 billion NTD, primarily allocated for expanding the server production line in Thailand. The company has already started receiving orders for the B100 AI server water-cooling project from NVIDIA and plans to enter the B200 product segment in the second half of next year.
Inventec’s statistics show that its server motherboard shipments account for 20% of the global total. This year, the focus has been on shipping H100 and A100 training-type AI servers, while next year, the emphasis will shift to the L40S inference-type AI servers. The overall project quantity for next year is expected to surpass this year’s.
(Photo credit: Google)
News
According to Liberty Times Net, Inventec, a prominent player in the realm of digital technology, is making significant strides in research and development across various domains, including artificial intelligence, automotive electronics, 5G, and the metaverse. The company has recently introduced a new all-aluminum liquid-cooled module for its general-purpose graphics processing units (GPGPU) powered by NVIDIA’s A100 chips. Additionally, this innovative technology is being applied to AI server products featuring AMD’s 4th Gen EPYC dual processors, marking a significant step towards the AI revolution.
Inventec has announced that their Rhyperior general-purpose graphics processors previously offered two cooling solutions: air cooling and air cooling with liquid cooling. The new all-aluminum liquid-cooled module not only reduces material costs by more than 30% compared to traditional copper cooling plates but also comes with 8 graphics processors (GPUs) and includes 6 NVIDIA NVSwitch nodes. This open-loop cooling system eliminates the need for external refrigeration units and reduces fan power consumption by approximately 50%.
Moreover, Inventec’s AI server product, the K885G6, equipped with AMD’s 4th Gen EPYC dual processors, has demonstrated a significant reduction in data center air conditioning energy consumption of approximately 40% after implementing this new cooling solution. The use of water as a coolant, rather than environmentally damaging and costlier chemical fluids, further enhances the product’s appeal, as it can support a variety of hardware configurations to meet the diverse needs of AI customers.
Inventec’s new facility in Mexico has commenced mass production, with plans to begin supplying high-end NVIDIA AI chips, specifically the H100 motherboards, in September. They are poised to increase production further in the fourth quarter. Additionally, in the coming year, the company is set to release more Application-Specific Integrated Circuit (ASIC) products, alongside new offerings from NVIDIA and AMD. Orders for server system assembly from U.S. customers (L11 assembly line) are steadily growing. The management team anticipates showcasing their innovations at the Taiwan Excellence Exhibition in Dongguan, China, starting on October 7th, as they continue to deepen their collaboration with international customers.
Press Releases
US-based CSPs have been establishing SMT production lines in Southeast Asia since late 2022 to mitigate geopolitical risks and supply chain disruptions. TrendForce reports that Taiwan-based server ODMs, including Quanta, Foxconn, Wistron (including Wiwynn), and Inventec, have set up production bases in countries like Thailand, Vietnam, and Malaysia. It’s projected that by 2023, the production capacity from these regions will account for 23%, and by 2026, it will approach 50%.
TrendForce reveals that Quanta, due to its geographical ties, has established several production lines in its Thai facilities centered around Google and Celestica, aiming for optimal positioning to foster customer loyalty. Meanwhile, Foxconn has renovated its existing facilities in Hanoi, Vietnam, and uses its Wisconsin plant to accommodate customer needs. Both Wistron and Wiwynn are progressively establishing assembly plants and SMT production lines in Malaysia. Inventec’s current strategy mirrors that of Quanta, with plans to build SMT production lines in Thailand by 2024 and commence server production in late 2024.
CSPs aim to control the core supply chain, AI server supply chain trends toward decentralization
TrendForce suggests that changes in the supply chain aren’t just about circumventing geopolitical risks—equally vital is increased control over key high-cost components, including CPUs, GPUs, and other critical materials. With rising demand for next-generation AI and Large Language Models, supply chain stockpiling grows each quarter. Accompanied by a surge in demand in 1H23, CSPs will become especially cautious in their supply chain management.
Google, with its in-house developed TPU machines, possesses both the core R&D and supply chain leadership. Moreover, its production stronghold primarily revolves around its own manufacturing sites in Thailand. However, Google still relies on cooperative ODMs for human resource allocation and production scheduling, while managing other materials internally. To avoid disruptions in the supply chain, companies like Microsoft, Meta, and AWS are not only aiming for flexibility in supply chain management but are also integrating system integrators into ODM production. This approach allows for more dispersed and meticulous coordination and execution of projects.
Initially, Meta heavily relied on direct purchases of complete server systems, with Intel’s Habana system being one of the first to be integrated into Meta’s infrastructure. This made sense since the CPU for their web-type servers were often semi-custom versions from Intel. Based on system optimization levels, Meta found Habana to be the most direct and seamless solution. Notably, it was only last year that Meta began to delegate parts of its Metaverse project to ODMs. This year, as part of its push into generative AI, Meta has also started adopting NVIDIA’s solutions extensively.
News
According to a report by Taiwan’s Economic Daily, Foxconn Group has achieved another triumph in its AI endeavors. The company has secured orders for over 50% of NVIDIA’s HGX GPU base boards, marking the first instance of such an achievement. Adding to this success, Foxconn had previously acquired an order for another NVIDIA DGX GPU base board, solidifying its pivotal role in NVIDIA’s two most crucial AI chip base board orders.
The report highlights that in terms of supply chain source, Foxconn Group stands as the exclusive provider of NVIDIA’s AI chip modules (GPU Modules). As for NVIDIA’s AI motherboards, the suppliers encompass Foxconn, Quanta, Inventec, and Super Micro.
Industry experts analyze that DGX and HGX are currently NVIDIA’s two most essential AI servers, and Foxconn Group has undertaken the monumental task of fulfilling the large order for NVIDIA’s AI chipboards through its subsidiary, Foxconn Industrial Internet (FII). Having previously secured orders for NVIDIA’s DGX base boards, Foxconn Group has now garnered additional orders from FII for the HGX base boards. This expanded supply constitutes more than half of the total, solidifying Foxconn Group’s role as a primary supplier for NVIDIA’s two critical AI chip base board orders.
Furthermore, Foxconn’s involvement doesn’t end with AI chip modules, base boards, and motherboards. The company’s engagement extends downstream to servers and server cabinets, creating a vertically integrated approach that covers the entire AI ecosystem.
(Photo credit: Nvidia)