News
AMD has encountered resistance from the United States when attempting to sell a custom-designed AI chip to the Chinese market, according to Bloomberg citing sources.
The same source further revealed that AMD sought approval from the U.S. Department of Commerce to sell this AI chip to Chinese customers. The chip’s performance is lower than AMD’s products sold in other regions of China, and its design complies with U.S. export restrictions.
However, U.S. officials informed AMD that the chip’s performance is still too robust, requiring the company to obtain approval from the U.S. Department of Commerce’s Bureau of Industry and Security before sales can proceed. It remains unclear whether AMD is currently seeking a license.
The U.S. government introduced initial export control measures in 2022 and bolstered a series of measures in October 2023, encompassing additional technologies and reforming potential sales that could undermine intermediary countries/regions.
The strict limitations on the sale of NVIDIA’s chips specifically designed for China align with the initial export regulations from 2022. Subsequently, the company developed new customized but lower-capacity products for the Chinese market, aligning with the restrictions imposed by the United States in 2023.
The U.S. ban in 2022 prevented both NVIDIA and AMD from selling their most powerful artificial intelligence chips to China. When the restrictions took effect in 2022, AMD anticipated that they would not substantially affect its operations.
While AMD has not publicly discussed its development of new AI chips for China, NVIDIA, on the other hand, immediately introduced improved products with reduced performance.
NVIDIA is set to commence pre-orders for its AI chip H20 specially designed for the Chinese market by the end of March this year in response to US export bans, according to sources cited by a report from STAR Market Daily.
Still, AMD is now actively expanding into the AI chip market. In December 2023, it launched the new MI300 series, challenging NVIDIA’s chips. According to sources cited by Bloomberg’s report, this customized product for China is known as the MI309. It is still unclear which Chinese customer wishes to purchase AMD’s AI chips, which could affect the company’s authorization.
Read more
(Photo credit: AMD)
News
The surge in demand for NVIDIA’s AI processors has made High Bandwidth Memory (HBM) a key product that memory giants are eager to develop. However, according to South Korean media DealSite cited by Wccftech on March 4th, the complex architecture of HBM has resulted in low yields, making it difficult to meet NVIDIA’s testing standards and raising concerns about limited production capacity.
The report has further pointed out that HBM manufacturers like Micron and SK Hynix are grappling with low yields. They are engaged in fierce competition to pass NVIDIA’s quality tests for the next-generation AI GPU.
The yield of HBM is closely tied to the complexity of its stacking architecture, which involves multiple memory layers and Through-Silicon Via (TSV) technology for inter-layer connections. These intricate techniques increase the probability of process defects, potentially leading to lower yields compared to simpler memory designs.
Furthermore, if any of the HBM chips are defective, the entire stack is discarded, resulting in inherently low production yields. As per the source cited by Wccftech, it has indicated that the overall yield of HBM currently stands at around 65%, and attempts to improve yield may result in decreased production volume.
Micron announced on February 26th the commencement of mass production of High Bandwidth Memory “HBM3e,” to be used in NVIDIA’s latest AI chip “H200” Tensor Core GPU. The H200 is scheduled for shipment in the second quarter of 2024, replacing the current most powerful H100.
On the other hand, Kim Ki-tae, Vice President of SK Hynix, stated on February 21st in an official blog post that while external uncertainties persist, the memory market is expected to gradually heat up this year. Reasons include the recovery in product demand from global tech giants. Additionally, the application of AI in devices such as PCs or smartphones is expected to increase demand not only for HBM3e but also for products like DDR5 and LPDDR5T.
Kim Ki-tae pointed out that all of their HBM inventory has been sold out this year. Although it’s just the beginning of 2024, the company has already begun preparations for 2025 to maintain its market-leading position.
Per a previous TrendForce press release, the three major original HBM manufacturers held market shares as follows in 2023: SK Hynix and Samsung were both around 46-49%, while Micron stood at roughly 4-6%.
Read more
(Photo credit: SK Hynix)
News
NVIDIA has yet to officially announce the exact release dates for its next-generation AI chip architectures, the Blackwell GPU, and the B100 chip. However, Dell’s Chief Operating Officer, Jeff Clarke, recently revealed ahead of schedule during Dell’s Q4 2024 Earnings Call that NVIDIA is set to introduce the Blackwell architecture next year, with plans to release not only the B100 chip but also another variant, the B200 chip.
Following Dell’s recent financial report, Clarke disclosed in a press release that NVIDIA is set to unveil the B200 product featuring the Blackwell architecture in 2025.
Clarke also mentioned that Dell’s flagship product, the PowerEdge XE9680 rack server, utilizes NVIDIA GPUs, making it the fastest solution in the company’s history. He expressed anticipation for NVIDIA’s release of the B100 and B200 chips. This news has sparked significant market interest, as NVIDIA has yet to publicly mention the B200 chip.
Clarke further stated that the B200 chip will showcase Dell’s engineering expertise in high-end servers, especially in liquid cooling systems. As for the progress of the B100 chip, NVIDIA has yet to disclose its specific parameters and release date.
NVIDIA’s current flagship H200 chip in the high-performance computing market adopts the Hopper GPU architecture paired with HBM3e memory chips, considered the most capable chip for AI computing in the industry.
However, NVIDIA continues to accelerate the development of its next-generation AI chip architectures. According to NVIDIA’s previously disclosed development roadmap, the next-generation product after the H200 chip is the B100 chip. Therefore, the expectation was that the B100 chip would be the highest-specification chip based on the Blackwell GPU architecture. Nevertheless, with the emergence of the B200 chip, it has sparked further speculation.
Previously, media speculation cited by the report from Commercial Times stated based on the scale of the H200 chip that the computational power of the B100 chip would be at least twice that of the H200 and four times that of the H100.
Read more
(Photo credit: NVIDIA)
News
The annual AI event, NVIDIA GTC (GPU Technology Conference), is set to take place on March 17th, as H200 and the next-generation B100 will reportedly be announced ahead of schedule to seize the market. According to Commercial Times’ report citing sources, H200 and the upcoming B100 will adopt TSMC’s 4-nanometer and 3-nanometer processes respectively. H200 is expected to be launched in the second quarter, while it’s rumored that orders for the B100 adopting Chiplet architecture have already been placed for production.
Sources cited by the report also indicate that NVIDIA’s orders are robust, pushing TSMC’s 3 and 4-nanometer production capacity to near full utilization, making the first quarter, traditionally a slow season, unexpectedly busy.
Regarding the matter of NVIDIA’s next-generation chip orders overwhelming TSMC’s advanced processes, TSMC stated that details regarding production capacity remain consistent with the previous earnings call and will not be elaborated further.
Still, Commercial Times further cited industry sources, revealing that TSMC, in response to anticipated capacity constraints by 2023, is accelerating its efforts. Particularly focusing on advanced packaging like CoWoS, they’ve not only relocated equipment from the Longtan facility but also swiftly activated the AP6 plant in Zhunan.
Another industry sources reportedly indicate that the planned construction of the Tongluo
facility, initially slated for the second half of this year, is now scheduled to commence in the second quarter. The aim is to ramp up 3D Fabric capacity to produce 110,000 12-inch wafers per month by the first half of 2027.
Meanwhile, TSMC’s advanced processes remain fully utilized, with capacity utilization exceeding 90% in February, driven by sustained AI demand.
NVIDIA, on the other hand, recently emphasized that computational-intensive tasks like Generative AI and large language models require multiple GPUs. From customer purchase to model deployment, it takes several quarters. Thus, this year’s inference applications stem from GPU purchases made last year. As model parameters grow, GPU demand is expected to expand.
In addition to increasing GPU quantities, NVIDIA’s GPU efficiency is poised for a significant boost this year. The Blackwell series, notably the B100, is hailed as NVIDIA’s next-generation GPU powerhouse by the market.
Not only is it the first to adopt TSMC’s 3-nanometer process, but it’s also the pioneer in Chiplet and CoWoS-L packaging among NVIDIA products. This tackles high power consumption and cooling issues, with projected single-card efficiency and transistor density expected to surpass AMD’s MI300 series set to debut in the first quarter.
Read more
(Photo credit: Kioxia)
News
NVIDIA, the global leader in artificial intelligence (AI) chips, is set to commence pre-orders for its AI chip H20 specially designed for the Chinese market by the end of March this year in response to US export bans, according to sources cited by a report from STAR Market Daily.
However, due to consecutive reductions in specifications, the cost-performance ratio has diminished, prompting cautious attitudes among Chinese distributors.
The report further mentions that by the end of 2022, the US Department of Commerce restricted the export of NVIDIA AI chips to China due to concerns about their potential military use. In response, NVIDIA has repeatedly reduced product performance to comply with US regulations. The H20 chip, derived from the H800, is specifically designed as a ‘special edition’ for the Chinese market.
Citing industry sources, STAR Market Daily‘s report also states that H20 will be available for pre-order following NVIDIA’s GTC 2024 conference (March 18th to 21st), with deliveries possible within a month. The sources note that H20’s performance is approximately one-fourth that of H100, resulting in a less favorable cost-performance ratio. Additionally, production capacity is unable to meet demand, with mass production expected to commence in the second half of this year.
A distributor in Beijing pointed out that currently, there is not significant demand for the H20 chip, primarily due to its relatively low cost-performance ratio. Chinese-made AI chips serve as viable alternatives.
The same distributor noted that most of the foundational technology for computing power providers is still supported by NVIDIA. The advantages of adopting the H20 lie in its compliance and low migration costs. However, the trend toward self-developed AI chip in China is a long-term certainty, presenting a choice between the two options.
The distributor further emphasized that NVIDIA’s introduction of the H20 is primarily aimed at stabilizing its presence in the Chinese market. As long as the product specifications slightly surpass those of domestically produced chips, it should suffice. However, whether there is demand for this chip still requires market validation.
Another distributor from the Shenzhen cited in the report also stated that it is uncertain whether they will stock the H20 chip, as their decision depends on subsequent market demand.
Regarding the need for H20, TrendForce believes Chinese companies will continue to buy existing AI chips in the short term. NVIDIA’s GPU AI accelerator chips remain a top priority—including H20, L20, and L2—designed specifically for the Chinese market following the ban.
At the same time, major Chinese AI firms like Huawei, will continue to develop general-purpose AI chips to provide AI solutions for local businesses. Beyond developing AI chips, these companies aim to establish a domestic AI server ecosystem in China.
TrendForce recognizes that a key factor in achieving success will come from the support of the Chinese government through localized projects, such as those involving Chinese telecom operators, which encourage the adoption of domestic AI chips.
Read more
(Photo credit: NVIDIA)