News
ByteDance, the parent company of TikTok, is said to be collaborating with TSMC, eyeing for the mass production of two self-developed AI chips by 2026, according to reports by Economic Daily News and The Information.
ByteDance’s AI chips are expected to be made with TSMC’s 5nm node, which would be one generation behind the foundry giant’s most advanced process, the reports suggest, making the move comply with the U.S. export regulations to China. The chips are similar to NVIDIA’s next-generation flagship AI chip, Blackwell, which are manufactured with TSMC’s 4NP node.
Citing sources familiar with the matter, the reports note that the tech giant in China aims to reduce its reliance on NVIDIA for AI model development. Though the chips are still in the design phase and the plan is subject to change, ByteDance’s self-designed chips could save billions of dollars compared to purchasing NVIDIA’s products, according to the reports.
The Information estimates that ByteDance’s spending on developing generative AI models has been increasing, and it is rumored that the company has ordered over 200,000 NVIDIA H20 chips this year, costing it over USD 2 billion, with some orders still pending delivery.
In response to US export bans, NVIDIA launched AI chip H20, L20 and L2, specially designed for the Chinese market earlier this year. According to a previous report by Wccftech, H20 GPU has 41% fewer Cores and 28% lower performance versus H100. Still, the product is reportedly seeing strong demand for AI servers among Chinese Cloud Service Providers (CSPs) and enterprises, including Huawei and Tencent.
However, due to its lower computing power, Chinese companies need to purchase more H20 chips to build clusters with equivalent computing capacity, which raises costs, Economic Daily News notes.
According to TSMC’s financial report in the second quarter, North American clients contributed 65% of its total revenue. While China, the second-largest market, contributed 16% of its quarterly revenue, with a significant jump from 9% in the first quarter and 12% during the same period last year.
Read more
(Photo credit: ByteDance)
News
According to a report from wccftech, AMD CEO Lisa Su has indicated that she believes the AI Supercycle has just started, and the company has accelerated its product development plans to meet the substantial market demand.
In addition to NVIDIA, AMD is a significant player in the AI market as well. This is not only due to its market impact but also because the company has significantly expanded its AI product portfolio over the past few quarters, attracting attention from major clients like Microsoft and Amazon.
While AMD has not yet replicated NVIDIA’s success in the market, the company remains optimistic about the future, which is why it believes the AI boom has only just begun.
A few months ago, AMD outlined its AI chip development roadmap for the next year. The “Advancing AI” event in this October will showcase the next-generation Instinct MI325X AI chip.
The flagship Instinct MI350 AI chip is scheduled for release in 2025, followed by the Instinct MI400 AI chip in 2026. Despite AMD’s advancements, there remains a generational gap, as competitor NVIDIA is poised to launch its Blackwell architecture AI chips in the coming months.
Moreover, per a report from Yahoo Finance, Su once stated that AMD could generate USD 4.5 billion in sales from the MI300 alone in 2024, a significant increase from around USD 100 million in AI-related chip revenue last year.
The company had previously projected MI300 sales at approximately USD 4 billion for this year. Su then added that, it’s the fastest-growing product in AMD’s history.
AMD recently announced that it will merge its consumer and data center architectures into a single unit known as “UDNA,” aiming to accelerate the development and optimization of both platforms.
This move is particularly noteworthy as AMD is focusing on competing with NVIDIA’s CUDA on the software front.
Read more
(Photo credit: AMD)
News
In a last-ditch effort before the upcoming board meeting this week, Intel is said to be seeking assistance from the U.S. government. The latest report by CNBC notes that Intel CEO, Pat Gelsinger, turned to Commerce Secretary Gina Raimondo recently, expressing his disappointment with the heavy dependence of U.S. companies on TSMC, the Taiwanese foundry heavyweight.
According to CNBC, after meeting with Intel, Raimondo followed up by meeting with several public market investors to emphasize the significance of U.S. chip manufacturing amid the rising geopolitical risks surrounding Taiwan. Her aim was to encourage shareholders in companies like NVIDIA and Apple to understand the economic advantages of having a U.S.-based foundry capable of producing AI chips, the sources cited by the report said.
Interesting enough, Jensen Huang, CEO of NVIDIA, mentioned yesterday that the U.S. chip giant heavily relies on TSMC for producing its most important chips, saying TSMC’s agility and ability to respond to demand are incredible. Thus, shifting orders to other suppliers could reportedly lead to a decline in chip quality.
Intel has introduced its Lunar Lake processors in early September, with the target to shake up the AI PC market. However, the chips are outsourced to TSMC, manufactured with the foundry giant’s 3nm node.
Getting stuck in its current situation, Intel is said to be pushing U.S. officials to expedite the release of funding, another report by Bloomberg notes. Earlier in April, Intel and Biden administration announced up to USD 8.5 billion in direct funding under the CHIPS Act.
The company is slated to receive USD 8.5 billion in grants and USD 11 billion in loans under the 2022 Chips and Science Act, but this funding is contingent on meeting key milestones and undergoing extensive due diligence, according to Bloomberg. Therefore, like other potential beneficiaries, Intel has not yet received any money.
An official cited by CNBC said that disbursements are anticipated by the end of the year.
Both the U.S. Commerce Department and an Intel spokesperson declined to comment, according to CNBC.
Regarding the latest development of Intel’s plan to shedding more than 15% of its workforce, a report by CTech notes that Intel may lay off over 1,000 employees in Israel as global cuts begin.
CTech states that Intel has been mindful of geopolitical factors and the timing of local holidays in Israel. Therefore, it would be rather unexpected for the company to initiate layoffs in the country before or during the holiday season, which begins in early October and extends through most of the month.
Citing Gelsinger’s remarks, the report notes that the restructuring process will continue until the end of the year, allowing Intel’s Israeli branch a window of time to start the layoffs after the holidays.
Read more
(Photo credit: Intel)
News
With AI giants in the Western world, such as OpenAI, Google and Meta, stealing the spotlight by their development of generative AI, some big names in China have introduced their AI models over the past 18 months, according to a latest report by CNBC.
Though trying to keep a low profile, five tech conglomerates in China, including Alibaba, Baidu, ByteDance, Huawei and Tencent, have launched their AI models lately, adding a new dimension to the competitive landscape of the ongoing tech rivalry between China and the U.S. Here is a brief introduction of the major Chinese AI models developed by the country’s leading tech firms, based on the information compiled by CNBC and other media reports.
Alibaba: Tongyi Qianwen
In November, 2023, a report by pandaily notes that Alibaba Cloud released its AI model, Tongyi Qianwen 1.0 a few months ago, while the 2.0 version was introduced later in the same year. Another report by South China Morning Post states that as of May, Alibaba reports that its Tongyi Qianwen AI models, often referred to as Qwen, are utilized by more than 90,000 corporate clients across China.
CNBC notes that in terms of Qwen, the company has developed various versions tailored to different functions. For instance, one model specializes in generating content or solving math problems, while another handles audio inputs and provides text-based responses.
It is worth noting that as some Qwen models are open-sourced, developers are allowed to download and use them under certain restrictions, according to CNBC.
Baidu: ERNIE
As one of China’s leading internet companies, Baidu was among the first to introduce generative AI applications in the country. A report by The Verge notes that the Ernie chatbot was available for download in August, 2023, after the approval by the Chinese government.
CNBC reports that Baidu intends to compete with OpenAI’s ChatGPT with Ernie Bot, as the company claims the bot to have 300 million users.
According to CNBC, ahead of the launch of its “Turbo” version, which took place in late June, Baidu stated that its Ernie 4.0 offers capabilities comparable to OpenAI’s GPT-4. According to Baidu, this foundational model has advanced understanding and reasoning abilities.
Similar to other companies, Baidu is offering access to its AI model through its cloud computing services, CNBC says.
ByteDance: Doubao
TikTok parent company ByteDance, though entered the AI race later than competitors like Baidu and Alibaba, has surprised the market with its low-cost Doubao model, which was launched in May, 2024.
According to a report by technode, the model can process 2 million Chinese characters, equivalent to 1.25 million tokens, for just RMB 1 (USD 0.14). In comparison, OpenAI’s latest multimodal model, GPT-4o, costs USD 5 per million input tokens.
CNBC notes that Doubao has various capabilities, including voice generation and coding support for developers.
Huawei: Pangu
Introduced by Huawei in 2021 as the world’s largest pre-trained Chinese large language models (LLMs) with over 100 billion parameters, the Pangu models are now entering their fourth iteration, according to Counterpoint. In May, 2024, the latest Pangu models are said to boast 230 billion parameters.
Interesting enough, Huawei has adopted a different strategy from its competitors with its Pangu AI models, CNBC remarks. The tech giant focuses on developing industry-specific models tailored to sectors like government, finance, manufacturing, mining, and meteorology.
For instance, Huawei claims that its Pangu Meteorology Model can predict a typhoon’s trajectory 10 days in advance in just 10 seconds, a task that previously took four to five hours, according to CNBC.
Tencent: Hunyuan
Last year, Tencent introduced its foundational model, Hunyuan, which is accessible through Tencent’s cloud computing services.
According to CNBC, Tencent has highlighted Hunyuan’s strong Chinese language processing abilities and advanced logical reasoning, supporting features like image generation and text recognition. The model is designed for use across industries such as gaming, social media, and e-commerce.
As the operator of China’s largest messaging app, WeChat, Tencent launched an AI chatbot this year based on the Hunyuan model. The AI assistant, named Yuanbao, can access information and content from WeChat, setting it apart from competitors, CNBC notes.
Notably, China’s large language models, just like its rivals in the West, rely on the strong computing power of AI chips. A previous report by Reuters in November, 2023, states that Tencent is said to have stockpiled a substantial reserve of AI chips from NVIDIA, as the company prepares in advance to train its Hunyuan AI models for the following generations.
How far will the tech giants in China be able to push the boundaries of AI models? The answer may lie in the development of the country’s domestic chips, as the U.S. authority already banned the export to China of AI chips.
Read more
(Photo credit: Baidu)
News
Amid concerns on the delay of NVIDIA’s Blackwell, CEO Jensen Huang spoke at the Goldman Sachs Communacopia and Technology Conference a few hours ago, trying to ease the doubts of the market by expressing his optimism on the demand of its products and the company’s future prospects.
“We have a lot of people on our shoulders, and everybody is counting on us,” said Huang, according to a report by Yahoo Finance. He even joked that as the world relies so much on the company, the engineers may have to spend more time on work. “Less sleep is fine. Three solid hours is all we need.”
Huang also elaborated on the demand for the delivery of NVIDIA’s components, technology, infrastructure and software, stating that it is so overwhelming that people may get “emotional,” as it has a direct impact on their revenue and competitiveness.
It is worth noting that Huang also mentioned that NVIDIA heavily relies on TSMC for producing its most important chips, as in many ways, according to a report by Economic Daily News. He said TSMC’s agility and ability to respond to demand are incredible.
Huang stated that most of NVIDIA’s technology is self-developed, and if necessary, orders could be shifted to other suppliers. However, such adjustments could lead to a decline in chip quality, according to the report.
According to a previous report from Commercial Times, NVIDIA has reportedly executed changes to the Blackwell series’ GPU mask. Therefore, the process can now proceed without re-taping out, with NVIDIA’s updated version of B200 expected to be completed by late October, allowing the GB200 to enter mass production in December.
Moreover, in his latest meeting with Goldman Sachs, Huang noted that the first trillion dollars of data centers is going to get accelerated, creating a new type of software, generative AI.
Citing Huang’s remarks, the report by Yahoo Finance stated that it matters a lot because generative AI is not just a tool but a “skill,” so for the first time, the AI chip giant is developing skills that will enhance human capabilities.
According to Yahoo Finance, Huang said that NVIDIA, along with cloud service providers (CSPs), build the infrastructure in the cloud so developers can access these machines to train, fine-tune, and safeguard models.
It is worth noting that Huang tried to materialize the benefit, saying that for every dollar a CSP spends with NVIDIA, it results in USD 5 worth of rentals. He also said while training AI models is resource-intensive, it pays off in the long run.
Citing Huang, the report stated that NVIDIA’s servers may seem expensive at first glance, as it potentially costs a couple of million dollars per rack. However, they replace thousands of nodes. What is remarkable is that the cost of cables for old, general-purpose computing systems is higher than consolidating everything into a single, dense rack, Huang said.
According to Yahoo Finance, Huang also noted that the days of software engineers writing every line of code are completely behind. In his vision, every software engineer will have digital companions working alongside them 24/7.
In addition, NVIDIA, with its 32,000 employees, hopes to be supported by “100 times more digital engineers” in the near future, the report noted.
Notably, there seems to be another piece of good news for the U.S. chip giant. According to a report by Reuters, the U.S. government is said to be mulling, allowing NVIDIA to export advanced chips to Saudi Arabia, which would enable the country to train and operate the most powerful AI models.
According to the report, Saudi Arabia expects to receive shipments of NVIDIA’s most advanced chips, the H200s, which were first used in OpenAI’s GPT-4o.
Read more
(Photo credit: NVIDIA)