Chinese semiconductor industry

Status
Not open for further replies.
D

Deleted member 24525

Guest
Please, Log in or Register to view URLs content!
lol, I found the CITIC securities report summary of data centers since 2021 and whose GPUs they used. Aside from Tencent, Alibaba & Baidu, which most likely will never use Huawei GPUs, almost everyone else used Huawei GPUs. Which would confirm the viewpoint by many that Ascend GPUs are best AI chips in China in production for the past couple of years. I guess with Cambrian winning a couple of times. The part that Huawei won works out to be almost 8 EFLOPS
View attachment 112051
View attachment 112052

2000 Ascend 910 GPU produces 640PFLOPS of FP16 computation need 2+ months, so each is 320 TFLOPS of FP16

Considering that A100 supports 300 TFLOPS, it really backs up the theory that Ascend-910 for the past couple of years was the closest China had to A100.


Again 8x Ascend 910 is 2.56 PFLOPS of FP16 computation. Used in Changsha & Chongqing


For example, if beijing was to expand from 100 to 500 FLOPS, would need 156 AI training servers of 8x Ascend-910 GPU

Also has this company that's core partner of Huawei which build intelligent server machines using Kunepng + Huawei chips to provide 128 core computation (so maybe 2x64 core Kunpeng-920 with 8 Ascend 910). Already migrated over 15000 of such cloud server and manage over 10000 such server. So looks like they've sold a lot of these GPUs and have large contracts to expand enough more. This part is not for smart city but rather just medium large enterprises.
Damn so Huawei has caught up with Nvidia for data center GPUs, at least on a technological level. That's great news.
I'm not very knowledgeable on the hpc business. Are GPUs the processors used in all data centers or are they specifically for ai applications?
 

staplez

New Member
Registered Member
Damn so Huawei has caught up with Nvidia for data center GPUs, at least on a technological level. That's great news.
I'm not very knowledgeable on the hpc business. Are GPUs the processors used in all data centers or are they specifically for ai applications?
Whoo that's a lot to unpack. So GPGPUs are kind of a misnomer. They are basically just data crunchers without too much to do with graphics. They're designed to take long running processes and speed them up. The most common application of this is AI.

They cannot replace the CPU, the CPU gives instructions to the rest of the computer. A GPGPU won't do that. But a computer with a CPU and GPGPU will out perform a machine with a more powerful CPU. As your off loading the heavy calculations to the GPGPU. Where as while a CPU can technically do it all, it would get overwhelmed pretty fast.

In a way, this is how China is forging ahead in AI and cloud computing. While Intel and AMD CPUs are far better than what China has,it doesn't affect too much because the GPUs are fairly close in performance. As anyone with a poor CPU and fast GPU for videogames can attest to.
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
Damn so Huawei has caught up with Nvidia for data center GPUs, at least on a technological level. That's great news.
I'm not very knowledgeable on the hpc business. Are GPUs the processors used in all data centers or are they specifically for ai applications?

Whoo that's a lot to unpack. So GPGPUs are kind of a misnomer. They are basically just data crunchers without too much to do with graphics. They're designed to take long running processes and speed them up. The most common application of this is AI.

They cannot replace the CPU, the CPU gives instructions to the rest of the computer. A GPGPU won't do that. But a computer with a CPU and GPGPU will out perform a machine with a more powerful CPU. As your off loading the heavy calculations to the GPGPU. Where as while a CPU can technically do it all, it would get overwhelmed pretty fast.

In a way, this is how China is forging ahead in AI and cloud computing. While Intel and AMD CPUs are far better than what China has,it doesn't affect too much because the GPUs are fairly close in performance. As anyone with a poor CPU and fast GPU for videogames can attest to.
I think he was just asking if china has caught up in ai and you. That's a reasonable question.

Recently, tencent came out with their hcc ai framework that promised to be much faster in training than previous incarnation of their ai framework.

It makes things clear that gpus are only part of the equation. The communication bandwidth between different chips and clusters really matters the dram speed and size really matters. The I/o speed to storage really matters. The CPU processing really matters. The chip to chip data rate matters. The software hardware layer really matters. The ai framework itself matters.

For a long time, I didn't think much about Huawei for ai because I thought they were sanctioned and they would only have enough chipset for their own usage. That turned out to be false. China built so many data centers as part of their smart city plans. Yet almost all of them picked Huawei/pangu with only a couple that picked Cambrian. Almost nobody picked a100. It seems like a100 is more common with other Chinese big cloud providers and inspir. Why is that?

Well, it would seem to me that Huawei has the most advanced full stack in terms of ai. Their pangu platform with ascend and kunpeng chipset along with their communication technology and software prowess is what's leading the industry.

Raw performance for ascend 910 is at about 1/3 computational power of br100 and at about the same level as a100. But in terms of actual ai application and training, the availability of ascend GPU and software hardware integration makes it the best option. The constant theme from Chinese social media is that ascend is the best Chinese GPU. When people say that, they are taking holistically. Biren chips still need to improve software hardware integration and integration with cloud service providers to achieve potential. They also need to stack up more gpus. Having everything in house allows Huawei to compete against a100 with cuda. Clearly, Chinese smart city planners think so.

So as we go forward, it's clear that the market has spoken. Huawei is the current ai leader in china. Not Baidu as I previously thought. And it appears to be that Huawei can get it's chips produced by smic. It won't be hurt that much if America sanctions it's cloud business. It will also be providing it's ai smart city solutions to arab countries. This is a major growing business for Huawei.

So from that perspective, I understand why they say the worst is over. The worst is over. Things should get better for them.

And as a supporter for china's ai industry, I would say it's a great thing they can now expand as much as they want. And as smic improve their 7nm process, Huawei will also get better chips produced. At this time, I would imagine smic has many limitations in what it is capable of producing. It's good to have a captive customer with money and design expertise to work with you.

Btw, when ascend first came out, it had spec of 256 tflops for 16fp. Now, hisilicon website shows 320. So looks like hisilicon design managed to overcome smic deficiencies and made chips even better than the original one. I guess that's to be expected after 3 years.
 

CMP

Senior Member
Registered Member
I think is more an Chicken and Egg complex situation, managers wanting to see quick results, over-trust on the global supply chain thinking that the US will not crazy enough to hurt their own companies and the double edge sword of foreign talent wanting to incorporate the tools that they are accustomed to use disregarding the overall development of the semiconductor supply chain. A complex situation with no easy fix.​
And that's why the more strategic decisions need to be made by CPC, c-suite, and board of directors. Leaving the important decisions to middle and lower management is a recipe for disaster.
 

vincent

Grumpy Old Man
Staff member
Moderator - World Affairs
For a long time, I didn't think much about Huawei for ai because I thought they were sanctioned and they would only have enough chipset for their own usage. That turned out to be false. China built so many data centers as part of their smart city plans. Yet almost all of them picked Huawei/pangu with only a couple that picked Cambrian. Almost nobody picked a100. It seems like a100 is more common with other Chinese big cloud providers and inspir. Why is that?

Well, it would seem to me that Huawei has the most advanced full stack in terms of ai. Their pangu platform with ascend and kunpeng chipset along with their communication technology and software prowess is what's leading the industry.
Not necessarily true. Chinese local governments may picked Huawei chips to keep the company affloat.
 

paiemon

Junior Member
Registered Member
And that's why the more strategic decisions need to be made by CPC, c-suite, and board of directors. Leaving the important decisions to middle and lower management is a recipe for disaster.
Strategic decisions are always set by upper management with the input of a board of directors (if everything functions as intended). I don't know of any companies where middle management gets to make those decisions. Middle and lower management are responsible for implementing those directions and leading their teams towards those goals. Secondly, if companies need to work towards political goals for the economy that are not profit related, those companies should have representatives of those non-profit related interests as part of the company management/board to provide related inputs. Take for example Volkswagen, which has appointees from the German states and workers unions on its board to represent the interests of the state and the workers even though they may clash with the profit motive.

The foundries operate as @hvpc stated because they are run first and foremost to generate profit, which requires them to operate around the clock at high throughput with minimal downtime. That's the goal and direction set by their management and is reflected in their choice of operations. And to be fair to the companies, that was the goal of the state as well for years so why would they operate any differently? If the state wants its goals and inputs to be represented in the companies, it needs to have a seat at the table which is why it bought the stakes in companies through investments and has representatives on the board. Its probably not a coincidence that YMTC has been one of the most aggressive at supply chain localization. You can fault companies for missing those risks, but the government also thought that things would never get to this point and everyone is now playing catchup albeit quickly.

As for integrating local equipment, keep in mind alot of the procurement will be forward looking towards expansions or new greenfield projects since the existing operations are already built out and they aren't going to rip out perfectly functional processes that can still be maintained. If they have to take an operational hit to assist with localization goals, it will be reflected in the costing along the line whether it be through the customers or the investors/shareholders (which includes the state). Someone has to pay for this, otherwise there is no incentive to work towards those goals.
 

tokenanalyst

Brigadier
Registered Member
And that's why the more strategic decisions need to be made by CPC, c-suite, and board of directors. Leaving the important decisions to middle and lower management is a recipe for disaster.
There should be division of labor, what I am saying is that at least when goverment money was involved there should have been taken into consideration the broader development of the semiconductor industry rather than a few sectors and groups.
 

KYli

Brigadier
Firstly, this article pretty confirmed that the ban has nothing to do with military applications. Secondly, they want to limit Chinese AI and other high tech development but at the same time not to push Chinese companies to the edge and go all in for domestic alternatives. Lastly, they failed to slow down Chinese AI development due to the fact that Chinese companies have adapted and have resources to keep up. In addition, companies take a lot of effort to slim down the systems to save costs that they ended up needing much less chips than expected which rendered the restrictions meaningless.
Please, Log in or Register to view URLs content!
By Stephen Nellis, Josh Ye and Jane Lanhee Lee

(Reuters) -U.S. microchip export controls imposed last year to freeze China's development of supercomputers used to develop nuclear weapons and artificial-intelligence systems like ChatGPT are having only minimal effects on China's tech sector.

The rules restricted shipments of Nvidia Corp and Advanced Micro Devices Inc chips that have become the global technology industry's standard for developing chatbots and other AI systems.

But Nvidia has created variants of its chips for the Chinese market that are slowed down to meet U.S. rules. Industry experts told Reuters the newest one - the Nvidia H800, announced in March - will likely take 10% to 30% longer to carry out some AI tasks and could double some costs compared with Nvidia's fastest U.S. chips.

Even the slowed Nvidia chips represent an improvement for Chinese firms. Tencent Holdings, one of China's largest tech companies, in April estimated that systems using Nvidia's H800 will cut the time it takes to train its largest AI system by more than half, from 11 days to four days.

"The AI companies that we talk to seem to see the handicap as relatively small and manageable," said Charlie Chai, a Shanghai-based analyst with 86Research.

The back-and-forth between government and industry exposes the U.S. challenge of slowing China's progress in high tech without hurting U.S. companies.

Part of the U.S. strategy in setting the rules was to avoid such a shock that the Chinese would ditch U.S. chips altogether and redouble their own chip-development efforts.

"They had to draw the line somewhere, and wherever they drew it, they were going to run into the challenge of how to not be immediately disruptive, but how to also over time degrade China's capability," said one chip industry executive who requested anonymity to talk about private discussions with regulators.

The export restrictions have two parts. The first puts a ceiling on a chip's ability to calculate extremely precise numbers, a measure designed to limit supercomputers that can be used in military research. Chip industry sources said that was an effective action.

But calculating extremely precise numbers is less relevant in AI work like large language models where the amount of data the chip can chew through is more important.

Nvidia is selling the H800 to China's largest technology firms, including Tencent, Alibaba Group Holding Ltd and Baidu Inc for use in such work, though it has not yet started shipping the chips in high volumes.

"The government isn’t seeking to harm competition or U.S. industry, and allows U.S. firms to supply products for commercial activities, such as providing cloud services for consumers," Nvidia said in a statement last week.

China is an important market for U.S. technology companies, and selling products there helps create jobs for both Nvidia and its U.S.-based partners, the company added.

"The October export controls require that we create products with an expanding gap between the two markets," Nvidia said last week. "We comply with the regulation while offering as-competitive-as-possible products in each market."

Bill Dally, Nvidia's chief scientist, said in a separate statement this week that “this gap will grow quickly over time as training requirements continue to double every six to 12 months."

A spokesperson for the Bureau of Industry and Security, the arm of the U.S. Commerce Department that oversees the rules, did not return a request for comment.

SLOWED BUT NOT STOPPED

The second U.S. limit is on chip-to-chip transfer speeds, which does affect AI. The models behind technologies such as ChatGPT are too large to fit onto a single chip. Instead, they must be spread over many chips - often thousands at a time - which all need to communicate with one another.

Nvidia has not disclosed the China-only H800 chip's performance details, but a specification sheet seen by Reuters shows a chip-to-chip speed of 400 gigabytes per second, less than half the peak speed of 900 gigabytes per second for Nvidia's flagship H100 chip available outside China.

Some in the AI industry believe that is still plenty of speed. Naveen Rao, chief executive of a startup called MosaicML that specializes in helping AI models to run better on limited hardware, estimated a 10-30% system slowdown.

"There are ways to get around all this algorithmically," he said. "I don't see this being a boundary for a very long time - like 10 years."

Money helps. A chip in China that takes twice as long to finish an AI training task than a faster U.S. chip can still get the work done. "At that point, you've got to spend $20 million instead of $10 million to train it," said one industry source who requested anonymity because of agreements with partners. "Does that suck? Yes it does. But does that mean this is impossible for Alibaba or Baidu? No, that's not a problem."

Moreover, AI researchers are trying to slim down the massive systems they have built to cut the cost of training products similar to ChatGPT and other processes. Those will require fewer chips, reducing chip-to-chip communications and lessening the impact of the U.S. speed limits.

Two years ago the industry was thinking AI models would get bigger and bigger, said Cade Daniel, a software engineer at Anyscale, a San Francisco startup that provides software to help companies perform AI work.

"If that were still true today, this export restriction would have a lot more impact," Daniel said. "This export restriction is noticeable, but it's not quite as devastating as it could have been."
 

CMP

Senior Member
Registered Member
Firstly, this article pretty confirmed that the ban has nothing to do with military applications. Secondly, they want to limit Chinese AI and other high tech development but at the same time not to push Chinese companies to the edge and go all in for domestic alternatives. Lastly, they failed to slow down Chinese AI development due to the fact that Chinese companies have adapted and have resources to keep up. In addition, companies take a lot of effort to slim down the systems to save costs that they ended up needing much less chips than expected which rendered the restrictions meaningless.
Please, Log in or Register to view URLs content!
By Stephen Nellis, Josh Ye and Jane Lanhee Lee

(Reuters) -U.S. microchip export controls imposed last year to freeze China's development of supercomputers used to develop nuclear weapons and artificial-intelligence systems like ChatGPT are having only minimal effects on China's tech sector.

The rules restricted shipments of Nvidia Corp and Advanced Micro Devices Inc chips that have become the global technology industry's standard for developing chatbots and other AI systems.

But Nvidia has created variants of its chips for the Chinese market that are slowed down to meet U.S. rules. Industry experts told Reuters the newest one - the Nvidia H800, announced in March - will likely take 10% to 30% longer to carry out some AI tasks and could double some costs compared with Nvidia's fastest U.S. chips.

Even the slowed Nvidia chips represent an improvement for Chinese firms. Tencent Holdings, one of China's largest tech companies, in April estimated that systems using Nvidia's H800 will cut the time it takes to train its largest AI system by more than half, from 11 days to four days.

"The AI companies that we talk to seem to see the handicap as relatively small and manageable," said Charlie Chai, a Shanghai-based analyst with 86Research.

The back-and-forth between government and industry exposes the U.S. challenge of slowing China's progress in high tech without hurting U.S. companies.

Part of the U.S. strategy in setting the rules was to avoid such a shock that the Chinese would ditch U.S. chips altogether and redouble their own chip-development efforts.

"They had to draw the line somewhere, and wherever they drew it, they were going to run into the challenge of how to not be immediately disruptive, but how to also over time degrade China's capability," said one chip industry executive who requested anonymity to talk about private discussions with regulators.

The export restrictions have two parts. The first puts a ceiling on a chip's ability to calculate extremely precise numbers, a measure designed to limit supercomputers that can be used in military research. Chip industry sources said that was an effective action.

But calculating extremely precise numbers is less relevant in AI work like large language models where the amount of data the chip can chew through is more important.

Nvidia is selling the H800 to China's largest technology firms, including Tencent, Alibaba Group Holding Ltd and Baidu Inc for use in such work, though it has not yet started shipping the chips in high volumes.

"The government isn’t seeking to harm competition or U.S. industry, and allows U.S. firms to supply products for commercial activities, such as providing cloud services for consumers," Nvidia said in a statement last week.

China is an important market for U.S. technology companies, and selling products there helps create jobs for both Nvidia and its U.S.-based partners, the company added.

"The October export controls require that we create products with an expanding gap between the two markets," Nvidia said last week. "We comply with the regulation while offering as-competitive-as-possible products in each market."

Bill Dally, Nvidia's chief scientist, said in a separate statement this week that “this gap will grow quickly over time as training requirements continue to double every six to 12 months."

A spokesperson for the Bureau of Industry and Security, the arm of the U.S. Commerce Department that oversees the rules, did not return a request for comment.

SLOWED BUT NOT STOPPED

The second U.S. limit is on chip-to-chip transfer speeds, which does affect AI. The models behind technologies such as ChatGPT are too large to fit onto a single chip. Instead, they must be spread over many chips - often thousands at a time - which all need to communicate with one another.

Nvidia has not disclosed the China-only H800 chip's performance details, but a specification sheet seen by Reuters shows a chip-to-chip speed of 400 gigabytes per second, less than half the peak speed of 900 gigabytes per second for Nvidia's flagship H100 chip available outside China.

Some in the AI industry believe that is still plenty of speed. Naveen Rao, chief executive of a startup called MosaicML that specializes in helping AI models to run better on limited hardware, estimated a 10-30% system slowdown.

"There are ways to get around all this algorithmically," he said. "I don't see this being a boundary for a very long time - like 10 years."

Money helps. A chip in China that takes twice as long to finish an AI training task than a faster U.S. chip can still get the work done. "At that point, you've got to spend $20 million instead of $10 million to train it," said one industry source who requested anonymity because of agreements with partners. "Does that suck? Yes it does. But does that mean this is impossible for Alibaba or Baidu? No, that's not a problem."

Moreover, AI researchers are trying to slim down the massive systems they have built to cut the cost of training products similar to ChatGPT and other processes. Those will require fewer chips, reducing chip-to-chip communications and lessening the impact of the U.S. speed limits.

Two years ago the industry was thinking AI models would get bigger and bigger, said Cade Daniel, a software engineer at Anyscale, a San Francisco startup that provides software to help companies perform AI work.

"If that were still true today, this export restriction would have a lot more impact," Daniel said. "This export restriction is noticeable, but it's not quite as devastating as it could have been."
It almost sounds like Nvidia's chief scientist is not in agreement with his peers in industry doing the actual AI training. It's also interesting to me that the headline is "barely slowed", while the one with a biggest stake in Nvidia sales claims "the gap will grow quickly" aka Nvidia sales has insane growth ahead of it. Meanwhile the startup CEO with the least to lose claims "I don't see this being a boundary for a very long time - like 10 years" and "No, that's not a problem".
 
Status
Not open for further replies.
Top