Artificial Intelligence thread

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
Please, Log in or Register to view URLs content!

Huawei Ascend AI platform has over 1.8 million developers - This is a huge number. This would be probably on the same magnitude as Cuda and they are all just writing applications for the GPU and integration with software

Please, Log in or Register to view URLs content!
中国工程院 using Huawei's Pengcheng lab to train a 200 Billion parameter NLP platform. Industry's first Chinese language NLP

Please, Log in or Register to view URLs content!
目前基于昇腾AI原生孵化了业界首个两千亿参数的中文NLP大模型鹏城。盘古、业界首个多模态大模型紫东.太初以及华为云盘古系列在内的20多个国产大模型。同时,昇腾AI也已适配支持了ChatGLM、LLaMA、GPT-3、BLOOM等数十个业界主流开源开放的大模型。据悉,中国大模型中约一半由昇腾AI支撑。
20+ domestic large models including Pangu & 紫东.太初 developed using Ascend AI platform. It has been adapted to support ChatGLM、LLaMA、GPT-3、BLOOM & other open source large model

Please, Log in or Register to view URLs content!
在论坛上,华为宣布将联合华为还联合达观数据、上海人工智能研究院、招商银行、网易伏羲、格灵深瞳、云天励飞、中国联通、科大讯飞、中国电信、达闼机器人、软通动力、面壁智能、云从科技、中国商飞上海飞机设计研究院等26家金融、运营商、互联网等行业领军企业,科研院所和高校,共同启动昇腾AI大模型联合创新。

同时,华为联合面壁智能、智谱AI、科大讯飞、云从科技四家伙伴发布昇腾AI大模型训推一体化解决方案,加速大模型在各行业应用落地,并有23家昇腾AI伙伴推出AI服务器、智能边缘与终端新品。值得一提,中国信息通信研究院、工业和信息化部电子第五研究所、华为等单位联合成立“大模型产业工作组”共同推进中国大模型应用落地及产业孵化。

此外,胡厚崑在2023世界人工智能大会开幕式上发表演讲时表示,通过架构创新,华为昇腾AI集群效率已提升10%。目前,集群规模从最初的4000卡集群扩展至16000卡,是业界首个万卡AI集群,拥有更快的训练速度和30天以上的稳定训练周期。

在发展生态方面,华为联合了5700+鲲鹏/昇腾合作伙伴,以及硬件合作伙伴30+,实现了国内大模型近一半创新使能,包括场景化系列AI硬件100+,孵化/适配大模型30+,鲲鹏/昇腾开发者380万+。在共建算力方面,华为已经在构建城市算力基础设施,帮助各地政府打造了25个昇腾人工智能计算中心。
more details here, Huawei working with Zhipu, Cloudwalk, iFlytek & ModelBest reelease Ascend AI large model training & pushed integration solution to accelerate the application of large models in industries

23 Ascend Partners launched AI servers & intelligent edge product.

CAICT, Huawei & other units established "Large model industry working group" to jointly promote the implementation of China's LLMs.

30+ LLMs using Ascend now, more than a couple of months ago
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
Screen Shot 2023-07-06 at 11.57.27 AM.png
Current size and scope of Ascend platform. A lot of developers, a lot of large models, research/lab partners & ecosystem partners

half of the large models in China developed using Ascend Platform.

They are building a 16 EFLOPS Pengcheng-3 project in Shenzhen and will work with universities to train AI talents. Remember 3 elements to AI:
computing power
data
Algo

That last part relies on AI talents
 

european_guy

Junior Member
Registered Member
Yeah, I've been posting about their Ascend data center successes for a while now

Here the million dollar question is of course: can Huawei currently manufacture/outsource Ascend chips?

I've not seen yet a clear confirmation of this.

Of course we have many hints that it can....but many hints do not make a confirmation.

Even US media, like WSJ, is quite silent on this topic. They usually are very quick to cry wolf as soon as CIA pass them info they want to spread...in many cases just fake news. But nevertheless we have nothing here.

ChatGLM2 is the number 1 trending model in huggingface.

The public released model and weights refer to the 6B version (6 billions parameter), instead the one topping
Please, Log in or Register to view URLs content!
is the big version, possibly the 130B version if, as it seems, is an improvement on the previous GLM-130B. Weights of this big version have not been released.
 
Last edited:

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
Here the million dollar question is of course: can Huawei currently manufacture/outsource Ascend chips?

I've not seen yet a clear confirmation of this.

Of course we have many hints that it can....but many hints do not make a confirmation.

Even US media, like WSJ, is quite silent on this topic. They usually are very quick to cry wolf as soon as CIA pass them info they want to spread...in many cases just fake news. But nevertheless we have nothing here.



The public released model and weights refer to the 6B version (6 billions parameter), instead the one topping
Please, Log in or Register to view URLs content!
is the big version, possibly the 130B version if, as it seems, is an improvement on the previous GLM-130B. Weights of this big version have not been released.
well, Huawei/Ascend guy said they will have a new Kunpeng & Ascend chip later this year. You can make your own guesses on where that's made. Maybe they will tell us, but most likely not

The specs for the GPU should be easy to derive.

PengCheng-3 lab is expected to be 16 EFLOPs (FP16 I think). Atlas-900 cluster can be as large as 16000 cards based on the unveiling from today. So each card should be doing about 1024 TFLOPS FP16. That's basically what Biren's BR100 is performing except that they are producing nerfed version of BR104 to stay under the radar. But fabbing it on 7nm should be doable.
 

tokenanalyst

Brigadier
Registered Member
Here the million dollar question is of course: can Huawei currently manufacture/outsource Ascend chips?

I've not seen yet a clear confirmation of this.

Of course we have many hints that it can....but many hints do not make a confirmation.

Even US media, like WSJ, is quite silent on this topic. They usually are very quick to cry wolf as soon as CIA pass them info they want to spread...in many cases just fake news. But nevertheless we have nothing here.



The public released model and weights refer to the 6B version (6 billions parameter), instead the one topping
Please, Log in or Register to view URLs content!
is the big version, possibly the 130B version if, as it seems, is an improvement on the previous GLM-130B. Weights of this big version have not been released.
The 130B model need server like hardware or a cloud service to run properly, the 6B models can run on consumer hardware and its seems to be pretty good.

Please, Log in or Register to view URLs content!
 
Top