Is ChatGLM2 by Tsinghua really so good? I read it is a open-source bilingual model.
20+ domestic large models including Pangu & 紫东.太初 developed using Ascend AI platform. It has been adapted to support ChatGLM、LLaMA、GPT-3、BLOOM & other open source large model目前基于昇腾AI原生孵化了业界首个两千亿参数的中文NLP大模型鹏城。盘古、业界首个多模态大模型紫东.太初以及华为云盘古系列在内的20多个国产大模型。同时,昇腾AI也已适配支持了ChatGLM、LLaMA、GPT-3、BLOOM等数十个业界主流开源开放的大模型。据悉,中国大模型中约一半由昇腾AI支撑。
more details here, Huawei working with Zhipu, Cloudwalk, iFlytek & ModelBest reelease Ascend AI large model training & pushed integration solution to accelerate the application of large models in industries在论坛上,华为宣布将联合华为还联合达观数据、上海人工智能研究院、招商银行、网易伏羲、格灵深瞳、云天励飞、中国联通、科大讯飞、中国电信、达闼机器人、软通动力、面壁智能、云从科技、中国商飞上海飞机设计研究院等26家金融、运营商、互联网等行业领军企业,科研院所和高校,共同启动昇腾AI大模型联合创新。
同时,华为联合面壁智能、智谱AI、科大讯飞、云从科技四家伙伴发布昇腾AI大模型训推一体化解决方案,加速大模型在各行业应用落地,并有23家昇腾AI伙伴推出AI服务器、智能边缘与终端新品。值得一提,中国信息通信研究院、工业和信息化部电子第五研究所、华为等单位联合成立“大模型产业工作组”共同推进中国大模型应用落地及产业孵化。
此外,胡厚崑在2023世界人工智能大会开幕式上发表演讲时表示,通过架构创新,华为昇腾AI集群效率已提升10%。目前,集群规模从最初的4000卡集群扩展至16000卡,是业界首个万卡AI集群,拥有更快的训练速度和30天以上的稳定训练周期。
在发展生态方面,华为联合了5700+鲲鹏/昇腾合作伙伴,以及硬件合作伙伴30+,实现了国内大模型近一半创新使能,包括场景化系列AI硬件100+,孵化/适配大模型30+,鲲鹏/昇腾开发者380万+。在共建算力方面,华为已经在构建城市算力基础设施,帮助各地政府打造了25个昇腾人工智能计算中心。
Yeah, I've been posting about their Ascend data center successes for a while now
ChatGLM2 is the number 1 trending model in huggingface.
well, Huawei/Ascend guy said they will have a new Kunpeng & Ascend chip later this year. You can make your own guesses on where that's made. Maybe they will tell us, but most likely notHere the million dollar question is of course: can Huawei currently manufacture/outsource Ascend chips?
I've not seen yet a clear confirmation of this.
Of course we have many hints that it can....but many hints do not make a confirmation.
Even US media, like WSJ, is quite silent on this topic. They usually are very quick to cry wolf as soon as CIA pass them info they want to spread...in many cases just fake news. But nevertheless we have nothing here.
The public released model and weights refer to the 6B version (6 billions parameter), instead the one topping is the big version, possibly the 130B version if, as it seems, is an improvement on the previous GLM-130B. Weights of this big version have not been released.
The 130B model need server like hardware or a cloud service to run properly, the 6B models can run on consumer hardware and its seems to be pretty good.Here the million dollar question is of course: can Huawei currently manufacture/outsource Ascend chips?
I've not seen yet a clear confirmation of this.
Of course we have many hints that it can....but many hints do not make a confirmation.
Even US media, like WSJ, is quite silent on this topic. They usually are very quick to cry wolf as soon as CIA pass them info they want to spread...in many cases just fake news. But nevertheless we have nothing here.
The public released model and weights refer to the 6B version (6 billions parameter), instead the one topping is the big version, possibly the 130B version if, as it seems, is an improvement on the previous GLM-130B. Weights of this big version have not been released.