, Huawei will release a 100-billion-level large-scale model product called "Pangu Chat" that directly targets ChatGPT!
According to Huawei's internal sources, it is expected that Huawei Pangu Chat will be released and internally tested at the Huawei Cloud Developer Conference held on July 7 this year. The products are mainly for government and enterprise customers.
This means that in the "arms race" of domestic large-scale models, after Ali and Baidu, another important technology giant has entered the game.
It is reported that the Pangu model was successfully established in Huawei Cloud in November 2020. For the positioning of the Pangu model, Huawei's internal team has established three key core design principles: first, the model must be large enough to absorb massive amounts of data; The generalization ability can truly be applied to work scenarios in all walks of life.
According to a paper released by Huawei, the parameters of Huawei's Pangu-Σ large model are at most 1.085 trillion, which is developed based on Huawei's self-developed MindSpore framework. Overall, the PanGu-Σ large model may be close to the level of GPT-3.5 in terms of dialogue.
According to public information, the Huawei Pangu model was officially released in April 2021, and was later upgraded to version 2.0 in April 2022. At present, the NLP large model, CV large model, and scientific computing large model (meteorological large model) in the AI large model have been marked as coming online.
According to reports, this is the first Chinese pre-training large model with 100 billion parameters, and the CV large model has reached 3 billion parameters for the first time. Pangu CV Large Model The largest CV large model in the industry, it is the first to achieve both discrimination and generation capabilities, and it is the industry's first in the small sample learning ability on ImageNet; the Pangu meteorological large model provides second-level weather forecasts; Zidong Taichu is the world's first map , text, and audio three-modal large model.
For the positioning of the Pangu model, Huawei's internal team has established three key core design principles: first, the model must be large enough to absorb massive amounts of data; The generalization ability can truly be applied to work scenarios in all walks of life.
According to the PPT information of Huawei Cloud executives' speeches, at present, the basic layer of Huawei's "Pangu series AI large model" mainly includes NLP large model, CV large model, and scientific computing large model, etc., and the upper layer is Huawei's industry large model developed with partners .
The Huawei Cloud official website shows that the Pangu large model is composed of multiple large models such as NLP large model, CV large model, multi-modal large model, and scientific computing large model. AI scale, industrialization problems, can support a variety of natural language processing tasks, including text generation, text classification, question answering system and so on.
Specifically, the Pangu NLP large model uses the Encoder-Decoder architecture for the first time, taking into account the comprehension and generation capabilities of the NLP large model, ensuring the flexibility of embedding the model in different systems. In downstream applications, only a small number of samples and learnable parameters are needed to complete the rapid fine-tuning and downstream adaptation of a large-scale model of 100 billion. This model has a good performance in intelligent public opinion and intelligent marketing.