💙 Gate Square #Gate Blue Challenge# 💙
Show your limitless creativity with Gate Blue!
📅 Event Period
August 11 – 20, 2025
🎯 How to Participate
1. Post your original creation (image / video / hand-drawn art / digital work, etc.) on Gate Square, incorporating Gate’s brand blue or the Gate logo.
2. Include the hashtag #Gate Blue Challenge# in your post title or content.
3. Add a short blessing or message for Gate in your content (e.g., “Wishing Gate Exchange continued success — may the blue shine forever!”).
4. Submissions must be original and comply with community guidelines. Plagiarism or re
Huawei's large model is finally here, my evaluation is: quite shocking
Original source: Bad review
Huawei, which has always been said to be lagging behind in the large-scale model competition, finally came with its guys this time.
No, at yesterday's Huawei Developer Conference 2023, Huawei showed off.
The nearly three-hour press conference still inherited Huawei's past hodgepodge style, which made Shichao dazzled.
However, summing it up actually highlights a theme: Pangu Large Model 3.0.
The most important thing is that its prediction accuracy even surpasses the IFS system of the European Meteorological Center, which is known as the world's strongest. It is the first AI prediction product that has won the traditional numerical prediction.
Moreover, the previous AI model will continue to accumulate iteration errors during the prediction process, which will easily affect the accuracy of the results.
Therefore, AI prediction methods have not been popular.
The Pangu meteorological large-scale model is awesome. They used a three-dimensional neural network called 3DEST to process meteorological data. If 2D can't do it, they can use 3D.
3DEST's Network Training and Inference Strategy
Although this word sounds easy to be fooled, it is actually very easy to understand.
For example, the previous AI weather forecasting model FourCastNet, before the typhoon comes, it will make a forecast 6 hours in advance, and during these 6 hours, the model will calculate repeatedly when the typhoon will come.
It may be calculated for 5 hours for a while, and 4 and a half hours for a while, and the error will be large if these results are added together.
But the Pangu Meteorological Large Model thought of a way to train 4 models with different forecast intervals, one iteration per 1 hour, and one iteration per 3 hours, 6 hours and 24 hours.
Then, according to the specific weather forecast requirements, select the corresponding model for iteration.
**The fewer iterations, the smaller the error. **
This wave of operations has brought weather forecasting to a new level.
However, some friends may have started to mutter. People’s large models are all generated images and texts. How did Huawei become a weather forecast?
One thing to say, this Pangu model is indeed different from the ChatGPT and Midjourney we have come into contact with before. They are doing business in the industry.
It is not the ChatGPT "nemesis" that everyone expects, but it is aimed at the To B market that is not usually accessible. **
Let’s not mention the difficulty or not, at least the enterprise customer resources that Huawei has accumulated over the years are really easy to cash out.
Moreover, Huawei's press conference this time not only brought the ruthless role of the weather forecasting model.
No new antibiotics have been discovered for more than 40 years, and the super antibacterial drug Drug X was found as soon as the Pangea drug molecular model came, and the drug development cycle was shortened from several years to several months, and the research and development costs were reduced by 70%.
You know, for a coal preparation plant with an annual output of 10 million tons of coking coal, every 0.1% increase in the clean coal production rate can increase the annual profit by 10 million.
**This is all white money. . . **
In fact, in addition to the weather forecasting, drug development and coal preparation mentioned above, the Pangea model has been used in many industries.
Huawei is able to mass-produce these large models of various industries, thanks to the 5+N+X three-layer architecture of Huawei Pangu Model 3.0.
Why do you say that?
Because AI is landing in the industry, data is a major difficulty.
Zhang Pingan said at the press conference, "Due to the difficulty in obtaining industry data and the difficulty in combining technology with the industry, the implementation of large models in the industry has been slow."
**Pangu is very ingenious, through the three-tier structure of 5+N+X, directly split this big problem into 3 small problems to solve. **
First of all, the five large models of Pangu's L0 layer learned hundreds of terabytes of text data such as encyclopedia knowledge, literary works, program codes, and billions of Internet images with text labels.
Then, the model in the second layer L1 is formed by learning the data of N related industries from a certain basic large model in L0. This is like the undergraduate stage of a university, where you need to choose a variety of majors to study.
But after all, one is a hospital and the other is a factory, and the usage scenarios are completely different. It will definitely not work to rely on the basic large model alone, but if the industry data is added, there may be surprises.
At the same time, Huawei has also added a feedback link, which is a bit like an internship in the company.
According to them, it usually took 5 months to develop a GPT-3 scale industry model in the past; with this set of tools, the development cycle can be shortened to 1/5 of the original.
At the same time, the limitations of small data sets in many industries can also be resolved. For example, a very detailed industry such as the manufacture of large aircraft can also have large models.
As we all know, we are really embarrassed in terms of AI computing power.
First, we cannot buy Nvidia's H100/A100, the core equipment of the AI industry. Second, even if Nvidia "intimately" released a replacement for the H800, we still have reservations. For example, the transmission rate has been cut a lot.
In the context of a large model that takes several months to train, it is easy to be overtaken by foreign counterparts with stronger computing power.
And this time, to solve this problem, Huawei still took out some real guys.
However, in practice, there are still some gaps. And the A100 isn't Nvidia's ultimate weapon either.
For example, according to the press conference, count the AI Ascend Cloud Computing Power Base and the computing framework CANN. . . In other aspects, Huawei's efficiency in training large models is 1.1 times that of mainstream GPUs in the industry.
Still quite impressive.
Moreover, Huawei also said that they now have nearly 4 million developers. This number is aligned with the NVIDIA CUDA ecosystem.
Generally speaking, after watching a Huawei press conference, the bad reviewers feel that Huawei’s layout in AI is very profound, and they have already begun to think about the question of "what AI can really bring us".
In the past six months, although the AI industry has received thunderous applause, it is somewhat embarrassing when it really falls to the industry level.
And this action of Huawei just confirmed what Ren Zhengfei said:
*" In the future, there will be a surge in AI large models, not just Microsoft. The direct contribution of artificial intelligence software platform companies to human society may be less than 2%, and 98% is the promotion of industrial society and agricultural society. " *
In the field of AI, the real big era is yet to come.