📢 Gate Square Exclusive: #WXTM Creative Contest# Is Now Live!
Celebrate CandyDrop Round 59 featuring MinoTari (WXTM) — compete for a 70,000 WXTM prize pool!
🎯 About MinoTari (WXTM)
Tari is a Rust-based blockchain protocol centered around digital assets.
It empowers creators to build new types of digital experiences and narratives.
With Tari, digitally scarce assets—like collectibles or in-game items—unlock new business opportunities for creators.
🎨 Event Period:
Aug 7, 2025, 09:00 – Aug 12, 2025, 16:00 (UTC)
📌 How to Participate:
Post original content on Gate Square related to WXTM or its
GPT-4 is a mixed model of 8x220 billion parameters? This gossip went crazy today
Source: Heart of the Machine
"The parameter volume of GPT-4 is as high as 100 trillion." I believe many people still remember the "heavy" news that swiped the screen at the beginning of this year and a chart that was spread virally.
In fact, many people believe and spread such rumors because the AI community has been increasing the parameter size of AI models in recent years. The Switch Transformer released by Google in January 2021 raised the parameters of the AI large model to 1.6 trillion. Since then, many institutions have successively launched their own trillion-parameter large models. Based on this, people have every reason to believe that GPT-4 will be a huge model with trillions of parameters, and 100 trillion parameters is not impossible.
Although Sam Altman's rumor helped us get rid of a wrong answer, the OpenAI team behind him has been tight-lipped about the real parameters of GPT-4, and even the official technical report of GPT-4 did not disclose any information.
Until recently, this mystery was suspected to be pierced by the "genius hacker" George Hotz.
George Hotz is famous for cracking the iPhone at the age of 17 and hacking the Sony PS3 at the age of 21. He is currently the boss of a company (comma.ai) that develops automatic driving assistance systems.
He was recently interviewed by an AI tech podcast called Latent Space. In the interview, he talked about GPT-4, saying that GPT-4 is actually a hybrid model. Specifically, it uses an ensemble system of 8 expert models, each with 220 billion parameters (slightly more than GPT-3's 175 billion parameters), and these models have been trained on different data and task distribution training.
As for the future trend, he believes that people will train smaller models and improve performance through long-term fine-tuning and discovering various tricks. He mentioned that compared with the past, the training effect has been significantly improved, although the computing resources have not changed, which shows that the improvement of the training method has played a big role.
At present, George Hotz's "breaking news" about GPT-4 has been widely spread on Twitter.