- AiNews.com
- Posts
- Meta to Increase Computing Power 10x for Llama 4 Training
Meta to Increase Computing Power 10x for Llama 4 Training
Meta to Increase Computing Power 10x for Llama 4 Training
Meta, the developer behind the foundational open-source large language model Llama, plans to significantly ramp up its computing power for future model training. According to Mark Zuckerberg, Meta will need ten times more computing power to train Llama 4 compared to Llama 3. This strategic move is aimed at maintaining a competitive edge in the AI space.
Zuckerberg's Statement on Increased Computing Needs
During Meta's second-quarter earnings call, Zuckerberg highlighted the increasing demands of AI model training. "The amount of computing needed to train Llama 4 will likely be almost 10 times more than what we used to train Llama 3, and future models will continue to grow beyond that," he stated. He emphasized the importance of building capacity ahead of demand to avoid falling behind competitors due to the long lead times for new inference projects.
Meta's Recent and Upcoming Model Releases
Meta released Llama 3 with 8 billion parameters in April and recently unveiled an upgraded version, Llama 3.1 405B, featuring 405 billion parameters. This makes it Meta's largest open-source model to date. The company's commitment to advancing its AI capabilities is evident in these substantial upgrades.
Investment in Data Centers and Infrastructure
Meta's CFO, Susan Li, mentioned that the company is considering various data center projects to support future AI model training. This investment is expected to increase capital expenditures in 2025. Meta's capital expenditures rose by nearly 33% in Q2 2024, reaching $8.5 billion, driven by investments in servers, data centers, and network infrastructure.
Comparative Costs in AI Training
Training large language models is a costly endeavor. A report from The Information indicated that OpenAI spends $3 billion on training models and an additional $4 billion on renting servers at a discounted rate from Microsoft. Meta's strategy involves scaling its generative AI training capacity to advance its foundation models, providing flexibility in how the infrastructure is used over time.
Global Reach and Market Insights
During the earnings call, Meta also discussed its consumer-facing AI products. India emerged as the largest market for Meta AI's chatbot. However, Li noted that the company does not expect generative AI products to significantly contribute to revenue in the near term.
Conclusion
Meta's proactive approach to increasing its computing power for AI model training underscores its commitment to staying at the forefront of AI development. By investing heavily in infrastructure and preparing for future demands, Meta aims to ensure its AI models remain competitive and capable of advancing the company's technological capabilities.