Llama 4 to be faster🦙


“We're training the Llama 4 models on a cluster that is bigger than 100,000 H100s, or bigger than anything that I've seen reported for what others are doing,” Zuckerberg said, referring to the Nvidia chips popular for training AI systems. “I expect that the smaller Llama 4 models will be ready first.”

Increasing the scale of AI training with more computing power and data is widely believed to be key to developing significantly more capable AI models. 

While Meta appears to have the lead now, most of the big players in the field are likely working toward using compute clusters with more than 100,000 advanced chips. 

On Wednesday, Zuckerberg declined to offer details on Llama 4’s potential advanced capabilities but vaguely referred to “new modalities,” “stronger reasoning,” and “much faster.” 



Comments

Popular posts from this blog

Perplexity

Aphorisms: AI

Hamza Chaudhry