Meta’s Llama 3 AI Models Outperform Google’s Gemini 1.5 Pro

#Meta has launched two new artificial intelligence (AI) models, Llama 3 with 8 billion and 70 billion parameters, surpassing Google’s Gemini 1.5 Pro in performance.

New Delhi : Meta introduced its latest AI models, Llama 3 8B and 70B, on Thursday, claiming enhanced capabilities over its predecessors. The company also implemented new training techniques to optimize model efficiency. Notably, whereas the largest Llama 2 model had 70 billion parameters, Meta now plans for its larger models to exceed 400 billion parameters. Last week, a report indicated Meta would unveil smaller AI models in April, followed by larger ones in the summer.

Meta Llama 3 availability
Meta is adopting a community-first approach with Llama 3, making the new foundation models open source, like its previous models. According to Meta’s blog post, “Llama 3 models will soon be available on various platforms including AWS, Databricks, Google Cloud, Hugging Face, Kaggle, IBM WatsonX, Microsoft Azure, NVIDIA NIM, and Snowflake, with hardware support from AMD, AWS, Dell, Intel, NVIDIA, and Qualcomm.”

These partnerships cover major cloud, hosting, and hardware platforms, facilitating accessibility for AI enthusiasts. Furthermore, Meta has integrated Llama 3 with its Meta AI accessible via Facebook Messenger, Instagram, and WhatsApp in supported regions.

Meta Llama 3 performance and architecture
Regarding performance, Meta shared benchmark scores for Llama 3’s pre-trained and instruct models. The pre-trained Llama 3 70B model outperformed Google’s Gemini 1.0 Pro in MMLU (79.5 vs. 71.8), BIG-Bench Hard (81.3 vs. 75.0), and DROP (79.7 vs. 74.1) benchmarks. Additionally, the 70B Instruct model surpassed Gemini 1.5 Pro in MMLU, HumanEval, and GSM-8K benchmarks based on company data.

Meta has adopted a decoder-only transformer architecture for the new AI models, making several improvements over its predecessor. Llama 3 now uses a tokenizer with a vocabulary of 128K tokens and incorporates grouped query attention (GQA) to enhance inference efficiency. GQA ensures that the AI’s attention remains within its designated context when answering queries. Meta claims to have pre-trained the models with over 15 trillion tokens sourced from publicly available data.

Related Posts

Is AI Coming for White-Collar Jobs?

The pace of artificial intelligence (AI) innovation is reshaping the labor marketplace at a historic scale. Once thought to affect largely blue-collar jobs and routine jobs, AI is now starting…

AI-Generated Ghibli-Style art sparks debate among artists and fans

Concerns rise over creativity, ethics, and the future of art A wave of AI-generated images mimicking the unique styles of Studio Ghibli in recent weeks has sparked a deep conversation…

Leave a Reply

Your email address will not be published. Required fields are marked *

You Missed

Did you know these films are produced by Anushka Sharma?

Did you know these films are produced by Anushka Sharma?

MEA rejects Bangladesh’s remarks on West Bengal situation, asks Bangladesh to ‘Fix Its Own House First’

MEA rejects Bangladesh’s remarks on West Bengal situation, asks Bangladesh to ‘Fix Its Own House First’

PM Modi misses TIME100 list again, but here are the State Honors Modi has received globally

PM Modi misses TIME100 list again, but here are the State Honors Modi has received globally

Guru Kashi University student Rishabh Yadav wins gold at Archery World Cup

Guru Kashi University student Rishabh Yadav wins gold at Archery World Cup

This Secret Adults-Only Resort in Jamaica Lets You Party Naked—And It’s Unlike Anything You’ve Ever Seen

This Secret Adults-Only Resort in Jamaica Lets You Party Naked—And It’s Unlike Anything You’ve Ever Seen

No, Indian Passport Ranking hasn’t fallen to 149 — It’s 85th with Visa-Free access to 57 Countries in 2025

No, Indian Passport Ranking hasn’t fallen to 149 — It’s 85th with Visa-Free access to 57 Countries in 2025