Meta releases Llama 4, a new crop of flagship AI models

Cyber Security, ICT, Most Popular, Trends News

No Comments

Photo of author

By Karla T Vasquez

WhatsApp Group Join Now
Telegram Group Join Now


Meta has released a new collection of Lama 4, AI models in her Lama family – Saturday, no less.

There are four new models in total: Lama 4 Scouts, Lama 4 Mavarick and Lama 4 Behemoth. Meta says that everyone was trained in “large -scale labeled text, images and video data” to give “wide visual understanding”.

The success of the Chinese AI Lab DIPSEC Open model, which is equal or better than the previous flagship Lama models of Meta, is reported to have kicked Llam Vikas in overdrive. It is said that there are meta scrambled war rooms to understand how models such as DPSIC R1 and V3 have reduced the cost of deploying and deploying.

Scout and Mavarick Lama.com and AI Dev Platform are publicly available from the meta partners with hugs, while Behemoth is still in training. Meta says its AI-powered assistant Meta AI AI has been updated to use Lama 4 in applications including WhatsApp, Messenger and Instagram. Multimodal features are now limited to English.

Some developer Lamas may issue issues with 4 licenses. The EU is forbidden to use or distribute users in the EU, perhaps the result of the need for administration imposed by the AI ​​and Data Privacy Act in this region. (In the past, Meta has denied these laws as an over -burden.) In addition, more than the previous Lama release, more than Million 1 million monthly active users must request a special license from the meta, which Meta may allow – or deny it on its only discretion.

“These Lama 4 models identify the beginning of a new era for the Lama ecosystem,” Meta wrote in a blog post. “It’s just starting for the collection of Lama 4″ ”

Meta says that the Lama 4 is a combination of the first models for using a mixture (MOE) architecture, which is more proficient for training and answering questions. MOE architecture basically breaks data processing tasks in subtasks and then represents small, special “specialist” models.

For example, Mavarick has 400 billion total parameters but only 17 billion Active Parameters across 128 “experts.” (Parameters are fairly adjusting to a model problem solving skills)) The scout has 17 billion active parameters, 16 experts and 109 billion total parameters.

According to Meta’s internal examination, Maverrick, which the company says that “general assistant and chat” cases like creative writing are good to use, specific coding, logic, multilingual, long -lone and image benchmarks like OPTA -4O and Google’s GPT -4O and Google’s Jemi 2.0. However, Mavarick does not fully measure the recent models such as GPT -4.5 of the GPT -4.5 of the Mavarick Google, Ethnic Clode 3.7 Sonnet and OpenAI.

The power of the scout is in the shortage of the document and the tasks like logic on the larger codebase. Uniquely, it has a very large context window: 10 million token. (Tokens present “bits of raw texts – such as” fantastic “the word” fan, “” TAS “and” tick. “Plain English can take up to the scout images and several million words, allowing it to process and work with extremely large documents.

The scout can run on a single NVEDIA H11 GPU, while an NVEDIA H1 100 DGX system is needed, according to the meta.

Meta’s unpublished Behemoth even needs Biffier hardware. According to the agency, Behmoth has 20 billion active parameters, 1 16 experts and about two trillion total parameters. As a result of the internal benchmarking of Meta, the stem efficiency such as solving the math problem has exceeded GPT -1.5, Claud 4.7 Sonnet and Gemi 2.5 Pro (but not 2.5 Pro).

Significantly, none of the Lama 4 models is a correct “argument” model in O1 and O3-Mini lines of O3-Mini. Reasonable models are answering their answers and usually reliably react to questions, but as a result it takes longer than the traditional, “non-resoning” model to provide answers.

The funny thing is, Meta says it often tunes the 4 model to refuse to answer the “controversial” question. According to the agency, Lama 4 responds to “controversial” to political and social issues that the previous crop of the Lama models is not. Furthermore, the agency says Lama 4 is “dramatically more balanced” with which it does not entertain flat-out.

“[Y]You can count [Lllama 4] Helpful without verdict, to provide practical reactions, “Meta spokesperson told TechCrunch.”[W]Continues to make Lama more responsive so that it answers more questions, react to different aspects of different aspects […] And not in favor of some opinion on others. “

These tweets come with the White House allies accused AI for political awakening.

Many presidents, including Elon Mask and Crypto and AI “Jazer” David Sacks, have complained that many AI chatbots censors the conservative aspect. The sack is the histor tihassically Out Especially the ChatzP of Openai is unpleasant about “programmed to wake up” and politically sensitive things.

In fact, bias in AI is a complex technical problem. Kastur’s own AI company, Jai, has fought to create a chatbot that does not support some political views on others.

It did not prevent companies from adjusting their AI models, including OPENA, that they could answer more questions on special questions about controversial political issues than ever before.

Leave a Comment