Meta’s next Llama models may have upgraded voice features

Nikesh Vaishnav
Disclosure: This website may contain affiliate links, which means I may earn a commission if you click on the link and make a purchase. I only recommend products or services that I personally use and believe will add value to my readers. Your support is appreciated!

Meta’s next major “open” AI model may have a voice focus, per a report in Financial Times.

According to the piece, Meta is planning to introduce improved voice features with Llama 4, the next flagship in its Llama model family, which is expected to arrive in “weeks.” Reportedly, Meta has been particularly focused on allowing users to interrupt the model mid-speech, similar to OpenAI’s Voice Mode for ChatGPT and Google’s Gemini Live experience.

In comments this week at a Morgan Stanley conference, Meta chief product officer Chris Cox said that Llama 4 will be an “omni” model, capable of natively interpreting and outputting speech as well as text and other types of data.

The success of open models from the Chinese AI lab DeepSeek, which perform on par or better than Meta’s Llama models, has kicked Llama development into overdrive. Meta is said to have scrambled to set up war rooms to decipher how DeepSeek lowered the cost of running and deploying models.

Share This Article
Leave a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *