Introducing Meta Omnilingual Automatic Speech Recognition (ASR), a suite of models providing ASR capabilities for over 1,600 languages, including 500 low-coverage languages never before served by any ASR system. While most ASR systems focus on a limited set of languages that are well-represented on the internet, this release marks a major step toward building a truly universal transcription system. 🔗 Learn more: https://go.meta.me/ff13fa Highlights include: - State-of-the-art performance with character error rates below 10 for 78% of supported languages. - First large-scale ASR framework with in-context learning, enabling extension to new languages with just a few audio samples. - A full suite of open source models and a dataset, including Omnilingual w2v 2.0, a 7B-parameter multilingual speech representation model, and Omnilingual ASR Corpus, a unique dataset spanning 350 underserved languages.
This could be transformative, especially for a country like India which has 1000+ languages and dialects, it can address problems such as making education more inclusive through native language learning tools, and improving healthcare communication in rural areas, especially with doctors from different states of the country. Truly a step towards a more "inclusive" future.
Head over to the Omnilingual demo to explore the languages in the dataset: https://aidemos.atmeta.com/omnilingualasr/language-globe
1,600 languages. 500 that have never had voice recognition before. Let that sink in. For decades, AI only spoke if you were privileged enough to speak English, Mandarin, or maybe 50 other languages. Billions of people were locked out. Their voices ignored. Their languages treated as irrelevant. Meta just shattered that barrier. This is not about better technology. This is about who gets to exist in the digital world. Communities speaking Tamasheq, Lingala, and hundreds of endangered languages now have a seat at the table. The in-context learning changes everything. A few audio samples and the system learns your language. No massive datasets required. No years of development. Just access. And they made it open source. Free for researchers, educators, and the communities who need it most. This is the AI revolution we should be talking about. Not the one that makes the rich richer. The one that gives voice to the voiceless. When 78 percent of 1,600 languages achieve error rates below 10, that is not a metric. That is millions of people finally being heard. What will you build with this?
Incredible work by the Meta team! The focus on low-coverage languages and the new in-context learning feature make this a game-changer for accessibility globally.
Truly impressive leap, Meta, expanding ASR to 1,600 languages isn’t just technological progress, it’s cultural inclusion at scale. Giving voice to underserved languages means giving identity back to communities that were digitally silent.
Outstanding progress toward universal, inclusive ASR. Exciting to see Meta pushing the boundaries of multilingual AI! 🚀
Radomir Popovic
For rich of personality, Frather ♣️🔥
That's great
Senior Researcher at Nvidia
1wSince when are Eastern and Western Armenian vulnerable or endangered languages ?