Meta’s AI translator can interpret unwritten languages

Nearly half of the world’s roughly 7,000 known languages four in ten of them exist without an accompanying written component. These unwritten languages pose a unique problem for modern machine learning translation systems, as they typically need to convert verbal speech to written words before translating to the new language and reverting the text back to speech, but one that Meta has reportedly addressed with its latest open-source language AI advancement.

As part of Meta’s Universal Speech Translator (UST) program which is working to develop real-time speech-to-speech translation so that can more easily interact (read: ). As part of this project, Meta researchers looked at Hokkien, an unwritten language spoken throughout Asia’s diaspora and one of Taiwan’s official languages.

Machine learning translation systems typically require extensive labelable examples of the language, both written and spoken, to train on — precisely what unwritten languages like Hokkien don’t have. To get around that, “we used speech-to-unit translation (S2UT) to convert input speech to a sequence of acoustic units directly in the path previously pioneered by Meta,” CEO Mark Zuckerberg explained in a Wednesday blog post. “Then, we generated waveforms from the units. In addition, UnitY was adopted for a two-pass decoding mechanism where the first-pass decoder generates text in a related language (Mandarin), and the second-pass decoder creates units.”

“We leveraged Mandarin as an intermediate language to build pseudo-labels, where we first translated English (or Hokkien) speech to Mandarin text, and we then translated to Hokkien (or English) and added it to training data,” he continued. Currently, the system allows for someone who speaks Hokkien to converse with someone who speaks English, albeit stiltedly. The model can only translate one full sentence at a time but Zuckerberg is confident that the technique can eventually be applied to more languages and will improve to the point of offering real-time translation.

In addition to the models and training data that Meta is already open-sourcing from this project, the company is also releasing a first-of-its-kind speech-to-speech translation benchmarking system based on a Hokkien speech corpus called Taiwanese Across Taiwan, as well as “the speech matrix, a large corpus of speech-to-speech translations mined with Meta’s innovative data mining technique called LASER,” Zuckerberg announced. This system will empower researchers to create speech-to-speech translation (S2ST) systems of their own.

All products recommended by Engadget are selected by our editorial team, independent of our parent company. Some of our stories include affiliate links. If you buy something through one of these links, we may earn an affiliate commission. All prices are correct at the time of publishing.

For all the latest Technology News Click Here 

 For the latest news and updates, follow us on Google News

Read original article here

Denial of responsibility! TechNewsBoy.com is an automatic aggregator around the global media. All the content are available free on Internet. We have just arranged it in one platform for educational purpose only. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials on our website, please contact us by email – [email protected]. The content will be deleted within 24 hours.