Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Bilingual Asr Model With Language Identification for Brazilian Portuguese and South-American Spanish

Version 1 : Received: 28 October 2022 / Approved: 31 October 2022 / Online: 31 October 2022 (10:06:34 CET)

How to cite: Farias, F.; Lobato, W.; Cruz, W.; Amadeus, M. Bilingual Asr Model With Language Identification for Brazilian Portuguese and South-American Spanish. Preprints 2022, 2022100480. https://doi.org/10.20944/preprints202210.0480.v1 Farias, F.; Lobato, W.; Cruz, W.; Amadeus, M. Bilingual Asr Model With Language Identification for Brazilian Portuguese and South-American Spanish. Preprints 2022, 2022100480. https://doi.org/10.20944/preprints202210.0480.v1

Abstract

This paper documents the development of a special case of multilingual Automatic Speech Recognition model, specifically tailored to attend two languages spoken by the majority of Latin America, Portuguese and Spanish. The bilingual model combines Language Identification and Speech Recognition developed with the Wav2Vec2.0 architecture and trained on several open and private speech datasets. In this model, the feature encoder is trained jointly for all tasks and different context encoders are trained for each task. The model is evaluated separately on two tasks: language identification and speech recognition. The results indicate that this model achieves good performance on speech recognition and average performance on language identification, training on a low quantity of speech material. The average accuracy of the language identification module on the MLS dataset is 66.75%. The average Word Error Rate in the same scenario is 13.89%, which is better than average 22.58% achieved by the commercial speech recognizer developed by Google.

Keywords

Speech Recognition; Automatic Speech Recognition; Language Identification; Wav2Vec2; Multilingual

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.