| Nome: | Descrição: | Tamanho: | Formato: | |
|---|---|---|---|---|
| 1.69 MB | Adobe PDF |
Orientador(es)
Resumo(s)
This paper presents a novel methodology for enhancing Automatic Speech Recognition (ASR) performance by utilizing contrastive learning to filter synthetic audio data. We address the challenge of incorporating synthetic data into ASR training, especially in scenarios with limited real-world data or unique linguistic characteristics. The method utilizes a contrastive learning model to align representations of synthetic audio and its corresponding text transcripts, enabling the identification and removal of low-quality samples that do not align well semantically. We evaluate the methodology on a medium-resource language across two distinct datasets: a general-domain dataset and a regionally specific dataset characterized by unique pronunciation patterns. Experimental results reveal that the optimal filtering strategy depends on both model capacity and dataset characteristics. Larger models, like Whisper Large V3, particularly benefit from aggressive filtering, while smaller models may not require such stringent
Descrição
Perezhohin, Y., Santos, T., Costa, V., Peres, F., & Castelli, M. (2024). Enhancing Automatic Speech Recognition: Effects of Semantic Audio Filtering on Models Performance. IEEE Access, 12, 155136 - 155150. https://doi.org/10.1109/ACCESS.2024.3482970 --- This work was supported by MyNorth AI Research. This work was supported by national funds through FCT (Fundação para a Ciência e a Tecnologia), under the project - UIDB/04152/2020 (DOI: 10.54499/UIDB/04152/2020) - Centro de Investigação em Gestão de Informação (MagIC)/NOVA IMS).
Palavras-chave
Automatic Speech Recognition Contrastive Learning Data Augmentation Embeddings Synthetic Data Filtering Text-to-Speech General Computer Science General Materials Science General Engineering SDG 8 - Decent Work and Economic Growth SDG 9 - Industry, Innovation, and Infrastructure
