Open Access. Powered by Scholars. Published by Universities.®

Computer Sciences Commons

Open Access. Powered by Scholars. Published by Universities.®

Artificial Intelligence and Robotics

Natural Language Processing Faculty Publications

2022

Embeddings

Articles 1 - 1 of 1

Full-Text Articles in Computer Sciences

Supervised Acoustic Embeddings And Their Transferability Across Languages, Sreepratha Ram, Hanan Aldarmaki Dec 2022

Supervised Acoustic Embeddings And Their Transferability Across Languages, Sreepratha Ram, Hanan Aldarmaki

Natural Language Processing Faculty Publications

In speech recognition, it is essential to model the phonetic content of the input signal while discarding irrelevant factors such as speaker variations and noise, which is challenging in low-resource settings. Self-supervised pretraining has been proposed as a way to improve both supervised and unsupervised speech recognition, including frame-level feature representations and Acoustic Word Embeddings (AWE) for variable-length segments. However, self-supervised models alone cannot learn perfect separation of the linguistic content as they are trained to optimize indirect objectives. In this work, we experiment with different pre-trained self-supervised features as input to AWE models and show that they work best …