Lightly supervised vs. semi-supervised training of acoustic model on Luxembourgish for low-resource automatic speech recognition
The result's identifiers
Result code in IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F00216305%3A26230%2F18%3APU130796" target="_blank" >RIV/00216305:26230/18:PU130796 - isvavai.cz</a>
Result on the web
<a href="https://www.isca-speech.org/archive/Interspeech_2018/abstracts/2361.html" target="_blank" >https://www.isca-speech.org/archive/Interspeech_2018/abstracts/2361.html</a>
DOI - Digital Object Identifier
<a href="http://dx.doi.org/10.21437/Interspeech.2018-2361" target="_blank" >10.21437/Interspeech.2018-2361</a>
Alternative languages
Result language
angličtina
Original language name
Lightly supervised vs. semi-supervised training of acoustic model on Luxembourgish for low-resource automatic speech recognition
Original language description
In this work, we focus on exploiting inexpensive data in order to to improve the DNN acoustic model for ASR. We explore two strategies: The first one uses untranscribed data from the target domain. The second one is related to the proper selection of excerpts from imperfectly transcribed out-of-domain public data, as parliamentary speeches. We found out that both approaches lead to similar results, making them equally beneficial for practical use. The Luxembourgish ASR seed system had a 38.8% WER and it improved by roughly 4% absolute, leading to 34.6% for untranscribed and 34.9% for lightlysupervised data. Adding both databases simultaneously led to 34.4% WER, which is only a small improvement. As a secondary research topic, we experiment with semi-supervised state-level minimum Bayes risk (sMBR) training. Nonetheless, for sMBR we saw no improvement from adding the automatically transcribed target data, despite that similar techniques yield good results in the case of cross-entropy (CE) training.
Czech name
—
Czech description
—
Classification
Type
D - Article in proceedings
CEP classification
—
OECD FORD branch
10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)
Result continuities
Project
<a href="/en/project/TJ01000208" target="_blank" >TJ01000208: Neural networks for speech signal processing and data mining</a><br>
Continuities
P - Projekt vyzkumu a vyvoje financovany z verejnych zdroju (s odkazem do CEP)
Others
Publication year
2018
Confidentiality
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Data specific for result type
Article name in the collection
Proceedings of Interspeech 2018
ISBN
—
ISSN
1990-9772
e-ISSN
—
Number of pages
5
Pages from-to
2883-2887
Publisher name
International Speech Communication Association
Place of publication
Hyderabad
Event location
Hyderabad, India
Event date
Sep 2, 2018
Type of event by nationality
WRD - Celosvětová akce
UT code for WoS article
—