Transcribing code-switched bilingual lectures using deep neural networks with unit merging in acoustic modeling
Journal
ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
Pages
220-224
Date Issued
2014
Author(s)
Yeh, C.-F.
Abstract
This paper considers the transcription of the widely observed yet less investigated bilingual code-switched speech: the words or phrases of the guest language are inserted within the utterances of the host language, so the languages are switched back and forth within an utterance, and much less data are available for the guest language. Two approaches utilizing the deep neural network (DNN) were tested and analyzed, including using DNN bottleneck features in HMM/GMM (BF-HMM/GMM) and modeling context-dependent HMM senones by DNN (CD-DNN-HMM). In both cases the unit merging (and recovery) techniques in acoustic modeling were used to handle the data imbalance problem. Improved recognition accuracies were observed with unit merging (and recovery) for the two approaches under different conditions. © 2014 IEEE.
Event(s)
2014 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2014
Subjects
Bilingual; Code-switching; Deep Neural Networks; Speech Recognition; Unit Merging
SDGs
Other Subjects
Codes (symbols); Computer system recovery; FORTH (programming language); Signal processing; Speech recognition; Transcription; Acoustic model; Bilingual; Bottleneck features; Code-switching; Context dependent; Data imbalance; Deep neural networks; Recognition accuracy; Merging
Type
conference paper
