More about HKUST
MULTI-TASK LEARNING FOR AUTOMATIC SPEECH RECOGNITION
PhD Thesis Proposal Defence
Title: "MULTI-TASK LEARNING FOR AUTOMATIC SPEECH RECOGNITION"
by
Mr. Dongpeng CHEN
Abstract:
It is well-known in machine learning that multi-task learning (MTL) can
help improve the generalization performance of singly learning tasks if
the tasks being trained in parallel are related, and the effect is more
prominent when the amount of training data is relatively small. Recently,
deep neural network (DNN) has been widely utilized as acoustic model in
ASR. We propose applying MTL on DNN to exploit extra information from the
training data, without requiring additional language resources, which is
great benefit when language resources are limited. In the first method,
phone and grapheme models are trained together within the same acoustic
model and the extra information is the phone-to-grapheme mappings, while
in the second method, a universal phone set (UPS) modeling task is learned
with language-specific triphones modeling tasks to help implicitly map the
phones of multiple languages. Although the methods were initially proposed
for low-resource speech recognition, we also generalized and applied the
idea to large vocabulary speech recognition tasks. Experiment results on
three low-resource South African languages in the Lwazi corpus, the TIMIT
English phone recognition task and the Wall Street Journal English reading
speech recognition tasks show the MLT-DNNs consistently outperform
single-task learning (STL) DNN.
Date: Tuesday, 12 May 2015
Time: 2:00pm - 4:00pm
Venue: Room 5503
lifts 25/26
Committee Members: Dr. Brian Mak (Supervisor)
Prof. Nevin Zhang (Chairperson)
Prof. Dit-Yan Yeung
Dr. Raymond Wong
**** ALL are Welcome ****