Date of Original Version

9-2012

Type

Conference Proceeding

Journal Title

Proceedings of INTERSPEECH

First Page

2586

Last Page

2589

Rights Management

Copyright 2012 ISCA

Abstract or Description

In this paper we present our latest investigation on initialization schemes for Multilayer Perceptron (MLP) training using multilingual data. We show that the overall performance of an MLP network improves significantly by initializing it with a multilingual MLP. We propose a new strategy called "open target language" MLP to train more flexible models for language adaptation, which is particularly suited for small amounts of training data. Furthermore, by applying Bottle-Neck feature (BN) initialized with multilingual MLP the ASR performance increases on both, on those languages which were used for multilingual MLP training, and on a new language. Our experiments show word error rate improvements of up to 16.9% relative on a range of tasks for different target languages (Creole and Vietnamese) with manually and automatically transcribed training data.

Share

COinS
 

Published In

Proceedings of INTERSPEECH, 2586-2589.