Tied Multitask Learning for Neural Speech Translation
Abstract
We explore multitask models for neural translation of speech, augmenting them in order to reflect two intuitive notions. First, we introduce a model where the second task decoder receives information from the decoder of the first task, since higher-level intermediate representations should provide useful information. Second, we apply regularization that encourages transitivity and invertibility. We show that the application of these notions on jointly trained models improves performance on the tasks of low-resource speech transcription and translation. It also leads to better performance when using attention information for word discovery over unsegmented input.
- Publication:
-
arXiv e-prints
- Pub Date:
- February 2018
- DOI:
- 10.48550/arXiv.1802.06655
- arXiv:
- arXiv:1802.06655
- Bibcode:
- 2018arXiv180206655A
- Keywords:
-
- Computer Science - Computation and Language
- E-Print:
- accepted at NAACL-HLT 2018