1. Ando, R.K., Zhang, T.: A framework for learning predictive structures from multiple tasks and unlabeled data. J. Mach. Learn. Res. 6, 1817–1853 (2005)
2. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: arXiv preprint arXiv:1409.0473 (2014)
3. Chen, Z., Watanabe, S.: Speech enhancement and recognition using multi-task learning of long short-term memory recurrent neural networks. In: InterSpeech’15 (2015)
4. Chung, J., Gucehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. In: arXiv preprint arXiv:1412.3555 (2014)
5. Collobert, R., Weston, J.: A unified architecture for natural languageprocessing: deep neural networks with multitask learning. In: Proceedings of the 25th International Conference on Machine Learning, pp. 160–167 (2008)