1. Beskow, J., Edlund, J., and Nordstrand, M. (2004). A model for multimodal dialogue system output applied to an animated talking head. In Minker, W., Bühler, D., and Dybkjær, L., editors, Spoken Multimodal Human-Computer Dialogue in Mobile Environments. Kluwer Academic Publishers, Dordrecht, The Netherlands. (this volume).
2. Bühler, D. and Minker, W. (2004). A reasoning component for information-seeking and planning dialogues. In Minker, W., Bühler, D., and Dybkjær, L., editors, Spoken Multimodal Human-Computer Dialogue in Mobile Environments. Kluwer Academic Publishers, Dordrecht, The Netherlands. (this volume).
3. Furui, S. (2004). Speech recognition technology in multimodal/ubiquitous computing environments. In Minker, W., Bühler, D., and Dybkjær, L., editors, Spoken Multimodal Human-Computer Dialogue in Mobile Environments. Kluwer Academic Publishers, Dordrecht, The Netherlands. (this volume).
4. Macherey, K. and Ney, H. (2004). Feature functions for tree-based dialogue course management. In Minker, W., Bühler, D., and Dybkjær, L., editors, Spoken Multimodal Human-Computer Dialogue in Mobile Environments. Kluwer Academic Publishers, Dordrecht, The Netherlands. (this volume).
5. Tamura, S., Iwano, K., and Furui, S. (2004). A robust multimodal speech recognition method using optical flow analysis. In Minker, W., Bühler, D., and Dybkjær, L., editors, Spoken Multimodal Human-Computer Dialogue in Mobile Environments. Kluwer Academic Publishers, Dordrecht, The Netherlands. (this volume).