Book Image

Voice Application Development for Android

Book Image

Voice Application Development for Android

Overview of this book

Speech technology has been around for some time now. However, it has only more recently captured the imagination of the general public with the advent of personal assistants on mobile devices that you can talk to in your own language. The potential of voice apps is huge as a novel and natural way to use mobile devices. Voice Application Development for Android is a practical, hands-on guide that provides you with a series of clear, step-by-step examples which will help you to build on the basic technologies and create more advanced and more engaging applications. With this book, you will learn how to create useful voice apps that you can deploy on your own Android device in no time at all. This book introduces you to the technologies behind voice application development in a clear and intuitive way. You will learn how to use open source software to develop apps that talk and that recognize your speech. Building on this, you will progress to developing more complex apps that can perform useful tasks, and you will learn how to develop a simple voice-based personal assistant that you can customize to suit your own needs. For more interesting information about the book, visit http://lsi.ugr.es/zoraida/androidspeechbook
Table of Contents (19 chapters)
Voice Application Development for Android
Credits
Foreword
About the Authors
Acknowledgement
About the Reviewers
www.PacktPub.com
Preface
Afterword
Index

Multimodality


An application may be considered multimodal if it uses several input and/or output modalities. In this sense, all the apps presented in this book are multimodal, as they use voice as well as a GUI either in the output or input.

However, in the previous examples the modalities were not synchronized or used to provide alternative ways for handling the same bits of information. For example, in the MusicBrain app in Chapter 5, Form-filling Dialogs the input was oral and the output visual, and in the SimpleParrot app described in this chapter, the input and output are oral (requiring speech recognition and synthesis) and visual (selecting a language and the push to speak button in the input and toasts for output), but they correspond to different elements in the interface.

In this section, we will describe how to develop multimodal applications in which the modalities can be seamlessly combined to input or output data, so the user can select the most convenient modality to use at...