A voice-user interface (VUI) enables spoken human interaction with computers, using speech recognition to understand spoken commands and answer questions, and typically text to speech to play a reply. A voice command device is a device controlled with a voice user interface.

Voice user interfaces have been added to automobiles, home automation systems, computer operating systems, home appliances like washing machines and microwave ovens, and television remote controls. They are the primary way of interacting with virtual assistants on smartphones and smart speakers. Older automated attendants (which route phone calls to the correct extension) and interactive voice response systems (which conduct more complicated transactions over the phone) can respond to the pressing of keypad buttons via DTMF tones, but those with a full voice user interface allow callers to speak requests and responses without having to press any buttons.

Newer voice command devices are speaker-independent, so they can respond to multiple voices, regardless of accent or dialectal influences. They are also capable of responding to several commands at once, separating vocal messages, and providing appropriate feedback, accurately imitating a natural conversation.[1]

Overview

edit

A VUI is the interface to any speech application. Only a short time ago, controlling a machine by simply talking to it was only possible in science fiction. Until recently, this area was considered to be artificial intelligence. However, advances in technologies like text-to-speech, speech-to-text, natural language processing, and cloud services contributed to the mass adoption of these types of interfaces. VUIs have become more commonplace, and people are taking advantage of the value that these hands-free, eyes-free interfaces provide in many situations.

VUIs need to respond to input reliably, or they will be rejected and often ridiculed by their users. Designing a good VUI requires interdisciplinary talents of computer science, linguistics and human factors psychology – all of which are skills that are expensive and hard to come by. Even with advanced development tools, constructing an effective VUI requires an in-depth understanding of both the tasks to be performed, as well as the target audience that will use the final system. The closer the VUI matches the user's mental model of the task, the easier it will be to use with little or no training, resulting in both higher efficiency and higher user satisfaction.

A VUI designed for the general public should emphasize ease of use and provide a lot of help and guidance for first-time callers. In contrast, a VUI designed for a small group of power users (including field service workers), should focus more on productivity and less on help and guidance. Such applications should streamline the call flows, minimize prompts, eliminate unnecessary iterations and allow elaborate "mixed initiative dialogs", which enable callers to enter several pieces of information in a single utterance and in any order or combination. In short, speech applications have to be carefully crafted for the specific business process that is being automated.

Not all business processes render themselves equally well for speech automation. In general, the more complex the inquiries and transactions are, the more challenging they will be to automate, and the more likely they will be to fail with the general public. In some scenarios, automation is simply not applicable, so live agent assistance is the only option. A legal advice hotline, for example, would be very difficult to automate. On the flip side, speech is perfect for handling quick and routine transactions, like changing the status of a work order, completing a time or expense entry, or transferring funds between accounts.

History

edit

Early applications for VUI included voice-activated dialing of phones, either directly or through a (typically Bluetooth) headset or vehicle audio system.

In 2007, a CNN business article reported that voice command was over a billion dollar industry and that companies like Google and Apple were trying to create speech recognition features.[2] In the years since the article was published, the world has witnessed a variety of voice command devices. Additionally, Google has created a speech recognition engine called Pico TTS and Apple released Siri. Voice command devices are becoming more widely available, and innovative ways for using the human voice are always being created. For example, Business Week suggests that the future remote controller is going to be the human voice. Currently Xbox Live allows such features and Jobs hinted at such a feature on the new Apple TV.[3]

Voice command software products on computing devices

edit

Both Apple Mac and Windows PC provide built in speech recognition features for their latest operating systems.

Microsoft Windows

edit

Two Microsoft operating systems, Windows 7 and Windows Vista, provide speech recognition capabilities. Microsoft integrated voice commands into their operating systems to provide a mechanism for people who want to limit their use of the mouse and keyboard, but still want to maintain or increase their overall productivity.[4]

Windows Vista

edit

With Windows Vista voice control, a user may dictate documents and emails in mainstream applications, start and switch between applications, control the operating system, format documents, save documents, edit files, efficiently correct errors, and fill out forms on the Web. The speech recognition software learns automatically every time a user uses it, and speech recognition is available in English (U.S.), English (U.K.), German (Germany), French (France), Spanish (Spain), Japanese, Chinese (Traditional), and Chinese (Simplified). In addition, the software comes with an interactive tutorial, which can be used to train both the user and the speech recognition engine.[5]

Windows 7

edit

In addition to all the features provided in Windows Vista, Windows 7 provides a wizard for setting up the microphone and a tutorial on how to use the feature.[6]

Mac OS X

edit

All Mac OS X computers come pre-installed with the speech recognition software. The software is user-independent, and it allows for a user to, "navigate menus and enter keyboard shortcuts; speak checkbox names, radio button names, list items, and button names; and open, close, control, and switch among applications."[7] However, the Apple website recommends a user buy a commercial product called Dictate.[7]

Commercial products

edit

If a user is not satisfied with the built in speech recognition software or a user does not have a built speech recognition software for their OS, then a user may experiment with a commercial product such as Braina Pro or DragonNaturallySpeaking for Windows PCs,[8] and Dictate, the name of the same software for Mac OS.[9]

Voice command mobile devices

edit

Any mobile device running Android OS, Microsoft Windows Phone, iOS 9 or later, or Blackberry OS provides voice command capabilities. In addition to the built-in speech recognition software for each mobile phone's operating system, a user may download third party voice command applications from each operating system's application store: Apple App store, Google Play, Windows Phone Marketplace (initially Windows Marketplace for Mobile), or BlackBerry App World.

Android OS

edit

Google has developed an open source operating system called Android, which allows a user to perform voice commands such as: send text messages, listen to music, get directions, call businesses, call contacts, send email, view a map, go to websites, write a note, and search Google.[10] The speech recognition software is available for all devices since Android 2.2 "Froyo", but the settings must be set to English.[10] Google allows for the user to change the language, and the user is prompted when he or she first uses the speech recognition feature if he or she would like their voice data to be attached to their Google account. If a user decides to opt into this service, it allows Google to train the software to the user's voice.[11]

Google introduced the Google Assistant with Android 7.0 "Nougat". It is much more advanced than the older version.

Amazon.com has the Echo that uses Amazon's custom version of Android to provide a voice interface.

Microsoft Windows

edit

Windows Phone is Microsoft's mobile device's operating system. On Windows Phone 7.5, the speech app is user independent and can be used to: call someone from your contact list, call any phone number, redial the last number, send a text message, call your voice mail, open an application, read appointments, query phone status, and search the web.[12][13] In addition, speech can also be used during a phone call, and the following actions are possible during a phone call: press a number, turn the speaker phone on, or call someone, which puts the current call on hold.[13]

Windows 10 introduces Cortana, a voice control system that replaces the formerly used voice control on Windows phones.

Apple added Voice Control to its family of iOS devices as a new feature of iPhone OS 3. The iPhone 4S, iPad 3, iPad Mini 1G, iPad Air, iPad Pro 1G, iPod Touch 5G and later, all come with a more advanced voice assistant called Siri. Voice Control can still be enabled through the Settings menu of newer devices. Siri is a user independent built-in speech recognition feature that allows a user to issue voice commands. With the assistance of Siri a user may issue commands like, send a text message, check the weather, set a reminder, find information, schedule meetings, send an email, find a contact, set an alarm, get directions, track your stocks, set a timer, and ask for examples of sample voice command queries.[14] In addition, Siri works with Bluetooth and wired headphones.[15]

Apple introduced Personal Voice as an accessibility feature in iOS 17, launched on September 18, 2023. [16] This feature allows users to create a personalized, machine learning-generated (AI) version of their voice for use in text-to-speech applications. Designed particularly for individuals with speech impairments, Personal Voice helps preserve the unique sound of a user's voice. It enhances Siri and other accessibility tools by providing a more personalized and inclusive user experience. Personal Voice reflects Apple's ongoing commitment to accessibility and innovation. [17] [18]

Amazon Alexa

edit

In 2014 Amazon introduced the Alexa smart home device. Its main purpose was just a smart speaker, that allowed the consumer to control the device with their voice. Eventually, it turned into a novelty device that had the ability to control home appliance with voice. Now almost all the appliances are controllable with Alexa, including light bulbs and temperature. By allowing voice control, Alexa can connect to smart home technology allowing you to lock your house, control the temperature, and activate various devices. This form of A.I allows for someone to simply ask it a question, and in response the Alexa searches for, finds, and recites the answer back to you.[19]

Speech recognition in cars

edit

As car technology improves, more features will be added to cars and these features could potentially distract a driver. Voice commands for cars, according to CNET, should allow a driver to issue commands and not be distracted. CNET stated that Nuance was suggesting that in the future they would create a software that resembled Siri, but for cars.[20] Most speech recognition software on the market in 2011 had only about 50 to 60 voice commands, but Ford Sync had 10,000.[20] However, CNET suggested that even 10,000 voice commands was not sufficient given the complexity and the variety of tasks a user may want to do while driving.[20] Voice command for cars is different from voice command for mobile phones and for computers because a driver may use the feature to look for nearby restaurants, look for gas, driving directions, road conditions, and the location of the nearest hotel.[20] Currently, technology allows a driver to issue voice commands on both a portable GPS like a Garmin and a car manufacturer navigation system.[21]

List of Voice Command Systems Provided By Motor Manufacturers:

Non-verbal input

edit

While most voice user interfaces are designed to support interaction through spoken human language, there have also been recent explorations in designing interfaces take non-verbal human sounds as input.[22][23] In these systems, the user controls the interface by emitting non-speech sounds such as humming, whistling, or blowing into a microphone.[24]

One such example of a non-verbal voice user interface is Blendie,[25][26] an interactive art installation created by Kelly Dobson. The piece comprised a classic 1950s-era blender which was retrofitted to respond to microphone input. To control the blender, the user must mimic the whirring mechanical sounds that a blender typically makes: the blender will spin slowly in response to a user's low-pitched growl, and increase in speed as the user makes higher-pitched vocal sounds.

Another example is VoiceDraw,[27] a research system that enables digital drawing for individuals with limited motor abilities. VoiceDraw allows users to "paint" strokes on a digital canvas by modulating vowel sounds, which are mapped to brush directions. Modulating other paralinguistic features (e.g. the loudness of their voice) allows the user to control different features of the drawing, such as the thickness of the brush stroke.

Other approaches include adopting non-verbal sounds to augment touch-based interfaces (e.g. on a mobile phone) to support new types of gestures that wouldn't be possible with finger input alone.[24]

Design challenges

edit

Voice interfaces pose a substantial number of challenges for usability. In contrast to graphical user interfaces (GUIs), best practices for voice interface design are still emergent.[28]

Discoverability

edit

With purely audio-based interaction, voice user interfaces tend to suffer from low discoverability:[28] it is difficult for users to understand the scope of a system's capabilities. In order for the system to convey what is possible without a visual display, it would need to enumerate the available options, which can become tedious or infeasible. Low discoverability often results in users reporting confusion over what they are "allowed" to say, or a mismatch in expectations about the breadth of a system's understanding.[29][30]

Transcription

edit

While speech recognition technology has improved considerably in recent years, voice user interfaces still suffer from parsing or transcription errors in which a user's speech is not interpreted correctly.[31] These errors tend to be especially prevalent when the speech content uses technical vocabulary (e.g. medical terminology) or unconventional spellings such as musical artist or song names.[32]

Understanding

edit

Effective system design to maximize conversational understanding remains an open area of research. Voice user interfaces that interpret and manage conversational state are challenging to design due to the inherent difficulty of integrating complex natural language processing tasks like coreference resolution, named-entity recognition, information retrieval, and dialog management.[33] Most voice assistants today are capable of executing single commands very well but limited in their ability to manage dialogue beyond a narrow task or a couple turns in a conversation.[34]

Privacy implications

edit

Privacy concerns are raised by the fact that voice commands are available to the providers of voice-user interfaces in unencrypted form, and can thus be shared with third parties and be processed in an unauthorized or unexpected manner.[35][36] Additionally to the linguistic content of recorded speech, a user's manner of expression and voice characteristics can implicitly contain information about his or her biometric identity, personality traits, body shape, physical and mental health condition, sex, gender, moods and emotions, socioeconomic status and geographical origin.[37]

See also

edit

References

edit
  1. ^ "Washing Machine Voice Control". Appliance Magazine.
  2. ^ Borzo, Jeanette (8 February 2007). "Now You're Talking". CNN Money. Retrieved 25 April 2012.
  3. ^ "Voice Control, the End of the TV Remote?". Bloomberg.com. Business Week. 9 December 2011. Archived from the original on December 8, 2011. Retrieved 1 May 2012.
  4. ^ "Windows Vista Built In Speech". Windows Vista. Retrieved 25 April 2012.
  5. ^ "Speech Operation On Vista". Microsoft.
  6. ^ "Speech Recognition Set Up". Microsoft.
  7. ^ a b "Physical and Motor Skills". Apple.
  8. ^ "DragonNaturallySpeaking PC". Nuance.
  9. ^ "DragonNaturallySpeaking Mac". Nuance.
  10. ^ a b "Voice Actions".
  11. ^ "Google Voice Search For Android Can Now Be "Trained" To Your Voice". 14 December 2010. Retrieved 24 April 2012.
  12. ^ "Using Voice Command". Microsoft. Retrieved 24 April 2012.
  13. ^ a b "Using Voice Commands". Microsoft. Retrieved 27 April 2012.
  14. ^ "Siri, The iPhone 3GS & 4, iPod 3 & 4, have voice control like an express Siri, it plays music, pauses music, suffle, Facetime, and calling Features". Apple. Retrieved 27 April 2012.
  15. ^ "Siri FAQ". Apple.
  16. ^ "How to use Personal Voice on iPhone with iOS 17". Engadget. 2023-12-06. Retrieved 2024-08-21.
  17. ^ Jason England (2023-07-13). "How to set up and use Personal Voice in iOS 17 — make your iPhone sound just like you". LaptopMag. Retrieved 2024-08-21.
  18. ^ "Advancing Speech Accessibility with Personal Voice". Apple Machine Learning Research. Retrieved 2024-08-21.
  19. ^ "How Amazon's Echo went from a smart speaker to the center of your home". Business Insider.
  20. ^ a b c d "Siri Like Voice". CNET.
  21. ^ "Portable GPS With Voice". CNET.
  22. ^ Blattner, Meera M.; Greenberg, Robert M. (1992). "Communicating and Learning Through Non-speech Audio". Multimedia Interface Design in Education. pp. 133–143. doi:10.1007/978-3-642-58126-7_9. ISBN 978-3-540-55046-4.
  23. ^ Hereford, James; Winn, William (October 1994). "Non-Speech Sound in Human-Computer Interaction: A Review and Design Guidelines". Journal of Educational Computing Research. 11 (3): 211–233. doi:10.2190/mkd9-w05t-yj9y-81nm. ISSN 0735-6331. S2CID 61510202.
  24. ^ a b Sakamoto, Daisuke; Komatsu, Takanori; Igarashi, Takeo (27 August 2013). "Voice augmented manipulation | Proceedings of the 15th international conference on Human-computer interaction with mobile devices and services": 69–78. doi:10.1145/2493190.2493244. S2CID 6251400. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  25. ^ Dobson, Kelly (August 2004). "Blendie | Proceedings of the 5th conference on Designing interactive systems: processes, practices, methods, and techniques": 309. doi:10.1145/1013115.1013159. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  26. ^ "Kelly Dobson: Blendie". web.media.mit.edu. Retrieved 2019-02-27.
  27. ^ Harada, Susumu; Wobbrock, Jacob O.; Landay, James A. (15 October 2007). "Voicedraw | Proceedings of the 9th international ACM SIGACCESS conference on Computers and accessibility": 27–34. doi:10.1145/1296843.1296850. S2CID 218338. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  28. ^ a b Murad, Christine; Munteanu, Cosmin; Clark, Leigh; Cowan, Benjamin R. (3 September 2018). "Design guidelines for hands-free speech interaction | Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services Adjunct": 269–276. doi:10.1145/3236112.3236149. S2CID 52099112. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  29. ^ Yankelovich, Nicole; Levow, Gina-Anne; Marx, Matt (May 1995). "Designing SpeechActs | Proceedings of the SIGCHI Conference on Human Factors in Computing Systems": 369–376. doi:10.1145/223904.223952. S2CID 9313029. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  30. ^ "What can I say? | Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices and Services". doi:10.1145/2935334.2935386. S2CID 6246618. {{cite journal}}: Cite journal requires |journal= (help)
  31. ^ Myers, Chelsea; Furqan, Anushay; Nebolsky, Jessica; Caro, Karina; Zhu, Jichen (19 April 2018). "Patterns for How Users Overcome Obstacles in Voice User Interfaces | Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems": 1–7. doi:10.1145/3173574.3173580. S2CID 5041672. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  32. ^ Springer, Aaron; Cramer, Henriette (21 April 2018). ""Play PRBLMS" | Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems": 1–13. doi:10.1145/3173574.3173870. S2CID 5050837. Retrieved 2019-02-27. {{cite journal}}: Cite journal requires |journal= (help)
  33. ^ Galitsky, Boris (2019). Developing Enterprise Chatbots: Learning Linguistic Structures (1st ed.). Cham, Switzerland: Springer. pp. 13–24. doi:10.1007/978-3-030-04299-8. ISBN 978-3-030-04298-1. S2CID 102486666.
  34. ^ Pearl, Cathy (2016-12-06). Designing Voice User Interfaces: Principles of Conversational Experiences (1st ed.). Sebastopol, CA: O'Reilly Media. pp. 16–19. ISBN 978-1-491-95541-3.
  35. ^ "Apple, Google, and Amazon May Have Violated Your Privacy by Reviewing Digital Assistant Commands". Fortune. 2019-08-05. Retrieved 2020-05-13.
  36. ^ Hern, Alex (2019-04-11). "Amazon staff listen to customers' Alexa recordings, report says". the Guardian. Retrieved 2020-05-21.
  37. ^ Kröger, Jacob Leon; Lutz, Otto Hans-Martin; Raschke, Philip (2020). "Privacy Implications of Voice and Speech Analysis – Information Disclosure by Inference". Privacy and Identity Management. Data for Better Living: AI and Privacy. IFIP Advances in Information and Communication Technology. Vol. 576. pp. 242–258. doi:10.1007/978-3-030-42504-3_16. ISBN 978-3-030-42503-6. ISSN 1868-4238.
edit