Much of today’s interactions involve computers, i.e., automated telephone systems, automobiles equipped with voice active accessories, interactive video games, ATMs, self checkout stations, etc. Many of these are equipped with voice recognizer systems. These are consumer-grade voice recognizers and are not designed to operate in the DC where unpredictable and loud extraneous noise is produced from many sources, such as forklifts whizzing by.
We mention this because as interest in voice technology for DC applications has rapidly expanded, some vendors, anxious to enter the market, are offering voice systems that are based on speaker-independent consumer-grade speech recognizers. However, for very noisy environments, as typically found in busy DCs, the speaker-dependent industrial-grade voice recognizer will deliver the most reliable performance.
Speaker-dependent industrial voice systems do have to be "trained" to recognize the particular pronunciation of every user. Training helps the system understand workers in noisy environments. Since many words, letters and numbers have a similar sound, very small differences make the difference between a system (or person) hearing "I left" or "five left." They include a preprogrammed vocabulary of common words to reduce the time it takes to train the system. Still, words unique to the application will require more training time (repetitions).
They also require "discrete" speech - that is, words cannot be slurred together. Some systems will, however, recognize a commonly used phrase such as "fifty-five" as "five five" or can be programmed to regard "loading dock" as a single word in order to facilitate recognition.
Speech systems have to be adaptable to cope with the increased diversity in the workplace. And, with more workers for whom English is not their native language, systems today can recognize non-English inputs as if they were English. Vendors also offer speech software in a number of the more common languages such as French and Spanish.
During training, speakers also learn certain commands that will put the system to "sleep" and "wake it up." This is necessary because there are times when an employee must talk but is not giving inputs to the speech system. Similar commands could be programmed to switch among different applications such as picking, shipping or inventory.
One very important aspect of the latest generation industrial speech systems is that they talk as well as listen - translating binary data in the host applications to verbal instructions.
Given the proper speech-based logistics software, irregular work situations (such as an aisle being temporarily blocked by equipment) or exceptions (such as material not being in the storage/picking location) can be handled efficiently. Here is where industrial speech system vendors feel their products shine, insisting that they allow workers to continually interact with the system by being directed to the next downstream task without interruption, and then once the problem is resolved, being directed to return to complete the task.
|