The gaming world has been revolutionized by games that track a
player's natural body movements and translates them into the virtual
environment. By using gesture and voice recognition, gaming consoles
such as the Xbox Kinect allow players to kick a ball, shoot an arrow,
and actively participate in the game by simply moving their bodies, no
controller required.
Watch out manufacturing, the revolution is
coming. Very soon, factory floors may start seeing gesture and voice
recognition systems, combined with biometrics, which allow workers to
control factory operations with natural body movements and voiced
commands. A simple example of this, according to a Machine Design.com
article, involves logging into workstations.
Currently, many
automated factories operate off of Graphic User Interfaces (GUI's),
where a worker would log in by clicking on an icon and entering a
username and password. In the future, the same worker could simply step
up to the work station, which would scan his retina and automatically
log him in. With a simple gesture the worker could command the computer
to start operations, and by holding up his hand in a "stop" gesture,
halt operations. The machine could be programmed to ask for
confirmations of these gestures, requiring a vocal "yes" from the
operator.
So how does this technology work? A color video camera
works with a depth sensor that provides a 3D perspective and a set of
microphones which isolates individual player's voices. Advanced software
tracks the layout of the room and player movement, monitoring movements
and responding accordingly.
A biometric natural user interface
(NUI) would be able to identify only the person logged into that
particular machine, responding singularly to that person's gestures and
movements while ignoring all other workers. Should a worker leave a
workstation, it would not respond to anyone else and can even be
programmed to shut down after a specified period of time.