Machine perception

Machine perception is the capability of a computer system to interpret data in a manner that is similar to the way humans use their senses to relate to the world around them.[1][2][3] The basic method that the computers take in and respond to their environment is through the attached hardware. Until recently input was limited to a keyboard, or a mouse, but advances in technology, both in hardware and software, have allowed computers to take in sensory input in a way similar to humans.[1][2]

Machine perception allows the computer to use this sensory input, as well as conventional computational means of gathering information, to gather information with greater accuracy and to present it in a way that is more comfortable for the user.[1] These include computer vision, machine hearing, and machine touch.

The end goal of machine perception is to give machines the ability to see, feel and perceive the world as humans do and therefore for them to be able to explain in a human way why they are making their decisions, to warn us when it is failing and more importantly, the reason why it is failing.[4]

Machine visionEdit

Computer vision is a field that includes methods for acquiring, processing, analyzing, and understanding images and high-dimensional data from the real world to produce numerical or symbolic information, e.g., in the forms of decisions. Computer vision has many applications already in use today such as facial recognition, geographical modeling, and even aesthetic judgment.[5]

Machine hearingEdit

Machine hearing, also known as machine listening or computer audition, is the ability of a computer or machine to take in and process sound data such as music or speech. This area has a wide range of application including music recording and compression, speech synthesis, and speech recognition.[6] Moreover, this technology allows the machine to replicate the human brain’s ability to selectively focus in a specific sound against many other competing sounds and background noise. This particular ability is called “auditory scene analysis”. The technology enables the machine to segment several streams occurring at the same time.[7] Many commonly used devices such as a smartphones, voice translators, and cars make use of some form of machine hearing.

Machine touchEdit

Machine touch is an area of machine perception where tactile information is processed by a machine or computer. Applications include tactile perception of surface properties and dexterity whereby tactile information can enable intelligent reflexes and interaction with the environment.

Machine olfactionEdit

Giving machines a sense of smell in the field of machine olfaction. Airborne chemicals can be sensed and classified with a device sometimes known as an electronic nose.

See alsoEdit


  1. ^ a b c d Malcolm Tatum (October 3, 2012). "What is Machine Perception".
  2. ^ a b c Alexander Serov (January 29, 2013). "Subjective Reality and Strong Artificial Intelligence". arXiv:1301.6359 [cs.AI].
  3. ^ "Machine Perception & Cognitive Robotics Laboratory". Retrieved 2016-06-18.
  4. ^ "Machine Perception Research - ECE - Virginia Tech". Retrieved January 10, 2018.
  5. ^ a b Dhar, Sagnik; Ordonez, Vicente; Berg, Tamara L. (2011). "High level describable attributes for predicting aesthetics and interestingness" (PDF). CVPR 2011. pp. 1657–1664. doi:10.1109/CVPR.2011.5995467. ISBN 978-1-4577-0394-2.
  6. ^ a b Lyon, Richard (2010). "Machine Hearing: An Emerging Field [Exploratory DSP". IEEE Signal Processing Magazine. 27 (5): 131–139. Bibcode:2010ISPM...27..131L. doi:10.1109/MSP.2010.937498.
  7. ^ "Machine Perception & Cognitive Robotics Laboratory". Retrieved January 10, 2018.
  8. ^ Turk, Matthew (2000). "Perceptive Media: Machine Perception and Human Computer Interaction" (PDF). Chinese Journal of Computers. 12. pages 1235-1244