Affective computing

From Wikipedia, the free encyclopedia

Jump to: navigation, search
Affective Computing is also the title of a textbook on the subject by Rosalind Picard.

Affective computing is a branch of the study and development of artificial intelligence that deals with the design of systems and devices that can recognize, interpret, and process human emotions. It is an interdisciplinary field spanning computer sciences, psychology, and cognitive science.[1] While the origins of the field may be traced as far back as to early philosophical enquiries into emotion,[2] the more modern branch of computer science originated with Rosalind Picard's 1995 paper[3] on affective computing.[4][5] A motivation for the research is the ability to simulate empathy. The machine should interpret the emotional state of humans and adapt its behaviour to them, giving an appropriate response for those emotions.

Contents

[edit] Areas of affective computing

[edit] Detecting and recognizing emotional information

Detecting emotional information begins with passive sensors which capture data about the user's physical state or behavior without interpreting the input. The data gathered is analogous to the cues humans use to perceive emotions in others. For example, a video camera might capture facial expressions, body posture and gestures, while a microphone might capture speech. Other sensors detect emotional cues by directly measuring physiological data, such as skin temperature and galvanic resistance.[6]

Recognizing emotional information requires the extraction of meaningful patterns from the gathered data. This is done by parsing the data through various processes such as speech recognition, natural language processing, or facial expression detection, all of which are dependent on the human factor vis-a-vis programming.[citation needed]

[edit] Emotion in machines

Another area within affective computing is the design of computational devices proposed to exhibit either innate emotional capabilities or that are capable of convincingly simulating emotions. A more practical approach, based on current technological capabilities, is the simulation of emotions in conversational agents[vague]. The goal of such simulation is to enrich and facilitate interactivity between human and machine[citation needed][vague]. While human emotions are often associated with surges in hormones and other neuropeptides, emotions in machines might be associated with abstract states associated with progress (or lack of progress) in autonomous learning systems[citation needed]. In this view, affective emotional states correspond to time-derivatives (perturbations) in the learning curve of an arbitrary learning system.[citation needed]

Marvin Minsky, one of the pioneering computer scientists in artificial intelligence, relates emotions to the broader issues of machine intelligence stating in The Emotion Machine that emotion is "not especially different from the processes that we call 'thinking.'"[7]

[edit] Technologies of affective computing

[edit] Emotional speech

Emotional speech processing recognizes the user's emotional state by analyzing speech patterns. Vocal parameters and prosody features such as pitch variables and speech rate are analyzed through pattern recognition.[8][9]

Emotional inflection and modulation in synthesized speech, either through phrasing or acoustic features is useful in human-computer interaction. Such capability makes speech natural and expressive. For example a dialog system might modulate its speech to be more puerile if it deems the emotional model of its current user is that of a child.[citation needed]

[edit] Facial expression

The detection and processing of facial expression is achieved through various methods such as optical flow, hidden Markov model, neural network processing or active appearance model.[citation needed]

[edit] Body gesture

Body gesture is the position and the changes of the body. There are many proposed methods[10] to detect the body gesture. Hand gestures have been a common focus of body gesture detection, apparentness[vague]methods[11] and 3-D modeling methods are traditionally used.

[edit] Potential applications

In e-learning applications, affective computing can be used to adjust the presentation style of a computerized tutor when a learner is bored, interested, frustrated, or pleased.[12] Psychological health services, i.e. counseling, benefit from affective computing applications when determining a client's emotional state.[citation needed] Affective computing sends a message via color or sound to express an emotional state to others.[citation needed]

Robotic systems capable of processing affective information exhibit higher flexibility while one works in uncertain or complex environments. Companion devices, such as digital pets, use affective computing abilities to enhance realism and provide a higher degree of autonomy.[citation needed]

Other potential applications are centered around social monitoring. For example, a car can monitor the emotion of all occupants and engage in additional safety measures, such as alerting other vehicles if it detects the driver to be angry.[citation needed] Affective computing has potential applications in human computer interaction, such as affective mirrors allowing the user to see how he or she performs; emotion monitoring agents sending a warning before one sends an angry email; or even music players selecting tracks based on mood.[citation needed]

Affective computing is also being applied to the development of communicative technologies for use by people with autism.[13]

[edit] Application examples

[edit] See also

[edit] References

  1. ^ Tao, Jianhua; Tieniu Tan (2005). "Affective Computing: A Review". Affective Computing and Intelligent Interaction LNCS 3784: 981–995, Springer. doi:10.1007/11573548. 
  2. ^ James, William (1884). "What is Emotion". Mind 9: 188–205.  Cited by Tao and Tan.
  3. ^ "Affective Computing" MIT Technical Report #321 (Abstract), 1995
  4. ^ Kleine-Cosack, Christian (October 2006). "Recognition and Simulation of Emotions" (PDF). http://ls12-www.cs.tu-dortmund.de//~fink/lectures/SS06/human-robot-interaction/Emotion-RecognitionAndSimulation.pdf. Retrieved on May 13 2008. "The introduction of emotion to computer science was done by Pickard (sic) who created the field of affective computing." 
  5. ^ Diamond, David (December 2003). "The Love Machine; Building computers that care.". Wired. http://www.wired.com/wired/archive/11.12/love.html. Retrieved on May 13 2008. "Rosalind Picard, a genial MIT professor, is the field's godmother; her 1997 book, Affective Computing, triggered an explosion of interest in the emotional side of computers and their users." 
  6. ^ Garay, Nestor; Idoia Cearreta, Juan Miguel López, Inmaculada Fajardo (April 2006). "Assistive Technology and Affective Mediation" (PDF). Human Technology: An Interdisciplinary Journal on Humans in ICT Environments 2 (1): 55–83. http://www.humantechnology.jyu.fi/articles/volume2/number1/2006/humantechnology-april-2006.pdf. Retrieved on 2008-05-12. 
  7. ^ Restak, Richard (2006-12-17). "Mind Over Matter". The Washington Post. http://www.washingtonpost.com/wp-dyn/content/article/2006/12/14/AR2006121401554.html. Retrieved on 2008-05-13. 
  8. ^ Dellaert, F., Polizin, t., and Waibel, A., Recognizing Emotion in Speech", In Proc. Of ICSLP 1996, Philadelphia, PA, pp.1970-1973, 1996
  9. ^ Lee, C.M.; Narayanan, S.; Pieraccini, R., Recognition of Negative Emotion in the Human Speech Signals, Workshop on Auto. Speech Recognition and Understanding, Dec 2001
  10. ^ J. K. Aggarwal, Q. Cai, Human Motion Analysis: A Review, Computer Vision and Image Understanding, Vol. 73, No. 3, 1999
  11. ^ Vladimir I. Pavlovic, Rajeev Sharma, Thomas S. Huang, Visual Interpretation of Hand Gestures for Human-Computer Interaction; A Review, IEEE Transactions on Pattern Analysis and Machine Intelligence, 1997
  12. ^ AutoTutor
  13. ^ Projects in Affective Computing

[edit] External links

Personal tools
Languages