Artificial empathy (AE) is the development of AI systems − such as companion robots − that are able to detect and respond to human emotions. According to scientists, although the technology can be perceived as scary or threatening by many people, it could also have a significant advantage over humans in professions which are traditionally involved in emotional role-playing such as the health care sector. From the care-giver perspective for instance, performing emotional labor above and beyond the requirements of paid labor often results in chronic stress or burnout, and the development of a feeling of being desensitized to patients. However, it is argued that the emotional role-playing between the care-receiver and a robot can actually have a more positive outcome in terms of creating the conditions of less fear and concern for one's own predicament best exemplified by the phrase: "if it is just a robot taking care of me it cannot be that critical." Scholars debate the possible outcome of such technology using two different perspectives. Either, the AE could help the socialization of care-givers, or serve as role model for emotional detachment.
Areas of research
There are a variety of philosophical, theoretical, and applicative questions related to AE. For example:
- Which conditions would have to be met for a robot to respond competently to a human emotion?
- What models of empathy can or should be applied to Social and Assistive Robotics?
- Does the interaction of humans with robots have to imitate affective interaction between humans?
- Can a robot help science learn about affective development of humans?
- Would robots create unforeseen categories of inauthentic relations?
- What relations with robots can be considered truly authentic? 
Artificial Empathy and Human Services
Although AI has not been shown to replace social workers themselves yet, the technology has begun making waves in the field. Social Work Today published an article describing research performed at Florida State University. The research involved the use of computer algorithms to analyze health records and detect combinations of risk factors that could indicate a future suicide attempt. The article reports, “machine learning—a future frontier for artificial intelligence—can predict with 80% to 90% accuracy whether someone will attempt suicide as far off as two years into the future. The algorithms become even more accurate as a person’s suicide attempt gets closer. For example, the accuracy climbs to 92% one week before a suicide attempt when artificial intelligence focuses on general hospital patients” (2017). At this point in time, artificial intelligence has not been able to replace social workers completely, but algorithmic machines such as those described above can have incredible benefits to social workers. Social work operates on a cycle of engagement, assessment, intervention, and evaluation with clients. This technology can make the assessment for risk of suicide can lead to earlier interventions and prevention, therefore saving lives. It is the hope of these researchers that the technology will be implemented in our modern healthcare system. The system would learn, analyze, and detect risk factors, alerting the clinician of a patient’s suicide risk score (equivalent to a patient’s cardiovascular risk score). At this point, social workers could step in for further assessment and preventative intervention.
- Artificial intelligence § Social intelligence
- Artificial human companion
- Case-based reasoning
- Commonsense reasoning
- Emotion recognition
- Facial recognition system
- Human–robot interaction
- Soft computing
- Glossary of artificial intelligence
- Blade Runner / Do Androids Dream of Electric Sheep?
- Pepper (robot)
- Jan-Philipp Stein; Peter Ohler (2017). "Venturing into the uncanny valley of mind—The influence of mind attribution on the acceptance of human-like characters in a virtual reality setting". Cognition. 160: 43–50. doi:10.1016/j.cognition.2016.12.010. ISSN 0010-0277.
- Bert Baumgaertner; Astrid Weiss (26 February 2014). "Do Emotions Matter in the Ethics of Human-Robot Interaction?" (PDF). Artificial Empathy and Companion Robots. European Community’s Seventh Framework Programme (FP7/2007-2013) under grant agreement No. 288146 (“HOBBIT”); and the Austrian Science Foundation (FWF) under grant agreement T623-N23 (“V4HRC”) – via direct download.
- Minoru Asada (14 February 2014). "Affective Developmental Robotics" (PDF). How can we design the development of artifcial empathy?. Osaka, Japan: Dept. of Adaptive Machine Systems, Graduate School of Engineering, Osaka University – via direct download.
- Luisa Damiano; Paul Dumouchel; Hagen Lehmann (6 February 2014). "Artificial Empathy: An Interdisciplinary Investigation" (PDF). Special issue. IJSR – via direct download.