Friday, November 29, 2019

The Risks of Using AI to Interpret Human Emotions - Mark Purdy, John Zealley, Omaro Maseli; Harvard Business Review

Because of the subjective nature of emotions, emotional AI is especially prone to bias. For example, one study found that emotional analysis technology assigns more negative emotions to people of certain ethnicities than to others. Consider the ramifications in the workplace, where an algorithm consistently identifying an individual as exhibiting negative emotions might affect career progression.
https://hbr.org/2019/11/the-risks-of-using-ai-to-interpret-human-emotions