{"id":133805,"date":"2025-09-05T03:38:07","date_gmt":"2025-09-05T03:38:07","guid":{"rendered":"https:\/\/www.newsbeep.com\/us\/133805\/"},"modified":"2025-09-05T03:38:07","modified_gmt":"2025-09-05T03:38:07","slug":"the-workplace-learns-to-feel-emotional-ai-enters-the-enterprise","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/us\/133805\/","title":{"rendered":"The workplace learns to feel: Emotional AI enters the enterprise"},"content":{"rendered":"<p>The promise of emotionally aware systems comes with risks. Bera warned that many emotionally responsive interfaces are essentially simulations, and users may not always realize what is real and what is not.<\/p>\n<p>\u201cOne major risk is empathy theater,\u201d he said. \u201cThat is when an AI system mimics emotion convincingly without actually understanding or caring. Just because a machine sounds empathetic, does not mean it is.\u201d<\/p>\n<p>The gap between performance and intention can lead to harm, especially when people place trust in systems that are not designed to offer care. Bera is particularly concerned about affective systems being used in hiring, feedback or mental health contexts without clear guardrails. \u201cWe must ask: are we amplifying human understanding or automating misunderstanding?\u201d he said.<\/p>\n<p>Signal misinterpretation is another danger, Bera said\u2014especially when emotional expressions can vary so widely across cultures and individuals. \u201cThere is the concern of misread signals,\u201d Bera said. \u201cEspecially across cultural or neurodiverse expressions of emotion.\u201d<\/p>\n<p>Haber raised a separate concern: what happens when AI systems are too agreeable? \u201cIt is important to have a conversation partner that can engage in sometimes seemingly adversarial ways,\u201d he said. \u201cIf I am talking with a chatbot about how I treated someone horribly and am trying to justify my actions, it can be really harmful if it tells me I was totally justified in doing that.\u201d<\/p>\n<p>Many large language models (LLMs) are trained to be affirming, not challenging, Haber said. The result is what Haber described as \u201cemotional echo chambers,\u201d where \u00a0harmful ideas get reinforced through AI feedback loops.<\/p>\n<p>\u201cWe can become uncritical, unreflective and more isolated,\u201d he said. \u201cThere has been a lot of talk about how large language model-based chatbots are sycophantic. That is just a small part of a much larger and more complex issue.\u201d<\/p>\n<p>Seif El Nasr added that emotionally responsive systems can create unintended psychological dependencies. \u201cThese include psychological issues such as overreliance, loneliness and anxiety,\u201d she said. \u201cThey can also lead to isolation from society.\u201d<\/p>\n<p>Privacy is another concern. Seif El Nasr cited applications like <a href=\"https:\/\/replika.com\/\" target=\"_blank\" rel=\"noopener noreferrer nofollow\">Replika<\/a>, which create intimate emotional experiences but operate with unclear data protections. \u00a0In her view, systems that deal with human emotion must be developed with interdisciplinary rigor. \u201cSuch systems need to be developed with extra care,\u201d she said. \u201cThat means grounding them in user research and involving social scientists, not just technologists.\u201d<\/p>\n<p>Bera said his lab is focused on designing systems that can be audited and explained. \u201cWe address these concerns in our lab by grounding emotional inference in multimodal explainable models and coupling them with rigorous ethical frameworks,\u201d he said. \u201cIt is not just about building better algorithms. It is about building trustworthy ones.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"The promise of emotionally aware systems comes with risks. Bera warned that many emotionally responsive interfaces are essentially&hellip;\n","protected":false},"author":2,"featured_media":133806,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[45],"tags":[182,66562,84247,181,507,74],"class_list":{"0":"post-133805","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-ai-ethics","10":"tag-ai-privacy","11":"tag-artificial-intelligence","12":"tag-artificialintelligence","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/133805","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/comments?post=133805"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/133805\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media\/133806"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media?parent=133805"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/categories?post=133805"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/tags?post=133805"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}