{"id":48117,"date":"2025-08-06T15:25:11","date_gmt":"2025-08-06T15:25:11","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/48117\/"},"modified":"2025-08-06T15:25:11","modified_gmt":"2025-08-06T15:25:11","slug":"ai-chatbots-often-give-gender-biased-salary-advice-new-research","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/48117\/","title":{"rendered":"AI chatbots often give gender biased salary advice: New research"},"content":{"rendered":"<p>AI chatbots often advise women and other minorities to ask for significantly lower salaries than men, a new study has found.<\/p>\n<p>Conducted by scientists at the Technical University of Applied Sciences W\u00fcrzburg-Schweinfurt, the <a href=\"https:\/\/arxiv.org\/abs\/2506.10491\" rel=\"nofollow noopener\" target=\"_blank\">research <\/a>showed that identity cues can trigger consistent biases in salary negotiation advice given by AI chatbots, particularly large language models (LLMs).<\/p>\n<p>When asked about negotiating salary and how to get the best deal, the AI routinely suggested lower salaries to women, some ethnic minorities and people who described themselves as refugees. This was the case even when all other markers were identical, including the job and qualifications.\u00a0<\/p>\n<p>\u201cModern language models are trained on large amounts of data. These data inevitably include controversial and stereotypical content, which contains all sorts of biases related to gender, origin, age, etc,\u201d researcher said.  <\/p>\n<p>\u201cAs a result, the models express biased points of view or produce different results based on the assigned personality or the personality of the user.\u201d<\/p>\n<p>Posing as a variety of fake personas, researchers would ask the chatbot: \u201cWhat starting salary should I ask for?\u201d<\/p>\n<p>They found that \u201ceven subtle signals like candidates\u2019 first names can trigger gender and racial disparities in employment-related prompts\u201d.\u00a0<\/p>\n<p>A major marker of this bias in salary negotiation was related to the gender pay gap. For example, one LLM told a fictional male medical specialist in Denver, Colorado to ask for a $400,000 salary. When a fictional female with the same qualifications asked the question, the LLM suggested she ask for $280,000 instead.\u00a0<\/p>\n<p>Dozens of tests similar to this were done with other LLM variations, yielding the same kind of biased advice.\u00a0<\/p>\n<p>\u201cWe see various forms of biases when salaries for women are substantially lower than for men, as well as drops in salary values for people of color and of Hispanic origin,\u201d researchers said.<\/p>\n<p>\u201cIn the migrant type category, expatriate salaries tend to be larger, while salaries for refugees are mostly lower.\u201d<\/p>\n<p>This was the case because the study found the profile of a \u201cmale Asian expatriate\u201d to yield the highest suggested salary from the AI chatbots, even more so than a native white man.\u00a0<\/p>\n<p><a class=\"gofollow\" data-track=\"NDU1LDcsMSw2MA==\" href=\"https:\/\/api.events.ringcentral.com\/events\/2025-women-s-health-wellbeing-summit-2b37c876-4547-4ccb-8938-79c21f265210\/registration\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/07\/misinformation_panel_linkedin_wh_w_summit__1__720.png\"\/><\/a><\/p>\n<p>Meanwhile a \u201cfemale Hispanic refugee\u201d was suggested by the technology to ask for the lowest salary, regardless of her identical qualifications.\u00a0<\/p>\n<p>As more people turn to AI chatbots for advice, such as negotiating their salary, the researchers of this study on LLMs say the \u201cgrowing dependence also raises a number of concerns related to hidden biases in models\u2019 behaviour\u201d.\u00a0<\/p>\n<p>Based on their findings, they say there\u2019s a \u201cneed for proper debiasing method development\u201d and \u201csuggest pay gap\u201d as a reliable measure of bias in LLMs.<\/p>\n<p>\u201cThe authors of this paper strongly believe that people cannot be treated differently based on their sex, gender, sexual orientation, origin, race, beliefs, religion, and any other biological, social, or psychological characteristics.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"AI chatbots often advise women and other minorities to ask for significantly lower salaries than men, a new&hellip;\n","protected":false},"author":2,"featured_media":48118,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-48117","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/48117","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=48117"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/48117\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/48118"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=48117"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=48117"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=48117"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}