{"id":396073,"date":"2026-01-29T02:35:07","date_gmt":"2026-01-29T02:35:07","guid":{"rendered":"https:\/\/www.newsbeep.com\/uk\/396073\/"},"modified":"2026-01-29T02:35:07","modified_gmt":"2026-01-29T02:35:07","slug":"south-koreas-world-first-ai-laws-face-pushback-amid-bid-to-become-leading-tech-power-south-korea","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/uk\/396073\/","title":{"rendered":"South Korea\u2019s \u2018world-first\u2019 AI laws face pushback amid bid to become leading tech power | South Korea"},"content":{"rendered":"<p class=\"dcr-130mj7b\">South Korea has embarked on a foray into the regulation of AI, launching what has been billed as the most comprehensive set of laws anywhere in the world, that could prove a model for other countries, but the new legislation has already encountered pushback.<\/p>\n<p class=\"dcr-130mj7b\">The laws, which will force companies to label AI-generated content, have been criticised by local tech startups, which say they go too far, and civil society groups, which say they don\u2019t go far enough.<\/p>\n<p class=\"dcr-130mj7b\">The AI basic act, which took effect on Thursday last week, comes amid <a href=\"https:\/\/www.theguardian.com\/technology\/2026\/jan\/27\/wake-up-to-the-risks-of-ai-they-are-almost-here-anthropic-boss-warns\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">growing global unease over artificially created media<\/a> and automated decision-making, as <a href=\"https:\/\/www.theguardian.com\/business\/2026\/jan\/25\/more-than-quarter-britons-fear-losing-jobs-ai-next-five-years\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">governments struggle to keep pace<\/a> with rapidly advancing technologies.<\/p>\n<p class=\"dcr-130mj7b\">The act will force companies providing AI services to:<\/p>\n<p class=\"dcr-130mj7b\">Add invisible digital watermarks for clearly artificial outputs such as cartoons or artwork. For realistic deepfakes, visible labels are required.<\/p>\n<p class=\"dcr-130mj7b\">\u201cHigh-impact AI\u201d, including systems used for <a href=\"https:\/\/www.theguardian.com\/technology\/ng-interactive\/2026\/jan\/24\/how-the-confident-authority-of-google-ai-overviews-is-putting-public-health-at-risk\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">medical diagnosis<\/a>, hiring and loan approvals, will require operators to conduct risk assessments and document how decisions are made. If a human makes the final decision the system may fall outside the category.<\/p>\n<p class=\"dcr-130mj7b\">Extremely powerful AI models will require safety reports, but the threshold is set so high that government officials acknowledge no models worldwide currently meet it.<\/p>\n<p class=\"dcr-130mj7b\">Companies that violate the rules face fines of up to 30m won (\u00a315,000), but the government has promised a grace period of at least a year before penalties are imposed.<\/p>\n<p class=\"dcr-130mj7b\">The legislation is being billed as the \u201cworld\u2019s first\u201d to be fully enforced by a country, and central to South Korea\u2019s ambition to become one of the world\u2019s three leading AI powers alongside the US and China. Government officials maintain the law is 80-90% focused on promoting industry rather than restricting it.<\/p>\n<p class=\"dcr-130mj7b\">Alice Oh, a computer science professor at the Korea Advanced Institute of Science and Technology (KAIST), said that while the law was not perfect, it was intended to evolve without stifling innovation. However <a href=\"https:\/\/en.yna.co.kr\/view\/AEN20251214000500320\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">a survey<\/a> in December from the Startup Alliance found that 98% of AI startups were unprepared for compliance. Its co-head, Lim Jung-wook, said frustration was widespread. \u201cThere\u2019s a bit of resentment,\u201d he said. \u201cWhy do we have to be the first to do this?\u201d<\/p>\n<p class=\"dcr-130mj7b\">Companies must self-determine whether their systems qualify as high-impact AI, a process critics say is lengthy and creates uncertainty.<\/p>\n<p class=\"dcr-130mj7b\">They also warn of competitive imbalance: all Korean companies face regulation regardless of size, while only foreign firms meeting certain thresholds \u2013 such as Google and OpenAI \u2013 must comply.<\/p>\n<p class=\"dcr-130mj7b\">The push for regulation has unfolded against a uniquely charged domestic backdrop that has left civil society groups worried the legislation does not go far enough.<\/p>\n<p class=\"dcr-130mj7b\"><a href=\"https:\/\/www.theguardian.com\/world\/2024\/sep\/13\/from-spy-cams-to-deepfake-porn-fury-in-south-korea-as-women-targeted-again\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">South Korea accounts<\/a> for 53% of all global deepfake pornography victims, according to a 2023 report by Security Hero, a US-based identity protection firm. In August 2024, an investigation <a href=\"https:\/\/www.theguardian.com\/world\/2020\/mar\/25\/outrage-in-south-korea-over-telegram-sexual-abuse-ring-blackmailing-women-and-girls\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">exposed massive networks of Telegram chatrooms<\/a> creating and distributing AI-generated sexual imagery of women and girls, foreshadowing <a href=\"https:\/\/www.theguardian.com\/technology\/2026\/jan\/22\/grok-ai-generated-millions-sexualised-images-in-month-research-says\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">the scandal<\/a> that would later erupt around Elon Musk\u2019s Grok chatbot.<\/p>\n<p class=\"dcr-130mj7b\">The law\u2019s origins, however, predate this crisis, with the first AI-related bill submitted to parliament in July 2020. It stalled repeatedly in part due to provisions that were accused of prioritising industry interests over citizen protection.<\/p>\n<p class=\"dcr-130mj7b\">Civil society groups maintain that the new legislation provides limited protection for people harmed by AI systems.<\/p>\n<p class=\"dcr-130mj7b\">Four organisations, including Minbyun, a collective of human rights lawyers, issued a joint statement the day after it was implemented arguing the law contained almost no provisions to protect citizens from AI risks.<\/p>\n<p class=\"dcr-130mj7b\">The groups noted that while the law stipulated protection for \u201cusers\u201d, those users were hospitals, financial companies and public institutions that use AI systems, not people affected by AI. The law established no prohibited AI systems, they argued, and exemptions for \u201chuman involvement\u201d created significant loopholes.<\/p>\n<p class=\"dcr-130mj7b\">The country\u2019s human rights commission <a href=\"https:\/\/www.humanrights.go.kr\/base\/board\/read?boardManagementNo=24&amp;boardNo=7611694&amp;searchCategory=&amp;page=2&amp;searchType=&amp;searchWord=&amp;menuLevel=&amp;menuNo=\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">has criticised<\/a> the enforcement decree for lacking clear definitions of high-impact AI, noting that those most likely to suffer rights violations remain in regulatory blind spots.<\/p>\n<p class=\"dcr-130mj7b\">In a statement, the ministry of science and ICT said it expected the law to \u201cremove legal uncertainty\u201d and build \u201ca healthy and safe domestic AI ecosystem\u201d, adding that it would continue to clarify the rules through revised guidelines.<\/p>\n<p class=\"dcr-130mj7b\">Experts said South Korea had deliberately chosen a different path from other jurisdictions.<\/p>\n<p class=\"dcr-130mj7b\">Unlike the EU\u2019s strict risk-based regulatory model, the US and UK\u2019s largely sector-specific, market-driven approaches, or China\u2019s combination of state-led industrial policy and detailed service-specific regulation, South Korea has opted for a more flexible, principles-based framework, said Melissa Hyesun Yoon, a law professor at Hanyang University who specialises in AI governance.<\/p>\n<p class=\"dcr-130mj7b\">That approach is centred on what Yoon describes as \u201ctrust-based promotion and regulation\u201d.<\/p>\n<p class=\"dcr-130mj7b\">\u201cKorea\u2019s framework will serve as a useful reference point in global AI governance discussions,\u201d she said.<\/p>\n","protected":false},"excerpt":{"rendered":"South Korea has embarked on a foray into the regulation of AI, launching what has been billed as&hellip;\n","protected":false},"author":2,"featured_media":396074,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[554,733,4308,86,56,54,55],"class_list":{"0":"post-396073","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-technology","12":"tag-uk","13":"tag-united-kingdom","14":"tag-unitedkingdom"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/396073","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/comments?post=396073"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/396073\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media\/396074"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media?parent=396073"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/categories?post=396073"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/tags?post=396073"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}