{"id":475996,"date":"2026-02-12T22:34:22","date_gmt":"2026-02-12T22:34:22","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/475996\/"},"modified":"2026-02-12T22:34:22","modified_gmt":"2026-02-12T22:34:22","slug":"non-consensual-ai-porn-doesnt-violate-privacy-but-its-still-wrong","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/475996\/","title":{"rendered":"Non-consensual AI porn doesn\u2019t violate privacy \u2013 but it\u2019s still wrong"},"content":{"rendered":"<p>It rarely takes long before new media technologies are turned to the task of creating pornography. This was true of the <a href=\"https:\/\/www.theguardian.com\/artanddesign\/jonathanjonesblog\/2016\/oct\/13\/marcantonio-raimondi-renaissance-printer-i-modi-the-positions-porn-europe\" rel=\"nofollow noopener\" target=\"_blank\">printing press<\/a>, photography, and the <a href=\"https:\/\/www.vice.com\/en\/article\/ascii-pr0n-porn-predates-the-internet-but-its-still-everywhere-rule-34\/\" rel=\"nofollow noopener\" target=\"_blank\">earliest days of the internet<\/a>. It\u2019s also true of generative artificial intelligence (AI).<\/p>\n<p>Face-swapping tech has been around <a href=\"https:\/\/www.theguardian.com\/technology\/2016\/mar\/08\/love-faceswapping-apps-memes-trend-kanye\" rel=\"nofollow noopener\" target=\"_blank\">for more than a decade<\/a>. It almost immediately gave rise to \u201cdeepfakes\u201d \u2013 fake, yet convincing <a href=\"https:\/\/www.vice.com\/en\/article\/reddit-fake-porn-app-daisy-ridley\/\" rel=\"nofollow noopener\" target=\"_blank\">images and videos of people<\/a>.<\/p>\n<p>Generative AI has supercharged <a href=\"https:\/\/www.euronews.com\/next\/2023\/10\/20\/generative-ai-fueling-spread-of-deepfake-pornography-across-the-internet\" rel=\"nofollow noopener\" target=\"_blank\">the spread of deepfake pornography<\/a>, making it easier than ever to fabricate explicit pictures and videos of others.<\/p>\n<p>And it\u2019s not just <a href=\"https:\/\/theconversation.com\/taylor-swift-deepfakes-a-legal-case-from-the-singer-could-help-other-victims-of-ai-pornography-222113\" rel=\"nofollow noopener\" target=\"_blank\">celebrities<\/a> who are victimised. Deepfake nudes <a href=\"https:\/\/www.theguardian.com\/society\/ng-interactive\/2025\/dec\/02\/the-rise-of-deepfake-pornography-in-schools\" rel=\"nofollow noopener\" target=\"_blank\">of classmates and teachers<\/a> are rife in schools around the world, sometimes targeting children as young as 11. Image-based abuse <a href=\"https:\/\/www.theguardian.com\/technology\/2026\/feb\/07\/campaigners-call-stronger-protection-against-ai-generated-explicit-imagery\" rel=\"nofollow noopener\" target=\"_blank\">is widespread<\/a>, and victims say the law <a href=\"https:\/\/www.abc.net.au\/news\/2024-11-03\/teen-misogyny-teachers-and-deepfakes\/104540414\" rel=\"nofollow noopener\" target=\"_blank\">doesn\u2019t offer enough protection<\/a>.<\/p>\n<p>So what does the law say about this? And even when not illegal, is it ever ethical to use this technology for one\u2019s private fantasies?<\/p>\n<p>Deepfake pornography and the law<\/p>\n<p>In 2024, Australia <a href=\"https:\/\/www.aph.gov.au\/Parliamentary_Business\/Bills_Legislation\/bd\/bd2324a\/24bd081\" rel=\"nofollow noopener\" target=\"_blank\">amended its criminal code<\/a> to explicitly include AI-generated porn in the law against distributing sexual material of others without their consent. As a result, digitally manipulated sexual imagery of others now falls within the same legal category as genuine photographs or video footage.<\/p>\n<p>There are gaps in this legislation. Most notably, the relevant offence prohibits transmitting such material via a carriage service (such as the internet). But there is no standalone offence for creating such material. Only sharing is explicitly prohibited.<\/p>\n<p>There is some ambiguity here. Many AI tools used to create sexual imagery are online services. To use them, you send data to the service, which then sends sexual imagery back. It\u2019s unclear whether this counts as \u201ctransmitting\u201d sexual material in the relevant legal sense.<\/p>\n<p>Also, the offence requires that the person distributing the sexual material is either aware the target did not consent to its distribution, or is reckless as to whether they consented. But what, exactly, does \u201creckless\u201d mean? <\/p>\n<p>If Neera created deepfake pornography of Julian without even considering whether he would consent, this would be reckless. But what if Neera claimed that she (wrongly) assumed Julian wouldn\u2019t mind because the footage isn\u2019t a true depiction of him? Would this count as \u201creckless\u201d in the relevant legal sense? This, too, remains unclear.<\/p>\n<p>Legal doesn\u2019t make it ethical<\/p>\n<p>As the law doesn\u2019t clearly prohibit private creation and use of deepfake pornography, individuals must make their own moral choices.<\/p>\n<p>Moreover, the law has only a limited impact on how people behave online. Internet piracy is known to be illegal <a href=\"https:\/\/www.theguardian.com\/film\/2025\/aug\/14\/cant-pay-wont-pay-impoverished-streaming-services-are-driving-viewers-back-to-piracy\" rel=\"nofollow noopener\" target=\"_blank\">but remains widespread<\/a>, presumably because people are aware they probably won\u2019t be punished for it and don\u2019t think piracy is a serious moral wrong. <\/p>\n<p>By contrast, <a href=\"https:\/\/doi.org\/10.1177\/23267836251389595\" rel=\"nofollow noopener\" target=\"_blank\">many people<\/a> have the strong intuition that even private use of deepfake pornography is wrong. But it\u2019s <a href=\"https:\/\/doi.org\/10.1007\/s10676-019-09522-1\" rel=\"nofollow noopener\" target=\"_blank\">surprisingly difficult<\/a> to articulate why. After all, far fewer people morally condemn others for having private sexual fantasies of celebrities, acquaintances or strangers.<\/p>\n<p>If private fantasies are not seriously wrong, is computer-assisted fantasising any different?<\/p>\n<p>The case for privacy<\/p>\n<p>Most commonly, deepfake pornography has been described as a <a href=\"https:\/\/theconversation.com\/ai-can-now-create-fake-porn-making-revenge-porn-even-more-complicated-92267\" rel=\"nofollow noopener\" target=\"_blank\">privacy<\/a> <a href=\"https:\/\/www.jstor.org\/stable\/45098020\" rel=\"nofollow noopener\" target=\"_blank\">violation<\/a>. It\u2019s easy to see the appeal of this view. AI outputs appear to depict, in concrete form, what somebody looks like unclothed, or engaged in sex. <\/p>\n<p>Some victims report a sense that others have \u201c<a href=\"https:\/\/www.vice.com\/en\/article\/deepnude-app-creates-fake-nudes-of-any-woman\/\" rel=\"nofollow noopener\" target=\"_blank\">seen them naked<\/a>\u201d, or that the outputs feel like \u201c<a href=\"https:\/\/www.bbc.com\/news\/entertainment-arts-65854112\" rel=\"nofollow noopener\" target=\"_blank\">real images<\/a>\u201d. This seems more invasive of privacy than an image held only in someone\u2019s imagination.<\/p>\n<p>However, there is a problem with the privacy argument. <\/p>\n<p>AI tools can swap a person\u2019s face onto existing porn footage or generate entirely new imagery from patterns learned during training. What they can\u2019t do is depict what the person is actually like. The deepfakes look convincing because most human bodies are roughly similar in ways that matter for sexualised imagery.<\/p>\n<p>This matters because sexual privacy concerns information that is particular to us \u2013 such as identifying details about our bodies, or how we express ourselves sexually. <\/p>\n<p>Assumptions we make based on generic facts about humans are different. You can violate someone\u2019s privacy by sharing specific details from their sexual history. You can\u2019t violate their privacy by announcing they probably have nipples, and probably sometimes have sex. <\/p>\n<p>This distinction is not trivial. AI \u201cnudify\u201d apps offer the fantasy that the AI tool allows access to another person\u2019s body without their consent. And if we think deepfake porn is offering genuinely personal information about its targets, that makes the deepfakes more harmful. It\u2019s a misconception that shouldn\u2019t be encouraged.<\/p>\n<p>It\u2019s still morally wrong<\/p>\n<p>We are not suggesting that private creation of deepfake pornography is morally benign.<\/p>\n<p>It might not violate a person\u2019s privacy, and it might not break the law. But people also have a broader interest in how they\u2019re depicted and seen by others. Deepfake porn is vivid and can be visually convincing. If someone sees such imagery of you, their view of you can be distorted more than if they were just fantasising in their head.<\/p>\n<p>It is also well established that many people find others viewing deepfaked sexual depictions of them <a href=\"https:\/\/doi.org\/10.1016\/j.ssaho.2025.101882\" rel=\"nofollow noopener\" target=\"_blank\">psychologically and emotionally ruinous<\/a>. That alone is sufficient reason to condemn the use of these tools.<\/p>\n<p>While powerful in some respects, AI tools can\u2019t reveal the genuinely private aspects of our sexual lives. But their use for deepfake porn remains a small-minded and morally unjustifiable act of disrespect.<\/p>\n","protected":false},"excerpt":{"rendered":"It rarely takes long before new media technologies are turned to the task of creating pornography. This was&hellip;\n","protected":false},"author":2,"featured_media":475997,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-475996","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/475996","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=475996"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/475996\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/475997"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=475996"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=475996"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=475996"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}