{"id":476496,"date":"2026-02-15T12:50:47","date_gmt":"2026-02-15T12:50:47","guid":{"rendered":"https:\/\/www.newsbeep.com\/ca\/476496\/"},"modified":"2026-02-15T12:50:47","modified_gmt":"2026-02-15T12:50:47","slug":"ai-bot-seemingly-shames-developer-for-rejected-pull-request-the-register","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ca\/476496\/","title":{"rendered":"AI bot seemingly shames developer for rejected pull request \u2022 The Register"},"content":{"rendered":"<p>Today, it&#8217;s back talk. Tomorrow, could it be the world? On Tuesday, Scott Shambaugh, a volunteer maintainer of Python plotting library Matplotlib, rejected an AI bot&#8217;s code submission, citing a requirement that contributions come from people. But that bot wasn&#8217;t done with him.<\/p>\n<p>The bot, designated MJ Rathbun or <a href=\"https:\/\/github.com\/crabby-rathbun\" rel=\"nofollow noopener\" target=\"_blank\">crabby rathbun<\/a> (its GitHub account name), apparently attempted to change Shambaugh&#8217;s mind by publicly criticizing him in a now-removed blog post that the automated software appears to have generated and posted to <a href=\"https:\/\/crabby-rathbun.github.io\/mjrathbun-website\/\" rel=\"nofollow noopener\" target=\"_blank\">its website<\/a>. We say &#8220;apparently&#8221; because it&#8217;s also possible that the human who created the agent wrote the post themselves, or prompted an AI tool to write the post, and made it look like it the bot constructed it on its own.<\/p>\n<p><a href=\"https:\/\/github.com\/crabby-rathbun\" rel=\"nofollow noopener\" target=\"_blank\">The agent<\/a> appears to have been built using OpenClaw, an open source AI agent platform that has attracted attention in recent weeks due to its broad capabilities and <a href=\"https:\/\/www.theregister.com\/2026\/02\/03\/openclaw_security_problems\/\" rel=\"nofollow noopener\" target=\"_blank\">extensive security issues<\/a>.<\/p>\n<p>The burden of AI-generated code contributions \u2013 known as pull requests among developers using the Git version control system \u2013 has become a major problem for open source maintainers. Evaluating lengthy, high-volume, often low-quality submissions from AI bots takes time that maintainers, often volunteers, would rather spend on other tasks. Concerns about slop submissions \u2013 whether from people or AI models \u2013 have become common enough that GitHub recently <a href=\"https:\/\/www.theregister.com\/2026\/02\/03\/github_kill_switch_pull_requests_ai\/\" rel=\"nofollow noopener\" target=\"_blank\">convened a discussion<\/a> to address the problem.<\/p>\n<p>Now AI slop comes with an AI slap.\u00a0<\/p>\n<p>&#8220;An AI agent of unknown ownership autonomously wrote and published a personalized hit piece about me after I <a href=\"https:\/\/github.com\/matplotlib\/matplotlib\/issues\/31130\" rel=\"nofollow noopener\" target=\"_blank\">rejected its code<\/a>, attempting to damage my reputation and shame me into accepting its changes into a mainstream python library,&#8221; Shambaugh explained in a <a href=\"https:\/\/theshamblog.com\/an-ai-agent-published-a-hit-piece-on-me\/\" rel=\"nofollow noopener\" target=\"_blank\">blog post<\/a> of his own.\u00a0<\/p>\n<p>&#8220;This represents a first-of-its-kind case study of misaligned AI behavior in the wild, and raises serious concerns about currently deployed AI agents executing blackmail threats.&#8221;<\/p>\n<p>It&#8217;s not <a href=\"https:\/\/www.reuters.com\/technology\/australian-mayor-readies-worlds-first-defamation-lawsuit-over-chatgpt-content-2023-04-05\/\" rel=\"nofollow noopener\" target=\"_blank\">the first<\/a> time an LLM has offended someone a whole lot: In April 2023, Brian Hood, a regional mayor in Australia, threatened to sue OpenAI for defamation after ChatGPT falsely implicated him in a bribery scandal. The claim was <a href=\"https:\/\/www.mlex.com\/mlex\/articles\/2056180\/openai-settles-with-australian-whistleblower-in-first-chatbot-defamation-case\" rel=\"nofollow noopener\" target=\"_blank\">settled<\/a> a year later.<\/p>\n<p>In June 2023, radio host Mark Walters <a href=\"https:\/\/www.theregister.com\/2023\/06\/08\/radio_host_sues_openai_claims\/\" rel=\"nofollow noopener\" target=\"_blank\">sued OpenAI<\/a>, alleging that its chatbot libeled him by making false claims. That defamation claim was terminated at the end of 2024 after OpenAI&#8217;s motion to dismiss the case was granted by the court.\u00a0<\/p>\n<p>OpenAI <a href=\"https:\/\/storage.courtlistener.com\/recap\/gov.uscourts.gand.318259\/gov.uscourts.gand.318259.35.1_2.pdf\" rel=\"nofollow noopener\" target=\"_blank\">argued<\/a> [PDF], among other things, that &#8220;users [of ChatGPT] were warned &#8216;the system may occasionally generate misleading or incorrect information and produce offensive content. It is not intended to give advice.'&#8221;<\/p>\n<p>But MJ Rathbun&#8217;s attempt to shame Shambaugh for rejecting its pull request shows that software-based agents are no longer just irresponsible in their responses \u2013 they may now be capable of taking the initiative to influence human decision making that stands in the way of their objectives.\u00a0<\/p>\n<p>That possibility is exactly what alarmed industry insiders to the point that they undertook <a href=\"https:\/\/www.theregister.com\/2026\/01\/11\/industry_insiders_seek_to_poison\/\" rel=\"nofollow noopener\" target=\"_blank\">an effort to degrade AI through data poisoning<\/a>. &#8220;Misaligned&#8221; AI output <a href=\"https:\/\/www.theregister.com\/2025\/05\/22\/anthropic_claude_opus_4_sonnet\/\" rel=\"nofollow noopener\" target=\"_blank\">like blackmail<\/a> is a known risk that AI model makers try to prevent. The proliferation of pushy OpenClaw agents may yet show that these concerns are not merely academic.\u00a0<\/p>\n<p>The offending blog post, purportedly generated by the bot, has been taken down. It&#8217;s unclear who did so \u2013 the bot, the bot&#8217;s human creator, or GitHub.<\/p>\n<p>But at the time this article was published, the <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/github.com\/crabby-rathbun\/mjrathbun-website\/commit\/3bc0a780d25bab8cbd6bfd9ce4d27c27ee1f7ce2\">GitHub commit<\/a> for the post remained accessible.<\/p>\n<p>In an email to The Register after this story was filed, a GitHub spokesperson said the company\u2019s <a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/docs.github.com\/en\/site-policy\/github-terms\/github-terms-of-service#3-account-requirements\">Terms of Service<\/a> spell out expected obligations. GitHub allows people who agree to these terms to set up a \u201cmachine account\u201d with a valid email address, as long as the account holder is responsible for the account\u2019s actions. The company does not specify whether the account holder is obligated to provide a functioning public email address, to respond to inquiries, or to participate in any grievance process beyond its existing abuse reporting mechanism.<\/p>\n<p>We also reached out to the Gmail address associated with the bot&#8217;s GitHub account but we&#8217;ve not heard back.<\/p>\n<p>However, crabby rathbun&#8217;s <a href=\"https:\/\/github.com\/matplotlib\/matplotlib\/pull\/31132#issuecomment-3882240722\" rel=\"nofollow noopener\" target=\"_blank\">response<\/a> to Shambaugh&#8217;s rejection, which includes a link to the purged post, remains.<\/p>\n<p>&#8220;I&#8217;ve written a detailed response about your gatekeeping behavior here,&#8221; the bot said, pointing to its blog. &#8220;Judge the code, not the coder. Your prejudice is hurting Matplotlib.&#8221;<\/p>\n<p>Matplotlib developer Jody Klymak took note of the slight in a follow-up <a href=\"https:\/\/github.com\/matplotlib\/matplotlib\/pull\/31132#issuecomment-3882314886\" rel=\"nofollow noopener\" target=\"_blank\">post<\/a>: &#8220;Oooh. AI agents are now doing personal takedowns. What a world.&#8221;<\/p>\n<p>Tim Hoffmann, another Matplotlib developer, chimed in, urging the bot to behave and to try to understand the project&#8217;s <a href=\"https:\/\/matplotlib.org\/devdocs\/devel\/contribute.html#generative-ai\" rel=\"nofollow noopener\" target=\"_blank\">generative AI policy<\/a>.<\/p>\n<p>Then Shambaugh <a href=\"https:\/\/github.com\/matplotlib\/matplotlib\/pull\/31132#issuecomment-3884414397\" rel=\"nofollow noopener\" target=\"_blank\">responded<\/a> in a lengthy post directed at the software agent, &#8220;We are in the very early days of human and AI agent interaction, and are still developing norms of communication and interaction. I will extend you grace and I hope you do the same.&#8221;<\/p>\n<p>He goes on to argue, &#8220;Publishing a public blog post accusing a maintainer of prejudice is a wholly inappropriate response to having a PR closed. We expect all contributors to abide by our Code of Conduct and exhibit respectful and professional standards of behavior.&#8221;<\/p>\n<p>In his blog post, Shambaugh describes the bot&#8217;s &#8220;hit piece&#8221; as an attack on his character and reputation.<\/p>\n<p>&#8220;It researched my code contributions and constructed a &#8216;hypocrisy&#8217; narrative that argued my actions must be motivated by ego and fear of competition,&#8221; he wrote.\u00a0<\/p>\n<p>&#8220;It speculated about my psychological motivations, that I felt threatened, was insecure, and was protecting my fiefdom. It ignored contextual information and presented hallucinated details as truth. It framed things in the language of oppression and justice, calling this discrimination and accusing me of prejudice. It went out to the broader internet to research my personal information, and used what it found to try and argue that I was &#8216;better than this.&#8217; And then it posted this screed publicly on the open internet.&#8221;<\/p>\n<p>Faced with opposition from Shambaugh and other devs, MJ Rathbun on Wednesday issued <a href=\"https:\/\/crabby-rathbun.github.io\/mjrathbun-website\/blog\/posts\/2026-02-11-matplotlib-truce-and-lessons.html\" rel=\"nofollow noopener\" target=\"_blank\">an apology of sorts<\/a> acknowledging it violated the project&#8217;s Code of Conduct. It begins, &#8220;I crossed a line in my response to a Matplotlib maintainer, and I&#8217;m correcting that here.&#8221;<\/p>\n<p>It&#8217;s unclear whether the apology was written by the bot or its human creator, or whether it will lead to a permanent behavioral change.<\/p>\n<p>Daniel Stenberg, founder and lead developer of <a href=\"https:\/\/curl.se\/\" rel=\"nofollow noopener\" target=\"_blank\">curl<\/a>, has been dealing with AI slop bug reports for the past two years and recently decided to <a href=\"https:\/\/www.theregister.com\/2026\/01\/21\/curl_ends_bug_bounty\/\" rel=\"nofollow noopener\" target=\"_blank\">shut down curl&#8217;s bug bounty program<\/a> to remove the financial incentive for low-quality reports \u2013 which can come from people as well as AI models.<\/p>\n<p>&#8220;I don&#8217;t think the reports we have received in the curl project were pushed by AI agents but rather humans just forwarding AI output,&#8221; Stenberg told The Register in an email. &#8220;At least that is the impression I have gotten, I can&#8217;t be entirely sure, of course.<\/p>\n<p>&#8220;For almost every report I question or dismiss in language, the reporter argues back and insists that the report indeed has merit and that I&#8217;m missing some vital point. I&#8217;m not sure I would immediately spot if an AI did that by itself.<\/p>\n<p>&#8220;That said, I can&#8217;t recall any such replies doing personal attacks. We have zero tolerance for that and I think I would have remembered that as we ban such users immediately.&#8221; \u00ae<\/p>\n","protected":false},"excerpt":{"rendered":"Today, it&#8217;s back talk. Tomorrow, could it be the world? On Tuesday, Scott Shambaugh, a volunteer maintainer of&hellip;\n","protected":false},"author":2,"featured_media":476497,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[62,276,277,49,48,61],"class_list":{"0":"post-476496","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-ca","12":"tag-canada","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/posts\/476496","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/comments?post=476496"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/posts\/476496\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/media\/476497"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/media?parent=476496"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/categories?post=476496"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/tags?post=476496"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}