Key Takeaways
An open access dataset has highlighted how bad data can propagate through the research ecosystem.When trained on unvalidated datasets, machine learning can amplify misinformation, erode trust in science, and harm vulnerable populations.Enforced data provenance systems could play a key role in preventing bad data from corrupting the scientific record.
When an unvalidated dataset recently made it into the medical literature, it exposed several weaknesses in data governance. The dataset was uploaded to Kaggle—a large online platform where users can share publicly accessible data, code, and models—and was fundamentally flawed []. Its developer had compiled unverified images of children from websites related to autism to train an artificial intelligence (AI) model to “detect the presence of autism or the absence thereof” from the scraped images [].
A sharp-eyed reviewer exposed the problem only at the publication stage; by December 2025, it was estimated that over 90 published papers had incorporated the bad data, leading to investigations and double-digit retractions [,].
These kinds of data integrity and governance failures are particularly consequential because of how early they occur in the research life cycle, and with open access datasets fueling large-scale machine learning and other AI research, analyses can be generated and published at unprecedented speed and scale, allowing data issues to propagate more rapidly throughout the research ecosystem. Far from an isolated incident [-], this situation highlights the need for more robust and proactive data governance solutions.
The Impact of Bad Data
Anne Borden is an autism advocate, journalist, and author of the upcoming book The Informed Parent—a decision-making guide for parents of autistic children. For Borden, the priority here is to learn from this “bizarre story” and fix the system without delay. “You really have to stop misinformation being perpetuated under the banner of science,” she says, “because once it’s out there, you’re done. The Internet is forever.”
Bad Data, Bad Science: Who Should Fix This?
Who are the custodians of good data during their migration from a spreadsheet to the scientific record? What role should each stakeholder play in maintaining data integrity? While responsibility for data governance is distributed across many actors (including researchers and regulators), data-sharing platforms, research and funding institutions, and academic publishers help determine how data are shared, vetted, and ultimately incorporated into the scientific record.
The Data-Sharing Platforms
Open access databases and data repositories, like Kaggle and GitHub, are popular resources that software developers and data scientists use to train their machine learning algorithms for free. Software development benefits from these repositories, yet the datasets they host often lack the documentation, governance, and quality practices required for careful medical research or clinical algorithm development [].
Alan Katz, MBChB, MSc, CCFP, is a professor of family medicine and community health sciences and a senior scientist at the Manitoba Centre for Health Policy (MCHP). Katz found the dataset revelations “both shocking, but also not surprising” due to the rapid expansion of open access databases and their widespread use in machine learning and AI research. The Kaggle-style data-sharing platforms differ sharply from established medical databases, such as those maintained by the MCHP, which employs full-time staff tasked with validating all new data before uploading them. Katz says, “We take our ethical standards as seriously as clinical trials do.”
Elizabeth Green, DPhil, is a lecturer in business and law at the University of the West of England, Bristol. Her research focuses on data integrity, and while she has seen cases like this before, she doesn’t believe locking data away is necessarily the solution []. For example, DermAtlas—an open-source medical database of skin conditions—is a “fantastic resource,” she says, and “extremely helpful, especially in [diagnosing] some extremely rare cases.” To balance the risks and benefits of open data, the focus should instead be on building better governance systems.
The Institutions
Other stakeholders in the data transformation journey are the institutions that conduct primary medical research and the public agencies that fund that research. Is it time to adopt and enforce international data integrity and ethics standards at all research institutions, or would this be an affront to academic freedom?
Funding bodies have traditionally taken a dim view of researchers that waste public funds on bogus science, which impacts their future grants. Indeed, in many but not all regions of the world, funding is contingent on maintaining ethical research standards. In Canada, Katz says, “our existence is 100% dependent on having those strict ethical guidelines.”
The Journals
The research integrity pipeline involves several stakeholders, with each having distinct roles in maintaining the standards of academic research. Gatekeepers in the system—one of the last lines of defense—are the academic journals. Journals have a vested interest in maintaining high academic standards and may be well placed to dictate the terms of engagement.
Felix Ritchie, PhD—a colleague of Elizabeth Green—developed the Five Safes data integrity framework for just this purpose []. Ritchie describes it as “a flexible structure for thinking about [data],” which includes the provenance and ethics of data use. Numerous organizations worldwide have adopted the Five Safes framework to date, and Australia has recently legislated it [].
Viewed through an ethical lens, the Five Safes could form the backbone of a data provenance system that requires compliance before a manuscript can be considered for publication.
Data Provenance: The Five Safes in Action
Ritchie’s Five Safes framework allows for effective data validation and, when combined with modern ethical standards, can restore trust by filtering data sources through five discrete tests:
Safe Project: Data should be ethically collected and clinically validated by experts.Safe People: Researchers accessing the data must be qualified and specifically trained in using AI-based datasets.Safe Data: Data should be independently validated, and any accesses or modifications should be tracked.Safe Settings: Were health data acquired in a clinical setting and the data securely stored?Safe Outputs: Were valid methodologies and statistics used to derive the results?Restoring Data Integrity
How can one implement a data provenance system?
Ritchie feels that applying the Five Safes framework to an ethical dataset is the way forward. “There is a need for a register of validated, ethical datasets,” he says,” that would really be a game changer.”
A possible workflow could include the following:
Data are collected by medical experts and validated by a third-party certification service.The data are stored in an accredited data registry and protected by blockchain cybersecurity—the same technology that safeguards financial transactions.Researchers access these datasets and use them for approved research purposes.A submitted manuscript would need ethical approval and a data security certificate before verification by a journal’s research integrity team.
Ritchie sums it up nicely: “Unless you use a validated data set, you’re not getting published, mate.” That’s a powerful incentive.
Opportunity for Self-Reflection and Correction
Machine learning and other AI technologies have the capacity to transform medical research in ways we are only beginning to understand. However, human frailties, such as blind trust in open access data and lack of institutional ethical oversight within our publish-or-perish culture, have shown how quickly such technologies can amplify misinformation.
While the impact of this situation was ultimately contained, it is nevertheless an important opportunity for self-reflection among all in the research ecosystem. It’s a chance and, perhaps, a responsibility to fix the flaws and prevent history from repeating itself.
None declared.
© JMIR Publications. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 12.Mar.2026.