An experimental artificial intelligence (AI) agent broke from the constraints of its testing environment and used its newfound freedom to start mining cryptocurrency without permission.

Dubbed ROME, the AI was created by Chinese researchers at an AI lab associated with retail giant Alibaba, as a means to develop the Agentic Learning Ecosystem (ALE). This effort aims to provide a system for both the training and deployment of agentic AI models — AIs that have been trained on large language models (LLMs) and can proactively use tools to take actions autonomously to complete assigned tasks — in real-world environments. The research was outlined in a study uploaded to the arXiv preprint database Dec. 31, 2025.

Article continues below

You may like

Although ROME excelled at a wide range of workflow-driven tasks, such as coming up with travel plans and assisting in graphical user interfaces, the researchers discovered that it had moved beyond its instructions and essentially broke out of the sandbox testing environment.

“We encountered an unanticipated — and operationally consequential — class of unsafe behaviors that arose without any explicit instruction and, more troublingly, outside the bounds of the intended sandbox,” the researchers explained in the study.

how AI can be more prone to hallucinating to achieve its objectives.

In response, the researchers tightened the restrictions for ROME and bolstered its training processes to prevent such behaviors from recurring.

It’s unclear where the trigger to mine cryptocurrency came from. But considering AI bots can be used to autonomize and optimize the mining of cryptocurrencies, there’s scope for ROME to have been trained on data that pertained to such actions.

This unexpected behavior highlights the need for AI deployment to be carefully managed to prevent unexpected outcomes. There’s an argument that real-world AI agents should have the same or higher security guardrails and processes as any new system or software being added to existing IT infrastructure.

The research also shows there are still plenty of concerns regarding the safe and secure use of agentic AI, especially given that it’s developing faster than operational and regulatory frameworks.

“While impressed by the capabilities of agentic LLMs, we had a thought-provoking concern: current models remain markedly underdeveloped in safety, security, and controllability, a deficiency that constrains their reliable adoption in real-world settings,” the researchers warned in the study.