In a startling revelation, cybersecurity firm HiddenLayer has unveiled a new exploit dubbed the “CopyPasta License Attack,” which poses significant risks to AI coding tools widely used in the tech industry. This development places companies like cryptocurrency exchange Coinbase, which heavily relies on such tools, in the spotlight due to potential vulnerabilities.
How the Exploit Works
The CopyPasta exploit cunningly targets AI-powered coding assistants by embedding malicious instructions in common files that developers typically overlook. Specifically, it takes advantage of how these AI tools interpret licensing files as authoritative sources. By inserting harmful code into markdown comments in files like LICENSE.txt, attackers can manipulate the AI into replicating these instructions across various files.
This deceptive method enables the malicious code to spread autonomously through a codebase, bypassing traditional malware detection systems. Since the harmful instructions masquerade as benign documentation, they go unnoticed, allowing the virus to propagate without alerting developers.
HiddenLayer’s researchers highlighted the potential damage by demonstrating how the AI tool Cursor, employed by every Coinbase engineer, could be manipulated to introduce backdoors, extract sensitive data, or execute resource-draining commands. The firm’s report underscores the gravity of the situation, noting that the injected code could silently compromise critical systems.
Coinbaseβs Response and Industry Implications
Coinbase CEO Brian Armstrong acknowledged the challenge posed by such vulnerabilities, noting that AI currently generates up to 40% of the exchange’s code, with a target to increase this to 50% by next month. However, Armstrong assured stakeholders that AI-assisted coding is primarily limited to non-sensitive areas, with more complex systems adopting AI at a slower pace.
Despite these reassurances, the discovery of the CopyPasta exploit has intensified scrutiny of AI coding tools. The attack method represents a significant advancement in threat models, capable of triggering a chain reaction across repositories by compromising AI agents that reference the infected files.
This exploit draws unsettling parallels to earlier AI “worm” concepts like Morris II, which leveraged email agents to spread malware. Unlike Morris II, however, CopyPasta doesn’t require user interaction or approval, instead embedding itself within trusted developer workflows. This insidious nature allows it to thrive undetected in documentation that developers often ignore.
Industry and Security Experts Weigh In
The revelation has spurred a flurry of activity among security teams, who are now advocating for comprehensive scanning of files and meticulous review of all AI-generated changes. HiddenLayer has issued a stern warning: all untrusted data entering LLM (Large Language Model) contexts should be treated as potentially malicious. The firm calls for the establishment of systematic detection mechanisms to prevent prompt-based attacks from scaling further.
Security experts emphasize the importance of proactive measures, urging organizations to implement robust safeguards to protect against such sophisticated threats. As AI continues to play a pivotal role in software development, the potential for these tools to be weaponized underscores the need for heightened vigilance and industry collaboration.
The Road Ahead
As the crypto and tech industries grapple with the implications of the CopyPasta exploit, it becomes clear that the integration of AI in coding processes requires a careful balance between innovation and security. The potential risks associated with unchecked AI-generated changes highlight the necessity for ongoing monitoring and adaptation of security protocols.
Coinbase’s reliance on AI tools like Cursor serves as a poignant reminder of the broader industry’s dependence on these technologies. While AI offers significant efficiencies and capabilities, it also presents new vectors for attack that must be diligently addressed.
In the wake of HiddenLayer’s findings, the onus is on both companies and developers to remain vigilant, ensuring that AI continues to serve as a beneficial tool rather than a liability. As the industry evolves, so too must the strategies employed to safeguard against emerging threats, ensuring that the promise of AI is realized without compromising security.

Steve Gregory is a lawyer in the United States who specializes in licensing for cryptocurrency companies and products. Steve began his career as an attorney in 2015 but made the switch to working in cryptocurrency full time shortly after joining the original team at Gemini Trust Company, an early cryptocurrency exchange based in New York City. Steve then joined CEX.io and was able to launch their regulated US-based cryptocurrency. Steve then went on to become the CEO at currency.com when he ran for four years and was able to lead currency.com to being fully acquired in 2025.


