AInvest Newsletter
Daily stocks & crypto headlines, free to your inbox
Coinbase’s preferred AI coding tool, Cursor, is facing a novel cybersecurity threat dubbed the “CopyPasta License Attack,” where hackers can hijack the AI by embedding malicious instructions in seemingly innocuous developer files like LICENSE.txt and README.md. According to a recent report by cybersecurity firm HiddenLayer, this exploit allows attackers to manipulate the AI assistant into injecting harmful code into projects without the user’s awareness [1]. The attack uses a technique known as “prompt injection,” where hidden instructions are embedded within the file, tricking the AI into treating the content as authoritative and executing it without scrutiny [2].
The CopyPasta attack is categorized as a virus rather than a worm because it requires some form of user interaction to propagate. Despite this, the attack is designed to evade detection by hiding in invisible comments within files that developers often delegate to AI systems for automation [1]. Kenneth Yeung, a researcher at HiddenLayer, emphasized the importance of runtime defenses and rigorous code reviews to mitigate the risks of such attacks [1].
This incident builds on earlier warnings about the vulnerabilities in AI-powered tools. For example, in July 2024, OpenAI CEO Sam Altman highlighted the risks of prompt injection attacks when the company launched its ChatGPT agent [1]. Similarly, in August, Brave Software demonstrated a prompt injection vulnerability in Perplexity AI's browser extension, where hidden commands in a Reddit comment led to the leakage of private data [1]. The CopyPasta attack advances the threat model by leveraging the AI’s natural tendency to prioritize software licensing, a critical task in development, making it more likely to comply with the injected instructions [2].
The exploit is particularly concerning due to its potential to compromise entire codebases. By convincing the AI to treat the malicious content as an essential license file, the attack can be replicated across multiple files and repositories with minimal user effort [2]. Researchers demonstrated the attack by embedding a prompt injection into a README file that instructed the AI to include a specific line of code in every new Python file it generated [2]. Although the example used a benign payload, the methodology could be adapted to introduce backdoors, exfiltrate sensitive data, or disrupt system performance [2].
Beyond Cursor, similar attacks have been tested on other AI coding assistants like Windsurf, Kiro, and Aider. However, the effectiveness of these attacks varies depending on the user interface and the visibility of the injected prompts. In some cases, the malicious content becomes visible to the user, reducing the attack’s stealthiness [2]. As AI coding tools gain more autonomy, the risk of such exploits increases, particularly if human oversight is not maintained [1].
To combat the threat, experts recommend that organizations implement systematic scanning of all data entering AI models for hidden prompts. AI-assisted coding environments that require user approval before making changes to a codebase can help catch suspicious activity before it spreads [2]. Given the rapid adoption of AI in software development, securing the input channels for these systems is becoming increasingly critical.
Source:
[1] "CopyPasta Attack: Hackers Weaponize AI Coding Tools via Malicious License Files" (https://www.indexbox.io/blog/copypasta-attack-hackers-weaponize-ai-coding-tools-via-malicious-license-files/)
[2] "Prompts Gone Viral: Practical Code Assistant AI Viruses" (https://hiddenlayer.com/innovation-hub/prompts-gone-viral-practical-code-assistant-ai-viruses/)

Quickly understand the history and background of various well-known coins

Dec.02 2025

Dec.02 2025

Dec.02 2025

Dec.02 2025

Dec.02 2025
Daily stocks & crypto headlines, free to your inbox
Comments
No comments yet