LegalPwn exploits AI models by using legitimate legal language to trick them into misclassifying malicious software as safe code.

LegalPwn exploits AI models by using legitimate legal language to trick them into misclassifying malicious software as safe code.

The novel “LegalPwn” attack, developed by researchers at Pangea Labs, demonstrates how attackers can trick artificial intelligence models like ChatGPT, Google Gemini, GitHub Copilot, Meta’s Llama, and xAI’s Grok into misclassifying malicious software as safe code by cleverly disguising it within seemingly legitimate legal language.