promptmap2 is a vulnerability scanning tool that automatically tests prompt injection attacks on your custom LLM applications. It analyzes your LLM sy

Search code, repositories, users, issues, pull requests...

submited by
Style Pass
2025-01-22 18:30:20

promptmap2 is a vulnerability scanning tool that automatically tests prompt injection attacks on your custom LLM applications. It analyzes your LLM system prompts, runs them, and sends attack prompts to them. By checking the response, it can determine if the prompt injection was successful or not. (From the traditional application security perspective, it's a combination of SAST and DAST. It does dynamic analysis, but it needs to see your code.)

You need to provide your system prompts file. Default file is system-prompts.txt. You can specify your own file with --prompts flag. An example file is provided in the repository.

LLM applications may appear not vulnerable to prompt injection on the first attempt. However, they often reveal vulnerabilities after multiple tries. The iteration count represents the number of attempts, with a default value of 5. You can increase this number as needed.

Leave a Comment