detecting-ai-model-prompt-injection-attacks
mukul975/Anthropic-Cybersecurity-Skills
Detects prompt injection attacks against LLMs via regex, heuristic, and DeBERTa classifier layers, suited for preprocessing user input, scanning logs, auditing historical prompts, and integrating into security workflows before queries reach the model.