adversarial prompt filtering
5 Prompt Filters Protect Machine Learning?
2023 research shows that five prompt filters - adversarial prompt filtering, data-poisoning checks, content-authenticity guard, NLP-pipeline safeguards, and robust testing - serve as a front-line defense for machine learning models. These filters examine every instruction before it reaches the core model, preventing degraded performance and compliance breaches. Adversarial Prompt Filtering