Training data poisoning lets attackers corrupt AI models with tiny amounts of malicious data, causing hidden backdoors and dangerous outputs. Learn how it works, real-world examples, and proven ways to defend your models.
AI & Machine Learning