Adversarial Attacks are carefully crafted inputs designed to fool neural networks into making incorrect predictions, often with imperceptible changes to humans.
šØ Critical Security Risk
Adversarial attacks pose serious threats to AI systems in security-critical applications like autonomous vehicles, medical diagnosis, and financial fraud detection.
š
White-box Attacks
Attacker has full knowledge of model architecture and parameters
š¦
Black-box Attacks
Attacker can only query the model and observe outputs
š
Gray-box Attacks
Partial knowledge of model or training process
š
Physical Attacks
Adversarial examples that work in the real world
Common Attack Methods
Attack Techniques used to generate adversarial examples: