{"iri":"https://folio.openlegalstandard.org/R8nzGmaooF9VdUO5jELI3Et","label":"Evasion Attack","sub_class_of":["https://folio.openlegalstandard.org/RCafxux1zDpXGZzqVBKNVpa"],"parent_class_of":[],"is_defined_by":null,"see_also":["https://folio.openlegalstandard.org/RBMj5dbvLFgFGrPZ3MvFGYl"],"comment":null,"deprecated":false,"preferred_label":"Adversarial Evasion","alternative_labels":["Deceptive Input Attack","Evasion","Misclassification Attack"],"translations":{},"hidden_label":null,"definition":"An evasion attack is one of the most common attacks on machine learning models performed during production. It involves designing an input that appears normal to a human but is misclassified by ML models, affecting their behavior; a typical example includes altering pixels in an image to cause an image recognition system to fail in classification.","examples":[],"notes":[],"history_note":null,"editorial_note":null,"in_scheme":null,"identifier":null,"description":null,"source":null,"country":null}