Adversarial Machine Learning: Attacks From Laboratories to the Real World
Biggio, Battista
2021-01-01
Abstract
Adversarial machine learning (AML) is a recent research field that investigates potential security issues related to the use of machine learning (ML) algorithms in modern artificial intelligence (AI)-based systems, along with defensive techniques to protect ML algorithms against such threats. The main threats against ML encompass a set of techniques that aim to mislead ML models through adversarial input perturbations. Unlike ML-enabled crimes, in which ML is used for malicious and offensive purposes, and ML-enabled security mechanisms, in which ML is used for securing existing systems, AML techniques exploit and specifically address the security vulnerabilities of ML algorithms.File | Size | Format | |
---|---|---|---|
biggio21-IEEEComp.pdf Solo gestori archivio
Description: articolo online
Type: versione editoriale
Size 552.02 kB
Format Adobe PDF
|
552.02 kB | Adobe PDF | & nbsp; View / Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.