On Model Inversion Attacks
Attacks on machine learning systems are defined as special actions on the ele- ments of the machine learning pipeline, which are designed to either prevent the normal operation of machine learning systems or ensure their special functioning, which is necessary for the attacker. Model inversion attacks aim to expose the private data used to train the model. Attacks that expose private information about machine learning systems are a big threat to machine learning as a service projects. In this article, we provide an overview of off-the-shelf software tools for performing model inversion attacks.