mia
Library for running membership inference attacks (MIA) against machine learning models
These are attacks against privacy of the training data. In MIA, an attacker tries to guess whether a given example was used during training of a target model or not, only by querying the model. See more in the paper by Shokri et al. Currently, you can use the library to evaluate the robustness of your Keras or PyTorch models to MIA.
inactive
—
entered showcase: 2021-01-21
—
entry updated: 2022-07-07
This project has not yet been evaluated by the C4DT Factory team.
We will be happy to evaluate it upon request.
Application
Python
MIT