A toolkit for tools and techniques related to the privacy and compliance of AI models. https://aip360.res.ibm.com
Find a file
abigailt a37ff06df8 Squashed commit of the following:
commit d53818644e
Author: olasaadi <92303887+olasaadi@users.noreply.github.com>
Date:   Mon Mar 7 20:12:55 2022 +0200

    Build the dt on all features anon (#23)

    * add param to build the DT on all features and not just on QI
    * one-hot encoding only for categorical features

commit c47819a031
Author: abigailt <abigailt@il.ibm.com>
Date:   Wed Feb 23 19:40:11 2022 +0200

    Update docs

commit 7e2ce7fe96
Merge: 7fbd1e4 752871d
Author: abigailt <abigailt@il.ibm.com>
Date:   Wed Feb 23 19:26:44 2022 +0200

    Merge remote-tracking branch 'origin/main' into main

commit 7fbd1e4b90
Author: abigailt <abigailt@il.ibm.com>
Date:   Wed Feb 23 19:22:54 2022 +0200

    Update version and docs

commit 752871dd0c
Author: olasaadi <92303887+olasaadi@users.noreply.github.com>
Date:   Wed Feb 23 14:57:12 2022 +0200

    add minimization notebook (#22)

    * add german credit notebook to showcase new features (minimize only some features and categorical features)

    * add notebook to show minimization data on a regression problem
2022-04-25 17:39:30 +03:00
apt Squashed commit of the following: 2022-04-25 17:39:30 +03:00
datasets Add data minimization functionality to the ai-privacy-toolkit (#3) 2021-07-12 15:56:42 +03:00
docs Squashed commit of the following: 2022-04-25 17:39:30 +03:00
notebooks Squashed commit of the following: 2022-04-25 17:39:30 +03:00
tests Squashed commit of the following: 2022-04-25 17:39:30 +03:00
.gitattributes Ignore Jupyter Notebooks in git language detection 2021-04-28 16:34:02 +03:00
.readthedocs.yaml Try to fix documentation 2021-06-07 17:01:21 +03:00
LICENSE Initial commit 2021-04-28 06:25:00 -04:00
pyproject.toml Files for pypi dist 2021-08-02 11:48:05 +03:00
README.md Add link to Slack 2021-11-02 14:19:22 +02:00
requirements.txt Remove redundant code. 2022-03-06 21:15:07 +02:00
setup.cfg Squashed commit of the following: 2022-04-25 17:39:30 +03:00

ai-privacy-toolkit


A toolkit for tools and techniques related to the privacy and compliance of AI models.

The anonymization module contains methods for anonymizing ML model training data, so that when a model is retrained on the anonymized data, the model itself will also be considered anonymous. This may help exempt the model from different obligations and restrictions set out in data protection regulations such as GDPR, CCPA, etc.

The minimization module contains methods to help adhere to the data minimization principle in GDPR for ML models. It enables to reduce the amount of personal data needed to perform predictions with a machine learning model, while still enabling the model to make accurate predictions. This is done by by removing or generalizing some of the input features.

Official ai-privacy-toolkit documentation: https://ai-privacy-toolkit.readthedocs.io/en/latest/

Installation: pip install ai-privacy-toolkit

For more information or help using or improving the toolkit, please contact Abigail Goldsteen at abigailt@il.ibm.com, or join our Slack channel: https://aip360.mybluemix.net/community.

Related toolkits:

ai-minimization-toolkit - has been migrated into this toolkit.

differential-privacy-library: A general-purpose library for experimenting with, investigating and developing applications in, differential privacy.

adversarial-robustness-toolbox: A Python library for Machine Learning Security. Includes an attack module called inference that contains privacy attacks on ML models (membership inference, attribute inference, model inversion and database reconstruction) as well as a privacy metrics module that contains membership leakage metrics for ML models.