Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
|
4 years ago | |
---|---|---|
.gitee | 5 years ago | |
.github | 5 years ago | |
docs | 4 years ago | |
examples | 4 years ago | |
mindarmour | 4 years ago | |
tests | 4 years ago | |
.gitignore | 4 years ago | |
LICENSE | 5 years ago | |
NOTICE | 5 years ago | |
README.md | 4 years ago | |
README_CN.md | 4 years ago | |
RELEASE.md | 4 years ago | |
package.sh | 5 years ago | |
requirements.txt | 4 years ago | |
setup.py | 4 years ago |
MindArmour focus on security and privacy of artificial intelligence. MindArmour can be used as a tool box for MindSpore users to enhance model security and trustworthiness and protect privacy data.
MindArmour contains three module: Adversarial Robustness Module, Fuzz Testing Module, Privacy Protection and Evaluation Module.
Adversarial robustness module is designed for evaluating the robustness of the model against adversarial examples,
and provides model enhancement methods to enhance the model's ability to resist the adversarial attack and improve the model's robustness.
This module includes four submodule: Adversarial Examples Generation, Adversarial Examples Detection, Model Defense and Evaluation.
The architecture is shown as follow:
Fuzz Testing module is a security test for AI models. We introduce neuron coverage gain as a guide to fuzz testing according to the characteristics of neural networks.
Fuzz testing is guided to generate samples in the direction of increasing neuron coverage rate, so that the input can activate more neurons and neuron values have a wider distribution range to fully test neural networks and explore different types of model output results and wrong behaviors.
The architecture is shown as follow:
Privacy Protection and Evaluation Module includes two modules: Differential Privacy Training Module and Privacy Leakage Evaluation Module.
Differential Privacy Training Module implements the differential privacy optimizer. Currently, SGD, Momentum and Adam are supported. They are differential privacy optimizers based on the Gaussian mechanism.
This mechanism supports both non-adaptive and adaptive policy. Rényi differential privacy (RDP) and Zero-Concentrated differential privacy(ZCDP) are provided to monitor differential privacy budgets.
The architecture is shown as follow:
Privacy Leakage Evaluation Module is used to assess the risk of a model revealing user privacy. The privacy data security of the deep learning model is evaluated by using membership inference method to infer whether the sample belongs to training dataset.
The architecture is shown as follow:
This library uses MindSpore to accelerate graph computations performed by many machine learning models. Therefore, installing MindSpore is a pre-requisite. All other dependencies are included in setup.py
.
git clone https://gitee.com/mindspore/mindarmour.git
$ cd mindarmour
$ python setup.py install
Pip
installationpip install mindarmour-{version}-cp37-cp37m-linux_{arch}.whl
No module named 'mindarmour'
when execute the following command:python -c 'import mindarmour'
Guidance on installation, tutorials, API, see our User Documentation.
Welcome contributions. See our Contributor Wiki for more details.
The release notes, see our RELEASE.
MindArmour关注AI的安全和隐私问题。致力于增强模型的安全可信、保护用户的数据隐私。主要包含3个模块:对抗样本鲁棒性模块、Fuzz Testing模块、隐私保护与评估模块。 对抗样本鲁棒性模块 对抗样本鲁棒性模块用于评估模型对于对抗样本的鲁棒性,并提供模型增强方法用于增强模型抗对抗样本攻击的能力,提升模型鲁棒性。对抗样本鲁棒性模块包含了4个子模块:对抗样本的生成、对抗样本的检测、模型防御、攻防评估。
Python Markdown Text other