ART 1.7.0
This release of ART v1.7.0 introduces many new evasion and inference attacks providing support for the evaluation of malware or tabular data classification, new query-efficient black-box (GeoDA) and strong white-box (Feature Adversaries) evaluation methods. Furthermore, this release introduces an easy to use estimator for Espresso ASR models to facilitate ASR research and connect Espresso and ART. This release also introduces support for binary classification with single outputs in neural networks classifiers and selected attacks. Many more new features and details can be found below:
Added
- Added LowProFool evasion attack for imperceptible attacks on tabular data classification in
art.attacks.evasion.LowProFool. (#1063) - Added Over-the-Air-Flickering attack in PyTorch for evasion on video classifiers in
art.attacks.evasion.OverTheAirFlickeringPyTorch. (#1077, #1102) - Added API for speech recognition estimators compatible with Imperceptible ASR attack in PyTorch. (#1052)
- Added Carlini&Wagner evasion attack with perturbations in L0-norm in
art.attacks.evasion.CarliniL0Method. (#844, #1109) - Added support for Deep Speech v3 in
PyTorchDeepSpeechestimator. (#1107) - Added support for TensorBoard collecting evolution of norms (L1, L2, and Linf) of loss gradients per batch, adversarial patch, and total loss and its model-specific components where available (e.g. PyTochFasterRCNN) in
AdversarialPatchPyTorch,AdversarialPatchTensorFlow,FastGradientMethod, and allProjectedGradientDescent*attacks. (#1071) - Added
MalwareGDTensorFlowattack for evasion on malware classification of portable executables supporting append based, section insertion, slack manipulation, and DOS header attacks. (#1015) - Added Geometric Decision-based Attack (GeoDA) in
art.attacks.evasion.GeoDAfor query-efficient black-box attacks on decision labels using DCT noise. (#1001) - Added Feature Adversaries framework-specific in PyTorch and TensorFlow v2 as efficient white-box attack generating adversarial examples imitating intermediate representations at multiple layers in
art.attacks.evasion.FeatureAdversaries*. (#1128, #1142, #1156) - Added attribute inference attack based on membership inference in
art.attacks.inference.AttributeInferenceMembership. (#1132) - Added support for binary classification with neural networks with a single output neuron in
FastGradientMethod, and allProjectedGradientDescent*attacks. Neural network binary classifiers with a single output require settingnb_classes=2and labelsyin shape (nb_samples, 1) or (nb_samples,) containing 0 or 1. Backward compatibility for binary classifiers with two outputs is guaranteed withnb_classes=2and labelsyone-hot-encoded in shape (nb_samples, 2). (#1118) - Added estimator for Espresso ASR models in
art.estimators.speech_recognition.PyTorchEspressowith support for attacks withFastGradientMethod,ProjectedGradientDescentandImperceptibleASRPyTorch. (#1036) - Added deprecation warnings for
art.classifiersandart.wrappersto be replace withart.estimators. (#1154)
Changed
- Changed
art.utils.load_iristo use Iris dataset fromsklearn.datasetsinstead ofarchive.ics.uci.edu. (#1097 ) - Changed
HopSkipJumpto check for NaN in the adversarial example candidates and return original (benign) sample if at least one NaN is detected. (#1124) - Changed
SquareAttackto accept user-defined loss and adversarial criterium definitions to enable black-box attacks on all machine learning tasks on images beyond classification. (#1127) - Changed
PyTorchFasterRCNN.loss_gradientsto process each sample separately to avoid issues with gradient propagation withtorch>=1.7. (#1138)
Removed
[None]