Please Tell Me More: Privacy Impact of
Explainability through the Lens of Membership Inference
Attack |
2024 |
SP |
Feature-based |
Membership Inference |
Differential Privacy, Privacy-Preserving Models,
DP-SGD |
- |
Towards a Game-theoretic Understanding of
Explanation-based Membership Inference Attacks |
2024 |
arXiv |
Feature-based |
Membership Inference |
Game Theory |
- |
On the Privacy Risks of Algorithmic
Recourse |
2023 |
AISTATS |
Counterfactual |
Membership Inference |
Differential Privacy |
- |
The Privacy Issue of Counterfactual
Explanations: Explanation Linkage Attacks |
2023 |
TIST |
Counterfactual |
Linkage |
Anonymisaion |
- |
Feature-based Learning for Diverse and
Privacy-Preserving Counterfactual Explanations |
2023 |
KDD |
Counterfactual |
- |
Perturbation |
[Code] |
Private Graph Extraction via Feature
Explanations |
2023 |
PETS |
Feature-based |
Graph Extraction |
Perturbation |
[Code]
|
Privacy-Preserving Algorithmic Recourse
|
2023 |
ICAIF |
Counterfactual |
- |
Differential Privacy |
- |
Accurate, Explainable, and Private Models:
Providing Recourse While Minimizing Training Data
Leakage |
2023 |
ICML-Workshop |
Counterfactual |
Membership Inference |
Differential Privacy |
- |
Probabilistic Dataset Reconstruction from
Interpretable Models |
2023 |
arXiv |
Interpretable Surrogates |
Data Reconstruction |
- |
[Code]
|
DeepFixCX: Explainable privacy-preserving
image compression for medical image analysis |
2023 |
WIREs-DMKD |
Case-based |
Identity recognition |
Anonymisation |
[Code] |
XorSHAP: Privacy-Preserving Explainable AI
for Decision Tree Models |
2023 |
Preprint |
Shapley |
- |
Multi-party Computation |
- |
DP-XAI |
2023 |
Github |
ALE plot |
- |
Differential Privacy |
[Code]
|
Inferring Sensitive Attributes from Model
Explanations |
2022 |
CIKM |
Gradient-based, Perturbation-based |
Attribute Inference |
- |
[Code]
|
Model explanations with differential
privacy |
2022 |
FAccT |
Feature-based |
- |
Differential Privacy |
- |
DualCF: Efficient Model Extraction Attack
from Counterfactual Explanations |
2022 |
FAccT |
Counterfactual |
Model Extraction |
- |
- |
Feature Inference Attack on Shapley
Values |
2022 |
CCS |
Shapley |
Attribute/Feature Inference |
Low-dimensional |
- |
Evaluating the privacy exposure of
interpretable global explainers, Privacy Risk of Global Explainers |
2022 |
CogMI |
Interpretable Surrogates |
Membership Inference |
- |
- |
Privacy-Preserving Case-Based Explanations:
Enabling Visual Interpretability by Protecting Privacy
|
2022 |
IEEE Access |
Example-based |
- |
Anonymisation |
- |
On the
amplification of security and privacy risks by post-hoc
explanations in machine learning models |
2022 |
arXiv |
Feature-based |
Membership Inference |
- |
- |
Differentially Private Counterfactuals via
Functional Mechanism |
2022 |
arXiv |
Counterfactual |
- |
Differential Privacy |
- |
Differentially Private Shapley Values for
Data Evaluation |
2022 |
arXiv |
Shapley |
- |
Differential Privacy |
[Code]
|
Exploiting Explanations for Model Inversion
Attacks |
2021 |
ICCV |
Gradient-based, Interpretable Surrogates |
Model Inversion |
- |
- |
On the Privacy Risks of Model
Explanations |
2021 |
AIES |
Feature-based, Shapley, Counterfactual |
Membership Inference |
- |
- |
Adversarial XAI Methods in Cybersecurity
|
2021 |
TIFS |
Counterfactual |
Membership Inference |
- |
- |
MEGEX:
Data-Free Model Extraction Attack against Gradient-Based
Explainable AI |
2021 |
arXiv |
Gradient-based |
Model Extraction |
- |
[Code]
|
Robust Counterfactual Explanations for
Privacy-Preserving SVM, Robust Explanations for Private Support
Vector Machines |
2021 |
ICML-Workshop |
Counterfactual |
- |
Private SVM |
[Code]
|
When
Differential Privacy Meets Interpretability: A Case
Study |
2021 |
RCV-CVPR |
Interpretable Models |
- |
Differential Privacy |
- |
Differentially Private Quantiles |
2021 |
ICML |
Quantiles |
- |
Differential Privacy |
[Code]
|
FOX: Fooling with Explanations : Privacy
Protection with Adversarial Reactions in Social Media
|
2021 |
PST |
- |
Attribute Inference |
Privacy-Protecting Explanation |
- |
Privacy-preserving generative adversarial
network for case-based explainability in medical image
analysis |
2021 |
IEEE Access |
Example-based |
- |
Generative Anonymisation |
- |
Interpretable and Differentially Private
Predictions |
2020 |
AAAI |
Locally linear maps |
- |
Differential Privacy |
[Code] |
Model
extraction from counterfactual explanations |
2020 |
arXiv |
Counterfactual |
Model Extraction |
- |
[Code] |
Model Reconstruction from Model
Explanations |
2019 |
FAT* |
Gradient-based |
Model Reconstruction, Model Extraction |
- |
- |
Interpret Federated Learning with Shapley
Values |
2019 |
__ |
Shapley |
- |
Federated |
[Code]
|
Collaborative Explanation of Deep Models with
Limited Interaction for Trade Secret and Privacy
Preservation |
2019 |
WWW |
Feature-based |
- |
Collaborative rule-based model |
- |
Model inversion attacks that exploit
confidence information and basic countermeasures |
2015 |
CCS |
Confidence scores |
Reconstruction, Model Inversion |
- |
- |