Attribute Attack should report confidence that training set is not more vulnerable than test #166
Labels
enhancement
New feature or request
waiting
This issue is waiting for something else to be completed (see issue for details)
At moment we effectively run a worst-case attack where a simulated attacker has the model which outputs probabilities, and has a record with the target label and with just the value for one feature missing.
A `competent' published model may increase the likelihood that an attacker can estimate the missing value for a record more reliably than they could without the model.
So this uses is, is this risk different for items that were in the training set than it is for the general population?
We assess this risk separately for each attribute - assuming the TRE may set a different risk appetite for each.
Procedure:
Currently we report the ratio of the two fractions$ \frac { p_{tr} }}{p_{te}}$
We should report the probability that the observed differences of proportions is significant
-- some code examples in metrics.py for pdf, or description here
using norm from scipy.stats,
Then for report we have to decide whether to use 95% or 99% confidence
The text was updated successfully, but these errors were encountered: