-
Notifications
You must be signed in to change notification settings - Fork 0
/
intro.tex
39 lines (34 loc) · 5.07 KB
/
intro.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
Machine learning models operate on the assumption, albeit incorrectly that they will be deployed on data distributed identically to what they were trained on.
The violation of this assumption is known as distribution shift and can often result in significant degradation of performance~\citep{cifar101,cifar10C,learnundercov, failloud, mindperfgap, trustuncert}.
There are several cases where a mismatch between training and deployment data results in very real consequences on human beings.
In healthcare, machine learning models have been deployed for predicting the likelihood of sepsis.
Yet, as~\citep{10.1001/jamainternmed.2021.3333} show, such models can be miscalibrated for large groups of individuals, directly affecting the quality of care they experience.
The deployment of classifiers in the criminal justice system~\citep{AIissend45:online}, hiring and recruitment pipelines~\citep{Amazonsc17:online} and self-driving cars~\citep{selfdrive} have all seen humans affected by the failures of learning models.
The need for methods that quickly detect, characterize and respond to distribution shift is, therefore, a fundamental problem in trustworthy machine learning.
For practitioners, regulatory agencies and individuals to have faith in deployed predictive models without the need for laborious manual audits, we need methods for the identification of distribution shift
that are \textit{sample-efficient} (identifying shifts from a small number of samples), \textit{informed} (identifying shifts relevant to the domain and learning algorithm), \textit{model-agnostic}
(identifying shifts regardless of the functional class of the predictive model) and \textit{statistically sound} (identifying true shifts while avoiding false positives with high-confidence).
In our work, we study a special case of distribution shift, commonly known as \textit{covariate shift}, which considers shifts only in the distribution of input data to a learning model $f: X\to Y$.
Considering a general factorization of the joint distribution of $X$ and $Y$ as $\Pc_{X,Y}=\Pc_X\times \Pc_{Y|X}$, covariate shift describes the set of distributions $\Qc_{X,Y}$ such that $\Pc_X \neq \Qc_X$ while the conditional distribution between the inputs and outputs remains fixed $\Pc_{Y|X} = \Qc_{Y|X}$\footnote{Shifts in the conditional distribution of $X|Y$ is commonly known as correlation or concept shift}.
In many standard machine learning settings we assume that $f$ is trained to learn $\Pc_{Y|X}$ using samples from $\Pc_{X,Y}$, but in deployment it observes only samples from $\Qc_X$.
In standard learning theory only relatively loose bounds can be made on the performance of $f$ when $\Pc_X\neq \Qc_X$ even in the absence of concept shift.
We build off recent progress in understanding model performance under covariate shift using the PQ-learning framework~\citep{pqlearn}, a framework for selective classifiers that may either predict on or abstain for a given sample.
PQ learning introduces a strong set of performance guarantees for such selective classifiers under arbitrary covariate shifts, we summarize the main results of PQ learning in \autoref{sec:selective-classification-and-pq-learning}.
Our work uses and extends PQ-learning to develop a practical, model-based hypothesis test, named \textit{the \method}, to identify potentially harmful covariate shifts given any existing classification model already in deployment.
\smallbreak\noindent
Our work makes the following key contributions:
\begin{itemize}
\item We show how to construct an ensemble of classifiers that maximize out-of-domain disagreement while behaving consistently in the training domain.
We propose the \emph{disagreement cross entropy} for models learned via continuous gradient-based methods (e.g., neural networks), as well as a generalization for those learned via discrete optimization (e.g., random forest).
\item We show that the rejection rate and the entropy of the learning ensemble can be used to define a model-aware hypothesis test for covariate shift, \textbf{the \method},\ that in idealized settings can provably detect covariate shift.
\item On high-dimensional image and tabular data, using both neural networks and gradient boosted decision trees, our method outperforms state-of-the-art techniques for detecting covariate shift, particularly when given access to as few as ten test examples.
\end{itemize}
\begin{figure}[!htb]
\centering
\includegraphics[width=\linewidth]{images/detectron.pdf}
\caption{\small \textbf{Detectron (PQ Learning for Covariate Shift Detection):} Starting with a base classifier trained on labeled samples from distribution $\Pc$ we
train new \textit{\textbf{C}onstrained \textbf{D}isagreement \textbf{C}lassifiers} (CDCs) on a small set of observed unlabeled samples from a new distribution $\Qc$.
CDCs aim to maximize classification disagreement on $\Qc$ while constrained to agree with the base classifier on $\Pc$. The rate $\phi$ that CDCs disagree,
as well as their entropy, on test data is a powerful and sample efficient statistic for identifying covariate shift $\Pc\neq \Qc$.}
\label{fig:detectron}
\end{figure}