In this talk we’ll show that under standard independence assumptions between classifier errors, this high dimensional data hides a simple low dimensional structure. In particular, we’ll present a novel spectral approach to address the above questions, and derive a novel unsupervised spectral meta-learner (SML). On both simulated and real data, SML typically achieves a higher accuracy than most classifiers in the ensemble and can provide a better starting point for iterative estimation of the maximum likelihood estimator than classical majority voting. Furthermore, SML is robust to the presence of small malicious groups of classifiers designed to veer the ensemble prediction away from the (unknown) ground truth.
Joint work with Fabio Parisi, Francesco Strino and Yuval Kluger (Yale).
Discovery Building, Orchard View Room
Boaz Nadler