## Document Type

Article

## Publication Date

2022

## Publication Source

Proceedings of the 2021 Undergraduate Mathematics Day

## Volume

7

## Inclusive pages

1-12

## Abstract

There are many types of statistical inferences that can be used today: Frequentist, Bayesian, Fiducial, and others. However, Vovk introduced a new version of statistical inference known as Conformal Predictions. Conformal Predictions were designed to reduce the assumptions of standard prediction methods. Instead of assuming all observations are drawn independently and identically distributed, we instead assume exchangeability. Meaning, all N! possible orderings of our N observations are equally likely. This is more applicable to fields such as machine learning where assumptions may not be easily satisfied. In the case of binary classification, Vovk provided the nearest neighbors (NN) measure which is a ratio of in-class versus out-of-class distance. Later on, Papodopolous introduced normalizing constants for NN for the regression case, we extend this work to the classification case. We provide an asymptotic guarantee which shows what is known empirically. The normalization of NN produces smaller confidence sets on average compared to standard NN. A small synthetic simulation is also presented to shown the viability in a non-asymptotic case.

## Keywords

Efficiency, conformal predictions, set prediction, asymptotic, confidence sets

## Disciplines

Mathematics

## eCommons Citation

Lovig, Maxwell, "Efficient Conformal Binary Classification under Nearest Neighbor" (2022). *Undergraduate Mathematics Day: Past Content*. 43.

https://ecommons.udayton.edu/mth_epumd/43

*Abstract only*

## Comments

Presented at University of Dayton Undergraduate Mathematics Day Nov. 6, 2021.