«THE BULLETIN OF IRKUTSK STATE UNIVERSITY». SERIES «MATHEMATICS»
«IZVESTIYA IRKUTSKOGO GOSUDARSTVENNOGO UNIVERSITETA». SERIYA «MATEMATIKA»
ISSN 1997-7670 (Print)
ISSN 2541-8785 (Online)

List of issues > Series «Mathematics». 2023. Vol 43

On the Properties of Bias-Variance Decomposition for kNN Regression

Author(s)
Victor M. Nedel’ko1

1Sobolev Institute of Mathematics SB RAS, Novosibirsk, Russian Federation

Abstract
When choosing the optimal complexity of the method for constructing decision functions, an important tool is the decomposition of the quality criterion into bias and variance.

It is generally assumed (and in practice this is most often true) that with increasing complexity of the method, the bias component monotonically decreases, and the variance component increases. The conducted research shows that in some cases this behavior is violated.

In this paper, we obtain an expression for the variance component for the kNN method for the linear regression problem in the formulation when the “explanatory” features are random variables. In contrast to the well-known result obtained for non-random “explanatory” variables, in the considered case, the variance may increase with the growth of 𝑘.

About the Authors
Victor M. Nedel’ko, Cand. Sci. (Phys.Math.), Assoc. Prof., Senior Researcher, Sobolev Institute of Mathematics SB RAS, Novosibirsk, 630090, Russian Federation, nedelko@math.nsc.ru
For citation
Nedel’ko V. M. On the Properties of Bias-Variance Decomposition for kNN Regression. The Bulletin of Irkutsk State University. Series Mathematics, 2023, vol. 43, pp. 110–121. https://doi.org/10.26516/1997-7670.2023.43.110
Keywords
bias-variance decomposition, machine learning, 𝑘-nearest neighbors algorithm, overfitting
UDC
519.246
MSC
68T10, 62H30
DOI
https://doi.org/10.26516/1997-7670.2023.43.110
References
  1. d’Ascoli S., Refinetti M., Biroli G., and Krzakala F. Double trouble in double descent: bias and variance(s) in the lazy regime. In Proceedings of the 37th International Conference on Machine Learning (ICML’20). JMLR.org, 2020, article 213, pp. 2280–2290.
  2. Belkin M., Hsu D., Ma S., Mandal S. Reconciling modern machine learning practice and the classical bias-variance trade-off. Proceedings of the National Academy of Sciences, 2019, vol. 116, no. 32, pp. 15849–1585. https://doi.org/10.1073/pnas.1903070116
  3. Berikov V. Semi-supervised classification using multiple clustering and low-rank matrix operations. Lecture Notes in Computer Science, 2019, vol. 11548, pp 529–540. https://doi.org/10.1007/978-3-030-22629-9_37
  4. Hastie T., Tibshirani R., Friedman H., Jerome. The Elements of Statistical Learning. 2009.
  5. Heskes T. Bias/variance decompositions for likelihood-based estimators. Neural Computation, 1998, vol. 10, no. 6, pp. 1425–1433. http://doi.org/10.1162/089976698300017232
  6. Kanevskiy D., Vorontsov K. Cooperative coevolutionary ensemble learning. Multiple Classifier Systems, 7th International Workshop, MCS 2007, Prague, Czech Republic, May, 23–25, 2007, pp. 469–478. http://dx.doi.org/10.1007/978-3-540-72523-7_47
  7. Kotsiantis S. Bagging and boosting variants for handling classifications problems: A survey. The Knowledge Engineering Review, 2014, vol. 29, no. 1, pp. 78–100. http://dx.doi.org/10.1017/S0269888913000313
  8. Lbov G., Startseva N. On some a concept of complexity of a strategy of nature in pattern recognition Data Analysis in Expert Systems.Computational systems, 1986, vol. 117, pp. 91–102. (in Russian).
  9. Lbov G.S. , Startseva N.G. Complexity of distributions in classification problems. Russ. Acad. Sci., Dokl. Math., 1994, vol. 50, no. 2.
  10. Lbov G., Startseva N. Logical Decision Functions and the Problem of Statistical Robustness of Solutions Novosibirsk: Institute of Mathematics SB RAS, 1999. (in Russian).
  11. Nakkiran P., Kaplun G., Bansal Y., Yang T., Barak B., Sutskever I. Deep double descent: where bigger models and more data hurt. Journal of Statistical Mechanics: Theory and Experiment, 2021.
  12. Neal B., Mittal S., Baratin A., Tantia V., Scicluna M., Lacoste-Julien S., Mitliagkas I. A Modern Take on the Bias-Variance Tradeoff in Neural Networks. 2018. https://doi.org/10.48550/arXiv.1810.08591
  13. Nedel’ko V. Some aspects of estimating a quality of decision functions construction methods. Tomsk state university. Journal of control and computer science, 2013, vol. 3, no. 24, pp. 123–132. (in Russian)
  14. Nedel’ko V. Statistical fitting criterion on the basis of cross-validation estimation. Pattern Recognition and Image Analysis, 2018, vol. 28, pp. 510–515. http://dx.doi.org/10.1134/S1054661818030148
  15. Nedel’ko V. On decompositions of decision function quality measure. Bulletin of Irkutsk State University. Series Mathematics, 2020, vol. 33, pp. 64–79. http://dx.doi.org/10.26516/1997-7670.2020.33.64
  16. Nedel’ko V. Tight risk bounds for histogram classifier. Proceedings of IFOST-2016 11th International Forum on Strategic Technology, 2016, pp. 267–271.
  17. Rachakonda A.R., Bhatnagar A. Aratio: Extending area under the roc curve for probabilistic labels. Pattern Recognition Letters, 2021, vol. 150, pp. 265–271. http://dx.doi.org/https://doi.org/10.1016/j.patrec.2021.06.023
  18. Rudakov K. Mathematical Foundations for Processing High Data Volume. Machine Learning and Artificial Intelligence, Pattern Recognit. Image Anal, 2019, vol. 29, pp. 339–343. http://dx.doi.org/10.1134/S1054661819030192
  19. Stuart G., Bienenstock E., Doursat R. Neural networks and the bias/variance dilemma. Neural Computation, 1992, vol. 4, no. 1, pp. 1–58. http://dx.doi.org/10.1162/neco.1992.4.1.1
  20. Yang Z., Yu Y., You C., Steinhardt J., Ma Y. Rethinking bias-variance trade-off for generalization of neural networks. International Conference on Machine Learning, 2020, pp. 10767–10777.

Full text (english)