«THE BULLETIN OF IRKUTSK STATE UNIVERSITY». SERIES «MATHEMATICS»
«IZVESTIYA IRKUTSKOGO GOSUDARSTVENNOGO UNIVERSITETA». SERIYA «MATEMATIKA»
ISSN 1997-7670 (Print)
ISSN 2541-8785 (Online)

List of issues > Series «Mathematics». 2020. Vol. 33

On Decompositions of Decision Function Quality Measure

Author(s)
V. M. Nedel’ko
Abstract

A comparative analysis of two approaches to the decomposition of quality criterion of decision functions is carried out.

The first approach is the bias-variance decomposition. This is the most well-known decomposition that is used in analyzing the quality of decision function construction methods, in particular for justifying some ensemble methods. This usually assumes a monotonous dependence of the bias and variance on the complexity. Recent studies show that this is not always true.

The second approach (G.S. Lbov, N.G. Startseva, 1989) is a decomposition into a measure of adequacy and a measure of statistical stability (robustness). The idea of the approach is to decompose the prediction error into approximation error and statistical error.

In this paper we propose a method of statistical estimation of the components of both decompositions on real data. We compare the dependencies of these components on the complexity of the decision function. Non-normalized margin is used as a general measure of complexity.

The results of the study and the experiments on UCI data show significant qualitative similarities in behavior of the bias and the adequacy measure and between the variance and the statistical stability measure. At the same time, there is a fundamental difference between the considered decompositions, in particular, with increasing complexity, the measure of adequacy cannot increase, while the bias first decreases, but at high enough values of complexity usually starts to grow.

About the Authors

Victor Nedel’ko, Cand. Sci. (Phys.–Math.), Sobolev Institute of Mathematics, 4, Koptjug av., 630090, Novosibirsk, Russian Federation, tel.: (383)3332793, email: nedelko@math.nsc.ru

For citation

Nedel’ko V.M. On Decompositions of Decision Function Quality Measur. The Bulletin of Irkutsk State University. Series Mathematics, 2020, vol. 33, pp. 64-79. https://doi.org/10.26516/1997-7670.2020.33.64

Keywords
machine learning, bias-variance decomposition, decision function complexity
UDC
519.246
MSC
68T10, 62H30
DOI
https://doi.org/10.26516/1997-7670.2020.33.64
References
  1. Baliuk A.S. Complexity Lower Bound for Boolean Functions in the Class of Extended Operator Forms. The Bulletin of Irkutsk State University. Series Mathematics, 2019, vol. 30, pp. 125-140. https://doi.org/10.26516/1997-7670.2019.30.125
  2. Berikov V. Semi-Supervised Classification Using Multiple Clustering And Low Rank Matrix Operations. Lecture Notes in Computer Science, 2019, vol. 11548, pp. 529-540.
  3. Domingos P. A Unified Bias-Variance Decomposition and its Applications. Proceedings of the Seventeenth International Conference on Machine Learning, 2000, Stanford, CA: Morgan Kaufmann, pp. 231-238.
  4. Dyakonov A.G. Bias and variance. https://dyakonov.org/page/2/, 2016. (In Russian)
  5. Heskes T. Bias/Variance Decompositions for Likelihood-Based Estimators. Neural Computation, 1998, vol. 10, no. 6, pp. 1425–1433.
  6. Kanevskiy D., Vorontsov K. Cooperative Coevolutionary Ensemble Learning. Multiple Classifier Systems, 7th International Workshop, MCS 2007, Prague, Czech Republic, May 23–25. 2007. Proceedings, pp. 469-478. https://doi.org/10.1007/978-3-540-72523-7_47.
  7. Khachay M., Khachay D. Attainable accuracy guarantee for the k-medians clustering in [0, 1]. Optimization Letters, 2019, vol. 13, no. 8, pp. 1837-1853. https://doi.org/10.1007/s11590-018-1305-3
  8. Kotsiantis S. Bagging and boosting variants for handling classifications problems: A survey. The Knowledge Engineering Review, 2014, no. 29(1), pp. 78-100. https://doi.org/10.1017/S0269888913000313
  9. Kuncheva L.I., Skurichina M., Duin R.P.W. An experimental study on diversity for bagging and boosting with linear classifiers. Information Fusion, 2002, no. 3, pp. 245-258.
  10. Kuncheva L., Vetrov D. Evaluation of stability of k-means cluster ensembles with respect to random initialization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2006, vol. 28, no. 11, pp. 1798-1808.
  11. Lbov G.S., Startseva N.G. On Some a Concept of Complexity of a Strategy of Nature in Pattern Recognition. Data Analysis in Expert Systems. Computational systems, Novosibirsk, 1986, issue 117, pp. 91-102. (In Russian).
  12. Lbov G.S., Startseva N.G. Complexity of Distributions in the Classification Problem. Doklady RAS, 1994, vol. 338, no. 5, pp. 592-594. (In Russian).
  13. Lbov G.S., Startseva N.G. Logicheskie reshajushhie funkcii i voprosy statisticheskoj ustojchivosti reshenij [Logical Decision Functions and the Problem of Statistical Robustness of Solutions]. Novosibirsk, Institute of Mathematics Publ., 211 p. (In Russian).
  14. Nedel’ko V.M. Some aspects of estimating a quality of decision functions construction methods. Tomsk State University Journal of Control and Computer Science, 2013, no. 3 (24), pp. 123-132. (In Russian).
  15. Nedel’ko V.M. On performance of boosting in classification problem. Vestn. Novosib. Gos. Univ., Ser. Mat. Mekh. Inform, 2015, vol. 15, no. 2, pp. 72-89. (In Russian).
  16. Nedel’ko V.M. Tight risk bounds for histogram classifier. Proceedings of IFOST2016 11th International Forum on Strategic Technology, 2016, pp. 267-271.
  17. Nedel’ko V.M. On the Maximization of Quadratic Weighted Kappa. The Bulletin of Irkutsk State University. Series Mathematics, 2018, vol. 23, pp. 36-45. https://doi.org/10.26516/1997-7670.2018.23.36 (In Russian).
  18. Nedel’ko V.M. Statistical Fitting Criterion on the Basis of CrossValidation Estimation. Pattern Recognition and Image Analysis, ISSN 1054-6618, Pleiades Publishing, Ltd., 2018, vol. 28, no. 3, pp. 510-515. https://doi.org/10.1134/S1054661818030148
  19. Rudakov K.V. Mathematical Foundations for Processing High Data Volume, Machine Learning, and Artificial Intelligence. Pattern Recognit. Image Anal, 2019, vol. 29, pp. 339-343. https://doi.org/10.1134/S1054661819030192
  20. Stuart G., Bienenstock E., Doursat R. Neural networks and the bias/variance dilemma. Neural Computation, 1992, vol. 4. https://doi.org/10.1162/neco.1992.4.1.1
  21. Zhuravlev Y.I., Ryazanov V.V., Aslanyan L.H. et al. On a Classification Method for a Large Number of Classes. Pattern Recognit. Image Anal, 2019, vol. 29, pp. 366- 376. https://doi.org/10.1134/S1054661819030246

Full text (english)