Abstract: The goal of the paper is to estimate misclassification probability for decision function by training
sample. Here are presented results of investigation an empirical risk bias for nearest neighbours, linear and
decision tree classifier in comparison with exact bias estimations for a discrete (multinomial) case. This allows to
find out how far Vapnik–Chervonenkis? risk estimations are off for considered decision function classes and to
choose optimal complexity parameters for constructed decision functions. Comparison of linear classifier and
decision trees capacities is also performed.
Keywords: pattern recognition, classification, statistical robustness, deciding functions, complexity, capacity,
overtraining problem.
ACM Classification Keywords:I.5.1 Pattern Recognition: Statistical Models
Link:
EVALUATING MISCLASSIFICATION PROBABILITY USING EMPIRICAL RISK1
Victor Nedel’ko
http://www.foibg.com/ijita/vol13/ijita13-3-p15.pdf