Prediction Models in Machine Learning by Classification and Regression
DOI:
https://doi.org/10.53555/nncse.v2i5.408Keywords:
Classification and regression tree algorithm, cross-validation, discriminant, linear model, prediction accuracy, recursive partitioning, selection bias, unbiased, QUEST, GUID, CRUISE, C4.5, RPARTAbstract
One of the machine-learning method for constructing prediction models from data is Classification and Regression. By partitioning the data space recursively these models are configuring and in each prediction model are fitting with a simple predictions. Finally, the partitioning can be represented pictorially as a decision tree. Finite number of unordered values are taken for the designing the classification trees and are designed for independent variables. And the prediction error are measured in terms of misclassification cost. Squared difference between the predicted and observed values are measured in regression trees, which are dependent variables that have ordered discrete values or continuous values. Here in this article reviewing and comparing some of the widely acceptable algorithms such as QUEST, GUIDE, CRUISE, C4.5 and RPART with their strengths, weakness and capabilities.
References
Fielding A, O’Muircheartaigh CA. Binary segmentation in survey analysis with particular reference to AID. The Statistician 1977, 25:17–28.
Messenger R, Mandell L. A modal search technique for predictive nominal scale multivariate analysis. J Am Stat Assoc 1972, 67:768–772.
R Development Core Team, R: a language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria, 2009.
Therneau TM, Atkinson B. RPART: recursive partitioning. R port by B. Ripley. R package version 3.1-41, 2008.
Loh WY, Vanichsetakul N. Tree-structured classification via generalized discriminant analysis (with discussion).
Kim H, Loh WY. Classification trees with unbiased multiway splits. J Am Stat Assoc 2001, 96:589–604.
Kim H, Loh WY. Classification trees with bivariate linear discriminant node models. J Comput Graphical Stat 2003, 12:512–530.
Loh WY, Chen C, Hordle W, Unwin A, eds. Improving the precision of classification trees. Ann Appl Stat 2009, 3:1710–1737.
Loh WY, Shih Y. Split selection methods for classification trees. Stat Sin 1997, 7:815–840.
Hothorn T, Hornik K, Zeileis A. Unbiased recursive partitioning: a conditional inference framework. J Comput Graphical Stat 2006, 15:651–674.
Kass GV. An exploratory technique for investigating large quantities of categorical data. Appl Stat 1980, 29:119–127.
Breiman L. Bagging predictors. Machine Learning 1996, 24:123–140.
Breiman L. Random forests. Machine Learning 2001, 45:5–32.
E. Alpaydin. Introduction to Machine Learning. 2nd ed. Boston: MIT Press; 2010.
J Am Stat Assoc 1988, 83:715–728.
Downloads
Published
Issue
Section
License

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
You are free to:
- Share — copy and redistribute the material in any medium or format for any purpose, even commercially.
- Adapt — remix, transform, and build upon the material for any purpose, even commercially.
- The licensor cannot revoke these freedoms as long as you follow the license terms.
Under the following terms:
- Attribution — You must give appropriate credit , provide a link to the license, and indicate if changes were made . You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use.
- No additional restrictions — You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.
Notices:
You do not have to comply with the license for elements of the material in the public domain or where your use is permitted by an applicable exception or limitation .
No warranties are given. The license may not give you all of the permissions necessary for your intended use. For example, other rights such as publicity, privacy, or moral rights may limit how you use the material.