Prediction Models in Machine Learning by Classification and Regression

Authors

  • Shamy Singh Research Scholar, Noorul Islam University, Tamil Nadu, India
  • J. Dheeba Associate Professor, CSE, Noorul Islam University, Tamil Nadu, India

DOI:

https://doi.org/10.53555/nncse.v2i5.408

Keywords:

Classification and regression tree algorithm, cross-validation, discriminant, linear model, prediction accuracy, recursive partitioning, selection bias, unbiased, QUEST, GUID, CRUISE, C4.5, RPART

Abstract

One of the machine-learning method for constructing prediction models from data is Classification and Regression. By partitioning the data space recursively these models are configuring and in each prediction model are fitting with a simple predictions. Finally, the partitioning can be represented pictorially as a decision tree. Finite number of unordered values are taken for the designing the classification trees and are designed for independent variables. And the prediction error are measured in terms of misclassification cost. Squared difference between the predicted and observed values are measured in regression trees, which are dependent variables that have ordered discrete values or continuous values. Here in this article reviewing and comparing some of the widely acceptable algorithms such as QUEST, GUIDE, CRUISE, C4.5 and RPART with their strengths, weakness and capabilities.

References

Fielding A, O’Muircheartaigh CA. Binary segmentation in survey analysis with particular reference to AID. The Statistician 1977, 25:17–28.

Messenger R, Mandell L. A modal search technique for predictive nominal scale multivariate analysis. J Am Stat Assoc 1972, 67:768–772.

R Development Core Team, R: a language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria, 2009.

Therneau TM, Atkinson B. RPART: recursive partitioning. R port by B. Ripley. R package version 3.1-41, 2008.

Loh WY, Vanichsetakul N. Tree-structured classification via generalized discriminant analysis (with discussion).

Kim H, Loh WY. Classification trees with unbiased multiway splits. J Am Stat Assoc 2001, 96:589–604.

Kim H, Loh WY. Classification trees with bivariate linear discriminant node models. J Comput Graphical Stat 2003, 12:512–530.

Loh WY, Chen C, Hordle W, Unwin A, eds. Improving the precision of classification trees. Ann Appl Stat 2009, 3:1710–1737.

Loh WY, Shih Y. Split selection methods for classification trees. Stat Sin 1997, 7:815–840.

Hothorn T, Hornik K, Zeileis A. Unbiased recursive partitioning: a conditional inference framework. J Comput Graphical Stat 2006, 15:651–674.

Kass GV. An exploratory technique for investigating large quantities of categorical data. Appl Stat 1980, 29:119–127.

Breiman L. Bagging predictors. Machine Learning 1996, 24:123–140.

Breiman L. Random forests. Machine Learning 2001, 45:5–32.

E. Alpaydin. Introduction to Machine Learning. 2nd ed. Boston: MIT Press; 2010.

J Am Stat Assoc 1988, 83:715–728.

Downloads

Published

2015-05-31

How to Cite

Singh, S., & Dheeba, . J. (2015). Prediction Models in Machine Learning by Classification and Regression. Journal of Advance Research in Computer Science & Engineering (ISSN 2456-3552), 2(5), 10-14. https://doi.org/10.53555/nncse.v2i5.408