Applied Predictive ModelingApplied Predictive Modeling covers the overall predictive modeling process, beginning with the crucial steps of data preprocessing, data splitting and foundations of model tuning. The text then provides intuitive explanations of numerous common and modern regression and classification techniques, always with an emphasis on illustrating and solving real data problems. The text illustrates all parts of the modeling process through many hands-on, real-life examples, and every chapter contains extensive R code for each step of the process. This multi-purpose text can be used as an introduction to predictive models and the overall modeling process, a practitioner’s reference handbook, or as a text for advanced undergraduate or graduate level predictive modeling courses. To that end, each chapter contains problem sets to help solidify the covered concepts and uses data available in the book’s R package.This text is intended for a broad audience as both an introduction to predictive models as well as a guide to applying them. Non-mathematical readers will appreciate the intuitive explanations of the techniques while an emphasis on problem-solving with real data across a wide variety of applications will aid practitioners who wish to extend their expertise. Readers should have knowledge of basic statistical ideas, such as correlation and linear regression analysis. While the text is biased against complex equations, a mathematical background is needed for advanced topics. |
Contents
1 | |
17 | |
Part II Regression Models | 93 |
Part III Classification Models | 244 |
Part IV Other Considerations | 461 |
Appendix | 546 |
Indicies | 569 |
Other editions - View all
Common terms and phrases
accuracy algorithm approach bagging boosting calculated caret package categorical predictors cell chapter chief investigators class probabilities classification models coefficients components computational confusion matrix correlation cost created cross-validation Cubist data frame data points data set distribution dummy variables error rate estimates evaluated example feature selection function grant data iterations large number linear models linear regression logistic regression Machine Learning matrix method metric model performance model tree neural networks nonlinear number of predictors number of samples NumCI optimal outcome over-fitting p-value partial least squares plot pre-processing predicted values predictive models random forest random forest model regression model regression trees relationship resampling response RMSE ROC curve rules scale scores Sect sensitivity shown shows solubility data specificity split Springer Science+Business Media statistic subset successful grants support vector machine SurfaceArea1 techniques test set training data training set trControl tuning parameters unsuccessful variable importance variance