Regression Methods Linear Regression Simple linear regression one
- Slides: 13
Regression Methods
Linear Regression § Simple linear regression (one predictor) § Multiple linear regression (multiple predictors) § Ordinary Least Squares estimation § Lasso regression § selects features by setting parameters to 0
Coefficient of Determination § Indicates how well a model fits the data § R 2 (R squared) § R 2 = 1−SSres/SStot § SSres = Σ(yi−fi)2 § SStot = Σ(yi−y)2 § between 0 and 1, if least squares model. Bigger range if other models are used § Explained variance § what percentage of the variance is explained by the model § linear least squares regression: R 2 = r 2
R Squared Source Wikipedia CC BY-SA 3. 0 § visual interpretation of R 2
Regression Trees § Regression variant of decision tree § Top-down induction § 2 options: § Constant value in leaf (piecewise constant) § Local linear model in leaf (piecewise linear)
M 5 algorithm (Quinlan, Wang) § M 5’, M 5 P in Weka (classifiers > trees > M 5 P) § Offers both regression trees and model trees § Model trees are default § -R option (build. Regression. Tree) for piecewise constant
M 5 algorithm (Quinlan, Wang) § Splitting criterion: Standard Deviation Reduction § SDR = sd(T) – Σ sd(Ti) |Ti|/|T| § Stopping criterion: § Standard deviation below some threshold (0. 05 sd(D)) § Too few examples in node (e. g. 4) § Pruning (bottom-up): § Estimate error: (n+v)/(n−v)×absolute error in node § n is examples in node, v is parameters in the model
Binary Splits § All splits are binary § Numeric as normal § Nominal: order all values according to average (prior to induction) § Introduce k-1 indicator variables in this order Example: database of skiing slopes avg(color = green) = 2. 5% avg(color = blue) = 3. 2% avg(color = red) = 7. 7% avg(color = black) = 13. 5% binary features: Green, Green. Blue. Red,
Regression tree on Servo dataset (UCI)
Model tree on Servo dataset (UCI) LM 1: 0. 0833 * motor=B, A + 0. 0682 * screw=B, A + 0. 2215 * screw=A + 0. 1315 * pgain=4, 3 + 0. 3163 * pgain=3 − 0. 1254 * vgain=1, 2 + 0. 3864
Regression in Cortana § Regression a natural setting in Subgroup Discovery § Local models, no prediction model § Subgroups are piecewise constant subsets h = 3600 h = 3100 h = 2200
Subgroup Discover: regression
Other regression models § Functions § Linear. Regression § Multi. Layer. Perceptron (artificial neural network) § SMOreg (Support Vector Machine) § Lazy § IBK § Rules § M 5 Rule (decision list)
- Simple multiple linear regression
- Survival analysis vs logistic regression
- Logistic regression vs linear regression
- Regression linear model
- Law of iterated expectations
- Linear regression with one regressor
- Simple linear regression hypothesis example
- Regression analysis excel 2007
- Useless regression chapter 16
- Simple linear regression
- Linear regression function
- Multiple linear regression spss
- Inlay wax pattern fabrication
- One empire one god one emperor