380 likes | 510 Vues
In this comprehensive guide on Multiple Regression Analysis (MRA), we explore key concepts including regression coefficients, semi-partial correlations, and diagnostics for identifying data issues. Learn about the assumptions behind MRA, how to detect multicollinearity, handle outliers, and ensure homoscedasticity. We also cover effective entry methods, the interpretation of standardized and unstandardized coefficients, and the importance of model building techniques. Enhance your understanding and application of multiple regression through practical examples and insightful strategies.
E N D
Multiple Regression Analysis: Part 2 Interpretation and Diagnostics
Learning Objectives • Understand regression coefficients and semi-partial correlations • Learn to use diagnostics to locate problems with data (relative to MRA) • Understand… • Assumptions • Robustness • Methods of dealing with violations • Enhance our interpretation of equations • Understand entry methods
Statistical Tests & Interpretation • Interpretation of regression coefficients • Standardized • Unstandardized • Intercept • Testing regression coefficients • t-statistic & interpretation • Testing R2
Output for MRA Run (coefficients) R2 = .558
Y A B X1 X2 Variance in Y Accounted for by two uncorrelated Predictors (A+B)/Y = R2, E (in Y circle) equals Error. Y E E B A X1 X2 Example #1: Small R2, A represents variance in Y accounted for by X1, B = variance in Y accounted for by X2. Example #2: Larger R2, A represents variance in Y accounted for by X1, B = variance in Y accounted for by X2.
Y Y A B A B C C X1 X2 D X1 X2 D Example #1: Small R2 Example #2: Larger R2 Variance in Y Accounted for by two correlated Predictors: sr2 and pr2 sr2 for X1 = pr2 for X1 =
A shortcoming to breaking down sr2 R2 = .120
Multicollinearity: One way it can all go bad! Y E A B C X1 X2 D
Ways to fix multicollinearity • Discarding Predictors • Combining Predictors • Using Principal Components • Parcelling • Ridge Regression
Outliers and Influential Observations:Another way it can all go bad! • Outliers on y • Outliers on x’s • Influential data points
Outliers • Outliers on y • Standardized Residuals • Studentized Residuals (df = N – k – 1) • Deleted Studentized Residuals • Outliers on x’s • Hat elements • Mahalanobis Distance
Outliers on y tcrit(21) = 2.08
Outliers on Xs (Leverage) χ2(crit) for Mahalanobis’ Distance = 7.82
Influential Observations • Cook’s Distance (cutoff ≈ 1.0) • DFFITs [cut-offs of 2 or 2*((k+1)/n)0.5] • DFBeta • Standardized DF Beta
Once more, with feeling R2 = .687
A cautionary tale:Some more ways it can all go bad! We will use X to predict y1, y2 and y3 in turn.
Homoscadasticity:Yet another way it can all go bad! • What is homoscedasticity? • Is it better to have heteroscedasticity? • The effects of violation • How to identify it • Strategies for dealing with it
Effect Size Multiple Correlation (R): SMC (R2):
Cross Validation • Why • Useful statistics and techniques • Conditions under which likelihood of cross-validation is increased
Assumptions of Regression • Sample Size • Absence of Outliers & Influential Observations • Absence of Multicollinearity and Singularity • Normality • Linearity • Homoscedasticity of Errors • Independence of Errors
Structure Coefficients • What are they? • Vs. pattern coefficients or “weights” • Why we may need both • When they would be used in MRA • Why they are not commonly used • How you get them in SPSS • CD sales example
Model Building in MRA:“Canned” procedures • Enter • Forward • Backward Selection (Deletion) • Stepwise • Hierarchical
Hierarchical – Example Predict employee satisfaction • Block 1: “Hygiene Factor” • Block 2: “Equity” • Block 3: “Organizational Commitment”
Interpretation revisited • In light of multicollinearity • Standardized or unstandardized? • Suppressor effects • Missing predictors • Correlated / uncorrelated predictors • Structure coefficients • Reliability of indicators • Mathematical maximization nature of MRA