1 / 54

Comprehensive Exam Review

Comprehensive Exam Review. Click the LEFT mouse key ONCE to continue. Research and Program Evaluation Part 4. Click the LEFT mouse key ONCE to continue. Overview of Statistics.

Télécharger la présentation

Comprehensive Exam Review

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Comprehensive Exam Review Click the LEFT mouse key ONCE to continue

  2. Research and Program Evaluation Part 4 Click the LEFT mouse key ONCE to continue

  3. Overview of Statistics

  4. Clearly, it is not possible to present a comprehensive review of all statistics here. Therefore, what follows is a general overview of major principles of statistics. There are “technical” exceptions to (or variations of) most of what is presented. However, the information provided here is adequate for and applicable to most of the research in the counseling profession.

  5. Parametric Statistics Use of so-called parametric statistics is based on assumptions including that: the data represent population characteristics that are continuous and symmetrical. the variable(s) has a distribution that is essentially normal in the population. the sample statistic provides an estimate of the population parameter.

  6. Recall that variables typically involved in research can be divided into the categories of “discrete” or “continuous.” Based on this distinction, (in general) all statistical analyses can be divided into: Analyses of Relationships among variables or Analyses of Differences based on variables

  7. In the context of this general overview, all of the variables involved in analyses of relationships are continuous. Similarly, for analyses of differences,at least one variable must be continuous and at least one variable must be discrete.

  8. Analyses of Relationships

  9. The simplest (statistical) relationship involves only two (continuous) variables. In statistics, a “relationship” between two variables is known as a “correlation.” Calculation of the correlation coefficient allows us to address the question, “What do we know (or can we predict) about Y given that we know X (or vice versa)?”

  10. The correlation coefficient: is used to indicate the relationship between two variables. is known more formally as the Pearson Product-Moment Correlation Coefficient. is designated by a lower case r. ranges in values from -1.00 through 0.00 to +1.00.

  11. When r = -1.00, there is a “perfect” negative, or inverse, relationship between the two variables. This means that as one variable is changing, the associated variable is changing in the opposite direction in a proportional manner. When r = +1.00, there is a “perfect” positive, or direct, relationship between the two variables.

  12. This means that as one variable is changing, the associated variable is changing in the same direction in a proportional manner. When r = 0.00, there is a “zero-order” relationship between the two variables. This means that change in one variable is unrelated to change in the associated variable.

  13. The question to be confronted is... “How do we know if the correlation coefficient calculated is any good?” In general, there are two major ways to evaluate a correlation coefficient. One method is in regard to “statistical significance.” Statistical significance has to do with the probability (likelihood) that a result occurred strictly as a function of chance.

  14. Evaluation based in probability is like a game of chance. The researcher decides whether it will be a “high stakes” or a “low stakes” situation, de-pending on the implications of being wrong. The results of the decision are operationalized in the “alpha level” selected for the study.

  15. In the language of statistics, the “alpha level” (e.g., .01 or .05), sometimes called the level of significance, represents the (proportionate) chance that the researcher will be wrong in rejecting the null hypothesis. That is, the alpha level also is the probability of making a Type I Error.

  16. In the language of statistics, the “p value” is the (exact) probability of obtaining the particular result for some statistical analysis. Technically, the p value is compared to the alpha level to determine statistical significance; if p is less than the alpha, the result is “statistically significant.”

  17. Most computer programs generate p values (i.e., exact probabilities) from statistical analyses. However, most journal articles report results as comparisons of p values to alpha levels; that is, they report, for example, *p < .05, rather than, for example, *p = .0471.

  18. There is at least one prominent limitation in the evaluation of a correlation coefficient based on statistical significance. This limitation is related to the “conditions” under which the statistical significance of the correlation coefficient is evaluated. The “critical value” is the value of the correlation coefficient necessary for it to be statistically significant at a given alpha level and for a given sample size.

  19. In statistics, sample size is usually expressed in regard to “degrees of freedom.” For example, the degrees of freedom for a correlation coefficient is given by: df = N - 2. For the correlation coefficient, there is an inverse relationship between the critical values and degrees of freedom. As the degrees of freedom (i.e., sample size) increase, critical values (needed for statistical significance) decrease.

  20. This means that a very small correlation coefficient can be “statistically significant” if the data are from a very large sample. Correlation coefficients cannot be evaluated as “good” or “bad” in an absolute sense; consideration must be given to the sample size from which the data were derived.

  21. Consider two variables: A and B A All (i.e., 100%) of the variance of A can be represented by a circle. Another way to evaluate a correlation coefficient is in terms of “shared variance.” By definition, if A is a variable, it has variance (i.e., not every person receives the same score on measure A).

  22. Similarly, because B is a variable, it has variance, and all (i.e., 100%) of the variance of B can be represented by a circle. B

  23. Of interest is how much variance variables A and B “share.” % ? B A

  24. The “percentage of shared variance” is equal to: B A r2 x 100

  25. The term “r2” is known as the “coefficient of determination.” The percentage of shared variance is how much of the variance of variable A is common to variable B, and vice versa. Another way to think of it is that the percentage of shared variance is the amount of the same thing measured by (or reflected in) both variables.

  26. The good news is that the shared variance method as a basis for evaluating a correlation coefficient is not dependent upon sample size. The bad news is that there is no way to determine what is an acceptable level of shared variance. Ultimately, the research consumer has to be the judge of what is a “good” correlation coefficient….

  27. r Z A The Pearson Product-Moment Correlation coefficient can be used to “predict” one variable from another. That’s helpful, but has limited application because only two variables are involved.

  28. Suppose we know of the relationships between Z and each of several other variables. C r B D r r Z r r E A

  29. In multiple correlation, one variable is predicted from a (combined) set of other variables. The capital letter R is used to indicate the relationship between the set of variables and the variable being predicted. The variable being predicted is known as the “criterion” variable, and the variables in the set are known as the “predictor” variables.

  30. A Z D Z B Z E Z C Z In computing a multiple correlation coefficient, the most desirable situation is what is known as the “Daisy Pattern.” In the hypothetical Daisy Pattern, each predictorhas a relatively high correlation with the criterion variable...

  31. A E B C D and each of the predictor variables has a relatively low correlation with each of the other predictor variables.

  32. If achieved, a true Daisy Pattern would look something like this. B C Z A D E

  33. The multiple correlation computational procedures lead to a weighted combination of (some of) the predictor variables and a specific correlation between the weighted combination and the criterion variable. The same two methods used to evaluate a Pearson Product-Moment Correlation coefficient can be used to evaluate a multiple correlation coefficient.

  34. The methods of evaluating R include: statistical significance, although the sample size limitation concern is less problematic if the sample is sufficient for the computa-tions. percentage of shared variance, where the expression R2 x 100 represents the sum of the intersections of the predictors with the criterion variable.

  35. A canonical correlation (Rc) represents the relationship between a set of predictor variables and a set of criterion variables. A canonical correlation is usually expressed as a lambda coefficient, often Wilk’s Lambda, which is the result of the statistical computations.

  36. B A D X C Y E Z Graphically, a canonical correlation might look like this:

  37. The statistical significance of the lambda coefficient can be readily determined. The percentage of shared variance also can be calculated. However, because the lambda coefficient can have a value greater than one, the calculation of shared variance involves more than just squaring the lambda coefficient.

  38. Predictors Criterion Statistic Variables Variable Calculated 1 1 r >1 1 R >1 >1 Rc The following chart summarizes the nature of the three preceding analyses of relationships.

  39. Factor analysis, a special type of analysis of relationships among variables, is a general family of data reduction techniques. It is intended to reduce the redundancy in a set of correlated variables and to represent the variables with a smaller set of derived variables (aka factors). Factor analyses may be computed within either of two contexts: exploratory or confirmatory.

  40. Factor analysis starts with input of the raw data. Next, an intervariable correlation matrix is generated from the input data. Then, using sophisticated matrix algebra procedures, an initial factor (loading) matrix is derived from the correlation matrix.

  41. There are three major components to the factor loading matrix. The first is the set of item numbers, usually arranged in sequence and hierarchical order. The second is the factor identifications, usually represented by Roman numerals. The third is the factor loadings, usually provided as hundredths - with or without the decimal point.

  42. I II III IV V VI VII 1 2 3 4 5 6 7 8 9 .01 -.86 -.21 .09 .17 -.29 -.08 -.79 .03 .11 .20 -.15 -.17 -.13 -.19 .75 .00 -.46 .65 .02 .26 .11 -.66 .14 .09 -.02 .24 -.07 .91 -.27 -.06 .12 .19 -.22 -.03 -.14 .10 .20 .69 .11 -.09 .04 .16 -.08 .12 .28 -.22 .13 .06 -.16 -.29 .04 .05 -.13 .20 -.11 .66 .15 -.03 .22 .12 -.19 .23 The result might look something like this:

  43. An important question is, “How do we know how many factors to retain?” In factor analysis, potentially there can be as many factors as items. However, usually one or some combination of three methods is used to decide how many factors to retain.

  44. One common method is to retain factors having “eigenvalues” greater than one. Each factor has an eigenvalue, which is the sum of the squared factor loadings for the factor. Retaining factors having eigenvalues greater than one also is known as applying the Kaiser Criterion.

  45. A second common method is to apply the scree test. The scree test is a visual, intuitive method of determining how many factors to retain by examining the graph of the eigenvalues from the initial factor loading matrix.

  46. A third possible method is based on how much of the total variance is to be accounted for by the retained factors. The total possible variance is equal to the number of items. Therefore, the variance percentage for any factor is the eigenvalue divided by the total number of items, times 100. Factors can be retained by summing these percentages until the desired percentage is reached.

  47. Another important question in factor analysis is, “How are the relationships among the factors to be conceptualized?” A factor is two things: Conceptually, a factor is a representation of a construct. However, in regard to mathematics, a factor is a vector in n-dimensional space.

  48. Factors as constructs may be separate and entirely distinct from one another or separate but conceptually related to one another. Factors as vectors reflect these possibilities by being positioned in n-dimensional space as either perpendicular to one another or as having an acute angle between them. The initial factor loading matrix is “rotated” to achieve the best mathematical represen-tation and clarity among the constructs.

  49. If the factors are assumed to be distinct (i.e., independent) from one another, the rotation is said to be “orthogonal.” An orthogonal rotation is one in which the angles between factors are maintained as right angles during and after the rotation. The most common orthogonal rotation is the “Varimax” procedure.

  50. If the factors are assumed to be related (i.e., dependent) to one another, the rotation is said to be “oblique.” An oblique rotation is one in which the angles between factors are maintained as less than right angles during and after the rotation. The most common oblique rotation is the “Oblimin” procedure.

More Related