Скачать книгу

In the language of sets, covxy must be a subset of the larger set represented by images.

      It is important to emphasize that a correlation of 0 does not necessarily represent the absence of a relationship. What it does represent is the absence of a linear one. Neither the covariance or Pearson's r capture nonlinear relationships, and so it is possible to have very strong relations in a sample or population yet still obtain very low values (even zero) for the covariance or Pearson r. Always plot your data to see what is going on before drawing any conclusions. Correlation coefficients should never be presented without an accompanying plot to characterize the form of the relationship.

      > cor(child, parent) [1] 0.4587624

      We can test it for statistical significance by using the cor.test function:

      > cor.test(child, parent) Pearson's product-moment correlation data: child and parent t = 15.7111, df = 926, p-value < 2.2e-16 alternative hypothesis: true correlation is not equal to 0 95 percent confidence interval: 0.4064067 0.5081153 sample estimates: cor 0.4587624

      We can see that observed t is statistically significant with a computed 95% confidence interval having limits 0.41 to 0.51, indicating that we can be 95% confident that the true parameter lies approximately between the limits of 0.41 and 0.51. Using the package ggplot2 (Wickham, 2009), we plot the relationship between parent and child (with a smoother):

      > library(ggplot2) > qplot(child, parent, data = Galton, geom = c("point", "smooth"))Graph depicts a plot using the package ggplot2 (Wickham, 2009).

      One drawback of such a simple plot is that the frequency of data points in the bivariate space cannot be known by inspection of the plot alone. Jittering is a technique that allows one to visualize the density of points at each parent–child pairing. By jittering, we can see where most of the data fall in the parent–child scatterplot (i.e., points are concentrated toward the center of the plot):

      Correlation coefficients, specifically the Pearson correlation, are employed in virtually all fields of study, and without the invention or discovery of correlation, most modern‐day statistics would simply not exist. This is especially true for the field of psychometrics, which is the science that deals with the measurement of psychological qualities such as intelligence, self‐esteem, motivation, among others. Psychometrics features the development of psychometric tests purported to measure the construct of interest. For an excellent general introduction to psychometrics, consult McDonald (1999).

      When developing psychometric instruments, two statistical characteristics of these tests are especially important: (1) validity, and (2) reliability. Validity of a test takes many forms, including face validity, criterion validity, and most notably, construct validity. Construct validity attempts to assess whether a purported psychometric test actually measures what it was designed to measure, and one way of evaluating construct validity is to correlate the newly developed measure with that of an existing measure that is already known to successfully measure the construct.

      The second area of concern, that of reliability, is just as important. Two popular and commonly used forms of reliability in psychometrics are those of test–retest and internal consistency reliability. Test–retest reliability evaluates the consistency of test scores across one or more measurement time points. For example, if I measured your IQ today, and the test was worth its salt, I should expect that a measurement of your IQ a month from now should, within a reasonable margin of error, generate a similar score, assuming it was administered under standardized conditions both times. If not, we might doubt the test's reliability. The Pearson correlation coefficient is commonly used to evaluate test–retest reliability, where a higher‐than‐not coefficient between testings is desirable. In addition to test–retest, we often would like a measure of what is known as the internal consistency of a measure, which, though having potentially several competing meanings (e.g., see Tang et al., 2014), can be considered to assess how well items on a scale “hang together,” which is informal language for whether or not items on a test are interrelated (Schmitt, 1996). For this assessment, we can compute Cronbach's alpha, which we will now briefly demonstrate in SPSS.

      As a very small‐scale example, suppose we have a test having only five items (items 1 through 5 in the SPSS data view), and would like to assess the internal consistency of the measure using Cronbach's alpha. Suppose the scores on the items are as follows:

Скачать книгу


Item_1 Item_2 Item_3 Item_4 Item_5
1 10.00 12.00 15.00 11.00 12.00
2 12.00 18.00 12.00 12.00 1.00
3 8.00 16.00 14.00 14.00 4.00
4 6.00 8.00 16.00 8.00 6.00
5 4.00 7.00 8.00 7.00 5.00