Main Page | See live article | Alphabetical index

Pearson product-moment correlation coefficient

In mathematics, and in particular statistics, the Pearson product-moment correlation coefficient (r) is a measure of how well a linear equation describes the relation between two variables X and Y measured on the same object or organism. It is defined as the sum of the products of the standard scores of the two measures divided by the degrees of freedom:

The result obtained is equivalent to dividing the covariance between the two variables by the product of their standard deviations. In general the quantity of a correlation coefficient is the square root of the coefficient of determination (r2), which is the ratio of explained variation to total variation:

where:

Y = a score on a random variable Y
Y' = corresponding predicted value of Y, given the correlation of X and Y and the value of X
= mean of Y

The correlation coefficient adds a sign to show the direction of the relationship. The formula for the Pearson coefficient conforms to this definition, and applies when the relationship is linear.

The coefficient ranges from -1 to 1. A value of 1 shows that a linear equation describes the relationship perfectly and positively, with all data points lying on the same line and with Y increasing with X. A score of -1 shows that all data points lie on a single line but that Y increases as X decreases. A value of 0 shows that a linear model is inappropriate – that there is no linear relationship between the variables.

The Pearson coefficient is a statistic which estimates the correlation of the two given random variables.

The linear equation that best describes the relationship between X and Y can be found by linear regression. If X and Y are both normally distributed, this can be used to "predict" the value of one measurement from knowledge of the other. That is, for each value of X the equation calculates a value which is the best estimate of the values of Y corresponding the specific value of X. We denote this predicted variable by Y.

Any value of Y can therefore be defined as the sum of Y and the difference between Y and Y:

The variance of Y is equal to the sum of the variance of the two components of Y:

Since the coefficient of determination implies that sy.x2 = sy2(1 − r2) we can derive the identity

The square of r is conventionally used as a measure of the strength of the association between X and Y. For example, if the coefficient is .90, then 81% of the variance of Y is said to be explained by the changes in X and the linear relation between X and Y.

r is a parametric statistic. It assumes that the variables being assessed are normally distributed. If this assumption is violated, a non-parametric alternative such as Spearman's ρ may be more successful in detecting a linear relationship.