33 Correlation: Karl Pearson’s Coefficient of Correlation, Spearman Rank Correlation ar

Dr Deependra Sharma

 

Learning Objectives

 

After reading this module , student will be able to –

 

1.      Clearly define the meaning of Correlation and its characteristics.

2.      Understand different types of correlation and their application in statistics.

3.      Define methods of studying correlation with suitable mathematical examples.

4.    Comprehend the concept of   ‘coefficient of determination’ and will be able to interpret it.

 

Correlation

 

Correlation refers to connection, in correlation analysis we study the connection or the relation between two or more variables. If two variables vary in such a way that the change in one variable is accompanied by changes in other variable, these variables are said to be correlated. For example, relationship between the height and weight of students in a class, there exists some relation between earning of family and amount spent on luxurious items, relationship between dose of insulin and the blood sugar, etc. There may be ‘n’ number of variables which may affect each other and relationship might exists among all of them for example there are three Variables X, Y and Z, X may be related to Z and Y both and Z may relate to Y, hence we can state that the correlation is a statistic tool to find out the relationship between two or more variables.

 

Definition:

 

Tuttle defined correlation as: “An analysis of the co-variation of two or more variables is usually called correlation”. Correlation is the degree of interrelatedness of two or more Variables. It is a process to determine the amount of relationship between variables with the help of tools and techniques provided by statistics. Many authors have given different definitions of correlation in simple terms correlation is the scale of relationship between variables.

 

Characteristics of Correlation:

 

  1. Although correlation analysis establishes the degree of relationship between variables but it fails to throw light on cause-effect relationship.
  2. Existence of correlation between the variables may be due to chance, especially when sample taken is small in number. If we write some data points or some observation of the change in a particular variable there is a possibility of some pattern or relationship in the observation which may not be intentional likewise it is also possible between different data sets or observations of different variables which may not be true in real situation. For example if we collect data sets of number of bikes sold in NCR and amount of rainfall.
Year Number of Bikes sold in NCR( In Lakhs) Total Rainfall in NCR(MM)
2001 25 120
2002 35 140
2003 45 160
2004 55 180
2005 65 200
2006 75 220
2007 85 240
2008 95 260

 

In the example given above we see a perfect positive relationship between Number of bikes sold in NCR and Total rainfall in NCR that is as the number of bikes sold is increasing the total rainfall is also increasing and the ration of change between two variables is the same, However we know that in real world there is no relation in both of the variables. To handle these situation researcher should always apply common sense.

 

3.  It should be noted that correlated variables may be influenced by one or more variables. It is close to impossible that we are able to find out complete cause and effect relationship between two or more variables, in most of the cases one variable is affected or influenced by many other variables in some proportion, even if our data is showing complete perfect relationship between two variables there may be third hidden variable which may be affecting the relationship, for example- if a television brand is trying to understand the buying decision of consumers and they are collecting data of both the husband and wives behavior and they stablish a relationship between the two observed data sets but is also a possibility that the purchase decision of television in the family may be somewhat effected by the number of children in the family for that data has not been collected.

 

Types of correlation:

 

Scholars have defined correlation in different types. Basically they all can be defined in three types-

In Type I correlation direction of change in the observation is important, it can be understood more once we differentiate between positive and negative correlation.

 

Positive Correlation:

 

Let us take two variables X and Y the change direction in both the variables is in such a way that if X decreases Y on an average also decreases and if X increases Y on an average also increases. This correlation is called Positive correlation. For Example:

Example 1 Example 2
X Y X Y
10 15 80 50
12 20 70 45
11 22 60 30
18 25 40 20
20 37 30 10

 

In both the cases we see that the change in X and Y is same direction. Examples of positive correlation can be height and weight, water consumption and temperature etc.

 

Negative Correlation:

 

Likewise positive correlation in negative correlation the direction of change is important but in opposite direction. If we take same example of X and Y variables and see that if X is increasing and Y on an average is decreasing or if X is decreasing and Y on an average is increasing, variables with this correlation are called negatively correlated, for example-

 

Situation 1 Situation 2
X Y X Y
20 40 100 10
30 30 90 20
40 22 60 30
60 15 40 40
80 16 30 50

In both the case above we see that the change in X and Y is in opposite direction. Law of demand is an example of negative correlation.

In Type II correlation number of variables studied is important. Variable Numbers decide whether they are simple or multiple, to understand this let us discus one by one-

 

Simple Correlation:

 

When in a problem we only study two variables, the problem is said a simple correlation problem. For example if we study the wheat production in an area and the degree of rainfall in that same area, it will be considered as an example of simple correlation.

 

Multiple Correlations:

 

When more than two variables are studied then the problem is of multiple correlations. For example if we study the production of rice in an area, the amount of rainfall and the amount of fertilizers used in the same area, it will be treated as a problem of this correlation.

 

It is of two types as per the influencing nature of Variables-

 

a) Partial Correlation:

 

Partial Correlation arrives where in a problem we identify more than two variables but consider only two variables to be influencing each other, assuming other variables as constant.

 

Eg. In the problem of wheat production, rainfall and temperature if correlation analysis is conducted between wheat production and rainfall and is limited to time periods when a certain average daily temperature existed.

 

b)   Total Correlation:

 

When we study all the existing variables it is a problem of total correlation. However it is normally impossible or close to impossible.

 

In Type III Correlation we differentiate the relationship of variable based on the ratio of change. Ratio of two variables decides whether both the variables are linearly correlated or Non-linearly.

 

Linear Correlation:

 

Linear correlation is based on the ratio of change and its consistency between the variables under study. Variables are said to be linearly correlated when the ratio is constant. In other words if the degree of change in one variable and amount of change in the other variable results in the same ratio, then the correlation is said to be linear.

 

If we draw a graph, of variables having linear relationship will always form a straight line.

For example: below are the examples of linear correlation-

a)X = 10 20 30 40 50 60 70 80
Y = 70 140 210 280 350 420 490 560

 

Non-Linear correlation:

 

When the amount of change in one variable does not hold a constant ratio to the amount of change in the other variables, it is said to be nonlinear correlation.

 

Eg.- Doubling the rainfall will not result in increasing the production of rice wheat by two times. This correlation is also referred as curvilinear relationship between the variables.

 

Various techniques of studying correlation:

 

Following are the techniques used to study correlation:

  1. Scatter diagram technique of correlation
  2. Karl Pearson’s Coefficient of correlation
  3. Spearman’s Rank Correlation

 

1.  Scatter diagram technique:

 

It is a easy and attractive technique of diagrammatic representation. Here, the given data are plotted on a graph sheet in the form of dots. The X variable is plotted on the horizontal axis and y variable on the vertical axis, now we can know the scatter or concentration of the various points. X and Y variables are generally referred as independent and dependent variables respectively.

(A. Low Degree correlation)                    (B. High Degree Correlation)                            (C. Zero correlation)

 

 

2. Karl Pearson’s Coefficient of correlation:

 

While we study two variables a few immediate questions come in mind, first is there any relationship between the behavior of these two variables second if there is any relationship between the variables whether it is statistically significant enough third what is the type of relationship (+ve or – ve) between the variables and the fourth is that can the value of one variable be predicted based on the change in other variables? If these questions can be answered correctly we can measure and predict the behavior of two variables. Correlation analysis answers the first three questions and the last question is answered by regression analysis.

 

It not only tells about the strength of the relationship among the two variables but also about their direction of change.

 

It is widely used in practice and is popularly known as Pearson’s coefficient of Correlation .It is denoted by ‘r’. It is a mathematical method used for measuring the degree of linear relationship between variables. It gives a number that states the strength and the direction of the relationship between the variables.

 

Karl Pearson’s coefficient of correlation can be measured in two ways:

 

1.   When deviation is taken from Actual Mean, the formula is-

r = r(x, y)= Σxy / √ Σx² Σy²

 

Illustration 2.1

 

Find the coefficient between the sales and expenses from the data given below-

Roll No. Marks in Subject A Marks in Subject B
1 48 45
2 35 20
3 17 40
4 23 25
5 47 45

 

Answer:

 

Let the marks in subject A be denoted by X and that in subject B by Y.

Since r = 0.429 it means that there is moderate positive correlation between the both the subjects A and B.

 

2.   When deviation is taken from Assumed mean, the formula is-

 

r =r(x,y)= (N Σdxdy – Σdx Σdy) /(√N Σdx²-( Σdx)² √N Σdy²-( Σdy)²) “r” is referred as coefficient of correlation.

 

Illustration 2.1: The above problem (illustration 2.1) can also be solved by taking assumed mean- Let the marks in subject A be denoted by X and that in subject B by Y and assumed mean of the marks in subject A and B is 35 and 40 respectively-


Coefficient of correlation:
Karl Pearson’s Coefficient of correlation has got following  observational properties like-

 

  1. The value of r is always in between -1 to +1 or -1 ≤ r ≥ +1.
  2. Degree of correlation is expressed by the value of r for example is the value of r is +1 variables are high positively correlated or perfect positive correlation, if it is -1 variables are high negatively correlated or perfect negative correlation and if it is 0 there is no correlation between the variables.
  3. Direction of change is indicated by sign (+ve) or (-ve).
  4. It is expressed as the geometric mean of two regression coefficient r= √ bxy * byx

 

Limitations of Karl Pearson’s Coefficient of correlation

 

There are a few limitations of Karl Pearson’s coefficient of correlation-

 

a)      It only defines the linear relationship between the variables.

b)      It is unduly affected by extreme values of two variable values.

c)      Computation of Pearson’s coefficient is sometimes cumbersome.

 

Coefficient of determination:

 

Coefficient of determination is denoted by ‘r2’. r2 measures the proportion of variation in one variable that is explained by the other. Both r and r2 are symmetric measures of association. In other words, the correlation of X and Y will not different as the correlation of Y and X. it means in both the variables X and Y whichever is taken as dependent variable and as independent variable it makes no difference, what matters is the relationship between both of them.

 

Degree of determination can be expressed as the ratio of explained variation and total variation- r 2 = Explained Variation / Total variation

 

The maximum value of r2 is 1 because it is possible to explain all of the variation in Y but it is not possible to explain more than all of it.

 

Interpretation of Coefficient of determination:

 

Coefficient of determination measures the relationship between two variables in terms of percentage. It is the proportion of explained variation in the value of response variables. For example if the value of r 2 is 0.82 it means 82 percent of the variation in response variable is explainable by explanatory variable and it does not state anything about rest of the 18 percent variation that might be because of any other factor or variable.

 

a)      If r 2 = 0 states there is no relationship between two variables.

b)      If r 2 = 1 is the perfect association between the variables.

c)      If 0 ≤ r 2 ≤1 then the degree of variation in response variable is due to variation in values of explanatory variables. Value of r2 close to zero shows low proportion of variation in response variable due to variation in values of explanatory variable. While the value of r2 close to 1 shows that the complete variation in response variable is due to variation in values of explanatory variable.

 

3. Spearman’s Rank Correlation Coefficient:

 

In case, it is not possible to measure the variables quantitatively but can be arranged in serial order then Karl Pearson’s coefficient of correlation is not applicable. For these situations, the coefficient of correlation had been given by a British psychologist Charles, Edward Spearman. In this method the individuals in the group are arranged in order thereby obtaining for each individual a number indicating its rank in the group. This type of correlation is often called as non-metric correlation as it is a measure of correlation for two non-metric variables that relies on ranking to compute the correlation.

 

Merits

 

This method is quite simple & is comparatively a easier way to establish a relationship between variables.

 

Spearman’s rank correlation coefficient can be calculated easily in case of non-quantitative or qualitative variables.

 

Through this method ordinal ranks can be given easily to establish a relationship between two qualitative variables.

 

Demerits

 

It is assumed that both the variables are normally distributed that may not always be true.

 

It only describes the linear relationship between the variables, does not interpret anything about nonlinear relationship.

 

In case of large data, giving ranks is a bit cumbersome.

 

When grouped data is available this method cannot be applied.

 

Calculation of Spearman’s Rank Correlation:

 

Rank Coefficient of correlation is calculated under three conditions-

 

A)  When Actual ranks are given the formula is-

 

R = 1- (6 ∑ D2) / N (N2 – 1) „

 

Where R = Rank correlation coefficient „ D = Difference of rank between paired item in two series.

 

= Total number of observation.

 

Calculation Method

 

Where actual ranks are given computation can be done by following these steps-Take the difference of the

two ranks (R1-R2) this difference is denoted by D.

 

Calculate ∑D2 by squaring the differences and take their total. Apply the formula-

R= 1 – 6∑  2

3−

 

Illustration 3.1:

 

The ranks of the ten students in two subjects A and B are given in the table below-

Ranks In Subject A Ranks in Subject B
3 4
5 6
4 3
8 9
9 10
7 7
1 2
2 1
6 5
10 8

 

Find out correlation coefficient by using spearman’s rank correlation formula-

Answer:

Step 1: Take the difference of the two ranks and square it-

Ranks In Subject A Ranks in Subject B D2
R1 R2 (R1-R2)
3 4 1
5 6 1
4 3 1
8 9 1
9 10 1
7 7 0
1 2 1
2 1 1
6 5 1
10 8 4

 

Step 2: Take a total of D2 that is ∑D2 = 12

Step 3: make the computation using spearman’s Rank coefficient formula-

R= 1 – 6∑  2

3−

R= 1 –  6 x 12 = 1- 72 = 0.927

103−10                         990

 

B)  When ranks are not given

 

In case ranks are not given we need to assign the ranks to the available observations. We can start giving ranks by giving first or last to highest or lowest but we must follow the same rule for all other available variables.

 

Illustration 3.2: Calculate the rank correlation coefficient for the following given marks of two Subjects A and B.

Marks of Subject A Marks of Subject B
92 86
89 83
87 91
86 77
83 68
77 85
71 52
63 82
53 37
50 57

Answer: Calculation of Rank Correlation Coefficient-

Marks of Subject A R1 Marks of Subject B R2 (R1-R2)2 = D2
92 10 86 9 1
89 9 83 7 4
87 8 91 10 4
86 7 77 5 4
83 6 68 4 4
77 5 85 8 9
71 4 52 2 4
63 3 82 6 9
53 2 37 1 1
50 1 57 3 4
N= 10 ∑D2=44

 

 

C)  When Ranks are equal

 

In some cases two or more data can be equal hence equal ranks are given, in those cases we assign average ranks to each of them for example if two individuals are equal at fourth place then each given the average rank 4+52 = 4.5 similarly if three are ranked equal at fourth place, they are given an average rank of 4+5+6 = 5.

3

 

In case of equal ranks formula for calculating Spearman’s coefficient of correlation is – Where m is number of items whose ranks are common.

 

Illustration 3.3:

 

Marks in the two subjects X and Y of eight applicants are shown below. Calculate rank coefficient of correlation-

Applicants Subject X Subject Y
A 15 40
B 20 30
C 28 50
D 12 30
E 40 20
F 60 10

In Subject X 20 is repeated two times hence m1 = 2 and in subject Y 30 is repeated three times hence m2= 3, putting these values in the formula-

6{81.5+ 1 (23−2)+ 1 (33−3)}

R= 1 –  12   12

83−8

=  1 – 6(81.5+0.5+2)

504

=     1 – 6  84

504

= 0

 

Since R= 0 it means there is no correlation between marks obtained in two subjects.

 

Summary

 

When two variables change in such a way that at any point the value of one variable is related with the value of another variable both are called correlated variables. Correlation analysis is a statistical method of measuring the degree of relationship between two or more variables. Correlation analysis establishes relationship between the variables but it does not say which variable is a cause and which variable is an effect or in other words Causal relationship between the variables cannot be established by correlation analysis.

 

Correlation can be of different types-

 

a)   Positive or negative

  1. b) Simple or multiple
  2. c) Linear or nonlinear

 

There are different methods of studying correlation, stated below-

 

1) Scatter Diagram Method

2) Karl Pearson’s Coefficient of correlation

3) Spearman’s Rank Correlation

 

In Correlation analysis a very important term is measured ‘coefficient of correlation’ and is denoted by ‘r’. The value of ‘r’ defines the strength of relationship between the variables. The value of coefficient of correlation can be between 0 and 1. As much r is closer to 1 the variable relationship is that much stronger also the sign of the numerical value of r shows positive or negative direction of the relationship. Coefficient of correlation is also used to determine the value of ‘coefficient of determination’ by just squaring it and is denoted by r2. Coefficient of determination r2 states the percentage of variation in one variable explained by another variable.

 

Learn More:

  1. Sharma, J K (2014), Business Statistics, S Chand & Company, N Delhi.
  2. Bajpai, N (2010) Business Statistics, Pearson, N Delhi.
  3. Trevor Hastie, Robert Tibshirani, Jerome Friedman (2009), The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition, Springer.
  4. Darrell Huff (2010), How to Lie with Statistics, W. W. Norton, California.
  5. K.R. Gupta (2012), Practical Statistics, Atlantic Publishers & Distributors (P) Ltd., N. Delhi.