Using Statistics to Examine Relationships



Using Statistics to Examine Relationships


imagehttp://evolve.elsevier.com/Grove/practice/


Correlational analyses identify relationships or associations among variables. There are many different kinds of statistics that yield a measure of correlation. All of these statistics address a research question or hypothesis that involves an association or a relationship. Examples of research questions that are answered with correlation statistics are as follows: “Is there an association between weight loss and depression?” “Is there a relationship between patient satisfaction and health status?” A hypothesis is developed to identify the nature (positive or negative) of the relationship between the variables being studied. For example, a researcher may hypothesize that better self-reported health is associated with higher levels of patient satisfaction (Cipher & Hooker, 2006).


This chapter presents the common analysis techniques used to examine relationships in studies. The analysis techniques discussed include the use of scatter diagrams before correlational analysis, bivariate correlational analysis, testing the significance of a correlational coefficient, spurious correlations, correlations between two raters or measurements, the role of correlation in understanding causality, and multivariate correlational procedure–factor analysis.



Scatter Diagrams


Scatter plots or scatter diagrams provide useful preliminary information about the nature of the relationship between variables (Corty, 2007; Munro, 2005). The researcher should develop and examine scatter diagrams before performing a correlational analysis. Scatter plots may be useful for selecting appropriate correlational procedures, but most correlational procedures are useful for examining linear relationships only. A scatter plot can easily identify nonlinear relationships; if the data are nonlinear, the researcher should select statistical alternatives such as nonlinear regression analysis (Zar, 1999). A scatter plot is created by plotting the values of two variables on an x-axis and y-axis. As shown in Figure 23-1, pain levels from 14 long-term care residents were plotted against their number of behavioral disturbances (Cipher, Clifford, & Roper, 2006). Specifically, each resident’s pair of values (pain score, behavior value) was plotted on the diagram. Pain was measured with the Geriatric Multidimensional Pain and Illness Inventory (GMPI), and behavioral disturbances were measured by the Geriatric Level of Dysfunction Scale (Clifford, Cipher, & Roper, 2005). The resulting scatter plot reveals a linear trend whereby higher levels of pain tend to correspond with higher behavioral disturbance values. The line drawn in Figure 23-1 is a regression line that represents the concept of least-squares. A least-squares regression line is a line drawn through a scatter plot that represents the smallest deviation of each value from the line (Cohen & Cohen, 1983). Regression analysis is discussed in detail in Chapter 24.




Bivariate Correlational Analysis


Bivariate correlational analysis measures the magnitude of linear relationship between two variables and is performed on data collected from a single sample (Munro, 2005). The particular correlation statistic that is computed depends on the scale of measurement of each variable. Correlational techniques are available for all levels of data: nominal (phi, contingency coefficient, Cramer’s V, and lambda), ordinal (Spearman rank order correlation coefficient, gamma, Kendall’s tau, and Somers’ D), or interval and ratio (Pearson’s product-moment correlation coefficient). Figure 21-7 in Chapter 21 illustrates the level of measurement for which each of these statistics is appropriate. Many of the correlational techniques (Kendall’s tau, contingency coefficient, phi, and Cramer’s V) are used in conjunction with contingency tables, which illustrate how values of one variable vary with values for a second variable. Contingency tables are explained further in Chapter 25.


Correlational analysis provides two pieces of information about the data: the nature or direction of the linear relationship (positive or negative) between the two variables and the magnitude (or strength) of the linear relationship. Correlation statistics are not an indication of causality, no matter how strong the statistical result.


In a positive linear relationship, the values being correlated vary together (in the same direction). When one value is high, the other value tends to be high; when one value is low, the other value tends to be low. The relationship between weight and blood pressure is considered positive because the more a patient weighs, usually the higher his or her blood pressure. In a negative linear relationship, when one value is high, the other value tends to be low. There is a negative linear relationship between calories consumed and weight loss because the more calories a person consumes, the less his or her weight loss. A negative linear relationship is sometimes referred to as an inverse linear relationship—the terms negative and inverse are synonymous in correlation statistics.


Sometimes the relationship between two variables is curvilinear, which reflects a relationship between the variables that changes over the range of both variables. For example, one of the most famous curvilinear relationships is that of stress and test performance. Test performance tends to be better as test-takers have more stress but only up to a point. When students experience very high stress levels, test performance deteriorates (Lupien, Maheu, Tu, Fiocco, & Schramek, 2007; Yerkes & Dodson, 1908). Analyses designed to test for linear relationships or associations between two variables, such as Pearson’s correlation, cannot detect a curvilinear relationship.



Pearson’s Product-Moment Correlation Coefficient


Pearson’s product-moment correlation was the first of the correlation measures developed and is the most commonly used (Corty, 2007; Munro, 2005). This coefficient (statistic) is represented by the letter r, and the value of r is always between −1.00 and +1.00. A value of zero indicates no relationship between the two variables. A positive correlation indicates that higher values of X are associated with higher values of Y, and lower values of X are associated with lower values of Y. A negative or inverse correlation indicates that higher values of X are associated with lower values of Y. The r value is indicative of the slope of the line (called a regression line) that can be drawn through a standard scatter plot of the values of two paired variables. The strengths of different relationships are identified in Table 23-1 (Aberson, 2010; Cohen, 1988). Figure 23-2 represents an r value approximately equal to zero, indicating no relationship or association between the two variables. An r value is rarely, if ever, equal to exactly zero. Figure 23-3 shows an r value equal to 0.50, which is a moderate positive relationship. Figure 23-4 shows an r value equal to −0.50, which is a moderate negative or inverse relationship.






As discussed earlier, Pearson’s product-moment correlation coefficient is used to determine the relationship between two variables measured at least at the interval level of measurement. The formula for Pearson’s correlation coefficient is based on the following assumptions:



Data that are homoscedastic are evenly dispersed above and below the regression line, which indicates a linear relationship on a scatter plot. Homoscedasticity reflects equal variance of both variables. In other words, for every value of X, the distribution of Y values should have equal variability. If the data for the two variables being correlated are not homoscedastic, inferences made during significance testing could be invalid (Cohen & Cohen, 1983).



Calculation


Pearson’s product-moment correlation coefficient is computed using one of several formulas; the following formula is considered the “computational formula” because it makes computation by hand easier (Zar, 1999).


r=nΣXYΣXΣY[nΣX2(ΣX)2][nΣY2(ΣY)2]


image


Table 23-2 displays how one would set up data to compute a Pearson’s correlation coefficient. This example includes a subset of data from a study introduced earlier in this chapter of long-term care residents (n = 14) with the most severe levels of dementia (Cipher et al., 2006). A subset of data was selected for this illustration so that the computation example would be small and manageable. In actuality, studies involving correlational analysis need to be adequately powered and involve a larger sample than is used here (Aberson, 2010; Cohen, 1988). However, all data presented in this chapter are actual, unmodified clinical data.



The first variable in Table 23-2 is the number of behavioral disturbances being exhibited by the residents. The second variable is the residents’ scores on the Pain and Suffering subscale of the GMPI (Clifford & Cipher, 2005). The GMPI is a comprehensive pain assessment instrument that was designed to be used by nurses, social workers, and psychologists to assess pain experienced by a person residing in a long-term care facility. The GMPI Pain and Suffering subscale is scored on a scale of 1 to 10, with higher numbers indicating more severe pain. Higher numbers are also indicative of more behavioral disturbances. The null hypothesis being tested is: There is no significant association between pain and behavioral disturbances among long-term care residents with severe dementia. The data in Table 23-2 are arranged in columns, which correspond to the elements of the formula. The summed values in the last row of Table 23-2 are inserted into the appropriate place in the formula:


r=14(226.6)(49)(55)[14(225)492][14(286)552]r=14(226.6)(49)(55)(749)(979)r=477.4856.3=0.56


image

Stay updated, free articles. Join our Telegram channel

Feb 17, 2017 | Posted by in NURSING | Comments Off on Using Statistics to Examine Relationships

Full access? Get Clinical Tree

Get Clinical Tree app for offline access