Evaluation of internal reliability in the presence of inconsistent responses
© Fong et al; licensee BioMed Central Ltd. 2010
Received: 10 September 2009
Accepted: 12 March 2010
Published: 12 March 2010
We aimed to assess the impact of inconsistent responses on the internal reliability of a multi-item scale by developing a procedure to adjust Cronbach's alpha.
A procedure for adjusting Cronbach's alpha when there are inconsistent responses was developed and used to assess the impact of inconsistent responses on internal reliability by evaluating the standard Chinese 12-item Short Form Health Survey in adolescents.
Contrary to common belief, random responses may inflate Cronbach's alpha when their mean differ from that of the true responses. Fixed responses inflate Cronbach's alpha except in scales with both positive and negative polarity items. In general, the bias in Cronbach's alpha due to inconsistent responses may change from negative to positive with an increasing number of items in a scale, but the effect of additional items beyond around 10 becomes small. The number of response categories does not have much influence on the impact of inconsistent responses.
Cronbach's alpha can be biased when there are inconsistent responses, and an adjustment is recommended for better assessment of the internal reliability of a multi-item scale.
Internal reliability is an attribute of a multi-item scale that refers to the extent to which items in the scale are related; it is very often evaluated to assess the reliability of patient-reported outcomes (PROs). The most common measure of internal reliability reported in psychometric studies of PROs is Cronbach's alpha , but unfortunately, it can be biased by the presence of inconsistent responses.
Inconsistent responding occurs when respondents complete a questionnaire without comprehending the items, typically in self-reported questionnaires when the participants are unmotivated or the questions are sensitive . Inconsistent responses are classified as random, when responses are given unsystematically, or fixed, when the same response is given to all items . Although the literature has not stipulated the impact of inconsistent responses on internal reliability, fixed responses by their nature would result in high association among the responses of the associated items and thus inflate the observed reliability in scales whose items have the same polarity. They can also diminish it in scales when that is not the case as the association among the item responses would be lower. Moreover, a substantial number of random responses would diminish the internal reliability by the independent nature of random responses, but what it means by substantial and such an effect in general are less certain.
In practice, inconsistent responses may not be easily identified since they can also be plausible responses. Random responses are particularly difficult to detect as they have no identifiable patterns. Nevertheless, there are tested personality scales, namely, the Minnesota Multiphasic Personality Inventory-2 (MMPI-2) and the Minnesota Multiphasic Personality Inventory-Adolescent (MMPI-A), that assess the level of inconsistency for a response [4, 5]. Both of them have a variable response inconsistency (VRIN) scale for assessing random responding and a true response inconsistency (TRIN) scale for assessing fixed responding. Cutoff values have also been established for the detection of random and fixed responses [4–6]. Depending on the instrument used, the VRIN scale comprises at least 50 item pairs and the TRIN at least 23 item pairs. As their length does not always allow for concurrent use with PRO instruments, we can only assess the sensitivity of internal reliability within an anticipated range of the proportion of inconsistent responses. However, to the best of our knowledge, no method is available for adjusting the internal reliability due to inconsistent responses.
In view of these, we aimed 1. to evaluate the impact of inconsistent responding on internal reliability; 2. to propose a method for adjusting Cronbach's alpha in the presence of inconsistent responses; and 3. to illustrate the use of the procedure in evaluating the internal reliability of the standard Chinese 12-item Short Form Health Survey (SF-12v2) for a large sample of adolescents.
Adjusting Cronbach's alpha for inconsistent responses
when all items have the same polarity. μR and are the mean and variance of the random responses and can be taken as and , respectively, for scales composing of items responded on a K-point Likert scale with each scored from 1 to K. μT is the mean of true responses and can be taken as [see Additional file 1].
Cronbach's alpha adjusted for inconsistent responding can be calculated from (1) after replacing the unknown quantities by the corresponding sample estimates. Note the adjustment assumes that both random and fixed responses to an item are uniformly distributed over the K-point Likert scale; i.e., there is no specific preference of a certain response category. Performance of the adjustment procedure is assessed by a small Monte-Carlo simulation study. Biases of the adjusted Cronbach's alpha are consistently smaller than those of the unadjusted alpha [see Additional file 2].
Assessing the impact of inconsistent responses on Cronbach'salpha
The influence of random responses is assessed by varying its proportion (pR) from 0 to 50% when pF is taken to be 0 or 5%. The mean difference between the true and random responses (μT-μR) is 0 or 1, and the scale has 5 positive polarity items, each responded on a 5-point Likert scale.
The influence of fixed responses is assessed by varying its proportion (pF) from 0 to 50%. The pR is taken to be 0 or 10%, and the number of positive polarity items is 5 or 3. Moreover, the mean difference between the true and random responses (μT-μR) is 0, and the scale has 5 items, each responded on a 5-point Likert scale.
The influence of the number of items is assessed by varying it from 2 to 20 when the proportion of positive polarity items is taken to be 0.5 or 1, and the mean difference between the true and random responses (μT-μR) is 0 or 1. Moreover, all items are responded on a 5-point Likert scale.
The influence of the number of item response categories (K) is assessed by varying it from 2 to 10 when the number of positive polarity items is 5 or 3, and the mean difference between the true and random responses is 0 or 0.2 K. Moreover, we assume that 20% and 5% of responses are random and fixed, respectively.
For each of the four scenarios, Cronbach's alpha based on the true responses is defined to be 0.4, 0.5, 0.6, 0.7 and 0.8.
A real example to illustrate the adjustment of inconsistent responses
As an example, we evaluate the internal reliability of the standard Chinese SF-12v2. The questionnaire consists of 12 items in eight scales. For the sake of illustration, we considered only the Physical functioning (PF), Role emotional (RE) and Mental health (MH) scales, each of which consists of two items. All items in the three scales are positively worded except one item in MH that is negatively worded. Items in the PF scale use a 3-point Likert scale, while the other items use a 5-point Likert scale. The original scale scores are standardized in the range of 0-100, but for convenience, we just considered the total score after reverse coding the responses of the negative polarity items. Note, however, that the internal reliability is invariant to this standardization.
Data in the standard Chinese SF-12v2 were collected from the Hong Kong Student Obesity Surveillance (HKSOS) project conducted in 2006-2007. This study was cross-sectional involving 42 high schools covering all 18 districts in Hong Kong. It administered a survey questionnaire that contained the SF-12v2. The project was approved by the Institutional Review Board of The University of Hong Kong and the Hospital Authority Hong Kong West Cluster.
The impact of inconsistent responses on Cronbach's alpha
Internal reliability of the standard Chinese SF-12v2
A summary of scales of the standard Chinese SF12v2 in adolescents
Role emotional (RE)
Physical functioning (PF)
Mental health (MH)
Number of response categories (K)
Number of items
Positive polarity (m+)
Negative polarity (m-)
Number of respondents
For the PF scale, K = 3, μR is estimated as 2 and as 0.67. The values of αa at different values of pR and pF are shown in Figure 5(b). While there remains an inflation of Cronbach's alpha when there are fixed responses, it is interesting to note a general decreasing trend of the true internal reliability after removing more random responses. In other words, the presence of random responses may also inflate Cronbach's alpha. A further examination of the scale items revealed that they were highly left skewed, with ceiling percentages of 80.4% and 81.5%, leading to 72.7% of the scale scores reaching the plausible maximum of 6 (Table 1). Indeed, random responses are systematically lower (μR= 2) than true responses (μT> 2). Thus, when there are random responses that uniformly spread over the plausible item values, small item values are more likely random responses than large item values. Consequently, individuals who gave random responses would more likely have small values in all items, and hence their presence would enhance the inter-item association. In fact, it can be shown that the presence of random responses increases the correlation between two positively worded items when the true correlation is below . This threshold increases when (μT-μR)2, which measures the skewness of the true responses from the mid-response, becomes large. In summary, the PF scale should have a Cronbach's alpha of at most 0.67 only, and its internal reliability could be unacceptably low given the anticipated range of inconsistent responses.
Figure 5(c) examines the impact of inconsistent responses on MH, which consists of a positive polarity and a negative polarity item. In contrast to the other two scales, the presence of both random and fixed responses would reduce the Cronbach's alpha of the MH scale. Thus, the reported Cronbach's alpha of 0.33 is indeed the minimum level, and the adjusted value could be as high as 0.66 given the anticipated range of inconsistent responses.
The presence of inconsistent responses may positively or negatively bias the Cronbach's alpha, making the assessment of internal reliability difficult. An adjustment was proposed to Cronbach's alpha for correcting the effects of inconsistent responses when one can estimate a possible range for the percentage of inconsistent responses. This enables a sensitivity analysis to assess the potential impact of inconsistent responses and facilitates a better understanding of the internal reliability of a multi-item scale.
As one would expect, the presence of fixed responses overestimates Cronbach's alpha for scales composed of items mostly worded in the same direction but would otherwise lead to an underestimation. However, it is interesting to observe that random responses may indeed inflate Cronbach's alpha when the distribution of true responses is skewed or, more precisely, when the true mean response deviates from the random/fixed mean response. This is contrary to the common intuition that random responses would dilute the association among items and hence reduce the internal reliability. Indeed, when the true item responses are skewed on the same side, the addition of random responses that scatter around the mid-response could strengthen association among the items if they are not too many. Thus, paradoxically, this kind of noise could inflate the internal reliability and hence Cronbach's alpha. Unfortunately, it is common for true responses to differ from random/fixed responses, on average, especially in patients whose quality of life has deteriorated due to their adverse conditions. Hence, we should be careful not to optimistically interpret Cronbach's alpha when there are random responses.
To determine random and fixed responses, tested personality scales such as the VRIN and TRIN scales of the MMPI-2 and MMPI-A may be considered . They are, however, rather lengthy, requiring at least 23 item pairs, and they may not be feasibly incorporated into large scale studies. Nevertheless, we need to have an estimate of the proportion of inconsistent responses in a sample before the proposed method can be effectively applied. While the determination of whether an individual was endorsing inconsistent responses can be a challenge, modification or addition of a few items for tracking potentially inconsistent responses will be helpful. As in our illustrative example, the response option of "none of the above" in items allowing multiple response choices could be easily incorporated to track for potential random responses. Fixed responses are more easily identified by the patterns that they follow. Incorporating items that would not likely receive the same response will be useful.
Cronbach's alpha of a scale has been known to be higher in scales with more items . We have found that, when there are inconsistent responses, scales with more items would also increase any upward bias in Cronbach's alpha. Although the increase diminishes and may become negligible when there are many items, it is better to keep the number of items minimal to avoid reporting an overly optimistic Cronbach's alpha. Nevertheless, there remains a chance of under-estimating Cronbach's alpha, and it is probably better to be conservative when assessing the internal reliability of a scale.
We have also shown that the number of response categories does not have much influence on the bias of Cronbach's alpha induced by the presence of inconsistent responses. There could be only a small positive increase in the bias for scales with items of 3 or fewer response categories. Previous studies have shown that scales with fewer response categories tend to have lower internal reliability and suggested the use of more than 3 response categories [8, 9]. This recommendation is indeed also good to minimize the impact of inconsistent responses. However, the choice of the number of response categories may largely depend on the actual content of the scale . Modern assessment of item characteristics utilizing item response theory is deemed more useful to determine an appropriate number of response categories .
We have illustrated how Cronbach's alpha can be adjusted for inconsistent responses by evaluating the standard Chinese SF-12v2 in a large sample of students. Note that each scale of the SF-12v2 consists of at most two items only. Although the Cronbach's alpha may in theory be used for scales of at least two items, its use for two-item scales has been criticized . The concern lies in whether two items are sufficient to represent the correspondingly larger domain comprising a much larger collection of items. Alternative forms of reliability that utilize more items in the same construct may be more desirable . Hence, the internal reliability of the SF-12v2 may require further study. It is used here to merely illustrate the impact of inconsistent responses on Cronbach's alpha.
The proposed adjustment to Cronbach's alpha for correcting the effects of inconsistent responses facilitates the assessment of the impact of inconsistent responses on internal reliability. In practice, as soon as respondents with inconsistent item-answer behavior had been identified, it would be simpler to exclude them from the calculation of Cronbach's alpha. However, when the identification of such responses is difficult and the anticipated range of inconsistent responses may be taken more conservatively than that of actually identified, the proposed adjustment may be used.
We assumed the random and fixed responses to an item are uniformly distributed over a K-point Likert scale. When an individual is endorsing a random or fixed response to an item without referencing to the actual content of the time, there would likely be no specific preference on endorsing a particular response category. Therefore, unless there are particular response categories that would be generally endorsed more often in the population, the assumption of uniform distribution appears to be reasonable. Nevertheless, a non-uniform distribution may also be incorporated. Indeed, the adjustment procedure depends on only the first two moments of the random and fixed responses. A different mean of random and fixed responses would either increase or decrease its difference from the mean of true responses (i.e. μT-μR), on which the influence has been examined in Figure 1. On the other hand, an increase of the variance of random and fixed responses would increase the proportion of variance in the total score that is due to inconsistent responses (i.e. /variance of S) which reduces the observed Cronbach's alpha.
We have not examined the impact of inconsistent responses on inference about Cronbach's alpha. However, it has been previously shown that the width of the corresponding confidence interval is negatively proportional to the estimated Cronbach's alpha [14, 15]. Thus a positively biased alpha would tend to result in a short confidence interval leading to a nominal coverage less than the required level. Hence, the false positive error rate for testing about the significance of Cronbach's alpha would also be increased.
Cronbach's alpha has been criticized on the grounds that is just a lower bound of reliability and that other measures may be considered as a better lower bound measure than the coefficient alpha . Moreover, it implicitly assumes the items are responded on an interval scale which limits its use in PRO instruments when items are categorically scored. Besides, it assumes a fixed level of reliability across the whole range of the measurement, and is not a measure of uni-dimensionality. Nevertheless, Cronbach's alpha may be interpreted as a measure of the proportion of the total score variance that can be attributed to true score variance that may be affected by the extent to which the items are associated. Hence, we believe that the impact of inconsistent responses could be applicable to the general evaluation of internal reliability of a scale. An analytical exploration of the impact of inconsistent responses would be desirable. A potential method was the modern psychometric assessment by item response theory which allows the examination of the response characteristics of individual items. It has gained much popularity but it has been reviewed and concluded to be relatively unsuccessful in identifying dissimulation [17, 18]. Further work may deem to be necessary.
Cronbach's alpha may be inflated by inconsistent responses when either the mean of true responses differ from that of the random/fixed responses or all items in the scale have the same polarity. The inflation in the former situation is due to the presence of random responses, while the latter is due to the presence of fixed responses. It should not be assumed that random responses always diminish Cronbach's alpha.
We thank Mr. KK Mak and Miss W.S. Lo, who coordinated and collected data in the HKSOS project. The HKSOS project was financially supported by The University of Hong Kong University Research Committee Strategic Research Theme on Public Health.
- Cronbach LJ: Coefficient Alpha and the Internal Structure of Tests. Psychometrika 1951, 16: 297–334. 10.1007/BF02310555View ArticleGoogle Scholar
- Siddiqui O, Mott JA, Anderson TL, Flay BR: Characteristics of inconsistent respondents who have "ever used" drugs in a school-based sample. Subst Use Misuse 1999, 34: 269–295. 10.3109/10826089909035646PubMedView ArticleGoogle Scholar
- Weiner IB, Freedheim DK, Schinka JA, Gallagher M, Healy AF, Millon T, Lerner RM, Reynolds WM, Stricker G, Nezu AM, et al.: Handbook of psychology. Hoboken, N.J.: John Wiley; 2003.View ArticleGoogle Scholar
- Handel RW, Arnau RC, Archer RP, Dandy KL: An evaluation of the MMPI-2 and MMPI-a true response inconsistency (TRIN) scales. Assessment 2006, 13: 98–106. 10.1177/1073191105284453PubMedView ArticleGoogle Scholar
- Baer RA, Ballenger J, Berry DTR, Wetter MW: Detection of random responding on the MMPI-A. Journal of Personality Assessment 1997, 68: 139–151. 10.1207/s15327752jpa6801_11PubMedView ArticleGoogle Scholar
- Baer RA, Kroll LS, Rinaldo J, Ballenger J: Detecting and discriminating between random responding and overreporting on the MMPI-A. Journal of Personality Assessment 1999, 72: 308–320. 10.1207/S15327752JP720213View ArticleGoogle Scholar
- Cortina JM: What Is Coefficient Alpha - an Examination of Theory and Applications. Journal of Applied Psychology 1993, 78: 98–104. 10.1037/0021-9010.78.1.98View ArticleGoogle Scholar
- Weng LJ: Impact of the number of response categories and anchor labels on coefficient alpha and test-retest reliability. Educational and Psychological Measurement 2004, 64: 956–972. 10.1177/0013164404268674View ArticleGoogle Scholar
- Preston CC, Colman AM: Optimal number of response categories in rating scales: reliability, validity, discriminating power, and respondent preferences. Acta Psychologica 2000, 104: 1–15. 10.1016/S0001-6918(99)00050-5PubMedView ArticleGoogle Scholar
- Halpin G, Halpin G, Arbet S: Effects of Number and Type of Response Choices on Internal Consistency Reliability. Perceptual and Motor Skills 1994, 79: 928–930.View ArticleGoogle Scholar
- Roberson-Nay R, Strong DR, Nay WT, Beidel DC, Turner SM: Development of an abbreviated Social Phobia and Anxiety Inventory (SPAI) using item response theory: The SPAI-23. Psychological Assessment 2007, 19: 133–145. 10.1037/1040-3522.214.171.124PubMedView ArticleGoogle Scholar
- Cudeck R: Cronbach's alpha on two-item scales. J Consum Psychol 2001, 10: 55–55. 10.1207/S15327663JCP1001&2_05View ArticleGoogle Scholar
- Ware JE, Turner-Bowker DM, Kosinski M, Gandek B: How to Score Version 2 of the SF-12® Health Survey. Lincoln, RI: QualityMetric; 2002.Google Scholar
- van Zyl JM, Neudecker H, Nel DG: On the distribution of the maximum likelihood estimator of Cronbach's alpha. Psychometrika 2000, 65: 271–280. 10.1007/BF02296146View ArticleGoogle Scholar
- Iacobucci D, Duhachek A: Advancing alpha: Measuring reliability with confidence. J Consum Psychol 2003, 13: 478–487. 10.1207/S15327663JCP1304_14View ArticleGoogle Scholar
- Sijtsma K: On the Use, the Misuse, and the Very Limited Usefulness of Cronbach's Alpha. Psychometrika 2009, 74: 107–120. 10.1007/s11336-008-9101-0PubMed CentralPubMedView ArticleGoogle Scholar
- Ferrando PJ, Chico E: Detecting dissimulation in personality test scores: A comparison between person-fit indices and detection scales. Educational and Psychological Measurement 2001, 61: 997–1012. 10.1177/00131640121971617View ArticleGoogle Scholar
- Reise SP, Flannery WP: Assessing person-fit on measures of typical performance. Appl Meas Educ 1996, 9: 9–26. 10.1207/s15324818ame0901_3View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.