How a well-grounded minimal important difference can enhance transparency of labelling claims and improve interpretation of a patient reported outcome measure
Health and Quality of Life Outcomes volume 4, Article number: 69 (2006)
The evaluation and use of patient reported outcome (PRO) measures requires detailed understanding of the meaning of the outcome of interest. The Food and Drug Administration (FDA) recently presented its draft guidance and view on the use of PRO measures as endpoints in clinical trials. One section of the guidance document specifically deals with advice about the use of the minimal important difference (MID) that we redefined as the smallest difference in score in the outcome of interest that informed patients or informed proxies perceive as important. The advice, however, is short, indeed much too short. We believe that expanding the section and making it more specific will benefit all stakeholders: patients, clinicians, other clinical decision makers, those designing trials and making claims, payers and the FDA.
There is no "gold standard" methodology of estimating the MID or achieving the meaningfulness of clinical trial results based on patient reported outcomes. There are many methods of estimating the MID usually grouped into two distinct categories: anchor-based methods, that examine the relationship between scores on the target instrument and some independent measure, and distribution-based methods resorting to the statistical characteristics of the obtained scores.
Estimation of an MID and interpretation of clinical trial results that present patient important outcomes is demanding but vital for informing the decision to recommend approve a given intervention. Investigators are encouraged to use reliable and valid methods to achieve meaningfulness of their results, preferably those that rely on patients to estimate what constitutes a minimal important, small, moderate, or large difference. However, acquiring the meaningfulness of PRO measures transcends beyond a concept of the MID and we advocate that dichotomizing the scores of patient-reported outcome measures facilitate interpretability of clinical trial results for those who need to understand trial results after a labelling claim has been granted. Irrespective of the strategy investigators use to estimate these values, from the individual patient perspective it is much more relevant if investigators report both the estimated thresholds and the proportion of patients achieving that benefit.
The Food and Drug Administration (FDA) presents its draft guidance and its view on the use of patient-reported outcome (PRO) measures as endpoints in clinical trials in this issue of Health and Quality of Life Outcomes . It includes information on how sponsors could use study results based on these measures to support claims in their product labelling that carries important implications for study design and interpretation of the findings . The advice, however, is short, indeed we believe much too short.
The evaluation and use of PRO measures requires detailed understanding of the meaning of the outcome of interest. Achieving this understanding presents a considerable challenge even for seemingly straightforward dichotomous outcomes such as stroke, myocardial infarction, or death [2, 3]. The complexity increases with the realization that no binary outcome is truly unambiguous: deaths can be painful or painless, strokes can be mild or severe, and myocardial infarctions can be large and complicated or small and uncomplicated. The way the investigators present the results of clinical trials also influences clinicians' willingness to undertake a specific action [4–7]. This problem becomes even more complex when one considers that different patients may place a different value on a particular benefit (inter-individual variation) or even the same patient may place a different value on the same benefit (intra-individual variation), depending on the circumstances. These difficulties occur despite the ease with which one can generally communicate an event such as stroke or death.
The challenges increase further, when one faces PRO scores expressed in unfamiliar ordinal or continuous scales. Even those familiar with the concept of PRO or health related quality of life (HRQL) assessment generally have no intuitive notion of the significance of a change in score of a particular magnitude on most instruments that investigators use to measure the severity of outcomes such as stroke or myocardial infarction.
Therefore, one can frame the problem as an issue of interpretability: what changes in score correspond to trivial, small, moderate, or large patient benefit or harm .
The FDA guidance dedicates section IV.C.4 (Choice of Methods for Interpretation) to this particular issue and describes "some of the methods that have helped sponsors and the FDA interpret clinical trial results based on PRO endpoints". We believe that expanding this section and making it more specific will benefit all stakeholders: patients, clinicians, other clinical decision makers, those designing trials and making claims, payers and the FDA.
The Authors of the guidance focused their attention on the minimal important difference (MID). Therefore, we will centre our attention on 4 questions related to the MID: 1) what is the MID; 2) why is the MID important; 3) how to estimate the MID; and 4) when the MID is the proper approach to claiming efficacy and how can it be used by clinicians to understand claims based on clinical trials using PRO measures. Acquiring the meaningfulness of PRO measures transcends beyond a concept of the MID, so we will supplement the discussion about questions 3 and 4 with a consideration on other related approaches investigators used to achieve interpretability of PRO instruments.
To place the interpretation of PRO measure scores in context, before we address the specific issues, we suggest, that the reader conceptualizes these scores as any continuous (e.g. visual analogue pain scale, height) or discrete, in particular ordinal (e.g. stage 1 through stage 4 cancer, severity of pain: none, mild, moderate, severe) variable. It may also be helpful to visualize the PRO score as a surrogate outcome measure that needs some "mapping" into another meaningful, patient-important outcome in order to gain interpretability.
What is the minimal important difference?
The FDA document does not provide a sensu stricto definition of the MID obtained with a PRO measure and confines itself to the notion of "meaningful change" or "effect that might be considered important" . We suggested that the MID be the smallest difference in score in the outcome of interest that informed patients or informed proxies perceive as important, either beneficial or harmful, and that would lead the patient or clinician to consider a change in the management [9, 10].
We place a greater weight on the preferences of informed patients than of informed proxies (including clinicians) in determining the MID. We should consider the MID estimates of informed proxies only if informed patients cannot make decisions about the management of their disease or if patients prefer informed proxies, including clinicians, to make these decisions. It is also important to bear in mind that any change in management will depend on the associated downsides, including harm, cost and inconvenience.
This definition of the MID precludes making its estimates for outcomes that are remote from those important, in themselves, to patients, such as spirometry or laboratory exercise capacity. It also suggests that only if one had reason to question the reliability or accuracy of data from patients would one rely on proxies to provide estimates of the MID. If one accepts that PRO measurement must be fundamentally patient-important, the first choice for establishing the MID should be a patient-based approach. Relative to patients, clinicians may overemphasize treatment effects  and agreement between patients and proxies in rating the PROs is far from perfect [12–15]. To be maximally informative, representative samples of informed patients or if necessary their proxies should provide estimates of the MID.
Why do we need a MID?
There are several reasons for which the concept of MID seems useful and investigators should derive it from patients. First, it appears easily understood by clinicians and investigators as a key concept when one considers the problem of interpretability of PRO scores. Second, it emphasizes the primacy of the patient's perspective and implicitly links that perspective to that of the interpretation by clinicians. Third, the choice of what constitutes a MID will inform judgments about the successfulness of an intervention. Fourth, it helps estimating the required sample size of clinical trials, and the very design of these studies. Fifth, an individual patient achieving the score equal or greater than the MID might be considered a beneficiary of the intervention, what would lead to the definition of a responder, as the authors of the guidance suggested. However, one should be cautious as it is certain that the MID varies across patients and possibly also across patient groups . Since the MIDs are usually derived from the groups of patients, the description of responders based on the MID should be used with great care and with full disclosure to readers how it was obtained.
How does one estimate the MID?
Unfortunately, there is no "gold standard" methodology of achieving the meaningfulness of PRO scores, estimating the MID, or interpreting these scores. This is part of the reason why interpreting PRO measures is indeed a demanding task. A possible and widely used technique would be to approach a group of experts and ask them whether the particular PRO score looks like a reasonable measure of what is important to patients, as they perceive it. This technique may be termed analogous to face validity. However, as described above this approach is based solely on the opinion of experts, and because the experts' perception of what is important to patients tends not to mirror what in fact it is [11, 15], this method must he regarded as a weak means to establish a score that would represent the MID for patients. Fortunately, less medico-centric techniques are available, although none of them is perfect. The authors of the FDA guidance name four examples of derivation of the MID: "mapping changes in PRO scores to clinically relevant and important changes in non-PRO measures of treatment outcome", "mapping changes in PRO scores to other PRO scores", "using an empirical rule", and "using a distribution-based approach". We think the users of the guidance would benefit from some explanation added to this presentation by giving specific examples or descriptions.
Patient versus population perspective
An important issue in shaping the interpretability of a PRO score is whether one makes inferences about patient important change with respect to individuals or populations . One frequently distinguishes between the significance of a particular change in score in an individual patient and a change of the same magnitude in the mean score of a group of patients . From the point of view of the individual, a worthwhile change may be the one that results in a meaningful reduction in symptoms or improvement in function. In contrast, a change in mean score of a magnitude that would be trivial in an individual (e.g., 2 mm Hg reduction in blood pressure), may translate into a large number of benefiting patients in a population (e.g. reduced strokes).
There are two reasons for this difference in interpretation. First, one might consider particular change in score (e.g. 2 mm Hg in blood pressure) in an individual trivial, that is within the error of measurement. In this sense, the change is trivial because one does not believe it is real. On the contrary, relatively modest improvements at the individual level may be rated as important when considered at the group level. The second reason for the distinction between interpretation of individual and group differences is that not every individual in the population does experience the same change in outcome – even groups with negligible mean changes in PRO scores (or any outcome expressed as a mean score) are likely to contain individual patients whose improvement is noteworthy and leads to a reduced stroke risk . From the group perspective individual variability is considered random variation associated with a measurement error. Therefore, from the individual patient perspective this very variability in individual response highlights the fundamental deficiency of summarizing treatment effects as a difference in means. Let us assume there is a threshold below which any change in status has no important consequences for the patient (i.e. the MID), and mean change in a group is below that threshold. If the distribution of change with treatment is narrow, it is possible that no patient will achieve an important benefit with treatment. On the other hand, if the distribution of change is large, it is likely that a substantial number of patients may achieve a benefit.
Inferences at the group or population level are likely to be informative with respect to the decisions regarding health care policy and inferences at the level of an individual are most relevant to the decisions concerning the management of individual patients.
Regardless of the chosen perspective, investigators have used two easily separable strategies to achieve an understanding of the meaning of scores of a given instrument . The first relies on anchor-based methods and examines the relationship between scores on the target instrument (the instrument for which interpretation is in question) and some independent measure, termed an anchor. The FDA guidance refers to this strategy as "mapping". The second strategy is based on the statistical characteristics of the obtained PRO scores and is termed distribution-based. These later methods differ from anchor-based approaches in that they interpret results in terms of the relation between the magnitude of effect and some measure of variability in results.
For an in-depth review of these methods we refer the Readers to the work by Crosby  and Guyatt . Herein we will present only the most general aspects of anchor- or distribution-based approaches.
Anchor-based approaches to estimating a meaningful change in PRO measure
Anchor-based methods compare PRO measures to an anchor that is itself interpretable, i.e. has a known relevance to patients. For example, a global rating of change [21–24], status on an important and easily understood measure of function , the presence of symptoms , mean scores of patients with a particular diagnosis [27–30], disease severity , response to treatment [31, 32], or the prognosis of future events such as mortality [26, 33, 34], job loss [26, 35, 36] or health care utilization  can provide a useful anchor. Anchor-based methods require at least moderate correlation of the change on the anchor with the change on the target instrument.
One can subclassify anchor-based approaches into those that solve the interpretability problem in a single step – presenting population differences in status on multiple anchors – which one may call a population-focused approach, and those, that require two separate steps – first establishing the MID and then examining the proportion of patients who achieved the MID – which one may call an individual-focused approach.
The population-focused approach classifies patients in terms of the population to which they belong and is analogous to establishing construct validity, in that multiple anchors are generally required. In contrast, the individual patient-focused strategy tends to focus on a single anchor that is usually designed to establish a MID, but not necessarily so. This approach is analogous to criterion validity.
Those taking the individual patient-based approach usually attempt to identify a threshold between a change in score that is trivial and a change that is important (i.e. the MID). Those taking the population-based approach most commonly avoid identifying such a threshold, but offer relationships between target measure and multiple anchors instead, implicitly acknowledging that the threshold may vary, depending on the population under study and the range and severity of the problems being measured by the PRO instrument in question.
Having chosen a single-anchor approach, investigators may use alternative analytic strategies that will lead to different estimates of the MID . The simplest and so far most widely used approach is to specify a result or a range of anchor instrument results that correspond to the MID and calculate the target score matching that value. The commonly used alternative is the use of receiver operating characteristic curves adopted from diagnostic testing [39–41]. This strategy classifies each patient according to the anchor instrument as experiencing an important change or not experiencing such a change. Investigators then test a series of cut-off points to determine the number of misclassifications. These misclassifications correspond to false-positive results (patients mistakenly categorized as changed) and false-negative results (patients mistakenly categorized as unchanged). The optimal cut-off point will minimize the number of misclassifications.
Distribution-based approaches to estimating a meaningful change in PRO score
Distribution-based methods interpret results in terms of the relation between the magnitude of effect and some measure of variability in results. Three categories of distribution-based approaches can be recognized . The first approach depends on statistical significance and rates the score change in relation to the probability that this change is a result of a random variation of scores. Paired t-statistic  and growth curve analysis  are the examples. A second approach evaluates the score change in relation to sample variation: baseline standard deviation of patients [44, 45], variation of change scores , and variation of change scores in a stable group . The third approach evaluates the score change in relation to measurement precision. Examples include standard error of the mean  and a reliable change index . As a measure of variability, investigators may choose between-patient variability (for example, the standard deviation of patients at baseline) or within-patient variability (for example, the standard deviation of change in the PRO that patients experienced during a study).
The most widely used distribution-based method is the between-person standard deviation, often referred to as effect size [44, 45]. The group from which it is drawn is typically the control group at baseline or the pooled standard deviation of the experimental and control groups at baseline. Cohen  provided a rough rule of thumb to interpret the magnitude of the effect sizes. Changes in the range of 0.2 standard deviation units represent small changes, 0.5 – moderate changes, and 0.8 – large changes. Some recent empirical studies suggest that Cohen's guideline may in fact be generally applicable , but other authors propose that the MID is in the range of 0.2 to 0.5 standard deviation unit  or corresponds with an effect size of 0.5 [51, 52].
The advantage of distribution-based methods is that the values are easy to generate in contrast with the work needed to generate an anchor-based interpretation. These methods have two basic limitations: estimates of variability differ from study to study and there is no intuitive meaning of the effect size (standard deviation units).
How does the MID help to make sense of the results of clinical trials?
Describing the choice of the methods for interpretation of PRO instruments the authors of the FDA guidance addressed only the issue of deriving the MID leaving the issue of the very interpretation of clinical trial results based on these instruments unanswered. We have advocated that dichotomizing the results of a PRO measure facilitates interpretation of the clinical trial utilizing HRQL instruments [53, 54]. Considering the above described approaches to achieve meaningfulness of PRO scores it is evident that one does not have to estimate the MID to grasp the meaning of particular scores.
Dichotomizing the distribution of scores
We have argued that one possibility is the use of intuitive thresholds to interpret PRO scores. To facilitate interpretability of clinical trial results, researchers can report thresholds that either refer to an absolute score (e.g. one can consider patients above a certain score as having achieved the outcome) or a change in score (e.g. one can consider patients' PRO measure as having improved or deteriorated if they achieve a certain change in score). For the absolute score, while interpreting the results of a trial, one could consider the proportion of patients who achieve a given mean score for which anchors exist before and after an intervention. For the change score approach, one could consider the proportion of patients who have changed by a certain score, for instance of 10. Researchers may report the results as a categorized distribution of the proportion of patients who achieved certain improvement in PRO measure. We also argued that using the example of the SF-36 instrument from the Medical Outcomes Study , the proportion of patients who are able, according to scores on the Physical Function scale (range 0–100), to walk a distance of one block (approximately 100 meters) without difficulty would be 32% for a score of 40, 50% for a score of 50, and 79% for a score of 60. Increasing the score from 40 to 50 indicates that 18% more people state that they can walk without serious limitations, and increasing it from 50 to 60 – that 29% more can walk one block, etc. From the group perspective, one could interpret a score of 50 as corresponding to approximately 50% of patients being able to walk one block. From an individual patient perspective, a score of 50 indicates a 50% chance that the patient is able to walk one block. If an intervention improved this score to 60, there would now be a 79% chance, or a 29% increase, of this patient's ability to walk one block. This interpretation is based on the assumption that the patient has similar characteristics to the population from whom these values are obtained.
Another example for the use of content-based interpretation of PRO measures is the construction of interpretation aids. Valderas et al. applied a specific model of item response theory (IRT) to an instrument measuring perceived visual function, the Visual Function Index (VF-14) . This instrument asks respondents to rate the difficulties they have with their vision during performance of 14 everyday activities. Valderas et al. developed simple interpretation aids, that may facilitate the understanding of a particular score. The items were ordered according to their difficulty and used in the construction of a 'ruler' aid. This aid indicates the expected performance of an average patient with a given score. The authors have chosen a VF-14 score at which 50% of respondents have no difficulty performing a given task. For instance, a score of 97 indicates that 50% of respondents can drive without difficulty at night in regard to their visual function. A score of 75 indicates that 50% of respondents have no difficulty reading small print, 48 – watching TV and seeing steps, 36 – recognizing people when they are close, etc. Obviously, the authors could have chosen a score at which any other proportion of respondents has no difficulty performing a given task, but using a cut-off of 50% simplifies interpretation because it implies a 1 to 1 chance. This method of developing interpretation aids could be applied to many other PRO instruments. The important contribution of interpretation aids developed utilizing the IRT is that it informs clinicians and patients what performance they can expect based on a score on a multi-item instrument.
Irrespective of the strategy used to estimate the MID, from the individual patient point of view it is relevant to present the clinical trial results as the proportion of patients achieving a particular benefit (e.g. a MID, or any other value for that matter, be it a small, moderate, or large difference), instead of reporting only a mean difference. To calculate the proportion who achieved a MID, one must consider not only the difference between groups in those who achieve that improvement but also the difference between groups in those who deteriorate by the same amount. These differences can also be transformed into a number needed to treat required to achieve an MID in one patient after a given time period.
Estimation of an MID and interpretation of clinical trial results that present patient important outcomes is as demanding as it is vital in informing the decision to recommend or not to recommend or approve a given intervention. Investigators should be encouraged to use reliable and valid methods to achieve meaningfulness of their results, preferably those that rely on patients to estimate the MID. Ideally, the different approaches to estimating the MID will produce similar results. If they do not, this should be explicitly labelled. The FDA will have to provide more specific guidance than what is offered in the current document as to which methods and approaches are preferred. Clinical investigators will benefit from such advice, since it will let them avoid designing or selecting approaches that are likely not to be valid and, therefore, not accepted by the regulators. We hope that patient-based approaches will prevail as the perspective of the patients or their informed proxies for conditions that render patient decisions difficult (e.g. end of life decisions). At a minimum all approaches should be patient-driven and involve scenarios and vignettes, but not solely a clinician's judgment. We agree with the authors of the parallel comment that demonstrating responsiveness is a key component of demonstrating appropriate measurement properties an instrument . We believe the MID of a generic instrument, however, should not vary by population and context because it questions the use of the PRO measure as a generic instrument . In regards to reporting of PRO measures it is advisable that investigators report the proportion of patients achieving that benefit.
Federal Drug Administration: Guidance for Industry. Patient-Reported Outcome Measures: Use in Medical Product Development to Support Labeling Claims. [http://www.fda.gov/cder/guidance/5460dft.pdf]
Feinstein AR: Indexes of contrast and quantitative significance for comparisons of two groups. Stat Med 1999,18(19):2557–2581. Publisher Full Text 10.1002/(SICI)1097-0258(19991015)18:19<2557::AID-SIM361>3.0.CO;2-R
Naylor CD, Llewellyn-Thomas HA: Can there be a more patient-centred approach to determining clinically important effect sizes for randomized treatment trials? J Clin Epidemiol 1994,47(7):787–795. 10.1016/0895-4356(94)90176-7
Bobbio M, Demichelis B, Giustetto G: Completeness of reporting trial results: effect on physicians' willingness to prescribe. Lancet 1994,343(8907):1209–1211. 10.1016/S0140-6736(94)92407-4
Hux JE, Levinton CM, Naylor CD: Prescribing propensity: influence of life-expectancy gains and drug costs. J Gen Intern Med 1994,9(4):195–201.
Naylor CD, Chen E, Strauss B: Measured enthusiasm: does the method of reporting trial results alter perceptions of therapeutic effectiveness? Ann Int Med 1992,117(11):916–921.
Redelmeier DA, Tversky A: Discrepancy between medical decisions for individual patients and for groups. New Engl J Med 1990,322(16):1162–1164.
Guyatt GH, Feeny DH, Patrick DL: Measuring health-related quality of life. Ann Int Med 1993,118(8):622–629.
Schünemann HJ, Guyatt GH: Commentary – goodbye M(C)ID! Hello MID, where do you come from? Health Serv Res 2005,40(2):593–597. 10.1111/j.1475-6773.2005.0k375.x
Schünemann HJ, Puhan M, Goldstein R, Jaeschke R, Guyatt GH: Measurement Properties and Interpretability of the Chronic Respiratory Disease Questionnaire (CRQ). J COPD 2005, 2: 81–89.
Puhan MA, Behnke M, Devereaux PJ, Montori VM, Braendli O, Frey M, Schünemann HJ: Measurement of agreement on health-related quality of life changes in response to respiratory rehabilitation by patients and physicians – a prospective study. Respir Med 2004,98(12):1195–1202. 10.1016/j.rmed.2004.04.011
Sneeuw KC, Sprangers MA, Aaronson NK: The role of health care providers and significant others in evaluating the quality of life of patients with chronic disease. J Clin Epidemiol 2002,55(11):1130–1143. 10.1016/S0895-4356(02)00479-1
Ubel PA, Loewenstein G, Jepson C: Whose quality of life? A commentary exploring discrepancies between health state evaluations of patients and the general public. Qual Life Res 2003,12(6):599–607. 10.1023/A:1025119931010
von Essen L: Proxy ratings of patient quality of life – factors related to patient-proxy agreement. Acta oncologica (Stockholm, Sweden) 2004,43(3):229–234. 10.1080/02841860410029357
Devereaux PJ, Anderson DR, Gardner MJ, Putnam W, Flowerdew GJ, Brownell BF, Nagpal S, Cox JL: Differences between perspectives of physicians and patients on anticoagulation in patients with atrial fibrillation: observational study. BMJ 2001,323(7323):1218–1222. 10.1136/bmj.323.7323.1218
Santanello NC, Zhang J, Seidenberg B, Reiss TF, Barber BL: What are minimal important changes for asthma measures in a clinical trial? Eur Respir J 1999,14(1):23–27. 10.1034/j.1399-3003.1999.14a06.x
Guyatt GH, Osoba D, Wu AW, Wyrwich KW, Norman GR: Methods to explain the clinical significance of health status measures. Mayo Clin Proc 2002,77(4):371–383.
Lydick E, Epstein RS: Interpretation of quality of life changes. Qual Life Res 1993,2(3):221–226. 10.1007/BF00435226
Samsa G, Edelman D, Rothman ML, Williams GR, Lipscomb J, Matchar D: Determining clinically important differences in health status measures: a general approach with illustration to the Health Utilities Index Mark II. PharmacoEcon 1999,15(2):141–155. 10.2165/00019053-199915020-00003
Crosby RD, Kolotkin RL, Williams GR: Defining clinically meaningful change in health-related quality of life. J Clin Epidemiol 2003,56(5):395–407. 10.1016/S0895-4356(03)00044-1
Deyo RA, Inui TS: Toward clinical applications of health status measures: sensitivity of scales to clinically important changes. Health Serv Res 1984,19(3):275–289.
Jaeschke R, Singer J, Guyatt GH: Measurement of health status. Ascertaining the minimal clinically important difference. Contr Clin Trials 1989,10(4):407–415. 10.1016/0197-2456(89)90005-6
Juniper EF, Guyatt GH, Willan A, Griffith LE: Determining a minimal important change in a disease-specific Quality of Life Questionnaire. J Clin Epidemiol 1994,47(1):81–87. 10.1016/0895-4356(94)90036-1
Stucki G, Liang MH, Fossel AH, Katz JN: Relative responsiveness of condition-specific and generic health status measures in degenerative lumbar spinal stenosis. J Clin Epidemiol 1995,48(11):1369–1378. 10.1016/0895-4356(95)00054-2
Thompson MS, Read JL, Hutchings HC, Paterson M, Harris ED Jr: The cost effectiveness of auranofin: results of a randomized clinical trial. J Rheumatol 1988,15(1):35–42.
Ware JE, Keller SD: Interpreting general health measures. In Quality of Life and Pharmacoeconomics in Clinical Trials. Edited by: Spilker B. Philadelphia, Pa: Lippincott-Raven Publishers; 1996:445–460.
Brooks WB, Jordan JS, Divine GW, Smith KS, Neelon FA: The impact of psychologic factors on measurement of functional status. Assessment of the sickness impact profile. Med Care 1990,28(9):793–804. 10.1097/00005650-199009000-00009
Deyo RA, Inui TS, Leininger JD, Overman SS: Measuring functional outcomes in chronic disease: a comparison of traditional scales and a self-administered health status questionnaire in patients with rheumatoid arthritis. Med Care 1983,21(2):180–192. 10.1097/00005650-198302000-00006
Fletcher A, McLoone P, Bulpitt C: Quality of life on angina therapy: a randomised controlled trial of transdermal glyceryl trinitrate against placebo. Lancet 1988,2(8601):4–8. 10.1016/S0140-6736(88)92942-X
McSweeny AJ, Grant I, Heaton RK, Adams KM, Timms RM: Life quality of patients with chronic obstructive pulmonary disease. Arch Intern Med 1982,142(3):473–478. 10.1001/archinte.142.3.473
King MT: The interpretation of scores from the EORTC quality of life questionnaire QLQ-C30. Qual Life Res 1996,5(6):555–567. 10.1007/BF00439229
Bergner M, Bobbitt RA, Carter WB, Gilson BS: The Sickness Impact Profile: development and final revision of a health status measure. Med Care 1981,19(8):787–805. 10.1097/00005650-198108000-00001
Mossey JM, Shapiro E: Self-rated health: a predictor of mortality among the elderly. Am J Pub Health 1982,72(8):800–808.
Idler EL, Angel RJ: Self-rated health and mortality in the NHANES-I Epidemiologic Follow-up Study. Am J Pub Health 1990,80(4):446–452.
Brook RH, Ware JE Jr, Rogers WH, Keeler EB, Davies AR, Donald CA, Goldberg GA, Lohr KN, Masthay PC, Newhouse JP: Does free care improve adults' health? Results from a randomized controlled trial. New Engl J Med 1983,309(23):1426–1434.
Fayers PM, Machin D: Quality of life: assessment, analysis and interpretation. Chichester: John Wiley & Sons; 2000.
Ware JE Jr, Manning WG Jr, Duan N, Wells KB, Newhouse JP: Health status and the use of outpatient mental health services. Am Psychol 1984,39(10):1090–1100. 10.1037/0003-066X.39.10.1090
Brant R, Sutherland L, Hilsden R: Examining the minimum important difference. Stat Med 1999,18(19):2593–2603. Publisher Full Text 10.1002/(SICI)1097-0258(19991015)18:19<2593::AID-SIM392>3.0.CO;2-T
Deyo RA, Centor RM: Assessing the responsiveness of functional scales to clinical change: an analogy to diagnostic test performance. J Chron Dis 1986,39(11):897–906. 10.1016/0021-9681(86)90038-X
Stratford PW, Binkley JM, Riddle DL, Guyatt GH: Sensitivity to change of the Roland-Morris Back Pain Questionnaire: part 1. Phys Ther 1998,78(11):1186–1196.
Ward MM, Marx AS, Barry NN: Identification of clinically important changes in health status using receiver operating characteristic curves. J Clin Epidemiol 2000,53(3):279–284. 10.1016/S0895-4356(99)00140-7
Husted JA, Cook RJ, Farewell VT, Gladman DD: Methods for assessing responsiveness: a critical review and recommendations. J Clin Epidemiol 2000,53(5):459–468. 10.1016/S0895-4356(99)00206-1
Speer DC, Greenbaum PE: Five methods for computing significant individual client change and improvement rates: support for an individual growth curve approach. J Consul Clin Psychol 1995,63(6):1044–1048. 10.1037/0022-006X.63.6.1044
Cohen J: Statistical Power Analysis for the Behavioral Sciences. 2nd edition. Hillsdale, NJ: Lawrence Erlbaum Associates; 1988.
Kazis LE, Anderson JJ, Meenan RF: Effect sizes for interpreting changes in health status. Med Care 1989,27(3 Suppl):S178–189. 10.1097/00005650-198903001-00015
Guyatt GH, Bombardier C, Tugwell PX: Measuring disease-specific quality of life in clinical trials. CMAJ 1986,134(8):889–895.
Wyrwich KW, Tierney WM, Wolinsky FD: Further evidence supporting an SEM-based criterion for identifying meaningful intra-individual changes in health-related quality of life. J Clin Epidemiol 1999,52(9):861–873. 10.1016/S0895-4356(99)00071-2
Jacobson NS, Truax P: Clinical significance: a statistical approach to defining meaningful change in psychotherapy research. J Consul Clin Psychol 1991,59(1):12–19. 10.1037/0022-006X.59.1.12
Redelmeier DA, Guyatt GH, Goldstein RS: Assessing the minimal important difference in symptoms: a comparison of two techniques. J Clin Epidemiol 1996,49(11):1215–1219. 10.1016/S0895-4356(96)00206-5
Osoba D, Rodrigues G, Myles J, Zee B, Pater J: Interpreting the significance of changes in health-related quality-of-life scores. J Clin Oncol 1998,16(1):139–144.
Best WR, Becktel JM: The Crohn's disease activity index as a clinical instrument. In Developments in Gastroenterology: Recent Advances in Crohn's Disease. Edited by: Pena AS, Weterman IT, Booth C, Strober W. Dordrecht, the Netherlands: Martinus Nijhoff; 1981:7–12.
Redelmeier DA, Guyatt GH, Goldstein RS: On the debate over methods for estimating the clinically important difference. J Clin Epidemiol 1996,49(11):1223–1224. 10.1016/S0895-4356(96)00208-9
Schünemann HJ, Akl EA, Guyatt GH: Interpreting the Results of Patient Reported Outcome Measures in Clinical Trials: The Clinician's Perspective. Health Qual Life Outcomes 2006, 4: 62. 10.1186/1477-7525-4-62
Guyatt GH, Juniper EF, Walter SD, Griffith LE, Goldstein RS: Interpreting treatment effects in randomised trials. BMJ Clinical research ed 1998,316(7132):690–693.
Stewart AL, Greenfield S, Hays RD, Wells K, Rogers WH, Berry SD, McGlynn EA, Ware JE Jr: Functional status and well-being of patients with chronic conditions. Results from the Medical Outcomes Study. JAMA 1989,262(7):907–913. 10.1001/jama.262.7.907
Valderas JM, Alonso J, Prieto L, Espallargues M, Castells X: Content-based interpretation aids for health-related quality of life measures in clinical practice. An example for the visual function index (VF-14). Qual Life Res 2004,13(1):35–44. 10.1023/B:QURE.0000015298.09085.b0
Revicki DA, Cella D, Hays RD, Sloan JA, Lenderking WR, Aaronson NK: Responsiveness and minimal important differences for patient reported outcomes. Health Qual Life Outcomes 2006, 4: 70. 10.1186/1477-7525-4-70
HJS and GHG are authors of the CRQ. McMaster University and a research account used by HJS and GHG receive licensing fees from the use of the CRQ. There are no other competing interests related to this work.
JB and HJS developed an outline of this article based on many discussions with GG. JB wrote the first draft of the article and HJS and GG critically revised it.
About this article
Cite this article
Brożek, J.L., Guyatt, G.H. & Schünemann, H.J. How a well-grounded minimal important difference can enhance transparency of labelling claims and improve interpretation of a patient reported outcome measure. Health Qual Life Outcomes 4, 69 (2006). https://doi.org/10.1186/1477-7525-4-69