- Open Access
Using the Rasch measurement theory to assess the psychometric properties of the Hopkins Symptom Checklist-10 in adolescents
Health and Quality of Life Outcomes volume 19, Article number: 248 (2021)
The Hopkins Symptom Checklist-10 (HSCL-10) is widely used to measure psychological distress in adolescents. To provide valid and reliable results and generate recommendations for practice and policies, instruments with sound psychometric properties are required. The purpose of this study is to use Rasch measurement theory to assess the psychometric properties of the HSCL-10 among adolescents aged 13–19.
In this cross-sectional study, 6445 adolescents responded to a web-based questionnaire. Data were collected from lower and upper secondary schools in Norway during 2018. The data were analysed using the partial credit parameterisation of the unidimensional Rasch model.
HSCL-10 was found to be unidimensional and to have acceptable reliability. One pair of items showed response dependency. The targeting of the instrument could have been better. All items had ordered thresholds. Three items under-discriminated and three displayed differential item functioning regarding gender.
HSCL-10 has potential for measuring psychological distress in adolescents, though there is room for improvement. To further improve this instrument, some items should be rephrased.
Mental health problems among adolescents are a leading cause of disease burden worldwide  and an important public health issue in Norway and other Western countries. Studies have reported an increase in health complaints and mental health problems among young individuals over the last decades [2,3,4], with girls have reporting higher psychological distress than boys [5, 6]. Hence, it is important to identify those at risk of mental health problems to provide adequate services.
In epidemiological studies, questionnaires measuring depression and anxiety symptoms are frequently used to collect self-reported data about psychological distress. Several instruments have been developed for this purpose, such as the Hopkins Symptom Checklist (HSCL) [7, 8], Hospital Anxiety and Depression Scale  and the General Health Questionnaire .
Hopkins Symptom Checklist-10 (HSCL-10) is commonly used to measure psychological distress in many Western countries [8, 11, 12]. An early version of HSCL was developed in the 1950s  but has since undergone several revisions . HSCL was originally intended to examine the efficacy of psychotropic drugs and contained questions pertaining to symptomatic behaviour of outpatients in adult populations . This instrument is available in different versions of different lengths (5–90 items). Among young individuals, the short versions primarily have been used (5, 10 and 25 items). Earlier studies have shown that HSCL-10 is suitable for the identification of psychological distress in adolescents [13,14,15].
Generally, measurements with sound psychometric properties are important for clinical and research purposes . Previous psychometric examinations of HSCL instruments mainly included Cronbach’s alpha coefficient and factor analysis. One study has examined the psychometric properties of HSCL-10 using Rasch measurement theory (RMT) among Norwegian adolescents aged 15–16 . However, this study was based on data from 2001 and 2009, and the analyses were based on data from only one region of Norway. Another recent study among adolescents in Norway assessed the psychometric properties of a six-item depressive symptom scale that includes five items from the HSCL-10 . Generally, the Rasch model facilitates the disclosure of measurement problems that may not be easily detected by traditional analyses, such as lack of invariance, commonly called differential item functioning (DIF). As adolescents’ mental health may have changed over the last decade and there might be differences in how well the instrument is suited for adolescents at different ages, there is a need to update knowledge about the psychometric properties of the HSCL-10. Therefore, the aim of this study is to use RMT to assess the psychometric properties of HSCL-10 among adolescents aged 13–19.
Data collection and study population
This study was based on Ungdata from 2018. Ungdata is a Norwegian national cross-sectional survey collecting data annually among adolescents in lower and upper secondary schools. The results from the Ungdata survey are frequently reported in the media and used when providing recommendations for practice and policies. All data collection is performed by Norwegian Social Research at Oslo Metropolitan University, in cooperation with all regional drug and alcohol competence centres. This survey is partially financed by the Norwegian Directorate of Health. All participants completed an anonymous web-based questionnaire at school. A teacher was present during data collection to help the participants if they had any questions. All parents were informed about the study in advance. Parents of adolescents aged 13–17 were informed that they can withdraw their children from the study at any time. All adolescents were informed that participation was voluntary. This study was ethically approved by the Norwegian Centre for Research Data .
Ungdata covers areas such as mental and physical health, relationships with peers and parents, drug use, leisure time activities and nutrition. The survey consists of a main module that all the respondents are supposed to answer, along with optional modules that the municipality has an opportunity to choose. Some of the items of the HSCL-10 are included in the main module, whereas others only in the optional one.
In this study, 6445 participants, aged 13–19, who answered all the HSCL-10 questions, were included. The analysis was, therefore, based on complete data. Six of the questions were from the main module, and four were from one of the optional modules.
In total, 73% of the respondents (aged 13–16) were recruited from lower secondary schools (Table 1). The sample comprised an approximately equal proportion of males and females.
Hopkins Symptom Checklist-10
HSCL-10 is a short version of the Hopkins Symptom Checklist-25 that has been developed to measure symptoms of anxiety and depression [8, 19]. It consists of 10 items in which adolescents are asked whether during the previous week they had any symptoms of anxiety or depression (Table 2). All items have four response categories: ‘Not been affected at all’ (1), ‘Not been affected much’ (2), ‘Been affected quite a lot’ (3) and ‘Been affected a great deal’ (4). Higher scores indicate higher levels of psychological distress .
Rasch measurement theory
In this study, the psychometric properties of HSCL-10 were analysed using the partial credit parameterisation  of the unidimensional Rasch model . If data from an instrument fit the unidimensional Rasch model, it is statistically defensible to sum the responses to each single item to a total score for each person . At a general level, dimensionality, response dependency and targeting were examined. Dimensionality was examined using t-tests on person estimates of items intended to measure depression and anxiety symptoms. A residual correlation between two items of < 0.3 was used as an indicator of response dependency between the two items (i.e. the answer of one item is dependent on the answer of another). In addition, residual correlations of items were assessed relative to each other .
Reliability was assessed using the Person Separation Index (PSI), which is based on non-linear transformation of raw scores and indicates a scale’s ability to differentiate individuals along the latent trait. The PSI is analogous to Cronbach’s alpha . Targeting indicates how well a scale captures the person estimates. A scale is considered well targeted if the mean person location values are around zero .
Analyses at a finer level included item fit, ordering of response categories and DIF. Chi-square statistics and standardised residuals based on comparisons between observed and expected values were used to analyse item fit. Chi-square probability values above Bonferroni’s adjusted 5% and fit residuals in the range ± 2.5 indicate adequate item fit . Item characteristic curves (ICCs) were inspected to assess item fit graphically.
Significance tests like chi-square are sensitive to sample size. Since this study included a rather large sample size, there was a risk of drawing false conclusions . The amend sample size function in RUMM was, therefore, used to draw a random sub-sample for further analyses concerning item fit and DIF. As recommended, sample size is calculated by multiplying the number of items (10) by the number of thresholds (3) with 30 persons per threshold , yielding a sample size of 900 (10 × 3 × 30), which can be deemed as adequate in these analyses.
A central requirement of measurement is that every item should work invariantly across levels of different person factors, such as gender and school level. To examine DIF, two-way analysis of variance of standardised residuals was used . Statistical significance was assumed at a Bonferroni-adjusted 5%, and graphical displays (i.e. ICCs) were used. DIF analyses were performed for the person factors gender, school level and grade. DIF can be handled either by resolving, deleting the item, or ignoring the DIF . Resolving DIF implies splitting the item into e.g. gender-specific items and treating the opposite person factor category as a non-response. To distinguish real from artificial DIF, the items were sequentially resolved, starting with the item having the highest F-value . Additionally, the mean person estimates, chi-square values and PSI were compared before and after splitting the items [26, 30]. Response categories were found to be ordered if the thresholds were significantly different and in the right order . All analyses were performed using the software RUMM2030Plus , that handles missing data through FIML (full information maximum likelihood).
The HSCL-10 items formed a unidimensional scale (the proportion of significant t-tests of the difference in person–location estimates between subsets of items was 3.48%) with acceptable reliability (PSI = 0.823, Cronbach’s alpha = 0.913). Items 1 and 2 showed evidence of response dependency (residual correlation = 0.42). Four other pairs of items also showed a positive residual correlation (items 7 and 8, 3 and 5, 5 and 8, and 7 and 10), but the value was 0.1 or lower. All other residual correlations were negative.
Comparing the distribution of person estimates to the item threshold estimates centred around zero revealed that the person thresholds showed a skewed distribution, with the main weight on the left. Hence, the instrument can be considered somewhat out of target as the mean person location was − 1.475 (Fig. 1). The person–item threshold distribution also indicated that the instrument was better targeted at females (mean value = − 0.937) than at males (mean value = − 2.029). The targeting was also better for adolescents in upper secondary schools (mean value = − 1.241) than for those in lower secondary schools (mean value = − 1.561).
Given the sample size, statistically significant chi-square values were observed for all items (Table 2).
Three items under-discriminated (items 3, 4 and 6), and seven over-discriminated (items 2, 5, 7, 8, 9 and 10; Table 2). When the sample size was reduced to 900, only items 6 and 7 had significant chi-square values. This was also valid when analysing respondents from lower and upper secondary schools separately. The graphical presentation of the observed compared to the expected values for item 6 shows that the curve is flatter than expected and that the item tends to under-discriminate (Fig. 2).
For all items the thresholds separating the categories were significantly different and in the correct order, indicating that the response categories worked well (Table 3).
Differential item functioning
When the sample size was reduced to 900, three items (items 2, 5 and 6) displayed DIF regarding gender. Item 6 had the greatest magnitude of gender DIF (F-value = 18.7). Figure 3 shows the DIF for the person factor gender for items 2, 5 and 6.
Females tended to score higher than males on items 2 and 5, despite the same location on the latent trait. Opposite results were found for item 6, with males scoring higher than females. Parallel ICCs for males and females indicate evidence of uniform DIF (Fig. 3). For the other person factors available, none of the items displayed DIF when the sample size was adjusted. When data from lower and upper secondary schools were analysed separately, similar results were found based on data from lower secondary schools, whereas only items 2 and 6 displayed gender DIF in data from upper secondary schools. Reducing sample size to 900 no items displayed DIF when it comes to school level or grade.
Gender DIF was resolved by first splitting item 6 (highest F-value) into two separate items, one for males and one for females (Table 4). After item 6 was split the PSI remained approximately the same (0.826). Given the adjusted sample size of 900, item 2, but not item 5, still displayed significant DIF. The same procedure was repeated for item 2, which did not affect the PSI. When items 2 and 6 were split, none of the other items displayed DIF.
When items 2 and 6 were resolved, the difference between the two genders was found to be 0.053 logit bigger than the magnitude of 1.092 in the original set of 10 items, a change of 4.9%. When only item 6 was resolved, the difference between the two genders was 0.076 logit bigger than the magnitude of 1.092 in the original set of 10 items. When item 2 was excluded, a PSI of 0.813 and a Cronbach’s alpha of 0.905 were found. Excluding items 2 and 6 resulted in a PSI of 0.795 and a Cronbach’s alpha of 0.90.
HSCL-10 shows satisfactory psychometric properties at an overall level. It was found to be unidimensional and to have acceptable reliability. However, HSCL-10 shows some weaknesses at a finer level related to response dependency and DIF.
The results indicate that HSCL-10 is somewhat out of target, meaning that adolescents may have better psychological health than the instrument can measure. This was expected because this instrument was originally developed for clinical purposes but was applied in this study on a healthy population. These results are in line with those of Kleppang and Hagquist  who nevertheless found that the instrument was even more off target. However, their study was based on data from 2001 and 2009. Since psychological distress has increased among adolescents over the last decade, the instrument is expected to be better targeted at today’s adolescents. In line with the results of Kleppang and Hagquist , the results indicate better targeting for females than for males, which might be due to females reporting more psychological complaints than males [6, 32].
Targeting is not considered a problem for this instrument if the intention is to measure psychological distress. However, if the intention is to measure the broader concept of mental health, questions assessing better, or positive mental health should be included. Bad targeting might imply decreased reliability , meaning that reliability can be strengthened if the instrument is applied to clinical samples.
Dimensionality and response dependency
Despite consisting of items intending to measure both symptoms of depression and anxiety, HSCL-10 was found to be unidimensional, although multi-dimensionality was expected from theoretical and medical perspectives. However, depression often co-occurs with anxiety in adolescents [33, 34]. Therefore, when measuring symptoms of depression in adolescents, indicators of anxiety should probably be included.
Contrary to the findings of Kleppang and Hagquist , response dependency was found between items 1 and 2, indicating the items to collect redundant information. Several items over-discriminated, including item 2, which might strengthen the evidence of dependency . As revealing response dependency may be dependent on the number of items, more item pairs are expected to show response dependency from more extended versions of the HSCL. As items 1 and 2 can be deemed as collecting redundant information, item 2 can be excluded as this also over-discriminates and displays DIF. A small decrease in PSI and Cronbach’s alpha was found when item 2 was excluded, which may strengthen the hypothesis that response dependency inflates reliability indexes. However, excluding item 2, which pertains to anxiety, creates a conceptually unbalanced construct. Response dependency may also result from translation. In Norwegian, ‘scared’ and ‘fear’ translate to one word close to ‘afraid’. Adolescents may not distinguish between being suddenly scared and feeling fearful. Therefore, the presence of response dependency should be investigated in other languages or translations of HSCL-10.
Item fit and differential item functioning
In line with Kleppang and Hagquist , we found that item 6 under-discriminated. Under-discriminating items tends to measure something else not correlated with the latent trait. Sleep problems and sleeplessness may be related to psychological distress. Schmalbach et al.  found a moderately high correlation between the HSCL and the Jenkins Sleep Scale. However, there might be other reasons for sleep problems than psychological distress, such as gaming and being active on social media late at night. The reason for the under-discrimination of this item might also be due to a translation error. The original wording of this item is ‘difficulties in falling asleep or staying asleep’. However, in Norwegian, this translates to ‘sleep problems’, which may be regarded as quite imprecise.
Item 6 also displayed gender DIF, in line with a previous study . In our case, the source of DIF might be explained by gender differences in sleep problems. Social jetlag, sleep deficiency [35, 36] and poor sleep quality  have been found to affect females more than males.
Deleting an item may improve the model fit and retain the invariance, but an important aspect of psychological distress may be lost, which may impact the scale’s validity and reliability . When measurement is constructed, items are selected given their relevance as well as representativeness . When resolving an item, the aspect is retained. However, resolving the item bring that the measurement is not invariant among genders, as it will have different difficulties in males and females . To decide whether to resolve DIF additional information about the sources of DIF may be required if the concept in question is not clearly defined .
Our results also showed that real DIF in items 2 and 6 affects the person measurement. However, only minor changes in group differences were observed.
Hence, whether an item should be excluded or resolved may be a trade-off between model fit and invariance. If the source of DIF is irrelevant to the variable, such as poor translation, resolving may be justified. When the source of DIF is relevant to the variable, it is not recommended to resolve the DIF as this might worsen the validity [28, 30]. When comparing psychological distress scores across genders, resolving the DIF will have the same result as deleting the item. However, regardless of the source of DIF, invariance should take precedence over fit and the item should not be resolved .
The same consideration described above for ‘sleeplessness’ should be adopted for item 2 (‘felt constant fear and anxiety’). The source of DIF for this item may be the result of females recognising the content differently from males. However, the source of DIF is unknown, and further research is encouraged to reveal potential sources of DIF for this item.
Strengths and limitations
This study comprises a large sample size, covers data from the whole country and provides an updated description of adolescents’ responses to HSCL-10. Items were selected from different parts of the questionnaire, some mandatory and some optional, which may influence the response pattern. We also found a discrepancy between the Norwegian wording and the original version of HSCL-10. Our analyses were based on self-reported data. Hence, there is a risk for potential response bias (e.g. social desirability), such as presenting favourable images of themselves. However, considering the large sample size of this study and that the questionnaire was fulfilled anonymously the potential random errors might be minimized .
The aim of our study was to assess the psychometric properties for measuring psychological distress in a general population of adolescents. As the HSCL-10 originally was developed for clinical purposes, assessing its clinimetric properties could also be appropriate. However, psychometric and clinimetric assessments could to a large extent be considered to overlap [40, 41]. Nevertheless, future studies should assess the instrument's utility and scalability in clinical practice . Information about respondents’ health status was not available in our study. Hence, future studies should also assess to what extent the items are invariant for people with varying degrees of psychological distress and how well the items discriminate between individuals with different mental health conditions.
HSCL-10 seems to be a suitable instrument for measuring psychological distress in adolescents. The potential weaknesses discussed here may be related to the imprecise translations of its items. We therefore suggest revising the wording of the Norwegian version to make it more in line with the original version. The instrument’s psychometric properties should also be assessed in other languages. To evaluate its properties in clinical practice clinimetric assessment could also be appropriate. HSCL-10 assesses psychological distress and is not a measure for general mental health. It is important that research employing this instrument uses the right concept. Mental health definitions are considered wider than those of psychological distress.
Availability of data and materials
Availability of data and materials in the Ungdata surveys are included in a national database administered by Norwegian Social Research (NOVA). Data is available for research purposes upon application. Information on the questionnaires can also be found from the web page (in Norwegian) (http://ungdata.no/).
Differential item functioning
Hopkins Symptom Checklist-10
Person Separation Index
Rasch measurement theory
Patton GC, Sawyer SM, Santelli JS, Ross DA, Afifi R, Allen NB, et al. Our future: a Lancet commission on adolescent health and wellbeing. Lancet. 2016;387(10036):2423–78.
Duffy ME, Twenge JM, Joiner TE. Trends in mood and anxiety symptoms and suicide-related outcomes among US undergraduates, 2007–2018: evidence from two national surveys. J Adolesc Health. 2019;65(5):590–8.
Potrebny T, Wiium N, Haugstvedt A, Sollesnes R, Torsheim T, Wold B, et al. Health complaints among adolescents in Norway: a twenty-year perspective on trends. PLoS ONE. 2019;14(1):e0210509.
Bor W, Dean AJ, Najman J, Hayatbakhsh R. Are child and adolescent mental health problems increasing in the 21st century? A systematic review. Aust N Z J Psychiatry. 2014;48(7):606–16.
Hoare E, Milton K, Foster C, Allender S. Depression, psychological distress and Internet use among community-based Australian adolescents: a cross-sectional study. BMC Public Health. 2017;17(1):365.
Moen ØL, Hall-Lord ML. Adolescents’ mental health, help seeking and service use and parents’ perception of family functioning. Nord J Nurs Res. 2019;39(1):1–8.
Parloff MB, Kelman HC, Frank JD. Comfort, effectiveness, and self-awareness as criteria of improvement in psychotherapy. Am J Psychiatry. 1954;111(5):343–52.
Derogatis LR, Lipman RS, Rickels K, Uhlenhuth EH, Covi L. The Hopkins Symptom Checklist (HSCL): a self-report symptom inventory. Behav Sci. 1974;19(1):1–15.
Zigmond AS, Snaith RP. The hospital anxiety and depression scale. Acta Psychiatr Scand. 1983;67(6):361–70.
French DJ, Tait RJ. Measurement invariance in the General Health Questionnaire-12 in young Australian adolescents. Eur Child Adolesc Psychiatry. 2004;13(1):1–7.
Schmalbach B, Zenger M, Tibubos AN, Kliem S, Petrowski K, Brähler E. Psychometric properties of two brief versions of the Hopkins Symptom Checklist: HSCL-5 and HSCL-10. Assessment. 2021;28(2):617–31.
Sirpal MK, Haugen W, Sparle K, Haavet OR. Validation study of HSCL-10, HSCL-6, WHO-5 and 3-key questions in 14–16 year ethnic minority adolescents. BMC Fam Pract. 2016;17(1):1–5.
Haavet OR, Sirpal MK, Haugen W, Christensen KS. Diagnosis of depressed young people in primary health care—a validation of HSCL-10. Fam Pract. 2011;28(2):233–7.
Kleppang AL, Hagquist C. The psychometric properties of the Hopkins Symptom Checklist-10: a Rasch analysis based on adolescent data from Norway. Fam Pract. 2016;33(6):740–5.
Søgaard AJ, Bjelland I, Tell GS, Røysamb E. A comparison of the CONOR Mental Health Index to the HSCL-10 and HADS. Nor Epidemiol. 2003;13(2):279–84.
Andrich D. Rasch models for measurement. Newbury Park: SAGE; 1988.
Kleppang AL, Steigen AM, Finbråten HS. Using Rasch measurement theory to assess the psychometric properties of a depressive symptoms scale in Norwegian adolescents. Health Qual Life Outcomes. 2020;18:1–8.
NOVA. Ungdata. 2019 [11.02.2019]. Available from http://www.ungdata.no/English.
Strand BH, Dalgard OS, Tambs K, Rognerud M. Measuring the mental health status of the Norwegian population: a comparison of the instruments SCL-25, SCL-10, SCL-5 and MHI-5 (SF-36). Nord J Psychiatry. 2003;57(2):113–8.
Masters GN. A Rasch model for partial credit scoring. Psychometrika. 1982;47(2):149–74.
Rasch G. Probabilistic models for some intelligence and attainment tests. Expanded. Chichago: The University of Chicago Press; 1960/80.
Marais I. Local dependence. In: Christensen KB, Kreiner S, Mesbah M, editors. Rasch models in health. Hoboken: Wiley; 2013. p. 111–30.
Hagquist C, Bruce M, Gustavsson JP. Using the Rasch model in nursing research: an introduction and illustrative example. Int J Nurs Stud. 2009;46(3):380–93.
Tennant A, Conaghan PG. The Rasch measurement model in rheumatology: what is it and why use it? When should it be applied, and what should one look for in a Rasch paper? Arthritis Care Res. 2007;57(8):1358–62.
Lantz B. The large sample size fallacy. Scand J Caring Sci. 2013;27(2):487–92.
Hagquist C, Andrich D. Recent advances in analysis of differential item functioning in health research using the Rasch model. Health Qual Life Outcomes. 2017;15(1):1–8.
Andrich D, Marais I. A course in Rasch measurement theory. Measuring in the educational, social and health sciences. Singapore: Springer; 2019.
El Masri YH, Andrich D. The trade-off between model fit, invariance, and validity: the case of PISA science assessments. Appl Meas Educ. 2020;33(2):174–88.
Andrich D, Hagquist C. Real and artificial differential item functioning in polytomous items. Educ Psychol Meas. 2014;75(2):185–207.
Hagquist C. Explaining differential item functioning focusing on the crucial role of external information—an example from the measurement of adolescent mental health. BMC Med Res Methodol. 2019;19(1):185.
Andrich D, Sheridan B. RUMM2030Plus. Duncraig: Rumm Laboratory Pty Ltd; 2019.
Hoare E, Milton K, Foster C, Allender S. Depression, psychological distress and Internet use among community-based Australian adolescents: a cross-sectional study. BMC Public Health. 2017;17(1):1–10.
Long EE, Young JF, Hankin BL. Temporal dynamics and longitudinal co-occurrence of depression and different anxiety syndromes in youth: evidence for reciprocal patterns in a 3-year prospective study. J Affect Disord. 2018;234:20–7.
Seligman LD, Ollendick TH. Comorbidity of anxiety and depression in children and adolescents: an integrative review. Clin Child Fam Psychol Rev. 1998;1(2):125–44.
Henderson SE, Brady EM, Robertson N. Associations between social jetlag and mental health in young people: a systematic review. Chronobiol Int. 2019;36(10):1316–33.
Hysing M, Pallesen S, Stormark KM, Lundervold AJ, Sivertsen B. Sleep patterns and insomnia among adolescents: a population-based study. J Sleep Res. 2013;22(5):549–56.
Galland BC, Gray AR, Penno J, Smith C, Lobb C, Taylor RW. Gender differences in sleep hygiene practices and sleep quality in New Zealand adolescents aged 15 to 17 years. Sleep Health. 2017;3(2):77–83.
Messick S. Validity of psychological assessment: validation of inferences from persons’ responses and performances as scientific inquiry into score meaning. Am Psychol. 1995;50(9):741.
Polit DF, Beck CT. Nursing research: generating and assessing evidence for nursing practice. Philadelphia: Wolters Kluwer Health. Lippincott Williams & Wilkins; 2012.
Kaiser T, Herzog P, Lutz W, Rief W, Brakemeier E-L, Zimmermann J. The distinction between psychometric and clinimetric approaches will not move the field forward. PsyArXiv. 2021 (preprint).
De Vet HC, Terwee CB, Bouter LM. Clinimetrics and psychometrics: two sides of the same coin. J Clin Epidemiol. 2003;56(12):1146.
Carrozzino D, Patierno C, Guidi J, Montiel CB, Cao J, Charlson ME, et al. Clinimetric criteria for patient-reported outcome measures. Psychother Psychosom. 2021;90(4):222–232.
The authors wish to thank the NOVA institute for their cooperation and for undertaking the data collection.
This study received no funding.
Ethics approval and consent to participate
Data were obtained from an already established data material (Ungdata survey) collected by Norwegian Social Research. The data collection was based on informed consent. Norwegian Centre for Research Data (NSD) and Norwegian Social Research institute (NOVA) have assessed that the data collection is in line with the privacy protections and regulations. Permission to access and use the data were given by NOVA at OsloMet .
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Finbråten, H.S., Kleppang, A.L. & Steigen, A.M. Using the Rasch measurement theory to assess the psychometric properties of the Hopkins Symptom Checklist-10 in adolescents. Health Qual Life Outcomes 19, 248 (2021). https://doi.org/10.1186/s12955-021-01884-9
- Hopkins Symptom Checklist
- Psychological distress
- Rasch measurement theory