- Research article
- Open Access
- Open Peer Review
Measurement properties of the Health Literacy Questionnaire (HLQ) among older adults who present to the emergency department after a fall: a Rasch analysis
BMC Health Services Researchvolume 17, Article number: 605 (2017)
Health literacy is an important concept associated with participation in preventive health initiatives, such as falls prevention programs. A comprehensive health literacy measurement tool, appropriate for this population, is required. The aim of this study was to evaluate the measurement properties of the Health Literacy Questionnaire (HLQ) in a cohort of older adults who presented to a hospital emergency department (ED) after a fall.
Older adults who presented to an ED after a fall had their health literacy assessed using the HLQ (n = 433). Data were collected as part of a multi-centre randomised controlled trial of a falls prevention program. Measurement properties of the HLQ were assessed using Rasch analysis.
All nine scales of the HLQ were unidimensional, with good internal consistency reliability. No item bias was found for most items (43 of 44). A degree of overall misfit to the Rasch model was evident for six of the nine HLQ scales. The majority of misfit indicated content overlap between some items and does not compromise measurement. A measurement gap was identified for this cohort at mid to high HLQ score.
The HLQ demonstrated good measurement properties in a cohort of older adults who presented to an ED after a fall. The summation of the HLQ items within each scale, providing unbiased information on nine separate areas of health literacy, is supported. Clinicians, researchers and policy makers may have confidence using the HLQ scale scores to gain information about health literacy in older people presenting to the ED after a fall.
This study was registered with the Australian New Zealand Clinical Trials Registry, number ACTRN12614000336684 (27 March 2014).
Falls represent the main cause of emergency department (ED) presentations for older adults . However, participation in falls prevention activities following presentation to the ED with a fall is suboptimal . Health literacy is an important concept associated with participation in preventive health initiatives . Health literacy is defined as “the cognitive and social skills which determine the motivation and ability of individuals to gain access to, understand and use information in ways which promote and maintain good health” .
Adults with sub-optimal health literacy are less likely to participate in preventive health programs, such as falls prevention programs, possibly due to lack of understanding of health information and education provided . Accurate measurement of health literacy prior to commencing a falls prevention program may guide clinicians to adapt provider-patient communication, such as provision of information related to falls risks and their management strategies, to match the patient’s level of health literacy. This may lead to increased participation in falls prevention activities, potentially resulting in improved outcomes for these individuals.
A range of health literacy measurement tools are available. However, most tools do not reflect the multidimensional definition of health literacy, and predominantly focus on reading comprehension, pronunciation and numeracy [6, 7]. The Health Literacy Questionnaire (HLQ) was developed to address the shortcomings of previous tools . The HLQ comprises nine independent scales related to the understanding of, engagement with, and use of health services, from both an individual and organisational perspective.
The measurement properties of the HLQ have been explored in depth using predominantly classical test theory (CTT) approaches [8,9,10,11] and qualitative approaches [8, 12]. The HLQ was originally validated using a sample from clinical, home and community care settings in Australia . A highly restrictive 9-factor confirmatory factor analysis (CFA) model fitted satisfactory, with each of the HLQ scales representing nine conceptually distinct areas of health literacy. Subsequent studies evaluating the psychometric properties of the HLQ, including German, Danish, and Slovakian versions, support these findings, with the HLQ demonstrating good model fit and reliability, as well as homogeneity of items within each of the HLQ scales [9,10,11, 13]. Diverse cohorts were used in these studies representing people with a range of health conditions, receiving a variety of health services. A recent study evaluated the measurement properties of the initial version of the HLQ among people at risk of cardiovascular disease, using Rasch methods . Similar to previous studies, each of the nine HLQ scales were found to measure nine separate constructs of health literacy with good internal consistency. Unclear distinction between some response categories in some HLQ scales was reported and the scales were deemed to be suboptimally targeted in relation to the particular cardiovascular cohort . With the HLQ version used in this study, some disordered thresholds among items in scales 6 to 9 were observed. Kolarcik et al. observed this effect as well and subsequently improved the response options which resulted in lower scores (better targeting), and improved model fit, with no disordered thresholds .
Rasch analysis is a modern and unique form of item response theory (IRT) . It involves testing an outcome scale against a mathematical model that operationalises the key principles of good measurement [15,16,17]. Rasch analysis allows for a unified approach to evaluating several measurement issues, such as unidimensionality, local dependency, response category ordering, item bias and targeting, producing rich data that complements and adds to CTT approaches [15,16,17,18]. Rasch analysis is widely accepted as the standard for modern psychometric evaluations of outcome scales [15, 19]. As such, this methodology was deemed to be the most appropriate for this study.
Previous studies provide robust evidence to guide the practical use of the HLQ among a variety of international community and clinical populations. However, the measurement properties of the HLQ have not previously been determined for older adults who have presented to an ED after a fall. The appropriateness of a tool may vary across settings, therefore it is imperative to analyse the HLQ in specific populations prior to applying the tool and interpreting scores [8, 12]. The aim of this study was to use Rasch methods to evaluate the measurement properties of the HLQ in a cohort of older adults who presented to a hospital ED after a fall.
This study was embedded within a multi-centre randomised controlled trial (RCT) of a patient-centred falls prevention program: RESPOND. RESPOND incorporates (1) a home-based assessment; (2) education, goal setting and telephone coaching for management of selected falls risk factors; and (3) healthcare provider communication and community linkage, delivered over 6 months . Ethical approval was obtained from Alfred Health (HREC 439/13) and Royal Perth Hospital (REG 13–128), Monash University Human Research Ethics Committee (HREC) (MUHREC CF13/3869–2013001975) and Curtin University HREC (HR 43/ 2014).
Participants and setting
Adults aged between 60 and 90 years who presented at two Australian EDs with a fall, and had a planned discharge home within 72 h, were eligible to participate in the RESPOND trial . Exclusion criteria were: current palliative care or terminal illness, requiring hands-on assistance to walk, needing an interpreter, a history of psychoses or social aggression, and cognitive impairment (Mini Mental State Examination (MMSE) <23) . A total of 438 patients were recruited to the RESPOND RCT and completed the HLQ. Of these participants, five withdrew prior to completion of the trial. Data from the remaining 433 participants were used for this study.
Demographic data were collected by members of the research team at the screening and recruitment phase at the participating hospitals, and the initial face-to-face assessment conducted at the participant’s home. The home visit was planned to occur within two weeks of discharge from hospital . The HLQ was self-administered by the participant either prior to or during the home visit.
The health literacy questionnaire (HLQ)
The HLQ comprises 44 items over nine independent scales, each representing a different element of the overall health literacy construct: (1) Feeling understood and supported by healthcare providers; (2) Having sufficient information to manage my health; (3) Actively managing my health; (4) Social support for health; (5) Appraisal of health information; (6) Ability to actively engage with healthcare providers; (7) Navigating the healthcare system; (8) Ability to find good health information; and (9) Understanding health information well enough to know what to do. There are four to six items in each scale. Depending upon the purpose of inquiry, the full instrument or selected scales can be used. The first five scales comprise items that ask the respondents to indicate their level of agreement on one of four response options (strongly disagree to strongly agree). The remaining scales (6–9) represent scales of self-reported capability and items within these scales are scored on one of five response options (cannot do; very difficult; quite difficult; quite easy; very easy). The full HLQ provides nine individual scores based on an average of the items within each of the nine scales. There is no overall total score for the HLQ as that could potentially mask individual needs in specific health literacy domains .
Socio-economic status (SES) was measured using The Index of Relative Socio-economic Advantage and Disadvantage (IRSAD) , a reliable and robust approach to assessing socio-economic status . Data are based on participant postcodes and take into consideration socio-economic factors such as income, education, employment, occupation and housing . The 20% most advantaged, according to their IRSAD score, were considered to be a relatively high socio-economic group for the purpose of this study. The remaining participants were combined into a second group representing lower socio-economic status.
Whether or not participants have private health insurance or live alone were self-report questions answered yes/no at the time of the initial face-to-face assessment. Falls risk status was measured at the face-to-face interview using a reliable assessment tool: the Falls Risk for Older People – Community setting (FROP-Com) . A FROP-Com score > 18 represented high falls risk .
Descriptive statistics were used to profile the cohort using SPSS v22.0 (IBM Corporation, Armonk, New York). Rasch analysis was conducted using the partial credit model, as this allows the thresholds to vary for each of the individual items , using RUMM2030 software (RUMM Laboratory Pty Ltd., Perth, Australia). In order to determine whether the HLQ scales fit the Rasch model, response patterns to HLQ items were evaluated against the model’s expectations . Three statistics were considered to determine the degree of fit for each HLQ scale: overall fit; individual person fit; and individual item fit . Adequate overall fit of the HLQ to the Rasch model was indicated by a non-significant Bonferroni adjusted Chi-square probability value  (p ≥ 0.0125 for four item scales (1 and 2); p ≥ 0.01 for five item scales (3, 4, 5, 6, 8 and 9); p ≥ 0.0083 for the six item scale (7)). Satisfactory overall item and individual fit for each scale was determined by a fit residual standard deviation (SD) value of ≤1.5 .
Individual items were further analysed to determine whether or not each of the four to six items comprising the nine HLQ scales fit the Rasch model requirements. Individual item fit was indicated by two statistics: fit residual values; and Chi-square probability values . Item fit residual values −2.5 to 2.5 indicated adequate fit . Above this range (underfit) suggests deviation from the model, below (overfit) suggests that some items in the scale are similar to each other . Consistent with overall fit, a non-significant Bonferroni adjusted Chi-square probability value (p > 0.0125 for scales 1 and 2; p > 0.01 for scales 3, 4, 5, 6, 8, and 9; and p > 0.0083 for scale 7) indicated adequate item fit .
In addition to model fit the following measurement properties were analysed: unidimensionality; internal consistency reliability; response format; item bias; and targeting. Measurement properties analysed, their definitions, statistical tests used and criteria for assessment are summarised in Table 1.
The mean age of participants was 73 years, 55% were female, and 42% of participants lived alone. Most had private health insurance (61%), and most were of high SES (62%). Approximately one third (34%) were classified as being at high risk of falls. Participant characteristics and HLQ scores are presented in Table 2.
Three of the nine scales: (5) Appraisal of health information; (8) Ability to find good health information; and (9) Understanding health information well enough to know what to do -demonstrated adequate overall fit to the Rasch model as indicated by a non-significant Bonferroni adjusted Chi-square probability value (p = 0.33; p = 0.02; p = 0.05 respectively) (Table 3). The remaining scales demonstrated some degree of misfit between the data and the Rasch model (scales 1 and 2 p < 0.0125; scales 3, 4 and 6 p < 0.01; scale 7 p < 0.0083). The majority of item misfit, as determined by a negative item fit residual value below −2.5 (17 items), suggested overfit (Table 4). A further seven items (one item from each of scales 1, 2, 3, 4, 6, 7, and 8) demonstrated underfit with a Chi-square probability below the adjusted alpha value (scale 1 and 2 p < 0.0125; scales 3, 4, 6, and 8 p < 0.01; and scale 7 p < 0.0083) (Table 4).
Good person fit was demonstrated for the majority of the scales (1, 2, 6, 7, 8, and 9) with a person fit residual SD < 1.5 indicating that overall people responded to items as expected. Minor person misfit was shown across three of the nine scales: (3) Actively managing my health; (4) Social support for health; and (5) Appraisal of health information, with a person fit residual SD >1.5 (Table 3). This suggest that some people responded in an unusual way to some items in these scales.
Unidimensionality is a critical property of good measurement and a prerequisite to the summation of items within a scale [15, 29]. Unidimensionality was demonstrated for all nine scales (Table 3) as determined by <5% significant t-tests (scales 1, 2, 4, 6, 7, and 8) or a 95% confidence interval (CI) including 5% where >5% significant t-tests were evident: scale (3) CI:0.04–0.09; scale (5) CI:0.04–0.08; and scale (9) CI:0.03–0.07. Local independence further supports the concept of unidimensionality . All nine scales demonstrated local independence with between-item residual correlations matrix values <0.2. The Person Separation Index (PSI) for all scales was >0.7 indicating good internal consistency reliability.
No item bias was evident for the majority of the HLQ items (43 out of 44), demonstrating that people with the same level of health literacy consistently responded to items in the same way, regardless of their gender or age group. Only one item: ‘Get health information by yourself’ from scale (8) Ability to find good health information, demonstrated item bias for gender as indicated by a probability value below the Bonferroni adjusted probability value (p < 0.005). This means that males and females responded differently to each other despite having the same level of health literacy (non-uniform DIF)  (Fig. 1).
Overall, the response format was found to be satisfactory for the ‘strongly disagree to strongly agree’ scales (scale 1 to 5) as indicated by the absence of disordered thresholds. Mild disordering was evident in scale (4) Social support for health, for the following item: ‘I have at least one person who can come to medical appointments with me’. Disordered thresholds predominantly occurred among the capability response categories (cannot do to very easy) for the following items: ‘discuss things with healthcare providers…’ and ‘Ask healthcare providers questions to get…’ from scale (6) Ability to actively engage with healthcare providers; ‘Find out what healthcare services you are…’ from scale (7) Navigating the healthcare system; ‘Find health information from several…’, Get information about health so you are…’, and ‘Get health information by yourself’ from scale (8) Ability to find good health information; and all items in scale (9) Understanding health information well enough to know what to do. On inspection of the category probability curves, the main issue participants had was choosing between ‘very difficult’ and ‘quite difficult’. The HLQ authors, however, recently changed the capability response options (scales 6–9) to include elements of frequency as well as difficulty, and this was found to be better than the original options .
In terms of targeting, a positive mean person location for all nine scales (0.89–2.99) suggested that participants found some of the items easy to endorse. Person-item distribution graphs plot item difficulty and the person’s level of health literacy along a common measure: logits. A logit is the unit of measurement that results when the Rasch model is used to transform raw scores from ordinal data to log odds ratios on a common scale . The value of zero is allocated to the mean of the item difficulty [16, 26]. There should be an even spread of HLQ items across the range of participants’ health literacy levels. On inspection of these graphs there were no items matching participants’ level of health literacy at approximately the one to two logit point (mid to high HLQ score) despite a number of participants at this ability level for each scale (Fig. 2).
This is the first study to assess the measurement properties of the HLQ among a cohort of older people who have presented to an ED after a fall. Health literacy is an important factor associated with participation in preventive health programs, such as falls prevention initiatives. Overall, the HLQ demonstrated good measurement properties. The summation of the HLQ items within each scale to provide scale summary scores, with each scale representing one distinct component of health literacy, is supported. This finding is consistent with previous validation studies of the HLQ [8,9,10,11, 14]. This indicates that each HLQ scale measures what it purports to measure, and nothing more, providing detailed information on nine separate areas of health literacy.
Absence of item bias is considered a fundamental principle of good measurement [15, 18]. It is important that items work consistently for individuals across different sub-groups, particularly if different demographic groups are to be compared . Almost all the items (43 of 44) did not demonstrate item bias for the covariates assessed, with minor bias demonstrated for only one item. This suggests that un-biased estimates of health literacy across gender and age groups can be obtained from the HLQ. This finding further supports previous studies that found both the English and Slovakian versions of the HLQ to be invariant across a number of key demographic groups [9, 13].
In this study, the majority of misfit suggests that the set of items within some scales may have overlapping content (overfit). Overfit does not compromise good measurement . A strong rationale for including the items is provided in the development of the tool. Multiple structured processes were undertaken to develop the HLQ items, guided by the revised Bloom’s taxonomy, to generate items of various difficulty. Detailed psychometric analyses were used to test and refine the items, leading to removal or re-wording of poorly performing items . Given the rigorous development process of the HLQ, deletion of misfitting items is not recommended. Doing so may compromise construct coverage and result in loss of some of the tool’s important items . Overall misfit to the Rasch model should be treated with caution. While Chi-square probability values are recommended to determine fit, these values are sensitive to sample size . Given a sufficiently large sample size (n = 433 in this study), even small deviations from model fit will be statistically significant .
All nine HLQ scales were found to be inadequately targeted for this sample, which is consistent with findings from Richtering et al. . It is important to note that the RESPOND cohort were not representative of the general population in several ways. Firstly, the cohort consisted of participants who were taking part in a clinical trial. Those who volunteer to participate in research projects may have levels of education, motivation and engagement that differ from those who decline to participate. Secondly, due to the exclusion criteria necessary for the purpose of the RCT, the sample was underrepresented for certain subgroups known to have lower levels of health literacy. For example, those born overseas or who speak languages other than English at home, those with lower education, no private health insurance, multiple chronic conditions, and women have been found to have lower health literacy on some HLQ scales . The RESPOND cohort had higher HLQ scores in seven of the nine HLQ scales (scales 1, 2, 4, 6, 7, 8, and 9), and similar levels of health literacy in two scales (3 and 5), when compared to a sample representing a diverse range of socio-economic and geographical characteristics . This may explain why the RESPOND cohort appeared to find some HLQ items easy to endorse. The measurement gap identified has implications for measurement precision, which decreases at the level corresponding with this gap . This means that a large change in health literacy is necessary in order to elicit a change in mid to high HLQ score for the RESPOND cohort.
The main strength of this study is that the sample was from a multi-centre trial, encompassing two geographically diverse areas of Australia. In terms of limitations, the sample size may have contributed to the significant Chi-square probability values . A further limitation was that the sample was under representative of a number of socio-economic groups, limiting generalisability of the results to the broader population of older adults who present to an ED after a fall.
The current study builds on previously established strong measurement properties of the HLQ and adds new knowledge specific to a population of older people who have presented to an ED after a fall. Overall, the HLQ was found to have good measurement properties among this cohort. The HLQ may be used to tailor falls prevention initiatives to allow for program components programs, such as provision of education, support and community linkage, to be delivered in a manner appropriate for individual health literacy ability. This may increase participation in falls prevention activities, potentially resulting in better health outcomes for these patients.
Confirmatory factor analysis
Classical test theory
Falls Risk for Older People – Community setting
Health Literacy Questionnaire
Index of Relative Socio-economic Advantage and Disadvantage
Item response theory
Mini Mental State Examination
Principal Components Analysis
Person Separation Index
Randomised controlled trial
Samaras N, Chevalley T, Samaras D, Gold G. Older patients in the emergency department: a review. Ann Emerg Med. 2010;56(3):261–9.
Russell MA, Hill KD, Day LM, Blackberry I, Schwartz J, Giummarra MJ, Dorevitch M, Ibrahim JE, Dalton AC, Dharmage SC. A randomized controlled trial of a multifactorial falls prevention intervention for older fallers presenting to emergency departments. J Am Geriatr Soc. 2010;58(12):2265–74.
Ishikawa H, Yano E. Patient health literacy and participation in the health-care process. Health Expect. 2008;11(2):113–22.
World Health Organization. Health promotion glossary. Health Promot Int. 1998;13(4):349–64.
Scott TL, Gazmararian JA, Williams MV, Baker DW. Health literacy and preventive health care use among Medicare enrollees in a managed care organization. Med Care. 2002;40(5):395–404.
Jordan JE, Osborne RH, Buchbinder R. Critical appraisal of health literacy indices revealed variable underlying constructs, narrow content and psychometric weaknesses. J Clin Epidemiol. 2011;64(4):366–79.
Haun JN, Valerio MA, McCormack LA, Sørensen K, Paasche-Orlow MK.Health literacy measurement: an inventory and descriptive summary of 51 instruments. J Health Commun. 2014;19(Suppl 2):302–33.
Osborne RH, Batterham RW, Elsworth GR, Hawkins M, Buchbinder R. The grounded psychometric development and initial validation of the health literacy questionnaire (HLQ). BMC Public Health. 2013;13(1):658.
Elsworth GR, Beauchamp A, Osborne RH. Measuring health literacy in community agencies: a Bayesian study of the factor structure and measurement invariance of the health literacy questionnaire (HLQ). BMC Health Serv Res. 2016;16(1):508.
Maindal HT, Kayser L, Norgaard O, Bo A, Elsworth GR, Osborne RH. Cultural adaptation and validation of the health literacy questionnaire (HLQ): robust nine-dimension Danish language confirmatory factor model. SpringerPlus. 2016;5(1):1232.
Nolte S, Osborne RH, Dwinger S, Elsworth GR, Conrad ML, Rose M, Härter M, Dirmaier J, Zill JM. German translation, cultural adaptation, and validation of the health literacy questionnaire (HLQ). PLoS One. 2017;12(2):e0172340.
Hawkins M, Gill SD, Batterham R, Elsworth GR, Osborne RH. The health literacy questionnaire (HLQ) at the patient-clinician interface: a qualitative study of what patients and clinicians mean by their HLQ scores. BMC Health Serv Res. 2017;17(1):309.
Kolarcik P, Cepova E, Geckova AM, Elsworth GR, Batterham RW, Osborne RH. Structural properties and psychometric improvements of the health literacy questionnaire in a Slovak population. Int J Public Health. 2017;62(5):591-604.
Richtering SS, Morris R, Soh S-E, Barker A, Bampi F, Neubeck L, Coorey G, Mulley J, Chalmers J, Usherwood T. Examination of an eHealth literacy scale and a health literacy scale in a population with moderate to high cardiovascular risk: Rasch analyses. PLoS One. 2017;12(4):e0175372.
Tennant A, Conaghan PG. The Rasch measurement model in rheumatology: what is it and why use it? When should it be applied, and what should one look for in a Rasch paper? Arthritis Care Res. 2007;57(8):1358–62.
Pallant JF, Tennant A. An introduction to the Rasch measurement model: an example using the hospital anxiety and depression scale (HADS). Br J Clin Psychol. 2007;46(1):1–18.
Misajon R, Pallant J, Bliuc A-M. Rasch analysis of the personal wellbeing index. Qual Life Res. 2016;25(10):2565–9.
Hagquist C, Bruce M, Gustavsson JP. Using the Rasch model in nursing research: an introduction and illustrative example. Int J Nurs Stud. 2009;46(3):380–93.
Prieto L, Alonso J, Lamarca R. Classical test theory versus Rasch analysis for quality of life questionnaire reduction. Health Qual Life Outcomes. 2003;1(1):27.
Barker A, Cameron P, Hill K, Flicker L, Haines T, Lowthian J, Waldron N, Arendts G, Redfern J, Forbes A. RESPOND–A patient-centred program to prevent secondary falls in older people presenting to the emergency department with a fall: protocol for a multi-centre randomised controlled trial. Inj Prev. 2015; 21:e1
Tombaugh TN, McIntyre NJ. The mini-mental state examination: a comprehensive review. J Am Geriatr Soc. 1992;40(9):922-35.
Batterham R, Hawkins M, Collins P, Buchbinder R, Osborne R. Health literacy: applying current concepts to improve health services and reduce health inequalities. Public Health. 2016;132:3–12.
Pink B. Technical Paper: Socio-economic indexes for areas (SEIFA) 2011. Australian Bureau of Statistics. Commonwealth of Australia. 2013. http://www.ausstats.abs.gov.au/ausstats/subscriber.nsf/0/22CEDA8038AF7A0DCA257B3B00116E34/$File/2033.0.55.001%20seifa%202011%20technical%20paper.pdf. Accessed 2 Dec 2015.
Australian Government: Comonwealth Grants Commission. Measuring socio-economic status: staff discussion paper: CGC 2012-03. 2012. https://cgc.gov.au/attachments/article/173/2012-03%20Measuring%20Socio-Economic%20Status.pdf. Accessed 2 Dec 2015.
Russell MA, Hill KD, Blackberry I, Day LM, Dharmage SC. The reliability and predictive accuracy of the falls risk for older people in the community assessment (FROP-com) tool. Age Ageing. 2008;37(6):634–9.
Bond TG, Fox CM. Applying the Rasch model: fundamental measurement in the human sciences. 3rd ed. New York: Routledge; 2015.
Shea TL, Tennant A, Pallant JF. Rasch model analysis of the depression, anxiety and stress scales (DASS). BMC Psychiatry. 2009;9(1):21.
Ramp M, Khan F, Misajon RA, Pallant JF. Rasch analysis of the multiple sclerosis impact scale (MSIS-29). Health Qual Life Outcomes. 2009;7(1):58.
Tennant A, McKenna SP, Hagell P. Application of Rasch analysis in the development and application of quality of life instruments. Value Health. 2004;7(Suppl 1):S22–6.
Ostini R, Nering M. Polytomous item response theory models. Thousand Oaks, CA: SAGE Publications Ltd; 2006.
Beauchamp A, Buchbinder R, Dodson S, Batterham RW, Elsworth GR, McPhee C, Sparkes L, Hawkins M, Osborne RH. Distribution of health literacy strengths and weaknesses across socio-demographic groups: a cross-sectional survey using the health literacy questionnaire (HLQ). BMC Public Health. 2015;15(1):678.
Hagell P, Broman JE. Measurement properties and hierarchical item structure of the Epworth sleepiness scale in Parkinson's disease. J Sleep Res. 2007;16(1):102–9.
Smith EV Jr. Detecting and evaluating the impact of multidimenstionality using item fit statistics and principal component analysis of residuals. J Appl Meas. 2002;3(2):205-31.
Darzins S, Imms C, Di Stefano M, Taylor NF, Pallant JF. Evaluation of the internal construct validity of the personal care participation assessment and resource tool (PC-PART) using Rasch analysis. BMC Health Serv Res. 2014;14(1):543.
Hendriks MR, Bleijlevens MH, van Haastregt JC, de Bruijn FH, Diederiks JP, Mulder WJ, Ruijgrok JM, Stalenhoef PA, Crebolder HF, van Eijk JT. A multidisciplinary fall prevention program for elderly persons: a feasibility study. Geriatr Nurs. 2008;29(3):186–96.
This study would not have been possible without the contributions of the RESPOND investigator team responsible for the development and implementation of the RESPOND RCT. RESPOND team members who are not co-authors in this paper, and their affiliations, are listed below: Associate Professor Caroline Brand, Health Services Research Unit, Department of Epidemiology and Preventive Medicine, Monash University, Melbourne, Victoria, Australia; Melbourne EpiCentre, University of Melbourne and Melbourne Health, Melbourne, Victoria, Australia; Professor Peter Cameron, Health Services Research Unit, Centre of Research Excellence in Patient Safety, Department of Epidemiology and Preventive Medicine, Monash University, Melbourne, Victoria, Australia; Dr. de Villiers Smit, Alfred Health, Melbourne, Victoria, Australia; Professor Leon Flicker, Western Australian Centre for Health & Ageing – WACHA, Centre for Medical Research; University of Western Australia, Royal Perth Hospital, Perth, Australia; Professor Andrew Forbes, Department of Epidemiology and Preventive Medicine, Monash University, Melbourne, Victoria, Australia; Terry Haines, Department of Physiotherapy, Monash University, and Allied Health Research Unit, Monash Health, Melbourne, Victoria, Australia; Associate Professor Peter Hunter, Alfred Health, Melbourne, Victoria, Australia; Central Clinical School, Monash University, Melbourne, Victoria, Australia; Dr. Samuel Nyman, Department of Psychology and Bournemouth University Dementia Institute, Faculty of Science and Technology, Bournemouth University, Poole, Dorset, UK; Dr. Nicholas Waldron, Health Networks Branch, System Policy and Planning, Department of Health, Government of Western Australia, Perth Australia; Department of Rehabilitation and Aged Care, Armadale Health, Perth, Western Australia, Australia. In addition, we thank representatives from our funding partner organisations who have contributed to development and implementation of the RESPOND project. The partner organisations include: (1) Health Strategy and Networks Branch, Strategic System Policy and Planning, Department of Health, Western Australia, Australia. (2) Aged and Continuing Care Directorate, Department of Health, Western Australia, Australia. (3) Royal Perth Hospital. (4) Curtin University. (5) The University of Western Australia. (6) The Royal Perth Hospital Medical Research Foundation. (7) Sir Charles Gairdner Hospital (SCGH) Area Rehabilitation and Aged Care Falls Specialist Program. (8) Injury Control Council of Western Australia (ICCWA). (9) The George Institute for Global Health. (10) The Alfred Hospital. (11) Monash University. (12) Integrated Care, Victorian Department of Health.
This project is funded under the Australian National Health and Medical Research Council (NHMRC) Partnership Projects funding scheme (project number APP1056802) with financial and in-kind contributions from the partner organisations listed in ‘Acknowledgements’. RLM is funded by an Australian Postgraduate Award (APA) scholarship, ALB is funded by a Career Development Fellowship funded by the NHMRC (1067236); JR is funded by a Career Development and Future Leader Fellowship co-funded by the NHMRC and National Heart Foundation (APP1061793). JAL is funded by an Early Career Fellowship funded by the NHMRC (1052442). RHO is funded in part through an NHMRC Senior Research Fellowship (APP1059122). RB is funded by an NHMRC Senior Principal Research Fellowship (APP1082138).
Availability of data and materials
The datasets generated and/or analysed during the current study are available from the corresponding author on reasonable request.
Ethics approval and consent to participate
Ethical approval was obtained from Alfred Health (HREC 439/13) and Royal Perth Hospital (REG 13–128), Monash University Human Research Ethics Committee (HREC) (MUHREC CF13/3869–2013001975) and Curtin University HREC (HR 43/ 2014). Written informed consent was obtained from all study participants.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.