Skip to main content

Main menu

  • Home
  • Content
    • Current issue
    • Past issues
    • Collections
  • About
    • General information
    • Staff
    • Editorial board
    • Open access
    • Contact
  • CMAJ JOURNALS
    • CMAJ
    • CJS
    • JAMC
    • JPN

User menu

Search

  • Advanced search
CMAJ Open
  • CMAJ JOURNALS
    • CMAJ
    • CJS
    • JAMC
    • JPN
CMAJ Open

Advanced Search

  • Home
  • Content
    • Current issue
    • Past issues
    • Collections
  • About
    • General information
    • Staff
    • Editorial board
    • Open access
    • Contact
  • RSS feeds
Getting new auth cookie, if you see this message a lot, tell someone!
Research
Open Access

Estimation of a Canadian preference-based scoring algorithm for the Veterans RAND 12-Item Health Survey: a population survey using a discrete-choice experiment

Nick Bansback, Logan Trenaman, Brendan J. Mulhern, Richard Norman, Rebecca Metcalfe, Richard Sawatzky, John E. Brazier, Donna Rowen and David G.T. Whitehurst
July 05, 2022 10 (3) E589-E598; DOI: https://doi.org/10.9778/cmajo.20210113
Nick Bansback
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Logan Trenaman
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Brendan J. Mulhern
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Richard Norman
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Rebecca Metcalfe
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Richard Sawatzky
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
John E. Brazier
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Donna Rowen
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
David G.T. Whitehurst
School of Population and Public Health (Bansback), University of British Columbia; Centre for Health Evaluation and Outcome Sciences (Bansback, Trenaman, Metcalfe, Sawatzky), Vancouver, BC; Centre for Health Economics Research and Evaluation (Mulhern), University of Technology Sydney, Sydney, New South Wales, Australia; School of Health and Related Research (Mulhern, Brazier, Rowen), University of Sheffield, Sheffield, UK; School of Public Health (Norman), Curtin University, Perth, Western Australia, Australia; School of Nursing (Sawatzky), Trinity Western University, Langley, BC; Faculty of Health Sciences (Whitehurst), Simon Fraser University, Burnaby, BC
PhD
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Tables
  • Related Content
  • Responses
  • Metrics
  • PDF
Loading

Abstract

Background: The Veterans RAND 12-Item Health Survey (VR-12) is a generic patient-reported outcome measure derived from the widely used 36- and 12-item Short Form Health Surveys. We aimed to estimate a Canadian preference-based scoring algorithm for the VR-12, enabling the derivation of health utility values for generating quality-adjusted life years (QALYs).

Methods: We conducted a discrete-choice experiment in a sample of the Canadian population in January and February 2019. Participants — recruited from a consumer research panel — completed an online survey, in English or French, that included 11 discrete-choice questions, each comprising 2 health profiles. We defined the health profiles using 8 VR-12 items and a duration attribute. Using conditional logit regressions, where each level of the respective VR-12 items was interacted with duration, we applied the coefficients to estimate health utility values interpretable on a scale of 0 (dead) to 1 (full health). Negative values reflect states considered worse than dead.

Results: A total of 3380 individuals completed the survey. Of these, 1688 (49.9%) were females, and 3101 (91.7%) completed the English version of the survey. Across all models, “feel downhearted and blue all of the time” and “pain interferes with your normal work extremely” were associated with the largest decrements in health utility. Excluding the 685 respondents (20.3%) who provided inconsistent responses had a negligible effect on the results. The recommended model, weighted to match population demographics, had health utility values ranging from −0.589 to 1.000.

Interpretation: Health utility values that reflect the preferences of the Canadian population can now be derived from responses to the VR-12. These values can be used to generate QALYs in future analyses.

Patient-reported outcome measures are used to collect information about aspects of respondents’ health status at a particular point in time. Generic patient-reported outcome measures (as opposed to condition-specific patient-reported outcome measures) can be applied across different populations. They comprise questions about health-related aspects of quality of life, each with defined levels of impact or severity, resulting in the definition of a multitude of health states.1 Preference-based patient-reported outcome measures include scoring algorithms that provide a “value set” — a set of index scores (often referred to as health utility values or preference weights) that provides a value for each unique instrument-defined health state.1 Health utility values are interpreted on a scale anchored at full health = 1 and dead = 0, with negative scores reflecting health states worse than dead.2–4 Health utility values can be used to calculate quality-adjusted life years (QALYs) to inform analyses such as economic evaluations and population health comparisons.

Health utility values can be obtained by eliciting preferences that capture willingness to trade extra years of life for improvements in health (i.e., trading quantity for quality). In several jurisdictions, including Canada, guidelines for the conduct of economic evaluation recommend that health utility values reflect the preferences of a representative sample of the general population.5–7 Currently, Canadian value sets are available for a number of generic, preference-based, patient-reported outcome measures, including the Health Utilities Index (HUI) Mark 2 (HUI2)8 and Mark 3 (HUI3)9 and the 3-level (EQ-5D-3L)10 and 5-level (EQ-5D-5L)11 versions of the EQ-5D. The availability of these value sets has the dual benefit of providing analysts with options when selecting outcome measures and ensuring that the preferences of Canadians are reflected in health care decision-making (e.g., in health technology appraisal processes).

Increasingly, the Veterans RAND 12-Item Health Survey (VR-12) is being used to measure health-related quality of life in routine data collection initiatives.12,13 The VR-12 is a shortened version of the Veterans RAND 36-item Health Survey,14 which was itself an adaptation of the RAND 36-item Health Survey (also known as the RAND SF-36).15 The VR-12 includes 12 questions that cover 8 domains, plus 2 questions that ask about health issues “compared with 1 year ago.”12 The VR-12 was not designed to describe health states for the purposes of deriving health utility values. Scoring of the VR-12 is based on psychometric measurement methods for deriving scales that correspond with the 8 domains, as well as overarching physical and mental component scores.12

Given increasing use of the VR-12 in routine population measurement (in part because it is free for the user, although permission from the developers is required),16 considerable VR-12 data are available to inform real-world analyses such as economic evaluations. To facilitate this utilization, in a manner aligned with national guidelines, a value set that reflects the preferences of Canadians is required. Accordingly, our objective was to develop a scoring algorithm, based on the preferences of the Canadian general public, that estimates health utility values for health states described by the VR-12.

Methods

Study design

We used an online survey to administer a discrete-choice experiment in January and February 2019. The experiment required participants to choose their preferred option when presented with 2 health profiles (called a “choice set”). Each health profile described living in a health state defined using attributes and levels from 8 VR-12 items, for a certain duration. Respondents completed 11 choice sets. With aggregation across the sample, such data can be used to estimate population-level trade-offs between different health states (representing aspects of quality of life) and length of life, resulting in a societal value set — an approach that has been used in previous studies.17–19 We piloted the survey to get feedback on participants’ understanding of the questions.

In the following sections (supplemented by further details in Appendix 1, available at www.cmajopen.ca/content/10/3/E589/suppl/DC1, including the Checklist for Reporting Results of Internet E-Surveys [CHERRIES]20), we describe development of the discrete-choice task and the survey, as well as the methods of analysis.

Participants

For the pilot phase of the project, we sought about 100 Canadians (recruited from the Amazon Mechanical Turk marketplace) to complete a survey hosted on the PrefApp platform located at the Centre for Health Evaluation and Outcome Sciences (British Columbia, Canada). We performed recruitment for the subsequent main survey (aiming for a sample of about 3000 respondents) through a larger study, in which half the participants were randomly assigned to complete the survey described here (again, on the PrefApp platform); the other half completed a different survey on an unrelated research question concerning the VR-12. A market research company (Ipsos) invited members of their national consumer research panel, primarily by email, to take part in “a study about a measure of quality of life.” Participants who clicked on the link were given further information about both surveys, including the type of questions and approximate completion time. Those who continued were given further study information and asked if they wanted to consent to proceed.

After providing consent and answering demographic questions, participants were randomly assigned to 1 of the 2 surveys. Participants who completed the survey were awarded points by the company, which could be used toward financial rewards. We used quota sampling to encourage representation of the Canadian general population in terms of age, biological sex and location (based on provinces). We checked Internet Protocol addresses for duplicates, to reduce the likelihood that individuals completed the survey more than once.

Data sources

We obtained data for the main analysis by asking participants to answer discrete-choice questions based on health states described using the VR-12. Not all items in the VR-12 are relevant for the purposes of defining and valuing a health state (e.g., questions asking “compared with 1 year ago”), and multiple items measure similar constructs; therefore, 3 of the authors (B.J.M., J.E.B., D.R.) used an established process to construct a health state classification system.21,22 The process comprises factor analysis to assess dimensionality, Rasch analysis to understand item performance and tests of differential item function. This resulted in the selection of 8 health attributes, each of which corresponded to a VR-12 item.

We used the following naming conventions for the VR-12-related attributes: physical functioning, role physical, role emotional, bodily pain, mental health — anxiety, mental health — depression, vitality and social functioning. We constructed the levels for each attribute on the basis of response options for the respective VR-12 item, with the exception of items that included “a good bit of the time.” This response option was deemed too similar to “a little of the time” and was removed from the classification system. The final classification system, presented in Appendix 2 (available at www.cmajopen.ca/content/10/3/E589/suppl/DC1), defines 234 375 (3*57) health states.

Following previous studies,17–19 we used a discrete-choice experiment to elicit trade-offs between health states described by the 8 VR-12 items and life duration. To do this, we added a ninth attribute, the duration attribute, termed “life years.” We described this attribute using 4 levels (“1 year,” “4 years,” “7 years” and “10 years”), an approach used in previous studies. 17,23 We chose an upper limit of 10 years for comparability with the time horizon in most national valuation studies.24–26

We developed a pilot version of the survey using feedback from 16 cognitive interviews (conducted by R.M.; details provided in Appendix 1, including Appendix Table 1). Feedback from the pilot survey (n = 103) informed some revisions to the wording and layout of the final survey (Appendix 3, available at www.cmajopen.ca/content/10/3/E589/suppl/DC1). The survey was then forward-and back-translated into French. The demographic characteristics of those who participated in the cognitive interviews and in the survey pilot testing are available in Appendix 1 (Appendix Tables 2 and 3, respectively).

View this table:
  • View inline
  • View popup
Table 1:

Characteristics of survey respondents, with comparable statistics for the Canadian general population, drawn from Statistics Canada 2016 Census28

View this table:
  • View inline
  • View popup
Table 2:

Results from the unanchored models

View this table:
  • View inline
  • View popup
Table 3:

Anchored disutility values for each attribute level of the preferred model (model 6)*

The final survey began with details about the study, followed by a consent form and questions to collect information about the characteristics and health history of participants. Participants were then asked to complete the VR-12 for themselves (with instruction that later questions would be based on this questionnaire), followed by a warm-up discrete-choice exercise and, finally, the discrete-choice questions.

Feedback from the cognitive interviews informed the development of the discrete-choice task in the final survey. Figure 1 provides an example of a choice set that was presented to participants in the final survey. One of the authors (R.N.) used experimental design theory (NGene, modified Federov algorithm) to select 200 choice sets (from the millions of possible choice sets) that would provide the most information for analysis purposes, i.e., a d-efficient design. These 200 choice sets were “blocked” into 20 groups of 10 questions to ensure that each block had representation of attributes and levels. We randomly assigned each participant to 1 of the blocks. To assess the consistency of participants’ responses, we repeated the warm-up discrete-choice question along with the set of 10. The repeated question (i.e., an 11th question) appeared fourth in the sequence, and we did not use the responses to this question in the modelling. We explored engagement with the task by timing how long participants took to answer all 11 discrete-choice questions.

Figure 1:
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1:

Example of a choice set presented to participants. Details about the use of different intensities of highlighting (i.e., the various shades of yellow) for 5 of the attributes are provided in Appendix 1.

Statistical analysis

Our analysis followed the methods reported in peer-reviewed studies that used discrete-choice experiments to elicit population value sets.17–19,23 We (N.B., B.J.M., R.N., J.E.B., D.R., D.G.T.W.) used conditional logit regression to model the data. This method estimates coefficients for each level of the VR-12-related attributes (dummy-coded, with the best level used as the reference category) interacted with the corresponding number of life years, along with a coefficient for the duration attribute as a main effect. The coefficient for the life years main effect is the value of living in full health for 1 year (expected to be positive). The coefficients for each level of the VR-12-related attributes reflect the disutility of living with the respective impairment for a duration of 1 year (expected to be negative). A random error term captured unobservable factors, as well as potential inconsistencies in decision-making. 27 We tested interactions between dummy-coded levels of duration and, if satisfied that they were sufficiently linear, coded them as continuous, both in the main effect and the set of interaction terms.

The estimated regression coefficients are on a latent scale: they do not have cardinal properties and, therefore, their magnitudes are not directly comparable (“unanchored” coefficients). We estimated coefficients anchored on the full health – dead scale as the marginal willingness to trade an extra year of life for each VR-12-related attribute level, calculated by dividing each interaction coefficient by the duration coefficient. Additive combinations of anchored coefficients provide the health utility values for the health states defined by the classification system.

Previous valuation studies have reported inconsistencies in the directionality of coefficients when valuing large descriptive systems (e.g., when an increase in severity leads to an increase rather than a decrease in utility). We used models involving combinations of attribute levels (merged on the basis of best fit) to overcome any inconsistencies, while also exploring issues of participant engagement and understanding regarding the survey. Because preferences may vary in relation to different sociodemographic characteristics, we weighted the “preferred” final model to ensure a representative population estimate and to account for non-response and noncoverage. We applied a raking approach using data from the 2016 Census28 and the 2015/16 Canadian Community Health Survey (https://www23.statcan.gc.ca/imdb/p2SV.pl?Function=getSurvey&Id=238854) to develop weights, and included these weights in the conditional logit models. We performed the analyses with Stata and SAS software.

Ethics approval

The study was approved by the University of British Columbia Behavioural Research Ethics Board (study no. H18-00969).

Results

A total of 8110 individuals clicked on the initial link to find out further information about the larger study (i.e., the 2 surveys), of whom 6731 (83.0%) provided consent to take part. The discrete-choice survey was randomly assigned to 3445 (51.2% of the 6731) potential participants; 3380 (98.1% of the 3445) completed all survey questions, which yielded 33 800 choice-set responses for analysis.

Table 1 reports the demographic characteristics of the survey respondents. Relative to the Canadian general population, differences included a higher proportion of study participants who identified as white, with lower reported annual income, with higher levels of education and having never legally married. A lower proportion of respondents chose to complete the survey in French, relative to the proportion of Canadians reporting French as their first official language spoken.

Of the 10 379 choice sets where the level of life duration varied between health profiles, the profile with the lower duration was selected on 2433 (23.4%) occasions.

Unanchored values

Results from the conditional logit models are provided in Table 2. Model 1 includes all respondents. Most coefficients (26 of 30) were negative, with the magnitude of the disutility tending to increase with more severe levels of impairment. Based on the coefficients of the worst levels, the most important attributes were mental health — depression (−0.145) and bodily pain (−0.127), whereas the least important were role physical (−0.047) and role emotional (−0.048). As expected, there were some directional inconsistencies; however, vitality level 4 was the only statistically significant inconsistency at the 5% level.

In models 2 to 5, attribute levels were combined to overcome directional inconsistencies (Table 2). Model 2 included all participants; model 3 excluded the 37 respondents (1.1%) who provided the same answer to all 10 discrete-choice questions (i.e., always selected health state A or always selected health state B); model 4 excluded the 685 respondents (20.3%) whose responses failed the consistency check; and model 5 excluded the 315 respondents (9.3%) who completed all questions in less than 2 minutes. Across all 4 consistent models, a similar pattern was observed when attributes were ordered according to the magnitude of the coefficients of the worst levels. Given the similarities across models 1 to 5, model 2 was considered the most defensible because of the directional consistency and inclusivity of participants. Model 2 was then weighted to adjust for the differences between the study sample and the Canadian general population (i.e., model 6 is a weighted version of model 2).

Anchored values

Figure 2 shows the anchored value sets produced by all models; the corresponding coefficients are presented in Appendix 1, Appendix Table 4. The anchored values for each attribute level of model 6 are provided in Table 3. These results provide the means to calculate health utility values for all health states defined by the classification system. The health utility values for model 6 ranged from 1.000 (no impairment on any attribute) to −0.589 (the worst level of impairment on all attributes); 5.9% of the health states had values less than or equal to 0. The distribution of health utility values for model 6 is shown in Figure 3.

Figure 2:
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2:

Anchored health utility values from the 6 models, presented as the health utility values for each model for each health state. The “best” health state (i.e., level 1 in all attributes, denoted as health state 11111111) is on the far left, with a health state equal to 1. The “worst” health state (i.e., the lowest level on each attribute, denoted as 35555555) is on the far right. The Pearson correlation with model 6 is shown in parentheses for each of the other models.

Figure 3:
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3:

Distribution of health utility values for model 6 (the preferred model) across all possible health states.

Interpretation

In this study, we developed a value set for the VR-12 based on the preferences of the Canadian general public. We suggest use of model 6 (Table 3) because it has consistently ordered levels, is based on the preferences of all respondents and is weighted to the demographic characteristics of the Canadian population. Further information and tools to facilitate the scoring of health utility values from VR-12 responses is available on the study website (www.bcpcm.ca/resources/VR-12/).

Many preference-based, patient-reported outcome measures are available for use,1 and 1 previous study has estimated health utility values for the VR-12 (based on a mapping exercise using values from the United Kingdom).29 Although there are myriad reasons why value sets differ across different preference-based instruments, recent studies in Canada and the UK have been consistent in identifying pain and mental health items as being associated with the largest utility decrements. 11,23,25 In the UK value set for the SF-6Dv2, the most severe level of pain had a disutility of 0.620,23 compared with 0.272 in the current study. The difference in magnitude is likely related to the items used in the respective studies, with the SF-6Dv2 using the SF-36v2 pain severity item,21 whereas the single pain item in the VR-12 is framed around interference “with your normal work.”

We recognize that there is no “best” preference-based patient-reported outcome measure, and health utility values across instruments are expected to differ,30,31 something that is reflected in Canadian guidelines for the conduct of economic evaluation.5 Instead, researchers are required to consider the relative merits of the available options for the evaluative purpose being pursued. For example, anxiety and depression, which are combined in a single item in the EQ-5D instruments, have been shown to be valued differently.32 The Canadian value set for the VR-12 separates anxiety from depression, meaning that it may better reflect changes in health outcome attributable to interventions that improve one but not the other. This is a context-specific empirical question and highlights the need for comprehensive comparative psychometric evaluation of the new classification system and value set (i.e., compared with other preference-based, patient-reported outcome measures).

Another empirical question is whether health state values from patient populations would differ from those elicited from the general public,33 and whether the use of societal preferences undervalues the quality of life of people with health conditions and disabilities.34,35 The approach taken in this study — elicitation of societal preferences — reflects the explicit expectations detailed in national guidelines for the economic evaluation of health technologies.5–7

Limitations

Preference elicitation tasks, such as discrete-choice experiments, are cognitively challenging,36,37 which can introduce bias into the sample and the subsequent responses and results. To mitigate the potential of such biases, we used cognitive interviews and pilot testing in designing our discrete-choice experiment, and we employed novel methods to simplify the task (Appendix 1). Such methods have been used successfully in other discrete-choice studies.19,37

Although our results showed some inconsistencies, these were resolved by combining adjacent levels. We do not know if these inconsistencies were a result of the choice sets presented (i.e., a “real” inconsistency), participants’ lack of comprehension of or attention to the task, or merely random. Regardless of the reason, a methodologic priority is to advance the design of preference elicitation tasks to address issues of inclusivity. Such advances must extend beyond recognition of individuals with lower cognitive abilities.36

Our study included the preferences of more Canadians than any other Canadian value set, yet it cannot be said to have explored preferences that reflect the full diversity of the population. For example, we did not include French speakers in the cognitive interview phase, and only 8.3% of the sample chose to complete the French version of the survey.

Our sample, recruited from a consumer research panel to complete an online survey, likely had higher literacy skill levels than the general population.

Conclusion

Health utility values reflecting Canadian preferences can now be generated from VR-12 responses. The new scoring algorithm provides a means to make further use of existing data sets, as well as an alternative to the HUI and EQ-5D instruments for the estimation of QALYs in economic evaluations involving primary data collection.

Acknowledgements

The authors acknowledge Judy Chiu for support in manuscript preparation. They also thank the respondents for taking part in the study and attendees of the 41st Annual North American Meeting of the Society for Medical Decision Making (Portland, Oregon), who provided valuable feedback.

Footnotes

  • Competing interests: None declared.

  • This article has been peer reviewed.

  • Contributors: Nick Bansback, Logan Trenaman, Richard Sawatzky and David Whitehurst conceived and designed the study; Nick Bansback, Logan Trenaman, Rebecca Metcalfe and David Whitehurst acquired the data; and Nick Bansback, Logan Trenaman, Brendan Mulhern, Richard Norman, Richard Sawatzky, John Brazier, Donna Rowen and David Whitehurst analyzed and interpreted the data. Nick Bansback and David Whitehurst drafted the manuscript. All of the authors revised the manuscript critically for important intellectual content, approved the final version for publication and agreed to be accountable for the work.

  • Funding: This study was funded by the British Columbia Patient-Centred Measurement Steering Committee.

  • Data sharing: Aggregated data presented in this manuscript are available to other investigators upon request to the corresponding author. Further analysis of individual-level data is also available upon request.

  • Supplemental information: For reviewer comments and the original submission of this manuscript, please see www.cmajopen.ca/content/10/3/E589/suppl/DC1.

This is an Open Access article distributed in accordance with the terms of the Creative Commons Attribution (CC BY-NC-ND 4.0) licence, which permits use, distribution and reproduction in any medium, provided that the original publication is properly cited, the use is noncommercial (i.e., research or educational use), and no modifications or adaptations are made. See: https://creativecommons.org/licenses/by-nc-nd/4.0/

References

  1. ↵
    1. Neumann PJ,
    2. Goldie SJ,
    3. Weinstein MC
    (2000) Preference-based measures in economic evaluation in health care. Annu Rev Public Health 21:587–611.
    OpenUrlCrossRefPubMed
  2. ↵
    1. Torrance GW
    (1987) Utility approach to measuring health-related quality of life. J Chronic Dis 40:593–603.
    OpenUrlCrossRefPubMed
    1. Torrance GW,
    2. Feeny D
    (1989) Utilities and quality-adjusted life years. Int J Technol Assess Health Care 5:559–75.
    OpenUrlCrossRefPubMed
  3. ↵
    1. Macran S,
    2. Kind P
    (2001) “Death” and the valuation of health-related quality of life. Med Care 39:217–27.
    OpenUrlCrossRefPubMed
  4. ↵
    (2017) CADTH methods and guidelines: guidelines for the economic evaluation of health technologies: Canada (Canadian Agency for Drugs and Technologies in Health, Ottawa) Available: https://www.cadth.ca/sites/default/files/pdf/guidelines_for_the_economic_evaluation_of_health_technologies_canada_4th_ed.pdf. accessed 2021 Apr. 6. 4th ed.
  5. (2013) Guide to the methods of technology appraisal 2013 (National Institute for Health and Care Excellence, London (UK)) Available: https://www.nice.org.uk/article/pmg9/chapter/Foreword. accessed 2021 Apr. 6.
  6. ↵
    1. Sanders GD,
    2. Neumann PJ,
    3. Basu A,
    4. et al.
    (2016) Recommendations for conduct, methodological practices, and reporting of cost-effectiveness analyses: second panel on cost-effectiveness in health and medicine. JAMA 316:1093–103.
    OpenUrlCrossRefPubMed
  7. ↵
    1. Torrance GW,
    2. Feeny DH,
    3. Furlong WJ,
    4. et al.
    (1996) Multiattribute utility function for a comprehensive health status classification system. Health Utilities Index Mark 2. Med Care 34:702–22.
    OpenUrlCrossRefPubMed
  8. ↵
    1. Feeny D,
    2. Furlong W,
    3. Torrance GW,
    4. et al.
    (2002) Multiattribute and singleattribute utility functions for the Health Utilities Index Mark 3 system. Med Care 40:113–28.
    OpenUrlCrossRefPubMed
  9. ↵
    1. Bansback N,
    2. Tsuchiya A,
    3. Brazier J,
    4. et al.
    (2012) Canadian valuation of EQ-5D health states: preliminary value set and considerations for future valuation studies. PLoS One 7:e31115.
    OpenUrlCrossRefPubMed
  10. ↵
    1. Xie F,
    2. Pullenayegum E,
    3. Gaebel K,
    4. et al.,
    5. Canadian EQ-5D-5L Valuation Study Group
    (2016) A time trade-off-derived value set of the EQ-5D-5L for Canada. Med Care 54:98–105.
    OpenUrlCrossRefPubMed
  11. ↵
    1. Selim AJ,
    2. Rogers W,
    3. Fleishman JA,
    4. et al.
    (2009) Updated U.S. population standard for the Veterans RAND 12-item Health Survey (VR-12). Qual Life Res 18:43–52.
    OpenUrlCrossRefPubMed
  12. ↵
    1. Health Outcomes Survey (HOS)
    (2021) Centres for Medicare and Medicaid Services, Available: https://www.cms.gov/Research-Statistics-Data-and-Systems/Research/HOS. accessed 2022 June 20.
  13. ↵
    1. Kazis LE
    (2000) The Veterans SF-36 health status questionnaire: development and application in the Veterans Health Administration. Med Outcomes Trust Monit 5:1–14.
    OpenUrl
  14. ↵
    1. Stewart A,
    2. Ware JE
    (1992) Measuring functioning and well-being: the Medical Outcomes Study approach (RAND Corporation, Santa Monica (CA)).
  15. ↵
    (2021) About the VR-36©, VR-12 and VR-6D© (Boston University School of Public Health, Boston) Availablehttp://www.bu.edu/sph/research/research-landing-page/vr-36-vr-12-and-vr-6d/about-the-vr-36-vr-12-and-vr-6d/. accessed 2015 Oct 12.
  16. ↵
    1. Bansback N,
    2. Brazier J,
    3. Tsuchiya A,
    4. et al.
    (2012) Using a discrete choice experiment to estimate health state utility values. J Health Econ 31:306–18.
    OpenUrlCrossRefPubMed
    1. Viney R,
    2. Norman R,
    3. Brazier J,
    4. et al.
    (2014) An Australian discrete choice experiment to value EQ-5D health states. Health Econ 23:729–42.
    OpenUrlCrossRefPubMed
  17. ↵
    1. Norman R,
    2. Viney R,
    3. Aaronson NK,
    4. et al.
    (2016) Using a discrete choice experiment to value the QLU-C10D: feasibility and sensitivity to presentation format [published erratum in Qual Life Res 2016;25:2401]. Qual Life Res 25:637–49.
    OpenUrl
  18. ↵
    1. Eysenbach G
    (2004) Improving the quality of Web surveys: the Checklist for Reporting Results of Internet E-Surveys (CHERRIES). J Med Internet Res 6:e34.
    OpenUrlCrossRefPubMed
  19. ↵
    1. Brazier JE,
    2. Mulhern BJ,
    3. Bjorner JB,
    4. et al.,
    5. SF-6Dv2 International Project Group
    (2020) Developing a new version of the SF-6D health state classification system from the SF-36v2: SF-6Dv2. Med Care 58:557–65.
    OpenUrl
  20. ↵
    1. Brazier JE,
    2. Rowen D,
    3. Mavranezouli I,
    4. et al.
    (2012) Developing and testing methods for deriving preference-based measures of health from condition-specific measures (and other patient-based measures of outcome). Health Technol Assess 16:1–114.
    OpenUrlCrossRefPubMed
  21. ↵
    1. Mulhern BJ,
    2. Bansback N,
    3. Norman R,
    4. et al.,
    5. SF-6Dv2 International Project Group
    (2020) Valuing the SF-6Dv2 classification system in the United Kingdom using a discrete-choice experiment with duration. Med Care 58:566–73.
    OpenUrlPubMed
  22. ↵
    1. Dolan P,
    2. Jones-Lee M
    (1997) The time trade-off: a note on the effect of lifetime reallocation of consumption and discounting. J Health Econ 16:731–9.
    OpenUrlCrossRefPubMed
  23. ↵
    1. Devlin NJ,
    2. Shah KK,
    3. Feng Y,
    4. et al.
    (2018) Valuing health-related quality of life: an EQ-5D-5L value set for England. Health Econ 27:7–22.
    OpenUrlCrossRefPubMed
  24. ↵
    1. Szende A,
    2. Oppe M,
    3. Devlin N
    , eds (2007) EQ-5D value sets: inventory, comparative review and user guide (Springer, Dordrecht (Netherlands)), pp 1–96.
  25. ↵
    1. Lancsar E,
    2. Fiebig DG,
    3. Hole AR
    (2017) Discrete choice experiments: a guide to model specification, estimation and software. PharmacoEconomics 35:697–716.
    OpenUrlPubMed
  26. ↵
    (2017) 2016 Census of population. Cat no 98-402-X2016005 (Statistics Canada, Ottawa) Availablehttps://www12.statcan.gc.ca/census-recensement/2016/index-eng.cfm. accessed 2022 June 16.
  27. ↵
    1. Selim AJ,
    2. Rogers W,
    3. Qian SX,
    4. et al.
    (2011) A preference-based measure of health: the VR-6D derived from the Veterans RAND 12-Item Health Survey. Qual Life Res 20:1337–47.
    OpenUrlPubMed
  28. ↵
    1. Whitehurst DGT,
    2. Bryan S
    (2011) Another study showing that two preference-based measures of health-related quality of life (EQ-5D and SF-6D) are not interchangeable. But why should we expect them to be? Value Health 14:531–8.
    OpenUrlCrossRefPubMed
  29. ↵
    1. Whitehurst DGT,
    2. Bryan S,
    3. Lewis M
    (2011) Systematic review and empirical comparison of contemporaneous EQ-5D and SF-6D group mean scores. Med Decis Making 31:E34–44.
    OpenUrlCrossRefPubMed
  30. ↵
    1. McDonald R,
    2. Mullett TL,
    3. Tsuchiya A
    (2020) Understanding the composite dimensions of the EQ-5D: an experimental approach. Soc Sci Med 265:113323.
    OpenUrl
  31. ↵
    1. Versteegh MM,
    2. Brouwer WBF
    (2016) Patient and general public preferences for health states: a call to reconsider current guidelines. Soc Sci Med 165:66–74.
    OpenUrl
  32. ↵
    1. Whitehurst DGT,
    2. Engel L
    (2018) Disability discrimination and misdirected criticism of the quality-adjusted life year framework. J Med Ethics 44:793–5.
    OpenUrlAbstract/FREE Full Text
  33. ↵
    1. Helgesson G,
    2. Ernstsson O,
    3. Åström M,
    4. et al.
    (2020) Whom should we ask? A systematic literature review of the arguments regarding the most accurate source of information for valuation of health states. Qual Life Res 29:1465–82.
    OpenUrlPubMed
  34. ↵
    1. Engel L,
    2. Bansback N,
    3. Bryan S,
    4. et al.
    (2016) Exclusion criteria in national health state valuation studies: a systematic review. Med Decis Making 36:798–810.
    OpenUrlCrossRefPubMed
  35. ↵
    1. Jonker MF,
    2. Donkers B,
    3. de Bekker-Grob E,
    4. et al.
    (2019) Attribute level overlap (and color coding) can reduce task complexity, improve choice consistency, and decrease the dropout rate in discrete choice experiments. Health Econ 28:350–63.
    OpenUrlPubMed
  • © 2022 CMA Impact Inc. or its licensors
PreviousNext
Back to top

In this issue

CMAJ Open: 10 (3)
Vol. 10, Issue 3
1 Sep 2022
  • Table of Contents
  • Index by author

Article tools

Respond to this article
Print
Download PDF
Article Alerts
To sign up for email alerts or to access your current email alerts, enter your email address below:
Email Article

Thank you for your interest in spreading the word on CMAJ Open.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Estimation of a Canadian preference-based scoring algorithm for the Veterans RAND 12-Item Health Survey: a population survey using a discrete-choice experiment
(Your Name) has sent you a message from CMAJ Open
(Your Name) thought you would like to see the CMAJ Open web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Citation Tools
Estimation of a Canadian preference-based scoring algorithm for the Veterans RAND 12-Item Health Survey: a population survey using a discrete-choice experiment
Nick Bansback, Logan Trenaman, Brendan J. Mulhern, Richard Norman, Rebecca Metcalfe, Richard Sawatzky, John E. Brazier, Donna Rowen, David G.T. Whitehurst
Jul 2022, 10 (3) E589-E598; DOI: 10.9778/cmajo.20210113

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Share
Estimation of a Canadian preference-based scoring algorithm for the Veterans RAND 12-Item Health Survey: a population survey using a discrete-choice experiment
Nick Bansback, Logan Trenaman, Brendan J. Mulhern, Richard Norman, Rebecca Metcalfe, Richard Sawatzky, John E. Brazier, Donna Rowen, David G.T. Whitehurst
Jul 2022, 10 (3) E589-E598; DOI: 10.9778/cmajo.20210113
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like

Related Articles

  • PubMed
  • Google Scholar

Cited By...

  • No citing articles found.
  • Google Scholar

Similar Articles

Collections

  • Clinical
    • Psychiatry
      • Anxiety disorders (including OCD and PTSD)
      • Mood disorders (including depression)
    • Mental health
  • Nonclinical
    • Patients
      • Patients' views
    • Statistics & Research Methods

Content

  • Current issue
  • Past issues
  • Collections

About

  • General Information
  • Staff
  • Editorial Board
  • Advisory Panel
  • Contact Us
  • Reprints
  • Copyright and Permissions
CMAJ Group

Copyright 2025, CMA Impact Inc. or its licensors. All rights reserved. ISSN 2291-0026

All editorial matter in CMAJ OPEN represents the opinions of the authors and not necessarily those of the Canadian Medical Association or its subsidiaries.

To receive any of these resources in an accessible format, please contact us at CMAJ Group, 500-1410 Blair Towers Place, Ottawa ON, K1J 9B9; p: 1-888-855-2555; e: [email protected].

CMA Civility, Accessibility, Privacy

 

 

Powered by HighWire