A discrete choice experiment, embed in an on-line survey, was designed to elicit the preferences of a sample of younger women for a model of service delivery for a breast cancer risk-prediction service. The DCE was designed and analysed following published methodological recommendations (15) and reported in line with a published checklist (16) (see Appendix 1). Ethical approval was obtained from The University of Manchester’s Proportionate Research Ethics Committee (reference: 2024-21125-37858).
Conceptualising the Choice Question
To conceptualise the choice question, the integrated screening action model (I-SAM) of cancer screening behaviour was used as a framework for considering the steps needed for a woman to take part in a breast cancer risk-prediction service to guide decisions about early intervention such as receiving earlier breast screening (17). This framework suggests that women have to go through multiple stages to take up the intervention on offer: becoming aware; becoming informed; deciding to act; acting; and repeating if necessarily.
When considering a breast cancer risk-prediction service there are likely to be more required steps for a woman to take up the service and experience health benefit. This will include becoming aware of the service, becoming informed about the service, making a decision to have their risk predicted, acting to have their risk predicted, deciding to receive their risk feedback, acting to receive their feedback, decided to act on their risk information to reduce their risk, actually acting to change their cancer risk.
As the BCAN-RAY study aimed to explore the feasibility of introducing a breast cancer risk-prediction service for younger women, this DCE focuses on women’s decision as to whether in principle they would like their risk to be predicted or not. It was decided that including questions to ascertain if women would then decide to receive their risk feedback and act on their risk information to reduce their risk (using strategies provided by the health service), would make a single survey too long to complete.
Firstly, for women to choose to receive risk-prediction, they must be aware of the service. As such, the sample to be recruited for this study was defined as women who would potentially receive the service: women between the ages of 30-39. Secondly, to decide to receive risk-prediction, women must be adequately informed about the service. As such, in the discrete choice experiment, information materials explaining the concepts of breast cancer risk-prediction were included at the start of the study. These were modelled as closely as possible on existing National Health Service leaflets for breast cancer screening (18).
Survey Design
The DCE was embedded into an online survey which was programmed in Qualtrics. The survey (Appendix 2) comprised 5 sections: (i) an introduction to the survey explaining what is involved with risk-prediction for breast cancer in younger women (referred to as ‘training materials’ in a DCE); (ii) the choice questions; (iii) questions regarding respondents’ views on the survey; (iv) attitudinal questions about their risk behaviour and healthcare decision-making and (v) sociodemographic questions about themselves.
DCE Design
The DCE was framed around the choice question: “If you had to choose between the following breast cancer risk-prediction services, which would you choose?”. The respondents were asked to choose between two unlabelled (generic) alternatives and an opt-out option. The alternatives and opt-out option were described using six attributes assigned levels (see Table 1). The opt out option was described with fixed text: “You would not have your breast cancer risk predicted, you would be invited to breast cancer screening at age 50, if you were worried about cancer before this you would visit your GP”. An infographic was also included showing that 0 out of 100 people would be identified at high risk.
The attributes and levels for this study were identified using seven focus groups (with 29 women) and eight semi-structured interviews conducted online with women aged 30-39 years for a breast cancer risk assessment (19). These semi-structured focus groups and interviews were designed with input from patient and public involvement. The qualitative data were used to generate a long list of 19 potential attributes. This long list was grouped into three categories: attributes of information about the risk-prediction service; attributes of the risk-prediction intervention itself; attributes of the process of returning risk information. A final list of six attributes was produced by the research team (see table 1). The research team focussed on defining attributes and levels that would describe a risk assessment service that was feasible to deliver.
Table 1: Attributes and levels included in the DCE
|
Attribute
|
Description
|
Levels
|
Attribute Type (coding for analysis)
|
|
How risk is predicted
|
The combination of interventions used to predict a woman’s risk of breast cancer
|
- A questionnaire only
- A questionnaire and mammographic breast density
- A questionnaire and radiofrequency breast density
- A questionnaire and genetic test
- A questionnaire, mammographic breast density, and genetic test
- A questionnaire, radiofrequency breast density, and genetic test
|
Categorical (Effects coded)
|
|
Appointments needed to predict risk
|
How many appointments would a woman need to attend to have her risk predicted
|
|
Categorical (Effects coded)
|
|
Location of appointment
|
Where the woman would need to go to have her risk predicted
|
- Home
- General Practitioner (GP)
- A mobile van
- Hospital
- Community Centre
|
Categorical (Effects coded)
|
|
Possible Times for the Appointment
|
Which days and what times of day appointments are available to book
|
- 9am-5pm weekdays
- 9am-5pm weekdays and evenings or weekends
|
Categorical (Effects coded)
|
|
How appointments are booked
|
What the woman needs to do to book an appointment to have her risk predicted
|
- You are sent a litter with a fixed date and time
- You can book a date and time yourself online or on the phone
|
Categorical (Effects coded)
|
|
The likelihood that you are predicted to be at high risk of breast cancer
|
The probability that the results suggest a woman should be classed as high risk and receive earlier interventions to reduce the risk of cancer or identify cancers at an earlier stage
|
- 5 out of every 100 (5%) people would be predicted to be high risk
- 10 out of every 100 (10%) people would be predicted to be high risk
- 15 out of every 100 (15%) people would be predicted to be high risk
- 20 out of every 100 (20%) people would be predicted to be high risk
|
Continuous (Linear in main analysis; with checks for non-linear functional forms)
|
The Experimental Design
Experimental design for discrete choice experiments is the creation of choice questions by combining attributes and levels in a way which maximises the probability that preferences for all of the attributes and levels can be estimated with the lowest level of statistical uncertainty (statistical efficiency) (20). A full factorial design would result in an unfeasible number of 921,600 potential combinations of attributes and levels in choice sets. A D-efficient, main effects design was created using the choiceDes package in the programming software R (21). Illogical combinations of attributes and levels such as having a mammogram at home were removed from the design informed by expert clinical advice. The final experimental design comprised three blocks of ten questions with each participant randomised to receive one block. As 5 out of 6 attributes were categorical, a dominance test question was not included in the DCE design.
Background questions
Background questions were included in the online survey to enable a description of the study sample and also for use when analysing for preference heterogeneity. The questions included were: sociodemographic questions including level of education, religion, ethnicity and whether they had children. Respondents were also asked about their attitude to risk and questions about their level of health information seeking or avoiding behaviour.
Piloting
The survey was quantitatively piloted using a purposive sample of younger women (n=50) adults recruited through an online panel-provider (Pureprofile). The results were then analysed using a conditional logit model to ensure that the coefficients for all attributes and levels could be estimated. The experimental design for the study was not updated using the results of the quantitative pilot.
Study Population and Sample
The relevant study population was framed around younger women (aged between30 and 39 years) who by definition, all have an as yet (undefined) risk of developing breast cancer at some point in their lives. Participants who had previously been diagnosed with breast cancer or had a close relative with breast cancer were also excluded as individuals with a family history of cancer are already potential eligible for earlier interventions in the NHS. The online survey was fielded to a sample of younger women living in the UK recruited using an online panel-provider (Pureprofile). There are no acceptable statistical approaches to set the required sample size for a DCE. This study used the Orme rule of thumb to calculate a minimum sample size of 150 participants needed.
Although a sample size of 150 was the minimum required to estimate the preferences of the sample, a final target sample size of 1000 was set to allow for understanding heterogeneity in preferences. Respondents were sent a link to the online survey and reminders were not used. Respondents who completed the survey in a time that was under 2 standard deviations from the median were defined as ‘speedsters’ and not engaging with the survey and removed from the dataset. These speedsters were then ‘replaced’ by a sample of further respondents identified by the panel-provider. Using Qualtrics also allowed the identification of responses which were likely from ‘bots’ completing the survey. These bots were ‘replaced’ by a sample of further respondents identified by the panel-provider.
Data Analysis
An analysis plan was created which specified that individuals who did not complete the survey and those who always chose the same alternative would be excluded. Speedsters and bots were replaced at the data collection phase. Descriptive statistics for sociodemographic characteristics, behavioural questions and survey feedback were produced for respondents in the final sample.
Following data cleaning, the choice data were analysed using conditional logit models in which the continuous attributes were specified as linear, continuous variables and categorial attributes effects coded. A single constant was included to represent the probability of opting in versus opting out.
Different model functional forms will be estimated whereby two constants are used to represent the probability of selecting hypothetical risk-prediction or feedback scenario A or scenario B. This serves as a test as to whether participants were always choosing scenario A or B regardless of the levels shown.
A series of regression models were then used to assess non-linearity in preferences for the probability of being identified as high risk attribute. All tests of model specification will be made by comparing the Bayesian Information Criterion (BIC) of the different models. If a model specification is found to result in a lower BIC value then this suggests that the model specification adds sufficient additional explanatory power for the number of additional parameters in the model.
When a final functional form was selected, a random parameter logit model was then estimated to determine if a model which allows for preference heterogeneity provided a better fit for the data. A two-step process was followed, with an uncorrelated random parameter logit estimated first and then a fully correlated random parameter logit estimated. The fully correlated model allows for both differences in error between participants as well as differences in preferences.
To better understand whether there were particular groups with similar preferences, a latent class model was also be estimated. The best number of classes was chosen using the BIC statistic. When the number of classes was chosen, a further model was estimated to determine if any demographic characteristics were correlated with membership of the classes. All of the collected demographic classes were tested for class membership prediction.
Coefficients and associated robust standard errors (SEs) from the best fitting model were used to calculate predicted uptake probabilities for different hypothetical risk-prediction services. The hypothetical services reported in this paper are the most and least preferred services based on the choice model for aggregated preferences as well as an exemplar service representing the risk-prediction approach used in the BCAN-RAY study. Differences in predicted uptake among the different predicted classes from the latent class analysis will be explored.
All analyses were conducted using the Apollo package (version 0.3.5) in the open source software R (22,23).