Objective To assess user preferences for different aspects of sexually transmitted infection (STI) testing services.
Design A discrete choice experiment.
Setting 14 centres offering tests for STIs in East Sussex, England.
Participants People testing for STIs.
Main outcome measure (Adjusted) ORs in relation to preferred service characteristics.
Results 3358 questionnaires were returned; mean age 26 (SD 9.4) years. 70% (2366) were recruited from genitourinary medicine (GUM) clinics. The analysis suggested that the most important characteristics to users were whether ‘staff had specialist STI knowledge’ compared with ‘staff without it’ (OR 2.55; 95% CI 2.47 to 2.63) and whether ‘tests for all STIs’ were offered rather than ‘some’ (OR 2.19; 95% CI 2.12 to 2.25). They remained the most important two service characteristics despite stratifying the analysis by variables such as age and sex. Staff levels of expertise were viewed as particularly important by people attending CASH centres, women and non-men who have sex with men. A ‘text or call to a mobile phone’ and ‘dropping in and waiting’ were generally the preferred methods of results reporting and appointment system, respectively.
Conclusions This study suggests that people testing for STIs place particular importance on testing for all infections rather than some and staff with specialist STI knowledge. Thus, targets based purely on waiting up to 48 h for an appointment are misguided from a user perspective.
- Sexually transmitted infections
- discrete choice experiment
- economic analysis
- evidence-based medicine
- qualitative research
- sexual behaviour
- opportunistic infections
- opportunistic INF
- antiretroviral therapy
- AIDS conference
This is an open-access article distributed under the terms of the Creative Commons Attribution Non-commercial License, which permits use, distribution, and reproduction in any medium, provided the original work is properly cited, the use is non commercial and is otherwise in compliance with the license. See: http://creativecommons.org/licenses/by-nc/3.0/ and http://creativecommons.org/licenses/by-nc/3.0/legalcode
Statistics from Altmetric.com
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.
- Sexually transmitted infections
- discrete choice experiment
- economic analysis
- evidence-based medicine
- qualitative research
- sexual behaviour
- opportunistic infections
- opportunistic INF
- antiretroviral therapy
- AIDS conference
Sexually transmitted infections (STIs) such as HIV, gonorrhoea and syphilis not only impact health at an individual level1 ,2 they also affect the health of communities as a whole and can be costly to treat.3–5 The last 15 years in the UK have seen significant increases in the incidence of STIs.6 They have consequently become a major public health concern, as has access to appropriate healthcare facilities with significant increases in related clinic attendances.6 Testing for infections is a key method of limiting disease-related morbidities, onward transmissions and costs. Traditionally, STI testing services in the UK have been provided through hospital outpatient facilities via genitourinary medicine (GUM) clinics,6 but this is no longer exclusively the case—an increasing array of services is being offered including primary care routes in order to provide increased access and choice.7
The recent white paper Equity and Excellence8 emphasises the importance of basing services around patient's needs rather than patient's needs around services. Recent British Association of Sexual Health and HIV endorsed standards of care also emphasise the importance of involving patients and the public in the design of future sexual health services.9 However, quantitative research with respect to assessing preferences for STI-testing services remains limited.10–14 It is unclear therefore whether changes to service configurations represent improvements from a user perspective, whether preferences vary by observable characteristics such as age and sex or how they relate to current policy targets such as appointments within 48 h at GUM clinics.15
Preferences for different aspects of STI-testing services were assessed using a discrete choice experiment (DCE).16–19 DCEs require respondents to choose between a number of competing service options that vary in terms of their design and outcomes they produce. Each service option is described in terms of a number of attributes (eg, the time it takes to receive test results) and levels (eg, the same day, 1 or 5 days) and is compared against at least one alternative. The overall results indicate the relative preference of each attribute and level.
Final decisions on all aspects of the questionnaire design were made by the study Steering Group (SG) in consultation with a user Advisory Group (AG). The penultimate version of the questionnaire was piloted to minimise the risk of error. Ethical approval was obtained from Brighton West Research Ethics Committee (08/H1111/86).
Choice of attributes and levels
The attributes and associated levels were determined using a three-staged approach. First, issues that the SG and AG considered to be important a priori were combined with issues from a literature review to generate a list of candidate attributes and levels. Second, this list was used to generate a protocol for a formative evaluation using a qualitative design,19 ,20 which is the subject of a separate publication.21 This involved 10 community focus groups consisting of a total of 65 people who had previously used STI-testing services. Groups were quota sampled based on age, sexual identity and sex. Lastly, candidate themes/attributes (table 1: web extra) were prioritised if potential policy solutions existed that could in theory at least be operationalised.20 For example, the qualitative study highlighted the importance of more comprehensive testing, which could be provided given increased resources. On the other hand, while ‘friendly staff’ was also identified as an important issue, it was excluded from the design as it is less clear how it can be achieved even with extra resources. Consideration was also given to the inclusion of a cost attribute since it allows the monetary valuation of benefits.16 However, it was also excluded from the final DCE design as strong objections to the notion of ‘cost’ in the context of STI testing were raised in most of the focus groups. By the end of the process, and with general regard to overall questionnaire burden, six attributes were chosen, three with two levels and three with four levels (table 1). The final format required participants to indicate which of two service options they preferred, ‘A’ or ‘B’ (table 2).
A fractional factorial design was used in order to limit the number of questions participants were required to complete.19 It was generated using the Kuhfeld SAS macro22 ensuring that the design had orthogonally balanced levels and minimal overlap,19 resulting in 16 different service options (known as choice sets). The alternative options to these choice sets were constructed using the ‘fold over’ approach, ensuring a 100% statistically efficient design.19 ,23 To reduce the number of questions, individuals were required to complete, the 16 choice sets were (non-randomly) halved using the SAS ‘mktblock’ macro.22 A second version of each questionnaire was also generated in which the questions appeared in a different randomly chosen order. Finally, a single test of consistency (in which one option was logically superior to the other) was added to each version of the questionnaire in order to gauge how difficult participants found the DCE task. This meant that each individual was required to answer a total of nine DCE questions, eight main DCE questions and a consistency test. While DCEs sometimes contain more than a single consistency question, one was considered proportional given the modest number of DCE questions. Respondents also provided general background information such as sex, age, maximum educational qualification, sexual preference, frequency of previous STI tests and presence of STI symptoms.
Non-commercial STI testing centres in East Sussex, England, participated in the study between January and June 2010. Fourteen centres agreed to recruit users, including four GUM clinics, six locally enhanced services (LES), a contraceptive advice and sexual health (CASH) service and three non-NHS ‘community-based’ centres (two young people's services and a Terrence Higgins Trust centre). CASH centres offer a broader range of sexual health services than STI testing such as family planning. LES's are typically primary care services that offer increased levels of STI testing with staff who are more likely to be GUM trained but not specialists.
Individuals seeking a STI test were verbally asked to participate in the study by either a researcher or member of staff and to complete the questionnaire before leaving. Participants had to be at least 16 years of age and provide written consent. Where centres held clinics at different times of the week, efforts were made to collect questionnaires at a variety of times. No attempt was made to calculate the number of users who declined to participate, as estimation methods proved unreliable given the number of clinics and different staff involved. Therefore, we recruited a convenience sample.
The data were analysed using random effects logistic regression analysis to account for multiple responses per participant, using STATA version 12.24 The six attributes were specified as 12 dummy variables as they were all categorical. The results are presented as unadjusted and adjusted ORs; the latter consisting of adjustments for demographic variables such as age, sex and employment status. The basic preference (or ‘utility’) function was assumed to be linear and additive.
A second objective was to assess whether preferences differed according to a number of basic demographic and background characteristics. To do this, a series of logistic regression analyses were undertaken in which each of the main DCE variables was interacted with the following: age, sex, presence of symptoms, whether CASH attendee, men who have sex with men (MSM), employment status and maximum educational qualifications. The results from these analyses were used to identify variables that independently predicted the levels on at least one DCE variable (data not shown but available on request). These associations are shown by reporting stratified ORs for a number of the subgroups of interest.
Three tests were undertaken in order to assess how ‘well/logically’ respondents completed the DCE questions. First, the direction of preference for each attribute was assessed against prior expectations. For example, if the questionnaires were answered logically, then all else equal individuals are more likely to prefer shorter to longer waiting times for test results. Second, the percentage of questionnaires in which dominance occurred was reported (ie, where respondents consistently appeared to base responses on the same level on one attribute rather than the levels on all six).16 Lastly, data for participants who answered the consistency question illogically were excluded in a sensitivity analysis.
A total of 3358 participants returned a questionnaire (table 3). Their mean age was 26 years (SD 9.4), 58% were women, 82% were heterosexual and almost 20% reported currently having a STI-related symptom. Over 70% of respondents were GUM attendees, and the majority of the remaining participants were from the CASH centre.
There were a number of differences in terms of the type of person who tested at each location. For example, CASH attendees were generally younger, more likely to be women with lower educational qualifications compared with GUM attendees. Logistic regression suggested that younger individuals were more likely to complete the questionnaire compared with older individuals (z=−1.97, p=0.05). No other demographic/location variables were found to be predictive of completeness of response.
The responses were generally consistent with logical expectations (table 1). For example, people preferred shorter to longer waiting times for test results. Ninety-two per cent of respondents answered the consistency question ‘logically’ and 99% of the DCE questions were completed. The tests for dominance revealed that 13% of individuals always chose to see specialist rather non-specialist staff and 13% always chose to ‘drop in and wait’ for an appointment. The percentage of dominant responses for the remaining attributes was negligible. The model predicted 74% of choices correctly with a McFadden's adjusted R2 of 0.22.
The results showed that the most important attributes to users were whether ‘staff had specialist STI knowledge’ compared with ‘not having specialist knowledge’ and the ‘comprehensiveness of testing’ (table 1). For example, the odds of respondents choosing clinics staffed by people with specialist STI knowledge was 2.55 (95% CI 2.47 to 2.63) times higher than the odds of choosing centres that were not, whereas the odds of choosing a site that tested for all STIs was 2.19 (95% CI 2.12 to 2.25) times higher than the odds of choosing a site that tested for some infections. The results also showed that, on average, ‘dropping in and waiting’ was the most preferred appointment system (OR 1.36; 95% CI 1.28 to 1.45) and that individuals preferred to receive test results on the same day (OR 1.61; 95% CI 1.52 to 1.72). A text or call to a mobile phone from the clinic was the most preferred results' notification option. Excluding questionnaires with a ‘failed’ consistency question and controlling for differences in variables such as age and sex had negligible effects on the results.
All the patient characteristics were significantly associated with at least one DCE variable. However, ‘staff knowledge’ and the ‘comprehensiveness of testing’ consistently remained the two most important attributes in terms of order, as they were always associated with the highest ORs (table 4). While they were of concern to all users, the results showed that women (OR 3.02; 95% CI 2.88 to 3.17), CASH attendees (OR 5.32; 95% CI 4.89 to 5.78) and non-MSM (OR 2.76; 95% CI 2.66 to 2.86) placed particular importance on staff with specialist knowledge compared with the remaining respondents, whereas the comprehensiveness of testing was particularly important to MSM (OR 2.94; 95% CI 2.67 to 3.22) compared with non-MSM (OR 2.13; 95% CI 2.06 to 2.21). The order of importance of the other DCE variables remained reasonably consistent across the different subgroups. For example, ‘dropping in and waiting’ and a ‘text or call to a mobile phone’ generally remained the preferred appointment and results reporting methods, respectively. However, some were shown to be of particular importance to specific subgroups. For example, there was some evidence to suggest that CASH attendees (OR 1.47; 95% CI 1.36 to 1.59) and women (OR 1.30; 95% CI 1.25 to 1.36) preferred receiving all test results rather than positive results only compared with non-CASH attendees (OR 1.20; 95% CI 1.16 to 1.24) and men (OR 1.17; 95% CI 1.12 to 1.23), respectively.
The results from this discrete choice experiment demonstrate that the most important issues to people testing for STIs are the comprehensiveness of testing and whether centres are staffed by STI specialists. While the absolute strength of preference for these attributes varied by subgroup, they remained the issues of most importance throughout. The results also suggest that users would generally prefer to ‘drop in and wait’ for tests compared with the remaining appointment systems and that there was a strong preference for receiving test results on the same day. While users also preferred to receive negative and positive test results rather than positive results alone (no news is good news), this was of particular importance to people testing at CASH centres and women. A text or call to a mobile phone from the test site was generally the most preferred method of receiving results.
The major strengths of this preference study are its large sample size and discrete choice design. This approach is more realistic than simply asking people what they prefer as it requires choices to be made, and it also allows the relative strength of different service characteristics to be assessed.
There are several limitations with the study. For example, the number of DCE attributes was limited to six. While the literature and qualitative analysis suggested that other issues were also important to people, such as ‘friendly staff’, including more attributes and levels would have significantly increased the number of questions, given the factorial design. Therefore, while the results are robust in terms of the relative preferences across the included attributes, it is less clear how they relate to other factors. Second, the focus was on people who were already engaged with STI testing services and the design meant that people were ‘forced’ to choose one of two service options. However, in a follow-on study using the same questionnaire but including an ‘opt-out’ option (data not shown) in a non-testing convenience sample (n=255), less than 1% of all responses indicated an unwillingness to test. Thus, there is some reason to believe that the service characteristics analysed in this study are unlikely to influence the likelihood of testing in the first instance. Third, while users indicated a strong preference for staff with specialist knowledge, the required policy response is less clear in so much that it could either involve increasing levels of staff training, ensuring users are fully informed about the skill levels of staff at different centres or both. Further research is required on this point. Fourth, a potential criticism is that the results could be influenced by individual's prior testing experiences. While this is possible, we think it unlikely to be important since the analysis adjusted for variables such as testing location, the DCE choices were between hypothetical services and over 50% of respondents had no prior testing experience. Moreover, even if this criticism is correct, the aim of the study was to estimate the strength of preference for different services characteristics rather than to assign some level of legitimacy to them. Lastly, the study does not take into account any personal or public health benefits of faster testing, other improvements to current services or their associated costs. Thus, while potential improvements in service provision have been identified from a ‘user’ perspective, they could be at odds with broader public health objectives such as reducing the likelihood of further transmissions. Further studies are required to evaluate these potential trade-offs.
Only a few studies have assessed preferences for STI testing services10 ,12–14 ,25 and only three have used DCE type designs. The first, however, was a US study that focused on HIV testing and methods of providing samples rather than broader service characteristics.10 Comparisons with the remaining two studies12 ,14 are more relevant as they are UK based and asked a number of similar questions. For example, in their sample of 746 current GUM users and the general population, Ross et al reported that individuals were indifferent to seeing GPs or STI specialists, although some ethnic groups expressed a preference for GPs. While these results are seemingly different from ours, we focused specifically on STI testing, not sexual healthcare more generally. Ross et al also reported that less than half of the respondents preferred to drop in and wait compared with the remainder who would rather have a booked appointment within 48 h. Our results are different in that users indicating that they would generally prefer to drop in and wait rather than waiting 48 h for an appointment. It is difficult to know exactly why the findings differ but Ross et al stipulated a 2 h wait at drop in centres, whereas we did not specify a length of time. However, the results from both studies are similar in that they both indicate an increased willingness to drop in and wait with increasing appointment waiting times.
The results showed that CASH participants had much stronger preferences for a number of attributes compared with other respondents, despite adjusting for differences in presenting characteristics such as age. For example, CASH participants expressed particularly strong preferences for centres staffed by STI specialists (OR 5.32; 95% CI 4.89 to 5.78). While the precise reasons for these differences are unclear, CASH centres by definition deal with other sexual health issues alongside STI testing, including family planning and reproductive health. Thus, while all CASH participants received a STI test, this might not have been their only or primary motivation for seeking healthcare, and in this sense, they represent a distinct user group.
Most sexual health services in primary care are either provided at GPs or LES. LES's were primarily established to provide increased choice to individuals given the excess demand for GUM appointments.26 Unlike GUM clinics, a typical LES will not test for all infections and is more likely to employ GUM-trained staff rather than specialists.27 However, these results suggest that unless they offer other advantages to users over and above shorter appointment waiting times, most individuals will continue to test at GUM clinics given a choice.
The time it takes to receive test results is likely to depend on a number of factors including local service arrangements, the type of tests taken and the infections that are being investigated. However, our results suggest that where feasible, all people testing for STIs would highly value receiving their results on the same day.
In 2005, the UKs Department of Health established a Public Service Agreement to ensure access to GUM services within 48 h.15 These results suggest that this objective might not be optimal from a user perspective as other appointment mechanisms were generally preferred and other issues were of greater concern.
In summary, the results from this study highlight particularly strong preferences for STI testing services that are staffed by people with specialist STI knowledge and sites that test for all infections rather than some. These findings remain unchanged despite adjusting for differences in characteristics such as age and where participants tested. They are perhaps of particular relevance if future policies continue to encourage more testing in primary care settings rather than GUM. They also suggest that a target based purely on a 48 h waiting time is misguided since it is neither the most favoured appointment mechanism nor the characteristic of most concern to users.
Users testing for STIs revealed particularly strong preferences for services staffed by specialist staff and sites testing for all infections.
While the strength of preference for these two service characteristics varied by patient subgroup, they consistently remained the issues of most concern to users.
Attempts to encourage more STI testing in primary care rather than GUM clinics should pay particular attention to these findings.
Service targets based purely on waiting up to 48 h for an appointment are misguided from a user perspective.
The authors would like to thank the staff at services who supported our research. In particular, we would like to thank consultants Dr Kazeem Aderogba (Avenue House, Eastbourne) and Dr Hitash Patel (Station Plaza Health Centre, Hastings) for their support. We would also like to thank our Advisory Group (Mark Cull, Jules and Ella Davies, Michael Hootman, Jonathon Roberts and Fern Terris-Prestholt), who provided invaluable comments throughout the project. Thanks to Georgina Frew, Katy Mason, Chloe Sakal and Sebastian Shaw who collected and entered data.
Funding This paper presents independent research commissioned by the National Institute for Health Research (NIHR) under its Research for Patient Benefit (RfPB) Programme (Grant Reference Number PB-PG-0407-13211). The views expressed are those of the author(s) and not necessarily those of the NHS, the NIHR, the Department of Health or the Advisory Group members. Service support costs were provided by the NIHR Primary Care Research Network-South East.
Competing interests None.
Patient consent Obtained.
Ethics approval Ethics approval was provided by Brighton West Research Ethics Committee.
Provenance and peer review Not commissioned; externally peer reviewed.