Objective International differences in colorectal cancer (CRC) survival and stage at diagnosis have been reported previously. They may be linked to differences in time intervals and routes to diagnosis. The International Cancer Benchmarking Partnership Module 4 (ICBP M4) reports the first international comparison of routes to diagnosis for patients with CRC and the time intervals from symptom onset until the start of treatment. Data came from patients in 10 jurisdictions across six countries (Canada, the UK, Norway, Sweden, Denmark and Australia).
Design Patients with CRC were identified via cancer registries. Data on symptomatic and screened patients were collected; questionnaire data from patients’ primary care physicians and specialists, as well as information from treatment records or databases, supplemented patient data from the questionnaires. Routes to diagnosis and the key time intervals were described, as were between-jurisdiction differences in time intervals, using quantile regression.
Participants A total of 14 664 eligible patients with CRC diagnosed between 2013 and 2015 were identified, of which 2866 were included in the analyses.
Primary and secondary outcome measures Interval lengths in days (primary), reported patient symptoms (secondary).
Results The main route to diagnosis for patients was symptomatic presentation and the most commonly reported symptom was ‘bleeding/blood in stool’. The median intervals between jurisdictions ranged from: 21 to 49 days (patient); 0 to 12 days (primary care); 27 to 76 days (diagnostic); and 77 to 168 days (total, from first symptom to treatment start). Including screen-detected cases did not significantly alter the overall results.
Conclusion ICBP M4 demonstrates important differences in time intervals between 10 jurisdictions internationally. The differences may justify efforts to reduce intervals in some jurisdictions.
- health policy
- organisation of health services
- international health services
- gastrointestinal tumours
This is an open access article distributed in accordance with the Creative Commons Attribution 4.0 Unported (CC BY 4.0) license, which permits others to copy, redistribute, remix, transform and build upon this work for any purpose, provided the original work is properly cited, a link to the licence is given, and indication of whether changes were made. See: https://creativecommons.org/licenses/by/4.0/.
Statistics from Altmetric.com
- health policy
- organisation of health services
- international health services
- gastrointestinal tumours
Strengths and limitations of this study
This is the first international study of this scale to use standardised survey methods to systematically examine key intervals from patients first noticing symptoms or bodily changes until the start of treatment for their colorectal cancer.
Questionnaire data were enriched and validated with registry data (cancer registry and screening programmes) and data rules were applied consistently to ensure validity.
As with all questionnaire based studies, there may be some response differences due to participant interpretation, cohort characteristics and sampling strategy, but we did not find obvious differences between study participants which could bias our results.
While our analyses adjusted for age, gender and comorbidity, we were unable to adjust for ethnicity and education due to different classification systems in participating countries.
Understanding variations in diagnostic and treatment intervals for patients with colorectal cancer may, in jurisdictions with longer intervals, signal the need for improvements in service configuration and patient pathways.
Colorectal cancer (CRC) is a leading cause of morbidity and mortality and places a major burden on health systems; worldwide 1.36 million new cases are diagnosed every year.1 CRC is the second most common cause of death due to cancer in Europe after female breast cancer, accounting for more than 200 000 deaths per year.2 Prognosis strongly depends on stage at diagnosis, and the disease can mostly be cured if diagnosed at an early stage. Survival has increased over the last several years in Europe.3 However, there remains substantial international variation in both 1-year and 5-year survival, with countries such as the UK and Denmark having significantly poorer survival than other countries such as Sweden, Canada and Australia (figure 1).4 Some of the variation derives from differences in stage at diagnosis which, in turn, is a result of the pathway to diagnosis and treatment.5 Therefore, it is crucial to investigate international differences in this pathway for CRC.
The International Cancer Benchmarking Partnership (ICBP) aims to not only quantify survival differences in comparable countries but to explore factors which may impact on observed differences.6 Module 4 (ICBP M4) focusses on the routes to diagnosis and length of diagnostic and treatment intervals as a means of understanding differences in cancer prognosis between countries. This may help shape policy and practice interventions in participating jurisdictions.6
Diagnosis of CRC can be difficult; the symptoms are often vague (eg, fatigue and non-specific abdominal pain), and this poses a significant diagnostic challenge for primary care, where most patients with CRC present.7–12 There is growing evidence that prolonged diagnostic and treatment intervals are associated with poorer outcomes in CRC.13 14 Access to investigations such as flexible sigmoidoscopy and colonoscopy is a further key issue; open access may expedite diagnosis and effect short diagnostic intervals.12 15
Many countries have implemented screening—typically faecal occult blood test (FOBT)-based—which can make a significant contribution to improved CRC outcomes.16 However, currently, the large majority of CRC diagnoses are based on symptomatic presentation—for example, seeking help in primary care or attending emergency services.7 17
This study aims to systematically compare the diagnostic routes and time intervals from first noticing symptoms to start of treatment in patients with CRC in 10 healthcare systems with broadly similar access to high quality treatment and valid cancer registration.6
The methods for ICBP M4 have been described.18 In brief, we recruited patients through cancer registries in 10 jurisdictions: Victoria (Australia); Manitoba and Ontario (Canada); Denmark; Norway; Sweden; England, Scotland, Northern Ireland and Wales (the UK). The target was to recruit 200 symptomatic recently diagnosed patients with CRC per jurisdiction and to measure the patient, primary care, diagnostic, treatment and total intervals (figure 2).
In defining these intervals we used principles articulated in the Aarhus Statement.19 Data were collected from patients, their primary care physician (PCP) and their cancer treatment specialists (CTSs) as well as cancer registries. When calculating the route and time intervals we used predefined rules including a data ‘hierarchy’ around these information sources (online supplementary file 1). Based on a standardised protocol, teams within each jurisdiction established data collection processes with registries; survey logistics and data management were adapted to each local setting.
Supplementary file 1
Data were transferred in anonymised format to the analysis team at Aarhus University—all data sources were combined into a single database.
Identification of study population
Eligible patients were consecutive patients aged 40 years or more with a first-diagnosis of CRC, International Classification of Diseases (ICD) 10 coded as C18.0-C18.9, C20.0 and C20.9.20 Patients who had had another non-index cancer earlier were eligible, but those with synchronous different primary cancers were excluded.18
Each jurisdiction used a registry-based identification to enhance validity. We aimed to recruit patients 3–6 months after diagnosis; this avoided approaching patients too soon after diagnosis, while minimising recall bias from a long period post-diagnosis.
Recruitment was via cancer registries; either through (1) sending a letter to the relevant healthcare professional, requesting a preaddressed envelope be forwarded to the patient on confirmation the person was aware of the diagnosis or (2) the local research team or registry directly sending a letter to the patient.18 Consent was required from all patients prior to surveys being sent to PCPs and CTSs and for data transfer.
Data from three questionnaires of eligible patients, their PCPs and CTSs (online supplementary files 2-4) were combined with information from participating cancer registries. Development, validation and implementation of these surveys is explained elsewhere.18
Questionnaires were developed collaboratively with all jurisdictions. For consented patients, based on practice lists or the patient’s response, a questionnaire was sent to the PCP with whom they were listed or who had been primarily involved in the diagnostic pathway. The patients and PCPs were asked about milestones, symptoms and route to cancer diagnosis. A questionnaire was sent to the CTSs who were first involved in the treatment. Jurisdictional differences in local recruitment processes are detailed in online supplementary files 5.
Supplementary file 5
To enhance complete and valid data on date of diagnosis, stage and screening status, data were collected through cancer registries wherever possible. Date of diagnosis was defined based on an established International Agency for Research on Cancer (IARC) hierarchy and stage was preferably given in tumour, node and metastasis and Duke’s.21 22
Local teams entered data and questionnaire responses. The records were checked for obvious errors (eg, dates in future) and queries were discussed and resolved with local contacts, who checked responses against original survey. All survey data underwent cleaning centrally (Aarhus University) to ensure that the same explicit rules were applied on the full dataset. Patients where age, date of diagnosis or date of consent were unknown were excluded.
As described, the data rules allowed the combination of data from different sources in a standardised way that ensured reproducibility and transparency (online supplementary files 1). The rules, based on the Aarhus Statement,19 employed a ‘hierarchy’ principle in terms of the order in which data sources (patient, PCP, CTS, registry) should take precedence where responses between sources differed, and included imputation rules based on the available data. The exact rule was guided by the measure in question—for example, patient interval was collected primarily from the patient questionnaire whereas primary care time-points were collected from the PCP questionnaire. All the measures were further validated using algorithms for outliers and out of range responses (eg, negative time intervals).
Although the protocol mandated contacting patients within a 3–6-month time window after diagnosis, some local registries needed to extend this period, primarily due to delays in recording the cancer diagnosis.
Measures of routes to diagnosis
We defined routes to diagnosis for CRC using categories derived from the Aarhus Statement check-list—the following categories were used in the analysis19:
Visit PCP and accident and emergency (A&E).
Investigation for another problem.
Other/unknown routes to diagnosis.
Measures of time intervals
To ensure international comparison, the time interval definition was adapted from the Aarhus Statement and included the following time-points19:
First onset of symptoms: the time-point when first bodily change(s) and/or symptom(s) are noticed by the patient.
First presentation to healthcare: the time-point at which it would be at least possible for the clinician seeing the patient to have started investigating.
First referral to secondary care: the time-point at which the PCP refers the patient (and responsibility of the patient) to secondary/specialist care.
Date of diagnosis: date the definite diagnosis was made, defined by the IARC hierarchy.21
Date of start of treatment: the date where the patient started curative or palliative treatment or a decision not to treat.
The time intervals were calculated as the number of days between these time-points (figure 2). For screen detected CRC, the patient and primary care interval were not applicable, with other intervals calculated using screening date as the first time-point. All time-points were validated manually and negative intervals were set to 0 days. Missing day was imputed based on specific rules to ensure that the direction of a possible misclassification bias was known (online supplementary files 1).
Establishing screening status
Patients with CRC were categorised using data rules as ‘screen-detected’, ‘symptomatic’ or ‘other presentation’. In some jurisdictions it was possible to identify screen-detected cancers from registries; in others this categorisation depended on questionnaire responses. Due to differences in the understanding and registration of screening across jurisdictions, we specified symptom-based detection should include all patients who reported symptoms or A&E/primary care presentation, even if the patient had indicated ‘screening’ as the diagnostic route (unless their PCP or CTS specified a screening route). For UK countries the distinction between a screen-detected and non-screen-detected CRC was validated using registry data on screen-detected cases identified through public programmes.
Health status was measured using the self-reported general health item from the 36-Item Short Form Health Survey.23 Comorbidity was assessed from the patient survey as presence of four diseases (stroke, diabetes, lung or heart diseases) and categorised into: ‘none’, ‘medium’ (one or two) or ‘high’ (three or four). Educational level was categorised as ‘low’ (vocational school or lower) and ‘high’ (university or higher). Symptoms reported were divided into two categories: ‘a CRC specific symptom’ or ‘other symptoms’. This was based on a symptom coding done independently by two PCP-authors (DW and PV) with the aim of identifying symptoms where clinical suspicion could be raised.24
Quantile regression was used to estimate differences in intervals between all jurisdictions.25 We compared the 50th, 75th and 90th percentiles. Wales was used as the reference jurisdiction as it had the lowest CRC survival according to the ICBP Module 1 cancer survival benchmark.4 Quantile regression allows a comparison on the interval scale with optimal information on differences. Counting days, we used the ‘qcount’ procedure proposed by Miranda.26 Parameters were calculated with 1000 jittered samples. The differences (in days) in intervals between jurisdictions (using Wales as a reference) were calculated as marginal effects after quantile regression by setting the continuous covariate age to its mean value and the categorical covariates (gender and comorbidity) to their modes. Significance level was set to 0.05 or less, and 95% CI were calculated when appropriate. Statistical analyses were carried out using STATA V.14 software.
Sensitivity and validity analyses
All analyses were undertaken using the 6 and 9 months cut-off criteria for allowable interval from diagnosis to questionnaire completion. To estimate the effect of using patient reported intervals only, a sensitivity analysis based solely on patient data was performed. The effect of excluding patients for whom at least one time interval hadn’t been reported was also investigated.
Kappa coefficient and overall agreement percentage assessed the agreement on routes to diagnosis (screening and symptomatic presentation) between the different data sources. Kappa coefficients were interpreted using Landis’ and Koch’s criteria27: 0.00–0.20=slight, 0.21–0.40=fair, 0.41–0.60=moderate, 0.61–0.80=substantial, above 0.80=almost perfect.
Agreement between the different data sources was also assessed by Lin’s concordance correlation coefficient (CCC).28 The ICBP M4 definition of screening-status was validated against registry data on screening where available, and assessed by kappa and overall agreement percentage.
The research questions for this survey drew on an extensive literature elating diagnosis and treatment delays leading to negative patient experiences. While patient experience was not a primary outcome measure for this study, patients were given the opportunity to comment on their experience through questionnaire free-text response options (under separate analysis). Patients were involved in the piloting of study instruments to ascertain if recruitment and questionnaire content and dissemination strategies were appropriate, described elsewhere.18 Each jurisdiction has committed to communicating the findings and local implications of this study to organisations representing their study participants.
Patient characteristics and participation
Of 14 664 eligible patients, 3881 returned completed questionnaires (a 31% response rate, ranging from 19% in Norway to 69% in Denmark). Of these, 2866 (95%) were included in the analyses after application of inclusion and exclusion criteria. The study flow with identification, exclusion and responses for each jurisdiction is seen in table 1.
The characteristics of the included patients are detailed in table 2. The patient questionnaire was completed at a median of 5 months following diagnosis. For tumour stage the proportion of missing stage was high in Norway due to lack of registry data.
Routes to diagnosis
Routes to diagnosis were broadly similar, except for screening; of all patients, 16.1% had a screen detected CRC, ranging from 6.3% in Norway to 31.4% in Manitoba (table 3). In Denmark and Sweden CRC-screening had not been implemented at the time of study and screening status in Norway was determined by information from local screening trials. The proportion of screen-detected cancers in Northern Ireland is lower as most were excluded in the recruitment process, with the local team actively including symptomatic patients in order to reach the target of 200 symptomatic patients. Overall (excluding Northern Ireland), most (82.2%) respondents presented with a symptomatic CRC. A high level of agreement was found between ICBP and registry data for screening status (table 3). PCP data indicated that the percentage of the symptomatic patients urgently referred with a suspicion of cancer was less than 50% in Ontario, Denmark, Manitoba, Norway (37.8%, 39.6%, 46.8%, 47.8%, correspondingly) and larger than 50% in Scotland, Northern Ireland, Victoria, England, Wales (51.5%, 57.1%, 57.6%, 67.6%, 69.3%, correspondingly).
Symptom prompting concern
The proportion (%) of patients and PCPs reporting symptoms are shown in table 4 (for Northern Ireland and Sweden, only patient data were collected). The most common symptom reported by PCP respondents was rectal bleeding (40% of respondents), followed by change in bowel habit. While every third patient indicated fatigue as a key symptom, it was rarely reported by PCPs.
The median patient interval varied from 21 days (Denmark) to 49 days (Wales) (table 5). Table 6 shows the adjusted patient median interval was 25 days shorter in Denmark than in Wales; none were longer compared with Wales (online supplementary files 6).
Supplementary file 6
The median primary care interval was 12 days in Norway (table 5), statistically significantly longer than Wales (table 6). For the 10% of patients waiting longest for referral, the longest intervals were observed in Manitoba, Victoria and Scotland (210, 128 and 93 days, respectively) (table 5). This interval at the 90th percentile was either 4 months (Manitoba) or 1 month (Victoria and Scotland) longer than in Wales (table 6).
The median diagnostic interval for symptomatic patients ranged from 27 to 28 days in Denmark and Victoria to 76 days in Manitoba. At the 90th percentile it ranged from 4 months in Denmark to 10 months in Ontario (table 5). All jurisdictions had shorter median diagnostic intervals compared with Wales, except Northern Ireland and Manitoba, where the intervals were 5 and 14 days longer, respectively (table 6).
The shortest median treatment intervals for all patients (about 2 weeks) were observed in Denmark, Victoria and Norway. In other jurisdictions this interval was 1 month or more (table 5). All jurisdictions had shorter treatment intervals compared with Wales, except Scotland and Manitoba (table 6).
The median total interval (from first symptom to treatment start) for all patients with CRC was between 74 days (Victoria) and 136 days (Manitoba) (table 5). In Scotland, Denmark, Norway and Victoria, this interval was statistically significantly shorter than in Wales (table 6).
Sensitivity and validity analyses
Changing the cut-off survey completion date from 9 months post-diagnosis to the per-protocol 6 months changed the number of included patients. However, the estimates of routes to diagnosis and time intervals were not significantly altered and the trend was the same as in the main analyses (results not shown). Sensitivity analysis based solely on patient data for those whose reporting on all time intervals was complete did not change the trend (results not shown). Comparing patient and PCP reported routes (screening and symptomatic presentation) and registry data on screening showed that agreement ranged from moderate to almost perfect (kappa 0.59–0.86).
Comparing the dates between the different data sources showed a high agreement between all data sources for all categories of dates (CCC ≥ 0.95 for date of diagnosis, CCC=0.94 for date of treatment and CCC=0.92 for date of first presentation to primary care). The analysis of the ICBP M4 definition of screening route compared with registry data showed an almost perfect agreement (kappa >0.80) in two jurisdictions and substantial (kappa >0.70) in two jurisdictions (table 3).
Our study showed marked variation in the proportion of screen-detected cancers, lengths of diagnostic and treatment (and total) intervals between jurisdictions. Patient intervals were shortest in Denmark and longest in Wales; longer primary care intervals were present in Norway, Scotland, Manitoba and Victoria. Differences in primary care intervals do not necessarily reflect PCP delay—they may arise from PCPs undertaking more investigations prior to referral. Overall, the differences are marked and suggest the need, in some jurisdictions, for revised diagnostic pathways to reduce the time taken for patients to be diagnosed and treated.
The interval differences did not show an obvious association with earlier reported survival differences—jurisdictions with poorer survival did not consistently show longer intervals, and vice versa.4 While this may question the validity of our findings, and/or the relationship between diagnostic intervals and survival, it is important to note these analyses were several years apart, and there may have been significant health system changes since the survival comparisons. Nevertheless, our study adds to a growing body of evidence on routes to diagnosis and time intervals; there are few similar examples involving multiple countries in the literature.29
Many factors underpin the differences observed between jurisdictions, such as structural differences in healthcare delivery (eg, care pathways, availability and accessibility of diagnostic and treatment facilities, etc.). Differences in routes to diagnosis were influenced by the presence or absence of screening programmes (CRC-screening had not been implemented in Denmark and Sweden at the time of study). Patient interval variation may reflect differences in symptom awareness or health-seeking behaviour. However, a study which compared awareness of cancer symptoms, attitudes towards cancer and barriers to attending a PCP did not demonstrate statistically significant differences.30 It is possible that other factors, such as culture, rurality, economic or patient-specific barriers and facilitators, influence this important part of the pathway.31
There were significant differences in primary care intervals; overall these intervals were much longer than those found in our breast cancer analyses.32 This raises important questions about diagnostic processes within primary care. There are widespread calls for PCPs to play a greater role in improving outcomes in CRC.33 Indeed, it is now widely acknowledged that primary care has a major role in cancer control at all stages of the patient journey.34 Nevertheless, there is some evidence based on observational associations that prolonged intervals might be associated with stronger ‘gatekeeper’ systems.35 Further, independent of the nature of symptoms, investigation in primary care has been noted to be associated with later referral for specialist assessment for CRC and other cancers—so a long primary care interval may mean that PCPs are doing more before they refer.36 Hence, PCP access to and use of investigations and differing national cancer referral guidance may influence primary care intervals. There were differences in diagnostic intervals, suggesting that once patients have been referred to secondary care there is considerable variation in their experiences; differences in treatment intervals were less marked. These variations suggest there is room for improvement in reducing the total interval and its various subintervals, and that guidance on optimal pathways should be better implemented. Each participating jurisdiction will likely draw unique conclusions about the most appropriate response to our findings.
Strengths and weaknesses
A major strength of this study is its use of standardised survey methods in a broad range of jurisdictions to systematically examine the various components of these intervals and to describe and compare, between countries, patient journeys to a cancer diagnosis and treatment. To ensure comparability across jurisdictions, our surveys drew on existing instruments and went through an extensive process of cognitive testing, piloting, and translation and adaptation.18
Data quality was enriched by information from national cancer registries and our algorithms showed very good agreement for jurisdictions where validation was possible. Using validated identification of patients with CRC minimised the risk of missing cancer cases during inclusion and of selection bias. Further, the use of registries made it possible to exclude patients with previous cancer in the same site, providing a homogeneous group of newly diagnosed patients with CRC in need of diagnostic work-up.
It was evident that there were subtle differences in the understanding of ‘screening’ between jurisdictions. Patients do not always distinguish between tests for screening and those for symptom-based diagnosis. Including data from registries and triangulating patient and PCP data enhanced the validity of ‘screen-detected versus non-screen-detected’ information, but the underlying factors varied between jurisdictions—for example, in Australia PCPs often provide screening FOBTs during consultations whereas this is rare in the UK and Scandinavia. To counter these inconsistencies, we applied our validated data rules which showed a high agreement with screening registries.
There are inevitable differences in questionnaire interpretation, characteristics of non-responders and availability of supplementary data for validation purposes. There are always considerations with questionnaire interpretation but the methodology and analysis of data sought to minimise or account for this as much as possible. Further, we used triangulation and comprehensive data rules to ensure validity, consistency and preserve statistical precision.18 We included registry data where possible (screening, stage, date of diagnosis) and developed reliable rules for imputation based on these registry data. To minimise misclassification from data entry and handling, data entry was internally audited by local teams and data interpretation was reduced to an absolute minimum and only performed centrally. Recall bias was minimised by the triangulation of different data sources and by ensuring that the patients received the questionnaire with a limited time window after the cancer diagnosis.
The overall response of 31% for the patient survey varied between jurisdictions. There were likely differences in the selection bias in individual jurisdictions; our patient sampling strategy will have led to some differences in the composition of our samples, as some patients were included directly from registries, some via PCPs and, in Northern Ireland, research nurses checked lists of potentially eligible patients to confirm eligibility. We have no mechanism to examine the direction of such possible selection bias. However, comparison of participating patients on a number of variables (including comorbidity, self-assessed health, smoking, stage, presenting symptom) did not show obvious differences with potential to bias our results. We also compared age, sex and stage of cancer among participants versus eligible patients and found no significant differences. There were different classification systems for ethnicity and education across jurisdictions which would lead to biased estimates if included in the regression model, even if mapped or harmonised—hence, they were excluded. There were few respondents from minority ethnic groups, limiting the generalisability of our findings; further work should target these groups as they are likely to have unique characteristics in their routes to diagnosis.
Confounding from aspects related to the diagnostic route for CRC was diminished by adjusting for age, gender and comorbidity. It is possible that there is some residual confounding which can bias the results in different directions. The statistical precision of the study was high as we were able to show clinically significant differences of 1 week in time intervals.
Comparison with other studies
Other studies have examined symptoms and routes to diagnosis for patients with CRC—although rarely in more than a single setting. A UK study on patients diagnosed in 2001–2002 and 2007–2008 reported median diagnostic intervals of 100 and 80 days, respectively.37 A Danish study showed median diagnostic intervals for CRC of between 31 and 55 days, depending on the timing of measurement in relation to implementation of pathway guidance.38 The difference between the present study and the former may stem from methodology issues, especially data sources (ie, databases vs surveys).
A study in Spain showed a symptom to diagnosis interval for CRC of 128 days and symptom to treatment interval of 155 days—these authors found that nature of symptoms, perceived seriousness of symptoms by patients, and place of first presentation influenced diagnostic and treatment intervals.39 Sampling strategies and survey differences will have influenced the results, making it difficult to compare these studies; nevertheless, they confirm that our results are broadly consistent with previous, single-jurisdiction studies.
The pattern of symptoms in ICBP participants was similar to other studies.40 CRC is known to be a cancer that clinically presents with either ‘alarm’ symptoms or more vague symptoms; there is evidence that doctors and patients respond less promptly to some symptoms of CRC than others—and that this can be influenced by the presence of co-morbid conditions.41 However, the minor differences in symptom patterns seen in table 4 are insufficient to explain the between-jurisdiction variation we have demonstrated in routes to diagnosis and diagnostic, treatment and total intervals.
Explaining observed differences between jurisdictions
The variation we see between jurisdictions mostly derives from differences in the extent to which healthcare systems support expedited CRC diagnosis and treatment; indeed, some health system characteristics, such as access and patient mobility between healthcare providers, may influence cancer outcomes—although these factors require further exploration.42 In Denmark there have been a number of reforms specifically designed to reduce diagnostic intervals.43 This study indicates a potential to optimise diagnostic routes for CRC in some jurisdictions. This should ideally be in conjunction with screening efforts which is gaining traction across many Western countries in response to policy and guideline initiatives.44
This study demonstrates considerable absolute and relative differences between jurisdictions in time intervals from first symptom until treatment for CRC. These differences do not demonstrate an obvious relationship with existing ICBP survival differences between the jurisdictions. ICBP phase 2 will report survival estimates to 2014, at which point it may be possible to explore the relationship between interval lengths and survival estimates in the population. Further work is ongoing to explore the outcome of patients included in this ICBP M4 study alongside the interval lengths observed.
The median total interval, which varied between 74 and 136 days, indicates that there is unrealised potential to optimise pathways for CRC. The main differences were found for structural parts of the pathway (eg, those not relating to patient behaviours/actions). Further, there is a ‘tail’ of patients waiting many months longer to start treatment for their cancer which may affect their outcomes. While our study highlights important international differences in routes to diagnosis, further research is needed to understand these differences and elucidate the contribution of patient pathway guidance and implementation, and health system structures. Nevertheless, the data provide important prompts for jurisdictions and suggest considerable room for improvement in some areas; they will also serve as a benchmark for measuring the effectiveness of future interventions.
Supplementary file 2
Supplementary file 3
Supplementary file 4
We thank our colleagues and participating patients, primary care physicians and cancer treatment specialists in all ICBP jurisdictions. Catherine Foot, Martine Bomb, Brad Groves, Irene Reguilon and Samantha Harrison of Cancer Research UK for managing the programme. John Butler, Royal Marsden NHS Foundation Trust and ICBP Clinical Lead, for his advice. The ICBP M4 academic reference group (online supplementary file 2) for providing independent peer review and advice for the study protocol and analysis plan development. Throughout the study, invaluable guidance was provide by the ICBP Programme Board. Membership of this board: Aileen Keel (Scottish Government, Edinburgh, Scotland); Anna Boltong (Cancer Council Victoria, Carlton, Australia); Anna Gavin (Northern Ireland Cancer Registry, Queens University, Belfast, UK); David Currow (Cancer Institute New South Wales, Sydney, Australia); Gareth Davies (Wales Cancer Network, Cardiff, UK); Gunilla Gunnarsson (Swedish Association of Local Authorities and Regions, Stockholm, Sweden); Heather Bryant (Canadian Partnership Against Cancer, Toronto, Canada); Jane Hanson (Welsh Cancer National Specialist Advisory Group, Cardiff, UK); Kathryn Whitfield (Department of Health, Victoria Australia); Linda Rabeneck (Cancer Care Ontario, Toronto, Canada); Michael A Richards (Care Quality Commission, London, UK); Michael Sherar (Cancer Care Ontario, Toronto, Canada); Nicola Quin (Cancer Council Victoria, Carlton, Australia); Nicole Mittmann (Cancer Care Ontario, Toronto, Canada); Robert Thomas (Department of Health and Human Services, Victoria, Melbourne, Australia); Sara Hiom (Cancer Research UK); Sean Duffy (NHS England, London, UK); Chris Harrison (NHS England, London, UK); Søren Brostrøm (Danish Health and Medicines Authority, Copenhagen, Denmark); Stein Kaasa (University Hospital of Trondheim, Trondheim, Norway).
UM and PV contributed equally.
Contributors DW, PV, UM, AZF and HJ planned the study design, data collection, carried out the analyses and wrote the draft manuscript. DW, PV, UM, HJ, AB, AKK, RB, DHB, VC, ATG, EG, EH, ML, R-JL, YL, MM, DT, RDN, VW, IR and SH were responsible for local data collection (alongside the ICBPM4WG), management and interpretation, and have participated in writing and approving the final manuscript version.
Funding This work was supported by: CancerCare Manitoba; Cancer Care Ontario; Canadian Partnership Against Cancer (CPAC); Cancer Council Victoria; Cancer Research Wales; Cancer Research UK; Danish Cancer Society; Danish Health and Medicines Authority; European Palliative Care Research Centre (PRC), Norwegian University of Science and Technology (NTNU); Northern Ireland Guidelines Audit and Implementation Network (GAIN); Macmillan Cancer Support; National Cancer Action Team; NHS England; Northern Ireland Cancer Registry, funded by the Public Health Agency NI; Norwegian Directorate of Health; Research Centre for Cancer Diagnosis in Primary Care (CaP), Aarhus University, Denmark; Scottish Government; Swedish Association of Local Authorities and Regions; University College London and NIHR Biomedical Research Centre at University College London NHS Foundation Trust; University of Edinburgh; Victorian Department of Health and Human Services; Welsh Government.
Disclaimer The funding bodies had no influence on the design of the study and collection, analysis and interpretation of data, in writing the manuscript or whether to publish the results.
Competing interests None declared.
Patient consent Not required.
Ethics approval For each local data collection, there were specific procedures and approvals which included anonymised data transfer to University College London and Aarhus University (online supplementary file 5). Approvals were received from the following institutions: Cancer Council Victoria Human Research Ethics Committee [HREC 1125]; Health Research Ethics Board, University of Manitoba [HS15227 (H2012:105)]; Research Resource Ethics Committee, CancerCare Manitoba [RRIC#28-2012]; University of Toronto Research Ethics Board ; The Danish Data Protection Agency [2013-41-2030]; Swedish Ethics Review Board, Uppsala [2013/306]; Norway Regional committees for medical and health research ethics [2013/136/REK nord]; England, Wales and Scotland, NRES Committee East Midlands – Derby 2, local R&D for each health board [11/EM/0420]; Northern Ireland ORECNI Ethical approval, local governance for each health Trust [11/EM/0420].
Provenance and peer review Not commissioned; externally peer reviewed.
Data sharing statement The data that support the findings of this study are available from the named authors from each ICBP jurisdiction, but restrictions apply to the availability of these data and so are not publicly available. Data are however available from the authors on reasonable request and with permission of the ICBP Programme Board. Please contact the ICBP Programme Management team, based at Cancer Research UK, with any queries (email@example.com).
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.