[Skip to Navigation]
Sign In
Figure 1.  Patient Selection and Exclusion
Patient Selection and Exclusion

AJCC indicates American Joint Commission on Cancer; SEER, Surveillance, Epidemiology, and End Results.

aPaget disease of the nipple without underlying tumor.

bInflammatory carcinoma (International Classification of Diseases for Oncology, 3rd edition, code 8530/3), diffuse, widespread—three-fourths or more of breast.

Figure 2.  Extent of Screening, Breast Cancer Incidence, and Mortality From Breast Cancer Among Women 40 Years and Older in 547 US Counties
Extent of Screening, Breast Cancer Incidence, and Mortality From Breast Cancer Among Women 40 Years and Older in 547 US Counties

Each circle presents data for a single county; circle area is proportional to the county population of women 40 years and older. Each fitted curve is a smoothing spline, presented with 95% confidence bands.

Figure 3.  Extent of Screening and Breast Cancer Incidence, Stratified by Tumor Size
Extent of Screening and Breast Cancer Incidence, Stratified by Tumor Size

A, Tumors ≤2 cm. B, Tumors >2 cm. Each circle presents data for a single county; circle area is proportional to the county’s population of women 40 years and older. Each fitted curve is a smoothing spline, presented with 95% confidence bands.

Figure 4.  Extent of Screening and Breast Cancer Incidence, Stratified by Stage and First-Course Surgical Treatment
Extent of Screening and Breast Cancer Incidence, Stratified by Stage and First-Course Surgical Treatment

A, Early disease indicates American Joint Commission on Cancer, 6th edition, stages 0-II and locally advanced and metastatic disease indicates American Joint Commission on Cancer, 6th edition, stages III-IV. We judged that locally advanced and metastatic disease is unambiguously progressive, in that overdiagnosed indolent disease would be unlikely among these cases, even if there were false-positive hormone status, lymph node involvement, or other test results. B, Breast-conserving indicates excisional biopsy, re-excision of biopsy site, lumpectomy, partial mastectomy, and segmental mastectomy. Non–breast-conserving indicates total or simple mastectomy, radical mastectomy, modified radical mastectomy, and subcutaneous mastectomy.

1.
Marmot  MG, Altman  DG, Cameron  DA, Dewar  JA, Thompson  SG, Wilcox  M.  The benefits and harms of breast cancer screening: an independent review.  Br J Cancer. 2013;108(11):2205-2240.PubMedGoogle ScholarCrossref
2.
Carter  JL, Coletti  RJ, Harris  RP.  Quantifying and monitoring overdiagnosis in cancer screening: a systematic review of methods.  BMJ. 2015;350:g7773.PubMedGoogle ScholarCrossref
3.
Barratt  A.  Overdiagnosis in mammography screening: a 45 year journey from shadowy idea to acknowledged reality.  BMJ. 2015;350:h867.PubMedGoogle ScholarCrossref
4.
Nelson  HD, Tyne  K, Naik  A, Bougatsos  C, Chan  BK, Humphrey  L; US Preventive Services Task Force.  Screening for breast cancer: an update for the US Preventive Services Task Force.  Ann Intern Med. 2009;151(10):727-737, W237-42.PubMedGoogle ScholarCrossref
5.
Humphrey  LL, Helfand  M, Chan  BKS, Woolf  SH.  Breast cancer screening: a summary of the evidence for the US Preventive Services Task Force.  Ann Intern Med. 2002;137(5, part 1):347-360.PubMedGoogle ScholarCrossref
6.
Biller-Andorno  N, Jüni  P.  Abolishing mammography screening programs? a view from the Swiss Medical Board.  N Engl J Med. 2014;370(21):1965-1967.PubMedGoogle ScholarCrossref
7.
Surveillance, Epidemiology, and End Results Program.  SEER*Stat databases: SEER 18 registries, November 2012 submission.http://seer.cancer.gov/data/seerstat/nov2012/. Accessed October 10, 2014.
8.
Surveillance, Epidemiology, and End Results Program.  SEER*Stat software, version 8.1.2.http://seer.cancer.gov/seerstat. October 10, 2014.
9.
Etzioni  R, Gulati  R, Mallinger  L, Mandelblatt  J.  Influence of study features and methods on overdiagnosis estimates in breast and prostate cancer screening.  Ann Intern Med. 2013;158(11):831-838.PubMedGoogle ScholarCrossref
10.
Greenlee  RT, Murray  T, Bolden  S, Wingo  PA.  Cancer statistics, 2000.  CA Cancer J Clin. 2000;50(1):7-33.PubMedGoogle ScholarCrossref
11.
Raghunathan  TE, Xie  D, Schenker  N,  et al.  Combining information from two surveys to estimate county-level prevalence rates of cancer risk factors and screening.  J Am Stat Assoc. 2007;102(478):474-486. doi:10.1198/016214506000001293.Google ScholarCrossref
12.
National Cancer Institute, Division of Cancer Control and Population Sciences, Statistical Methodology & Applications Branch.  Small area estimates for cancer risk factors & screening behaviors.http://sae.cancer.gov. Released May 2010. Accessed October 10, 2014.
13.
Breast Cancer Surveillance Consortium, National Cancer Institute.  Distribution of key variables.http://breastscreening.cancer.gov/data/variables/. Accessed February 23, 2015.
14.
Chu  KC, Miller  BA, Feuer  EJ, Hankey  BF.  A method for partitioning cancer mortality trends by factors associated with diagnosis: an application to female breast cancer.  J Clin Epidemiol. 1994;47(12):1451-1461.PubMedGoogle ScholarCrossref
15.
Kalager  M, Zelen  M, Langmark  F, Adami  H-O.  Effect of screening mammography on breast-cancer mortality in Norway.  N Engl J Med. 2010;363(13):1203-1210.PubMedGoogle ScholarCrossref
16.
Singletary  SE, Allred  C, Ashley  P,  et al.  Staging system for breast cancer: revisions for the 6th edition of the AJCC Cancer Staging Manual.  Surg Clin North Am. 2003;83(4):803-819.Google ScholarCrossref
17.
Carlson  RW, Anderson  BO, Bensinger  W,  et al; National Comprehensive Cancer Network.  NCCN practice guidelines for breast cancer.  Oncology (Williston Park). 2000;14(11A):33-49.PubMedGoogle Scholar
18.
Fritz  A, Ries  L, eds.  SEER Extent of Disease 1998, Codes and Coding Instructions (No. T-899; Pub. No. 98-4354). Bethesda, MD: National Cancer Institute; 1998.
19.
Seiffert  JE, ed.  SEER Program: Comparative Staging Guide for Cancer, Version 1.1. Bethesda, MD: National Cancer Institute; 1993.
20.
Wood  SN.  Thin plate regression splines.  J R Stat Soc Series B Stat Methodol. 2003;65(1):95-114. doi:10.1111/1467-9868.00374.Google ScholarCrossref
21.
Greenland  S, Robins  J.  Invited commentary: ecologic studies—biases, misconceptions, and counterexamples.  Am J Epidemiol. 1994;139(8):747-760.PubMedGoogle Scholar
22.
Wood  S.  Generalized Additive Models: An Introduction With R. Boca Raton, FL: CRC Press; 2006.
23.
Wood  SN.  Fast stable restricted maximum likelihood and marginal likelihood estimation of semiparametric generalized linear models.  J R Stat Soc Series B Stat Methodol. 2011;73(1):3-36. doi:10.1111/j.1467-9868.2010.00749.x.Google ScholarCrossref
24.
Wood  S.  mgcv: Mixed GAM computation vehicle with GCV/AIC/REML smoothness estimation. R package, version 1.7. http://cran.r-project.org/web/packages/mgcv/index.html. Accessed September 1, 2014.
25.
Marra  G, Wood  SN.  Coverage properties of confidence intervals for generalized additive model components.  Scand J Stat. 2012;39(1):53-74. doi:10.1111/j.1467-9469.2011.00760.x.Google ScholarCrossref
26.
Pasek  J, Tahk  A, Culter  G, Schwemmle  M.  Weights: Weighting and Weighted Statistics. http://cran.r-project.org/web/packages/weights/index.html. Accessed February 20, 2015.
27.
Rodgers  JL, Nicewander  WA.  Thirteen ways to look at the correlation coefficient.  Am Stat. 1988;42(1):59-66. doi:10.1080/00031305.1988.10475524.Google ScholarCrossref
28.
Taylor  R.  Interpretation of the correlation coefficient: a basic review.  J Diagn Med Sonogr. 1990;6(1):35-39. doi:10.1177/875647939000600106.Google ScholarCrossref
29.
Esserman  L, Shieh  Y, Thompson  I.  Rethinking screening for breast cancer and prostate cancer.  JAMA. 2009;302(15):1685-1692.PubMedGoogle ScholarCrossref
30.
Bleyer  A, Welch  HG.  Effect of three decades of screening mammography on breast-cancer incidence.  N Engl J Med. 2012;367(21):1998-2005.PubMedGoogle ScholarCrossref
31.
Anderson  WF, Jatoi  I, Devesa  SS.  Assessing the impact of screening mammography: breast cancer incidence and mortality rates in Connecticut (1943-2002).  Breast Cancer Res Treat. 2006;99(3):333-340.PubMedGoogle ScholarCrossref
32.
Greenland  S, Morgenstern  H.  Ecological bias, confounding, and effect modification.  Int J Epidemiol. 1989;18(1):269-274.PubMedGoogle ScholarCrossref
33.
Wakefield  J.  Ecologic studies revisited.  Annu Rev Public Health. 2008;29:75-90.PubMedGoogle ScholarCrossref
34.
Puliti  D, Duffy  SW, Miccinesi  G,  et al; EUROSCREEN Working Group.  Overdiagnosis in mammographic screening for breast cancer in Europe: a literature review.  J Med Screen. 2012;19(suppl 1):42-56.PubMedGoogle ScholarCrossref
35.
Njor  S, Nyström  L, Moss  S,  et al; Euroscreen Working Group.  Breast cancer mortality in mammographic screening in Europe: a review of incidence-based mortality studies.  J Med Screen. 2012;19(suppl 1):33-41.PubMedGoogle ScholarCrossref
36.
Jonsson  H, Johansson  R, Lenner  P.  Increased incidence of invasive breast cancer after the introduction of service screening with mammography in Sweden.  Int J Cancer. 2005;117(5):842-847.PubMedGoogle ScholarCrossref
37.
Esserman  LJ, Thompson  IM, Reid  B,  et al.  Addressing overdiagnosis and overtreatment in cancer: a prescription for change.  Lancet Oncol. 2014;15(6):e234-e242.PubMedGoogle ScholarCrossref
3 Comments for this article
EXPAND ALL
Age and Screening
Rabeb Khlifi | CRCHUM, Université de Montréal (UdeM), Montreal, Quebec,Canada
This is a very interesting study examining the association between the mammography screening rates at the county level and breast cancer incidence and mortality. Given the considerably controversy surrounding the optimal age at which women should be screened, it would be interesting to present separately the analysis for women 50 and over and those under 50 years of age.
CONFLICT OF INTEREST: None Reported
Suppose Another Conclusion is Possible?
Marybeth Lambe MD FAAFP | Watergrass Hill
This will sound rash & foolish in contrast with this brilliant and rather heroic research, but I am holding out hope to know further imagined data. Especially since this has set off such a boisterous clamor against screening mammograms in my community. Previous discussion has touched overtreatment but always centered around false positives & harm vs benefit so this was hugely surprising. I am just a lowly local physician, yet I was thinking of how carefully I send women for screening & just to whom & where I send them. I wondered if there was any chance this lovely analysis could be dissected down very small. If tiny subsets could be pulled out & examined—is there any chance to look at centers of excellence to know, well, if excellence matters in this? Are we still left with the uncomfortable truth of no correlation between screening and breast cancer mortality change?

In my practice & then in reading the research, I worry about who is reading the mammogram, or who then the ultrasound, how very experienced will be the patient’s cautious, wise breast specialist next. Even here in Seattle, I use much caution--many sites, many readers, many surgeons would be hopeless and overcall, over biopsy—so I never send them there. Multiple this by a million...
Is there any way of parceling out the data gathered & sorting it by excellence of the mammogram site, mammogram reader, sagacity of breast specialists used ? Would that alter the data; worse—would it not?

Frankly, the inequity, the unevenness of health care, of health prevention, in this country is appalling & quite well known. I know this research appears in a series of “Less is More” but I am old & I was taught to always first brainstorm all the possible diagnoses. Sometimes, there is more before we can clinch the case as less. Would we find instead a need for cloning the very few GOOD mammogram centers, the thoughtful mammogram radiology readers, the cautious breast surgeons, & caring referral people? Or worse, still to be wrong—for all the best & attentive specialists to make no difference. What I wonder, if the data could be pulled apart like that, the large US cohort, whirled down to small intersections; likely impossible. But, oh the value of knowing would make all the difference! Does great care alter these results?
Should the epilogue be instead? Be Careful Where, & with Whom, You Trust Your Breasts
CONFLICT OF INTEREST: None Reported
READ MORE
increased screening=increased detection
leong | private practice, former prof dartmouth
as expected, the study showed that increased screening results in increased incidence of breast cancer detection, particularly small cancer. unfortunately, the study was retrospective and was not specific regarding how frequently these patients were screened, other than to state that screening was done within 2 years. to claim that increased screening results in over diagnosis is a stretch unless there is a vertical study or prospective study where all patients are followed long term. since this is cross sectional study, it is expected that each year, new cancer diagnoses will occur with fewer large cancers than small. a yearly cross section for 10 years may not be enough to tease out any effect screening has on the diagnosis of larger cancers since the screening rate is not 100% and the population is not stable.
breast cancer mortality rates are due to a number of factors, of which size of tumor at time of diagnosis is only a part of the equation. tumor biology and receptor status plays a very important role and is not detectable via screening.
the study does raise interesting issues about the utility of screening mammography in mortality, but until we can obtain 100% screening rates, we cannot be certain that it is not effective.
CONFLICT OF INTEREST: None Reported
READ MORE
Original Investigation
Less Is More
September 2015

Breast Cancer Screening, Incidence, and Mortality Across US Counties

Author Affiliations
  • 1Currently in private practice, Seattle, Washington
  • 2Department of Physics, Harvard University, Cambridge, Massachusetts
  • 3Department of Medicine, The Dartmouth Institute for Health Policy and Clinical Practice, Hanover, New Hampshire
  • 4Harvard School of Engineering and Applied Sciences, Harvard University, Cambridge, Massachusetts
JAMA Intern Med. 2015;175(9):1483-1489. doi:10.1001/jamainternmed.2015.3043
Abstract

Importance  Screening mammography rates vary considerably by location in the United States, providing a natural opportunity to investigate the associations of screening with breast cancer incidence and mortality, which are subjects of debate.

Objective  To examine the associations between rates of modern screening mammography and the incidence of breast cancer, mortality from breast cancer, and tumor size.

Design, Setting, and Participants  An ecological study of 16 million women 40 years or older who resided in 547 counties reporting to the Surveillance, Epidemiology, and End Results cancer registries during the year 2000. Of these women, 53 207 were diagnosed with breast cancer that year and followed up for the next 10 years. The study covered the period January 1, 2000, to December 31, 2010, and the analysis was performed between April 2013 and March 2015.

Exposures  Extent of screening in each county, assessed as the percentage of included women who received a screening mammogram in the prior 2 years.

Main Outcomes and Measures  Breast cancer incidence in 2000 and incidence-based breast cancer mortality during the 10-year follow-up. Incidence and mortality were calculated for each county and age adjusted to the US population.

Results  Across US counties, there was a positive correlation between the extent of screening and breast cancer incidence (weighted r = 0.54; P < .001) but not with breast cancer mortality (weighted r = 0.00; P = .98). An absolute increase of 10 percentage points in the extent of screening was accompanied by 16% more breast cancer diagnoses (relative rate [RR], 1.16; 95% CI, 1.13-1.19) but no significant change in breast cancer deaths (RR, 1.01; 95% CI, 0.96-1.06). In an analysis stratified by tumor size, we found that more screening was strongly associated with an increased incidence of small breast cancers (≤2 cm) but not with a decreased incidence of larger breast cancers (>2 cm). An increase of 10 percentage points in screening was associated with a 25% increase in the incidence of small breast cancers (RR, 1.25; 95% CI, 1.18-1.32) and a 7% increase in the incidence of larger breast cancers (RR, 1.07; 95% CI, 1.02-1.12).

Conclusions and Relevance  When analyzed at the county level, the clearest result of mammography screening is the diagnosis of additional small cancers. Furthermore, there is no concomitant decline in the detection of larger cancers, which might explain the absence of any significant difference in the overall rate of death from the disease. Together, these findings suggest widespread overdiagnosis.

Introduction

Quiz Ref IDThe goal of screening mammography is to reduce breast cancer mortality by detecting and treating cancer early in the course of the disease. If screening detects tumors early, the diagnosis of smaller and more treatable cancers should increase, while the diagnosis of larger and less treatable cancers should decrease. The associations between screening, incidence, mortality, and tumor size can be investigated at the population level by comparing areas of the United States that have different rates of screening.

One may expect that these associations are already understood. However, there are increasing concerns that screening unintentionally leads to overdiagnosis by identifying small, indolent, or regressive breast tumors that would not otherwise become clinically apparent.1-3 In addition, although screening mammography showed favorable efficacy in most randomized trials,4,5 these trials were conducted decades ago. There are concerns that the benefits and harms may have changed as treatments improved and screening was applied in general practice.6

Local data on rates of mammography screening and breast cancer diagnosis are available for approximately one-fourth of the US population.7,8 We used these data to examine the associations between rates of modern screening mammography and the incidence of breast cancer, mortality from breast cancer, and tumor size.

Methods
Data Source

We analyzed all regions reporting to the Surveillance, Epidemiology, and End Results (SEER) cancer registries from January 1 to December 31, 2000,7,8 with the following exceptions: Louisiana, because Hurricane Katrina (2005) affected data quality; the Alaska Native Tumor Registry because it does not include counties; and Kalawoa County, Hawaii, because of insufficient data (population of <200 individuals). The 547 remaining counties serve as the units of our analysis, which was performed between April 2013 and March 2015. The Dartmouth College Institutional Review Board has deemed studies using de-identified, publicly available data to be exempt from review.

Participants

Because screening mammography is not generally recommended for younger women, our study population was restricted to the 16 120 349 women who were 40 years or older in the SEER counties in 2000. Elderly women were not excluded; doing so would overlook any reductions in diagnoses at older ages owing to cancers that had been detected early by screenings at younger ages.9

In the 547 counties, 55 809 women 40 years or older were diagnosed with breast cancer in 2000 (approximately one-fourth of all cases diagnosed in the United States10), of whom 53 207 (95.3%) had 10 years of follow-up and were included in the primary analysis (Figure 1). During follow-up, fewer deaths were attributed to breast cancer (7729 [42.4%]) than to other causes (10 511 [57.6%]).

Exposure

The study exposure was the percentage of women 40 years or older in each county who had a mammogram in the past 2 years, as of 2000 (median, 62.2%; range, 39.1%-77.8%). For all counties, estimates of screening mammography were published by the National Cancer Institute’s Small Area Estimates (NCI-SAE) for Screening Behaviors program.11,12 The NCI-SAE combined self-reported mammogram histories from the National Health Interview Survey and Behavioral Risk Factors Surveillance System survey, as supplemented by census data and synthesized using hierarchical modeling to adjust for survey nonresponse and noncoverage errors. To provide estimates for the year 2000, we took the mean of the NCI-SAE estimates for the periods 1997-1999 and 2000-2003. During this time, screening in counties increased by a median of 6 percentage points (interquartile range, 3-10 percentage points). Although NCI-SAE values potentially include some misclassified diagnostic mammograms, the practical consequences would be small, as Breast Cancer Surveillance Consortium data on women 40 years or older show that 91% of those with any breast imaging during 1998-2001 received routine screening.13

Main Outcome Measures

The primary analysis evaluated the incidence of breast cancer, including ductal carcinoma in situ, among women 40 years or older in each county. For context, we also evaluated 10-year incidence-based mortality, which is defined as the proportion of women 40 years or older in a given county who were diagnosed with breast cancer in 2000 and died of the disease during the 10-year follow-up period. Forms of incidence-based mortality were first developed at the NCI in the early 1990s.14 Incidence-based mortality is common in studies of screening because, unlike the conventional form of mortality, it excludes cases diagnosed before the study period and, unlike 10-year survival, it is unaffected by overdiagnoses.1,15 Despite these differences, conventional and 10-year incidence-based mortality have similar magnitudes: in the 547 counties, the overall 10-year incidence-based mortality was 47.2 per 100 000 for cases diagnosed in 2000 and conventional mortality was 53.3 per 100 000 for deaths occurring in 2000-2010.

In stratified analyses, we investigated breast cancer incidence by tumor size (≤2 cm vs >2 cm), American Joint Commission on Cancer stage (0-II vs III-IV),16 and surgical treatments17 during first-course therapy (breast-conserving vs non–breast-conserving therapy). The American Joint Commission on Cancer stage was derived from the extent of disease data by SEER.18,19 In all analyses, incidence and incidence-based mortality in each county were age-adjusted to the year 2000 US population by the direct method.

Statistical Analysis

Spline methods20 were used to model smooth, curving associations between screening and cancer rates. We believed it would be inappropriate to assume that associations were linear, especially since nonlinear associations often arise in ecological data.21 In detail, univariate thin-plate regression splines20 (negative binomial model to accommodate overdispersion, log link, and person-years as offset) were specified in the framework of generalized additive models22 and fitted via restricted maximum likelihood,23 as implemented in the mgcv package24 in R. Quantile-quantile plots showed good agreement with some deviation in the tails, and residuals appeared patternless.

To summarize cross-sectional changes in incidence and mortality, we evaluated the mean rate differences and geometric mean relative rates (RRs) associated with a 10–percentage point increase in the extent of screening across the range of data (39%-78% screening). The 95% CIs were calculated by directly simulating from the posterior distribution of the model coefficients (50 000 replicates conditional on smoothing parameters).25

Although the statistical analysis focused on nonparametric regression, we also calculated weighted Pearson correlation coefficients via population-weighted covariance.26 Pearson correlation coefficients have many intricacies and shortcomings21,27,28 but are popular because they are familiar to readers from all medical disciplines. To counterbalance, we include spline regression results with each correlation.

Results
Primary Analysis

Figure 2 shows a clear correlation between the extent of screening mammography and breast cancer incidence across US counties (weighted r = 0.54; P < .001). Quiz Ref IDHowever, there is no evident correlation between the extent of screening and 10-year breast cancer mortality (weighted r = 0.00; P = .98).

We used the regressions shown in Figure 2 to summarize the mean change in incidence or mortality that accompanies an absolute increase of 10 percentage points in screening at the county level (eg, from 60% to 70% or 65% to 75%). A 10–percentage point increase in screening is associated with a 16% mean increase in breast cancer incidence (RR, 1.16; 95% CI, 1.13-1.19, or 35-49 cases per 100 000 as an absolute difference [AD]). However, there is no commensurate change in 10-year breast cancer mortality (RR, 1.01; 95% CI, 0.96-1.06, or –2 to +3 deaths per 100 000 as an AD).

Tumor Size

By stratifying on tumor size, Figure 3 demonstrates that Quiz Ref IDthe association between the extent of screening mammography and breast cancer incidence is largely confined to small cancers (≤2 cm), with no reduction in women presenting with larger cancers (>2 cm). A 10–percentage point increase in screening is accompanied by a 25% mean increase in the incidence of small cancers (RR, 1.25; 95% CI, 1.18-1.32, and 26-40 cases per 100 000 as an AD) and a 7% increase in the incidence of larger cancers (RR, 1.07; 95% CI, 1.02-1.12, and 2-9 cases per 100 000 as an AD). Examining results for individual centimeters of tumor size (eg, ≤0.9, 1-1.9, 2-2.9 cm) shows that our conclusions are not sensitive to the dichotomization into tumors of 2 cm or less vs those more than 2 cm (eAppendix 1 in the Supplement). At the county level, screening mammography is associated with an increased incidence of small cancers and no significant reduction is seen in the incidence of breast cancer at any tumor size.

Disease Stage and Surgery

Since even some small tumors are aggressive, we considered disease stage as a proxy for aggressiveness (Figure 4A). A 10–percentage point increase in screening is associated with increased incidence of early disease (stage 0–II: RR, 1.22; 95% CI, 1.17-1.28) and no change in the incidence of locally advanced and metastatic disease (stage III–IV: RR, 1.02; 95% CI, 0.97-1.07). Given that one of the rationales for screening is to spare women more aggressive surgical procedures, we also examined first-course surgical treatment (Figure 4B). Quiz Ref IDWhile a 10–percentage point increase in screening is associated with more breast-conserving surgical procedures (RR, 1.24; 95% CI, 1.15-1.34), there is no concomitant reduction in non–breast-conserving surgical procedures (eg, total and radical mastectomies).

Discussion

For the individual, screening mammography should ideally detect harmful breast cancers early, without prompting overdiagnosis. Therefore, screening mammography ideally results in increased diagnosis of small cancers, decreased diagnosis of larger cancers (such that the overall risk of diagnosis is unchanged), and reduced mortality from breast cancer. Across US counties, the data show that the extent of screening mammography is indeed associated with an increased incidence of small cancers but not with decreased incidence of larger cancers or significant differences in mortality. In addition, although it has been hoped that screening would allow breast-conserving surgical procedures to replace more extensive mastectomies, we saw no evidence supporting this change.

What explains the observed data? Quiz Ref IDThe simplest explanation is widespread overdiagnosis, which increases the incidence of small cancers without changing mortality, and therefore matches every feature of the observed data. Indeed, our cross-sectional findings are supported by the prior longitudinal analyses of Esserman et al29 and others,30,31 in which excess incidences of early-stage breast cancer were attributed to overdiagnosis. However, 4 alternatives are also logically possible: lead time, reverse causality, confounding, and ecological bias.

In the absence of overdiagnosis, periods of increasing screening could result in rising incidence as cancer diagnoses advance in time. For example, it is well known that first screenings are especially likely to catch many presymptomatic cancers, temporarily increasing incidence. However, once screening is in a steady state, these lead-time effects on incidence diminish and then disappear. To consider the contribution of lead time, we compared counties where screening had increased from 1997-1999 to 2000-2003 with counties where screening rates had been stable (eAppendix 2 in the Supplement) while controlling for the current extent of screening. Incidence was not elevated in counties where the extent of screening had recently increased, suggesting that lead-time effects do not explain our results.

Counties could also have very different incidence rates of true breast cancers, which might be associated with the extent of screening because high-risk regions are especially targeted for screening (reverse causation) or because risk factors for true breast cancer happen to be associated with screening (confounding)—for example, through income’s associations with both age at first birth and participation in screening. Either way, the existence of more true breast cancers would result in deaths—yet we observe no increase in breast cancer mortality at the county level. Even where there are 1.8 times as many cancers being diagnosed, mortality is the same (Figure 2, counties at right). To maintain that these additional diagnoses are nevertheless true breast cancers, we would have to suppose that counties with more screening have better outcomes and, oddly, that the counterbalance is so precise that no (r= 0.00; RR = 1.01) association between screening and mortality remains. We see no reason for such a balance except coincidence. Therefore, although the balance is not impossible, we consider it improbable. It seems additionally improbable to see more than twice as many diagnoses of true breast cancers at earlier stages without concomitant changes to the incidence of locally advanced and metastatic disease (Figure 4A).

Finally, our conclusions are inherently vulnerable to ecological biases.32,33 However, when we evaluated the association between screening and incidence within individual areas, we found similar results in all 9 states and 2 metropolitan areas (eAppendix 3 in the Supplement). Furthermore, the association between screening and breast cancer does not vary substantially by county population (eAppendix 4 in the Supplement). In summary, although ecological bias cannot be excluded without individual data, neither of the analyses suggested threats to validity.

To investigate the limits of the 10-year follow-up, we also analyzed cumulative mortality rates. However, there was no evidence to suggest that results observed during the first 10 years after diagnosis would change substantially thereafter (eAppendix 5 and eAppendix 6 in the Supplement). Other study limitations include loss to follow-up; missing tumor sizes; the absence of analyses of adjuvant therapy, risk factors, ultrasound, and spatial autocorrelation; the use of NCI-SAE estimates; and the quality of cause of death adjudication in registries.

Clinicians are correct to be wary of ecological studies because of the ecological fallacy. Ultimately, however, decisions must be made based on the evidence that is available, not unachievable ideals. Ecological studies are especially suitable for investigating overdiagnosis because overdiagnosis is currently not observable in individuals, only in populations. Indeed, the recent methodologic review by Carter et al2 concluded that the best designs for investigating overdiagnosis are high-quality ecological and cohort studies with multiple settings. Here, we examined 547 counties with diverse screening rates. It appears that no previous study of breast cancer overdiagnosis compared 12 or more counties, nations, or other regions.2,34-36 Other researchers consider mathematical modeling and simulation studies more reliable,9 but of course clinicians are also right to be wary of the untested assumptions required to model the fundamental unknown—the natural history of screen-detected breast cancers.2

Conclusions

Our analysis shows that, when directed toward the general US population, the most prominent effect of screening mammography is overdiagnosis. Nonetheless, we do not believe that the right rate of screening mammography is zero. As is the case with screening in general, the balance of benefits and harms is likely to be most favorable when screening is directed to those at high risk, provided neither too frequently nor too rarely, and sometimes followed by watchful waiting instead of immediate active treatment.37

Beyond these conclusions, the county data show 2 other troubling features, which we have avoided discussing heretofore because they are more tentative. First, screening is not associated with reduced presentation of larger breast cancers. However, it is not clear whether this county-level result means that screening is failing to catch true breast cancers before they become large, or if reductions in the presentation of large true breast cancers are concealed by increased presentation of large overdiagnoses. Second, and perhaps relatedly, screening was not associated with reduced mortality from breast cancer during the 10-year follow-up. However, observed mortality from breast cancer may be too rare and too noisy to reliably detect the 20% reduction at 13 years of follow-up that was estimated in a comprehensive meta-analysis of screening mammography trials.1 In summary, both of these features are promising topics for future research. This is also the right time to begin investigating whether all women undergoing screening mammography have the same risk of overdiagnosis, or if overdiagnosis is especially likely in some groups.

Back to top
Article Information

Accepted for Publication: April 4, 2015.

Corresponding Author: Richard Wilson, DPhil, Department of Physics, Harvard University, Jefferson Physical Laboratory, Room 257, Cambridge, MA 02138 (wilson@physics.harvard.edu).

Published Online: July 6, 2015. doi:10.1001/jamainternmed.2015.3043.

Author Contributions: Mr Harding had full access to all the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis.

Study concept and design: Harding, Pompei, Burmistrov, Wilson.

Acquisition, analysis, or interpretation of data: Harding, Pompei, Welch, Abebe, Wilson.

Drafting of the manuscript: Harding, Welch.

Critical revision of the manuscript for important intellectual content: Harding, Pompei, Burmistrov, Abebe, Wilson.

Statistical analysis: Harding, Welch.

Obtained funding: Pompei, Wilson.

Administrative, technical, or material support: Harding, Pompei, Wilson.

Study supervision: Pompei, Wilson.

Conflict of Interest Disclosures: Dr Pompei is chief executive officer and founder of Exergen Corp. Mr Harding reported receiving funding from Exergen Corp for work on this study. No other disclosures were reported.

Additional Contibutions: Steven H. Lamm, MD, DTPH, Consultants in Epidemiology and Occupational Health, LLC, provided helpful comments on the manuscript. He received no compensation for this assistance.

References
1.
Marmot  MG, Altman  DG, Cameron  DA, Dewar  JA, Thompson  SG, Wilcox  M.  The benefits and harms of breast cancer screening: an independent review.  Br J Cancer. 2013;108(11):2205-2240.PubMedGoogle ScholarCrossref
2.
Carter  JL, Coletti  RJ, Harris  RP.  Quantifying and monitoring overdiagnosis in cancer screening: a systematic review of methods.  BMJ. 2015;350:g7773.PubMedGoogle ScholarCrossref
3.
Barratt  A.  Overdiagnosis in mammography screening: a 45 year journey from shadowy idea to acknowledged reality.  BMJ. 2015;350:h867.PubMedGoogle ScholarCrossref
4.
Nelson  HD, Tyne  K, Naik  A, Bougatsos  C, Chan  BK, Humphrey  L; US Preventive Services Task Force.  Screening for breast cancer: an update for the US Preventive Services Task Force.  Ann Intern Med. 2009;151(10):727-737, W237-42.PubMedGoogle ScholarCrossref
5.
Humphrey  LL, Helfand  M, Chan  BKS, Woolf  SH.  Breast cancer screening: a summary of the evidence for the US Preventive Services Task Force.  Ann Intern Med. 2002;137(5, part 1):347-360.PubMedGoogle ScholarCrossref
6.
Biller-Andorno  N, Jüni  P.  Abolishing mammography screening programs? a view from the Swiss Medical Board.  N Engl J Med. 2014;370(21):1965-1967.PubMedGoogle ScholarCrossref
7.
Surveillance, Epidemiology, and End Results Program.  SEER*Stat databases: SEER 18 registries, November 2012 submission.http://seer.cancer.gov/data/seerstat/nov2012/. Accessed October 10, 2014.
8.
Surveillance, Epidemiology, and End Results Program.  SEER*Stat software, version 8.1.2.http://seer.cancer.gov/seerstat. October 10, 2014.
9.
Etzioni  R, Gulati  R, Mallinger  L, Mandelblatt  J.  Influence of study features and methods on overdiagnosis estimates in breast and prostate cancer screening.  Ann Intern Med. 2013;158(11):831-838.PubMedGoogle ScholarCrossref
10.
Greenlee  RT, Murray  T, Bolden  S, Wingo  PA.  Cancer statistics, 2000.  CA Cancer J Clin. 2000;50(1):7-33.PubMedGoogle ScholarCrossref
11.
Raghunathan  TE, Xie  D, Schenker  N,  et al.  Combining information from two surveys to estimate county-level prevalence rates of cancer risk factors and screening.  J Am Stat Assoc. 2007;102(478):474-486. doi:10.1198/016214506000001293.Google ScholarCrossref
12.
National Cancer Institute, Division of Cancer Control and Population Sciences, Statistical Methodology & Applications Branch.  Small area estimates for cancer risk factors & screening behaviors.http://sae.cancer.gov. Released May 2010. Accessed October 10, 2014.
13.
Breast Cancer Surveillance Consortium, National Cancer Institute.  Distribution of key variables.http://breastscreening.cancer.gov/data/variables/. Accessed February 23, 2015.
14.
Chu  KC, Miller  BA, Feuer  EJ, Hankey  BF.  A method for partitioning cancer mortality trends by factors associated with diagnosis: an application to female breast cancer.  J Clin Epidemiol. 1994;47(12):1451-1461.PubMedGoogle ScholarCrossref
15.
Kalager  M, Zelen  M, Langmark  F, Adami  H-O.  Effect of screening mammography on breast-cancer mortality in Norway.  N Engl J Med. 2010;363(13):1203-1210.PubMedGoogle ScholarCrossref
16.
Singletary  SE, Allred  C, Ashley  P,  et al.  Staging system for breast cancer: revisions for the 6th edition of the AJCC Cancer Staging Manual.  Surg Clin North Am. 2003;83(4):803-819.Google ScholarCrossref
17.
Carlson  RW, Anderson  BO, Bensinger  W,  et al; National Comprehensive Cancer Network.  NCCN practice guidelines for breast cancer.  Oncology (Williston Park). 2000;14(11A):33-49.PubMedGoogle Scholar
18.
Fritz  A, Ries  L, eds.  SEER Extent of Disease 1998, Codes and Coding Instructions (No. T-899; Pub. No. 98-4354). Bethesda, MD: National Cancer Institute; 1998.
19.
Seiffert  JE, ed.  SEER Program: Comparative Staging Guide for Cancer, Version 1.1. Bethesda, MD: National Cancer Institute; 1993.
20.
Wood  SN.  Thin plate regression splines.  J R Stat Soc Series B Stat Methodol. 2003;65(1):95-114. doi:10.1111/1467-9868.00374.Google ScholarCrossref
21.
Greenland  S, Robins  J.  Invited commentary: ecologic studies—biases, misconceptions, and counterexamples.  Am J Epidemiol. 1994;139(8):747-760.PubMedGoogle Scholar
22.
Wood  S.  Generalized Additive Models: An Introduction With R. Boca Raton, FL: CRC Press; 2006.
23.
Wood  SN.  Fast stable restricted maximum likelihood and marginal likelihood estimation of semiparametric generalized linear models.  J R Stat Soc Series B Stat Methodol. 2011;73(1):3-36. doi:10.1111/j.1467-9868.2010.00749.x.Google ScholarCrossref
24.
Wood  S.  mgcv: Mixed GAM computation vehicle with GCV/AIC/REML smoothness estimation. R package, version 1.7. http://cran.r-project.org/web/packages/mgcv/index.html. Accessed September 1, 2014.
25.
Marra  G, Wood  SN.  Coverage properties of confidence intervals for generalized additive model components.  Scand J Stat. 2012;39(1):53-74. doi:10.1111/j.1467-9469.2011.00760.x.Google ScholarCrossref
26.
Pasek  J, Tahk  A, Culter  G, Schwemmle  M.  Weights: Weighting and Weighted Statistics. http://cran.r-project.org/web/packages/weights/index.html. Accessed February 20, 2015.
27.
Rodgers  JL, Nicewander  WA.  Thirteen ways to look at the correlation coefficient.  Am Stat. 1988;42(1):59-66. doi:10.1080/00031305.1988.10475524.Google ScholarCrossref
28.
Taylor  R.  Interpretation of the correlation coefficient: a basic review.  J Diagn Med Sonogr. 1990;6(1):35-39. doi:10.1177/875647939000600106.Google ScholarCrossref
29.
Esserman  L, Shieh  Y, Thompson  I.  Rethinking screening for breast cancer and prostate cancer.  JAMA. 2009;302(15):1685-1692.PubMedGoogle ScholarCrossref
30.
Bleyer  A, Welch  HG.  Effect of three decades of screening mammography on breast-cancer incidence.  N Engl J Med. 2012;367(21):1998-2005.PubMedGoogle ScholarCrossref
31.
Anderson  WF, Jatoi  I, Devesa  SS.  Assessing the impact of screening mammography: breast cancer incidence and mortality rates in Connecticut (1943-2002).  Breast Cancer Res Treat. 2006;99(3):333-340.PubMedGoogle ScholarCrossref
32.
Greenland  S, Morgenstern  H.  Ecological bias, confounding, and effect modification.  Int J Epidemiol. 1989;18(1):269-274.PubMedGoogle ScholarCrossref
33.
Wakefield  J.  Ecologic studies revisited.  Annu Rev Public Health. 2008;29:75-90.PubMedGoogle ScholarCrossref
34.
Puliti  D, Duffy  SW, Miccinesi  G,  et al; EUROSCREEN Working Group.  Overdiagnosis in mammographic screening for breast cancer in Europe: a literature review.  J Med Screen. 2012;19(suppl 1):42-56.PubMedGoogle ScholarCrossref
35.
Njor  S, Nyström  L, Moss  S,  et al; Euroscreen Working Group.  Breast cancer mortality in mammographic screening in Europe: a review of incidence-based mortality studies.  J Med Screen. 2012;19(suppl 1):33-41.PubMedGoogle ScholarCrossref
36.
Jonsson  H, Johansson  R, Lenner  P.  Increased incidence of invasive breast cancer after the introduction of service screening with mammography in Sweden.  Int J Cancer. 2005;117(5):842-847.PubMedGoogle ScholarCrossref
37.
Esserman  LJ, Thompson  IM, Reid  B,  et al.  Addressing overdiagnosis and overtreatment in cancer: a prescription for change.  Lancet Oncol. 2014;15(6):e234-e242.PubMedGoogle ScholarCrossref
×