[Skip to Navigation]
Sign In
Figure.  Precision Recall and Area Under the Receiver Operating Characteristic Curve for the Optimal Model
Precision Recall and Area Under the Receiver Operating Characteristic Curve for the Optimal Model

A, Because of the low incidence of atrial fibrillation systemwide, most decision thresholds did not have a high recall (positive predictive value). See text for details. B, Area under the receiver operating characteristic curve was 0.80.

Table 1.  UCHealth Population by AF Diagnosis
UCHealth Population by AF Diagnosis
Table 2.  Comparison of Resampling Strategiesa
Comparison of Resampling Strategiesa
Table 3.  Comparison of Machine Learning Approachesa
Comparison of Machine Learning Approachesa
1.
Majeed  A, Moser  K, Carroll  K.  Trends in the prevalence and management of atrial fibrillation in general practice in England and Wales, 1994-1998: analysis of data from the general practice research database.  Heart. 2001;86(3):284-288. doi:10.1136/heart.86.3.284PubMedGoogle ScholarCrossref
2.
Benjamin  EJ, Blaha  MJ, Chiuve  SE,  et al; American Heart Association Statistics Committee and Stroke Statistics Subcommittee.  Heart disease and stroke statistics, 2017 update: a report from the American Heart Association.  Circulation. 2017;135(10):e146-e603. doi:10.1161/CIR.0000000000000485PubMedGoogle ScholarCrossref
3.
Benjamin  EJ, Wolf  PA, D’Agostino  RB, Silbershatz  H, Kannel  WB, Levy  D.  Impact of atrial fibrillation on the risk of death: the Framingham Heart Study.  Circulation. 1998;98(10):946-952. doi:10.1161/01.CIR.98.10.946PubMedGoogle ScholarCrossref
4.
Benjamin  EJ, Levy  D, Vaziri  SM, D’Agostino  RB, Belanger  AJ, Wolf  PA.  Independent risk factors for atrial fibrillation in a population-based cohort: the Framingham Heart Study.  JAMA. 1994;271(11):840-844. doi:10.1001/jama.1994.03510350050036PubMedGoogle ScholarCrossref
5.
Singer  DE, Hughes  RA, Gress  DR,  et al; Boston Area Anticoagulation Trial for Atrial Fibrillation Investigators.  The effect of low-dose warfarin on the risk of stroke in patients with nonrheumatic atrial fibrillation.  N Engl J Med. 1990;323(22):1505-1511. doi:10.1056/NEJM199011293232201PubMedGoogle ScholarCrossref
6.
 Stroke Prevention in Atrial Fibrillation Study: final results.  Circulation. 1991;84(2):527-539. doi:10.1161/01.CIR.84.2.527PubMedGoogle ScholarCrossref
7.
 Warfarin versus aspirin for prevention of thromboembolism in atrial fibrillation: Stroke Prevention in Atrial Fibrillation II Study.  Lancet. 1994;343(8899):687-691.PubMedGoogle Scholar
8.
Petersen  P, Boysen  G, Godtfredsen  J, Andersen  ED, Andersen  B.  Placebo-controlled, randomised trial of warfarin and aspirin for prevention of thromboembolic complications in chronic atrial fibrillation: the Copenhagen AFASAK study.  Lancet. 1989;1(8631):175-179. doi:10.1016/S0140-6736(89)91200-2PubMedGoogle ScholarCrossref
9.
Ezekowitz  MD, Bridgers  SL, James  KE,  et al; Veterans Affairs Stroke Prevention in Nonrheumatic Atrial Fibrillation Investigators.  Warfarin in the prevention of stroke associated with nonrheumatic atrial fibrillation.  N Engl J Med. 1992;327(20):1406-1412. doi:10.1056/NEJM199211123272002PubMedGoogle ScholarCrossref
10.
Connolly  SJ, Laupacis  A, Gent  M, Roberts  RS, Cairns  JA, Joyner  C.  Canadian Atrial Fibrillation Anticoagulation (CAFA) Study.  J Am Coll Cardiol. 1991;18(2):349-355. doi:10.1016/0735-1097(91)90585-WPubMedGoogle ScholarCrossref
11.
 Risk factors for stroke and efficacy of antithrombotic therapy in atrial fibrillation: analysis of pooled data from five randomized controlled trials.  Arch Intern Med. 1994;154(13):1449-1457. doi:10.1001/archinte.1994.00420130036007PubMedGoogle ScholarCrossref
12.
Hart  RG, Pearce  LA, Aguilar  MI.  Meta-analysis: antithrombotic therapy to prevent stroke in patients who have nonvalvular atrial fibrillation.  Ann Intern Med. 2007;146(12):857-867. doi:10.7326/0003-4819-146-12-200706190-00007PubMedGoogle ScholarCrossref
13.
van Walraven  C, Hart  RG, Singer  DE,  et al.  Oral anticoagulants vs aspirin in nonvalvular atrial fibrillation: an individual patient meta-analysis.  JAMA. 2002;288(19):2441-2448. doi:10.1001/jama.288.19.2441PubMedGoogle ScholarCrossref
14.
Cooper  NJ, Sutton  AJ, Lu  G, Khunti  K.  Mixed comparison of stroke prevention treatments in individuals with nonrheumatic atrial fibrillation.  Arch Intern Med. 2006;166(12):1269-1275. doi:10.1001/archinte.166.12.1269PubMedGoogle ScholarCrossref
15.
Connolly  SJ, Ezekowitz  MD, Yusuf  S,  et al; RE-LY Steering Committee and Investigators.  Dabigatran versus warfarin in patients with atrial fibrillation.  N Engl J Med. 2009;361(12):1139-1151. doi:10.1056/NEJMoa0905561PubMedGoogle ScholarCrossref
16.
Patel  MR, Mahaffey  KW, Garg  J,  et al; ROCKET AF Investigators.  Rivaroxaban versus warfarin in nonvalvular atrial fibrillation.  N Engl J Med. 2011;365(10):883-891. doi:10.1056/NEJMoa1009638PubMedGoogle ScholarCrossref
17.
Connolly  SJ, Eikelboom  J, Joyner  C,  et al; AVERROES Steering Committee and Investigators.  Apixaban in patients with atrial fibrillation.  N Engl J Med. 2011;364(9):806-817. doi:10.1056/NEJMoa1007432PubMedGoogle ScholarCrossref
18.
Jaakkola  J, Mustonen  P, Kiviniemi  T,  et al.  Stroke as the first manifestation of atrial fibrillation.  PLoS One. 2016;11(12):e0168010. doi:10.1371/journal.pone.0168010PubMedGoogle Scholar
19.
Curry  SJ, Krist  AH, Owens  DK,  et al; US Preventive Services Task Force.  Screening for atrial fibrillation with electrocardiography: US Preventive Services Task Force recommendation statement.  JAMA. 2018;320(5):478-484. doi:10.1001/jama.2018.10321PubMedGoogle ScholarCrossref
20.
Jonas  DE, Kahwati  LC, Yun  JDY, Middleton  JC, Coker-Schwimmer  M, Asher  GN.  Screening for atrial fibrillation with electrocardiography: evidence report and systematic review for the US Preventive Services Task Force.  JAMA. 2018;320(5):485-498. doi:10.1001/jama.2018.4190PubMedGoogle ScholarCrossref
21.
Freedman  B, Camm  J, Calkins  H,  et al; AF-Screen Collaborators.  Screening for atrial fibrillation: a report of the AF-SCREEN international collaboration.  Circulation. 2017;135(19):1851-1867. doi:10.1161/CIRCULATIONAHA.116.026693PubMedGoogle ScholarCrossref
22.
Kocher  R, Emanuel  EJ, DeParle  NA.  The Affordable Care Act and the future of clinical medicine: the opportunities and challenges.  Ann Intern Med. 2010;153(8):536-539. doi:10.7326/0003-4819-153-8-201010190-00274PubMedGoogle ScholarCrossref
23.
Dexheimer  JW, Talbot  TR, Sanders  DL, Rosenbloom  ST, Aronsky  D.  Prompting clinicians about preventive care measures: a systematic review of randomized controlled trials.  J Am Med Inform Assoc. 2008;15(3):311-320. doi:10.1197/jamia.M2555PubMedGoogle ScholarCrossref
24.
Sadoughi  F, Erfannia  L.  Health information system in a cloud computing context.  Stud Health Technol Inform. 2017;236:290-297.PubMedGoogle Scholar
25.
Jensen  RE, Rothrock  NE, DeWitt  EM,  et al.  The role of technical advances in the adoption and integration of patient-reported outcomes in clinical care.  Med Care. 2015;53(2):153-159. doi:10.1097/MLR.0000000000000289PubMedGoogle ScholarCrossref
26.
Casey  JA, Schwartz  BS, Stewart  WF, Adler  NE.  Using electronic health records for population health research: a review of methods and applications.  Annu Rev Public Health. 2016;37:61-81. doi:10.1146/annurev-publhealth-032315-021353PubMedGoogle ScholarCrossref
27.
Huang  Y, Lee  J, Wang  S, Sun  J, Liu  H, Jiang  X.  Privacy-preserving predictive modeling: harmonization of contextual embeddings from different sources.  JMIR Med Inform. 2018;6(2):e33. doi:10.2196/medinform.9455PubMedGoogle Scholar
28.
Jiang  G, Kiefer  RC, Sharma  DK, Prud’hommeaux  E, Solbrig  HR.  A consensus-based approach for harmonizing the OHDSI Common Data Model with HL7 FHIR.  Stud Health Technol Inform. 2017;245:887-891.PubMedGoogle Scholar
29.
Jiang  G, Evans  J, Oniki  TA,  et al.  Harmonization of detailed clinical models with clinical study data standards.  Methods Inf Med. 2015;54(1):65-74. doi:10.3414/ME13-02-0019PubMedGoogle ScholarCrossref
30.
Ocaña  K, de Oliveira  D.  Parallel computing in genomic research: advances and applications.  Adv Appl Bioinform Chem. 2015;8:23-35.PubMedGoogle Scholar
31.
Korb  O, Finn  PW, Jones  G.  The cloud and other new computational methods to improve molecular modelling.  Expert Opin Drug Discov. 2014;9(10):1121-1131. doi:10.1517/17460441.2014.941800PubMedGoogle ScholarCrossref
32.
Rosenberg  MA, Samuel  M, Thosani  A, Zimetbaum  PJ.  Use of a noninvasive continuous monitoring device in the management of atrial fibrillation: a pilot study.  Pacing Clin Electrophysiol. 2013;36(3):328-333. doi:10.1111/pace.12053PubMedGoogle ScholarCrossref
33.
Diederichsen  SZ, Haugan  KJ, Køber  L,  et al.  Atrial fibrillation detected by continuous electrocardiographic monitoring using implantable loop recorder to prevent stroke in individuals at risk (the LOOP study): rationale and design of a large randomized controlled trial.  Am Heart J. 2017;187:122-132. doi:10.1016/j.ahj.2017.02.017PubMedGoogle ScholarCrossref
34.
LeCun  Y, Bengio  Y, Hinton  G.  Deep learning.  Nature. 2015;521(7553):436-444. doi:10.1038/nature14539PubMedGoogle ScholarCrossref
35.
Stallkamp  J, Schlipsing  M, Salmen  J, Igel  C.  Man vs. computer: benchmarking machine learning algorithms for traffic sign recognition.  Neural Netw. 2012;32:323-332.PubMedGoogle ScholarCrossref
36.
Kooi  T, Litjens  G, van Ginneken  B,  et al.  Large scale deep learning for computer aided detection of mammographic lesions.  Med Image Anal. 2017;35:303-312. doi:10.1016/j.media.2016.07.007PubMedGoogle ScholarCrossref
37.
Minarro-Giménez  JA, Marín-Alonso  O, Samwald  M.  Exploring the application of deep learning techniques on medical text corpora.  Stud Health Technol Inform. 2014;205:584-588.PubMedGoogle Scholar
38.
Zhu  Q, Li  X, Conesa  A, Pereira  C.  GRAM-CNN: a deep learning approach with local context for named entity recognition in biomedical text.  Bioinformatics. 2018;34(9):1547-1554. doi:10.1093/bioinformatics/btx815PubMedGoogle ScholarCrossref
39.
Moravčík  M, Schmid  M, Burch  N,  et al.  DeepStack: expert-level artificial intelligence in heads-up no-limit poker.  Science. 2017;356(6337):508-513. doi:10.1126/science.aam6960PubMedGoogle ScholarCrossref
40.
Schaeffer  J, Burch  N, Björnsson  Y,  et al.  Checkers is solved.  Science. 2007;317(5844):1518-1522. doi:10.1126/science.1144079PubMedGoogle ScholarCrossref
41.
Silver  D, Huang  A, Maddison  CJ,  et al.  Mastering the game of Go with deep neural networks and tree search.  Nature. 2016;529(7587):484-489. doi:10.1038/nature16961PubMedGoogle ScholarCrossref
42.
Ramírez  J, Monasterio  V, Mincholé  A,  et al.  Automatic SVM classification of sudden cardiac death and pump failure death from autonomic and repolarization ECG markers.  J Electrocardiol. 2015;48(4):551-557. doi:10.1016/j.jelectrocard.2015.04.002PubMedGoogle ScholarCrossref
43.
Beaulieu-Jones  BK, Greene  CS; Pooled Resource Open-Access ALS Clinical Trials Consortium.  Semi-supervised learning of the electronic health record for phenotype stratification.  J Biomed Inform. 2016;64:168-178. doi:10.1016/j.jbi.2016.10.007PubMedGoogle ScholarCrossref
44.
Luo  Y, Li  Z, Guo  H,  et al.  Predicting congenital heart defects: a comparison of three data mining methods.  PLoS One. 2017;12(5):e0177811. doi:10.1371/journal.pone.0177811PubMedGoogle Scholar
45.
Amini  P, Ahmadinia  H, Poorolajal  J, Moqaddasi Amiri  M.  Evaluating the high risk groups for suicide: a comparison of logistic regression, support vector machine, decision tree and artificial neural network.  Iran J Public Health. 2016;45(9):1179-1187.PubMedGoogle Scholar
46.
Churpek  MM, Yuen  TC, Winslow  C, Meltzer  DO, Kattan  MW, Edelson  DP.  Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards.  Crit Care Med. 2016;44(2):368-374. doi:10.1097/CCM.0000000000001571PubMedGoogle ScholarCrossref
47.
Johnson  KW, Torres Soto  J, Glicksberg  BS,  et al.  Artificial intelligence in cardiology.  J Am Coll Cardiol. 2018;71(23):2668-2679. doi:10.1016/j.jacc.2018.03.521PubMedGoogle ScholarCrossref
48.
Collins  GS, Reitsma  JB, Altman  DG, Moons  KG.  Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement.  BMJ. 2015;350:g7594. doi:10.1136/bmj.g7594PubMedGoogle ScholarCrossref
49.
Makadia  R, Ryan  PB.  Transforming the Premier Perspective hospital database into the Observational Medical Outcomes Partnership (OMOP) Common Data Model.  EGEMS (Wash DC). 2014;2(1):1110.PubMedGoogle Scholar
50.
Chawla  NV, Bowyer  KW, Hall  LO, Kegelmeyer  WP.  SMOTE: synthetic minority over-sampling technique.  J Artif Intell Res. 2002;16:321-357. doi:10.1613/jair.953Google ScholarCrossref
51.
Chai  KE, Anthony  S, Coiera  E, Magrabi  F.  Using statistical text classification to identify health information technology incidents.  J Am Med Inform Assoc. 2013;20(5):980-985. doi:10.1136/amiajnl-2012-001409PubMedGoogle ScholarCrossref
52.
Davis  J, Goadrich  M.  The relationship between precision-recall and ROC curves.  In:  ICML ’06 Proceedings of the 23rd International Conference on Machine Learning. New York, NY: Association for Computing Machinery; 2006:233-240. doi:10.1145/1143844.1143874Google Scholar
53.
Hripcsak  G, Rothschild  AS.  Agreement, the f-measure, and reliability in information retrieval.  J Am Med Inform Assoc. 2005;12(3):296-298. doi:10.1197/jamia.M1733PubMedGoogle ScholarCrossref
54.
Wang  Q, Garrity  GM, Tiedje  JM, Cole  JR.  Naive Bayesian classifier for rapid assignment of rRNA sequences into the new bacterial taxonomy.  Appl Environ Microbiol. 2007;73(16):5261-5267. doi:10.1128/AEM.00062-07PubMedGoogle ScholarCrossref
55.
Mansoor  H, Elgendy  IY, Segal  R, Bavry  AA, Bian  J.  Risk prediction model for in-hospital mortality in women with ST-elevation myocardial infarction: A machine learning approach.  Heart Lung. 2017;46(6):405-411.PubMedGoogle ScholarCrossref
56.
Qu  Y, Adam  BL, Yasui  Y,  et al.  Boosted decision tree analysis of surface-enhanced laser desorption/ionization mass spectral serum profiles discriminates prostate cancer from noncancer patients.  Clin Chem. 2002;48(10):1835-1843.PubMedGoogle Scholar
57.
Penny  W, Frost  D.  Neural networks in clinical medicine.  Med Decis Making. 1996;16(4):386-398. doi:10.1177/0272989X9601600409PubMedGoogle ScholarCrossref
58.
Agresti  A, Coull  BA.  Approximate is better than “exact” for interval estimation of binomial proportions.  Am Stat. 1998;52(2):119-126. http://users.stat.ufl.edu/~aa/articles/agresti_coull_1998.pdf. Accessed December 2, 2019.Google Scholar
59.
Lakhani  P, Langlotz  CP.  Automated detection of radiology reports that document non-routine communication of critical or significant results.  J Digit Imaging. 2010;23(6):647-657. doi:10.1007/s10278-009-9237-1PubMedGoogle ScholarCrossref
60.
Nattel  S, Shiroshita-Takeshita  A, Brundel  BJ, Rivard  L.  Mechanisms of atrial fibrillation: lessons from animal models.  Prog Cardiovasc Dis. 2005;48(1):9-28. doi:10.1016/j.pcad.2005.06.002PubMedGoogle ScholarCrossref
61.
Nattel  S, Harada  M.  Atrial remodeling and atrial fibrillation: recent advances and translational perspectives.  J Am Coll Cardiol. 2014;63(22):2335-2345. doi:10.1016/j.jacc.2014.02.555PubMedGoogle ScholarCrossref
62.
Rosenberg  MA, Manning  WJ.  Diastolic dysfunction and risk of atrial fibrillation: a mechanistic appraisal.  Circulation. 2012;126(19):2353-2362. doi:10.1161/CIRCULATIONAHA.112.113233PubMedGoogle ScholarCrossref
63.
Alonso  A, Krijthe  BP, Aspelund  T,  et al.  Simple risk model predicts incidence of atrial fibrillation in a racially and geographically diverse population: the CHARGE-AF consortium.  J Am Heart Assoc. 2013;2(2):e000102. doi:10.1161/JAHA.112.000102PubMedGoogle Scholar
64.
Li  DC, Hu  SC, Lin  LS, Yeh  CW.  Detecting representative data and generating synthetic samples to improve learning accuracy with imbalanced data sets.  PLoS One. 2017;12(8):e0181853. doi:10.1371/journal.pone.0181853PubMedGoogle Scholar
65.
Colborn  KL, Bronsert  M, Amioka  E, Hammermeister  K, Henderson  WG, Meguid  R.  Identification of surgical site infections using electronic health record data.  Am J Infect Control. 2018;46(11):1230-1235. doi:10.1016/j.ajic.2018.05.011PubMedGoogle ScholarCrossref
66.
Blagus  R, Lusa  L.  SMOTE for high-dimensional class-imbalanced data.  BMC Bioinformatics. 2013;14:106. doi:10.1186/1471-2105-14-106PubMedGoogle ScholarCrossref
67.
Lee  JG, Jun  S, Cho  YW,  et al.  Deep learning in medical imaging: general overview.  Korean J Radiol. 2017;18(4):570-584. doi:10.3348/kjr.2017.18.4.570PubMedGoogle ScholarCrossref
68.
Agarwalla  S, Sarma  KK.  Machine learning based sample extraction for automatic speech recognition using dialectal Assamese speech.  Neural Netw. 2016;78:97-111.PubMedGoogle ScholarCrossref
69.
Chen  Y, Ghosh  J, Bejan  CA,  et al.  Building bridges across electronic health record systems through inferred phenotypic topics.  J Biomed Inform. 2015;55:82-93. doi:10.1016/j.jbi.2015.03.011PubMedGoogle ScholarCrossref
70.
Pham  T, Tran  T, Phung  D, Venkatesh  S.  Predicting healthcare trajectories from medical records: a deep learning approach.  J Biomed Inform. 2017;69:218-229. doi:10.1016/j.jbi.2017.04.001PubMedGoogle ScholarCrossref
71.
Yang  H, Pan  Z, Tao  Q.  Robust and adaptive online time series prediction with long short-term memory.  Comput Intell Neurosci. 2017;2017:9478952. doi:10.1155/2017/9478952PubMedGoogle Scholar
72.
Zhang  Z.  Parametric regression model for survival data: Weibull regression model as an example.  Ann Transl Med. 2016;4(24):484. doi:10.21037/atm.2016.08.45PubMedGoogle ScholarCrossref
73.
Rosenberg  MA, Maziarz  M, Tan  AY,  et al.  Circulating fibrosis biomarkers and risk of atrial fibrillation: the Cardiovascular Health Study (CHS).  Am Heart J. 2014;167(5):723-8.e2.PubMedGoogle ScholarCrossref
74.
Patton  KK, Ellinor  PT, Heckbert  SR,  et al.  N-terminal pro-B-type natriuretic peptide is a major predictor of the development of atrial fibrillation: the Cardiovascular Health Study.  Circulation. 2009;120(18):1768-1774. doi:10.1161/CIRCULATIONAHA.109.873265PubMedGoogle ScholarCrossref
75.
Rosenberg  MA, Gottdiener  JS, Heckbert  SR, Mukamal  KJ.  Echocardiographic diastolic parameters and risk of atrial fibrillation: the Cardiovascular Health Study.  Eur Heart J. 2012;33(7):904-912. doi:10.1093/eurheartj/ehr378PubMedGoogle ScholarCrossref
Original Investigation
Health Informatics
January 17, 2020

Assessment of a Machine Learning Model Applied to Harmonized Electronic Health Record Data for the Prediction of Incident Atrial Fibrillation

Author Affiliations
  • 1Colorado Center for Personalized Medicine, University of Colorado School of Medicine, Aurora
  • 2Colorado School of Public Health, Department of Biostatics and Informatics, University of Colorado Denver, Aurora
  • 3Children’s Hospital Colorado, Cancer Center Biostatistics Core, Department of Pediatrics, University of Colorado, Aurora
  • 4Division of Cardiology and Cardiac Electrophysiology, University of Colorado School of Medicine, Aurora
JAMA Netw Open. 2020;3(1):e1919396. doi:10.1001/jamanetworkopen.2019.19396
Key Points

Question  Can machine learning approaches applied to harmonized electronic health record data identify patients at risk of 6-month incident atrial fibrillation with greater accuracy than standard risk factors?

Findings  This diagnostic study used electronic health record data from more than 2 million individuals to classify patients diagnosed with incident atrial fibrillation within a 6-month period, comparing several approaches to data management. A strategy that included the use of the 200 most common electronic health record features, random oversampling, and a single-layer neural network provided optimal classification of 6-month incident atrial fibrillation; however, this model was only marginally better than a logistic regression model with age, sex, and known risk factors for atrial fibrillation.

Meaning  Machine learning approaches applied to electronic health record data hold promise for predicting clinical outcomes, such as incident atrial fibrillation, but this model was not substantially more accurate than a logistic regression model with standard risk factors.

Abstract

Importance  Atrial fibrillation (AF) is the most common sustained cardiac arrhythmia, and its early detection could lead to significant improvements in outcomes through the appropriate prescription of anticoagulation medication. Although a variety of methods exist for screening for AF, a targeted approach, which requires an efficient method for identifying patients at risk, would be preferred.

Objective  To examine machine learning approaches applied to electronic health record data that have been harmonized to the Observational Medical Outcomes Partnership Common Data Model for identifying risk of AF.

Design, Setting, and Participants  This diagnostic study used data from 2 252 219 individuals cared for in the UCHealth hospital system, which comprises 3 large hospitals in Colorado, from January 1, 2011, to October 1, 2018. Initial analysis was performed in December 2018; follow-up analysis was performed in July 2019.

Exposures  All Observational Medical Outcomes Partnership Common Data Model–harmonized electronic health record features, including diagnoses, procedures, medications, age, and sex.

Main Outcomes and Measures  Classification of incident AF in designated 6-month intervals, adjudicated retrospectively, based on area under the receiver operating characteristic curve and F1 statistic.

Results  Of 2 252 219 individuals (1 225 533 [54.4%] women; mean [SD] age, 42.9 [22.3] years), 28 036 (1.2%) developed incident AF during a designated 6-month interval. The machine learning model that used the 200 most common electronic health record features, including age and sex, and random oversampling with a single-layer, fully connected neural network provided the optimal prediction of 6-month incident AF, with an area under the receiver operating characteristic curve of 0.800 and an F1 score of 0.110. This model performed only slightly better than a more basic logistic regression model composed of known clinical risk factors for AF, which had an area under the receiver operating characteristic curve of 0.794 and an F1 score of 0.079.

Conclusions and Relevance  Machine learning approaches to electronic health record data offer a promising method for improving risk prediction for incident AF, but more work is needed to show improvement beyond standard risk factors.

Introduction

Atrial fibrillation (AF) is the most common sustained cardiac arrhythmia, and its prevalence is increasing1; approximately 5.1 million individuals had AF in 2010, and an increase to between 9.3 and 12.1 million is anticipated by 2030.2 Importantly, the increased risk of mortality with AF is almost entirely because of an increased risk of thromboembolic stroke.3,4 This risk could be reduced if a patient with AF and moderate or high risk used oral anticoagulation medication.5-17 A major challenge in the management of patients with AF is that stroke is often the first presentation of AF,18 indicating that simply waiting for a patient to develop AF may not be the optimal approach to limiting the risk of stroke. On the other hand, population-wide screening for AF is not currently recommended,19-21 although some suggest that targeted screening may be useful.21 A model that could predict risk of 6-month incident AF could be applied to target screening and identify a patient with AF before their next clinic visit.

The promise of electronic health record (EHR) data has included the potential to leverage so-called big data analytic approaches to predict clinical outcomes in a real-world context. However, despite widespread adoption of EHRs as mandated under the Patient Protection and Affordable Care Act,22 there are limited examples of practical applications of EHR data to predict a meaningful clinical outcome.23-26 In addition to the technical limitations of working with data at the scale of the EHR, there are also challenges in performing external validation across health care systems.27-29 Nonetheless, with the increasing availability of cloud computing platforms24 and data storage30,31 as well as scalable computational models that can be developed and potentially shared across health care systems, opportunities to apply EHR data to clinical decision-making are emerging. For identifying patients at risk of AF, if an automated prediction algorithm could be applied to patient EHR data, then a clinical decision support system could be developed to guide clinicians toward aggressive screening approaches using standard wearable32 or implantable telemetry33 devices.

A great deal of enthusiasm has accompanied the potential application of deep learning34 and artificial intelligence to outperform humans in image recognition,35,36 text recognition,37,38 and games,39 including checkers40 and Go.41 However, within the health care setting, the so-called black box characteristic of machine learning (ML) has caused hesitancy in application. In certain situations, ML approaches, such as support vector machines42 or random forests,43 have been found to produce greater predictive performance than standard regression models.44-46 More recently, there has been an increased recognition that deep learning models,34,47 which are composed of multiple hidden layers of a neural network rather than a single layer, are better equipped to handle the large amount of data that exists in EHRs. However, to understand how these approaches can be applied to a clinical situation, such as the prediction of incident AF, additional study is needed.

In this investigation, we developed and tested an ML model to predict 6-month incidence of AF using EHR data. This time horizon provided a clinically relevant period of prediction, during which patients could undergo screening tests with wearable monitors or electrocardiograms before a follow-up visit. We conducted a systematic examination of EHR data sampled from 2.2 million individuals, in which we harmonized 26 000 features, including diagnoses and medications, under the Observational Medical Outcomes Partnership Common Data Model (OMOP-CDM). During this developmental process, we also identified the appropriate data resampling method to manage data set imbalance and developed a classification algorithm based on training time and accuracy.

Methods

The code used for the analyses as well as the model weights and mapping (ie, OMOP-CDM input codes) for the final model are available in the eAppendix in the Supplement. The study protocol was approved for analysis of deidentified data (ie, limited data set with dates included) by the University of Colorado institutional review board. Informed consent was waived for practical reasons because of the sample size and deidentification of data. This report adheres to the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) reporting guideline.48

Study Population and Case Ascertainment

The UCHealth hospital system includes 3 large regional centers (North, Central, and South) in the front range of Colorado that share a single EHR, which allows data from all centers to be pooled into a single data warehouse, a copy of which is located on a cloud platform. This warehouse of data was queried using BigQuery (Google) to create a data set and conduct analyses directly on the cloud platform, where an array of ML tools can be run on virtual machines. To create our study data set, we applied a classification approach based on predicting risk of incident AF during a 6-month period. We performed a structured query language query on the UCHealth EHR for patients with a new diagnosis of AF obtained during a 6-month period. To identify cases, we filtered out all patients with prevalent AF on their first encounter and then at 6-month intervals from each encounter. Patients were assigned a case classification if they had AF diagnosed by International Classification of Diseases, Ninth Revision (ICD-9) code 427.31 or ICD-10 code I48.91 within that interval. Once a patient was designated a case, they were removed from the pool, and all patients without AF were designated as controls. Data were available in the EHR from January 1, 2011, to October 1, 2018, for 2 252 219 patients.

CDM and Data Splitting

To offer others the opportunity to validate the findings of this study, we used a CDM for EHR data based on the Observational Health Data Sciences and Informatics collaboration, which uses the OMOP-CDM.49 The OMOP-CDM maps raw EHR data to a harmonized data set; for this investigation, we used this CDM with 26 000 variables (ie, features) from the EHR, including age, sex, diagnoses, procedures, and medications. Additional data, including laboratory values and relevant cardiovascular studies (eg, electrocardiograms), have not been mapped to OMOP-CDM in our system and were excluded from this investigation. These values were time-stamped with the time of entry into the medical record, which was used to associate the value with the timing of the outcome of interest. Features were encoded using 1-hot encoding, which assigned a value of 1 if that feature was present for that individual and 0 otherwise, and were collected cumulatively from the time of first encounter until diagnosis of AF (for cases) or until end of follow-up (for controls). Preliminary studies identified a substantial decrease in analytical time using the 200 most common concepts among the complete EHR. These were used as input into prediction models (eTable 1 and eTable 2 in the Supplement). The final data set was composed of 2 252 219 million records, which were then split into training (1 801 775 records [80.0%]) and testing (450 444 records [20.0%]) sets to compare the models developed in this investigation.

Model Development

For all models, hyperparameter tuning to identify the optimal values for parameters that are not learned during the training process (ie, learning rate) was performed using iterative random sampling of 10 000 records for manual grid search (for neural networks) and 10-fold cross-validation for automated grid search (for other ML approaches). Further details appear in the eMethods in the Supplement.

Because of the relative infrequency of the outcome (ie, 6-month incident AF) across the data set, there was an imbalance between the cases and controls. In the presence of imbalance, many classification algorithms, which often base classification on a probability of disease of greater than 50%, will take longer or even fail to achieve convergence. Approaches to rebalance data, which match the number of cases to the number of controls, include oversampling (ie, adding cases to match the number of controls), undersampling (ie, reducing the number of controls to match the number of cases), and hybrid approaches. We examined several strategies for resampling, including random oversampling, the synthetic minority oversampling technique,50 random undersampling, and cluster centroid undersampling. To identify the best resampling approach, we used a deep neural network (7 layers with 100 neurons/layer) because pilot analyses using a smaller data set suggested this approach might be superior to other ML approaches. Each resampling approach was applied to the training data set and then compared using the testing set. We also compared the resampled training set with a model using no resampling (ie, an imbalanced set).

Once we identified an optimal resampling approach, we compared several classification algorithms, including naive Bayesian classification, regularized logistic regression, random forest classification, gradient boosted classification, 1-layer fully connected neural networks (shallow), and multiple-layer fully connected neural networks (deep). Model comparison was based on area under the receiver operating characteristic curve and the F1 statistic.51-53 Naive Bayesian classification algorithms54 apply Bayes theorem to predict the probability of disease classification based on the assumption of independence (ie, naive) between predictors. Regularized logistic regression (ie, ridge regression) applies a shrinkage penalty to predictors to avoid overfitting. Random forest classification55 algorithms create a multitude of decision trees using different subsets of the predictors, from which the appropriate class is selected as the mode of the collection of trees. Gradient boosted classification56 algorithms are similar to random forests, except they use a stepwise process to reduce the misclassification error of prior models. Neural networks,34,47,57 which can be shallow (ie, 1 layer) or deep (ie, multiple layers), are interconnected nodes, each of which is essentially a nonlinear regression equation, all of which are combined to minimize a loss function; in this study, the loss applied is the cross-entropy loss. Computation time includes all prior data sampling and algorithm performance. Once an optimal model and resampling approach were identified, we conducted sensitivity analyses using several alternative resampling and modeling approaches in combination to ensure that the combination (ie, dimensionality reduction, resampling, and classification algorithm) identified was indeed optimal. Precision recall and receiver operating characteristic curves as well as feature importance plots were created for the optimal model for manual inspection.

Validation of Developed Model

The optimal model was compared with an unregularized logistic regression model based on the presence of known clinical predictors of AF. These included age, sex, hypertension (ICD-9 code, 401.x; ICD-10 code, I10), obesity (ICD-9 code, 278.0; ICD-10 code, E66.9), diabetes (ICD-9 code, 250.0; ICD-10 code, E11.9), coronary artery disease (ICD-9 code, 414.01; ICD-10 code, I25.1x), mitral valve disease (ICD-9 code, 394.0 or 424.0; ICD-10 code, I34.2 or I34.0), heart failure (ICD-9 code, 428.0; ICD-10 code, I50.9), and chronic kidney disease (ICD-9 code, 585.9; ICD-10 code, N18.9).

Statistical Analysis

Model comparison was performed using the area under the receiver operating characteristic curve for classification accuracy of 6-month incident AF and the F1 score, defined as the harmonic mean of precision (ie, positive predictive value) and recall (ie, sensitivity), with perfect precision and recall at an F1 score of 1 and the worst precision and recall at 0. All analyses were run on the Google Cloud Platform (Google), using 96 central processing units and 620 gigabytes of random access memory. Scripts were composed in Python version 3 (Python) and were run on Jupyter Notebook (Project Jupyter) with Tensorflow platform (Google) on the Google Cloud Platform. Machine learning packages included scikit-learn and keras. Confidence intervals were calculated using the Wald method,58,59 although almost all were within the rounding error of the estimates because of the large testing sample size (ie, approximately 440 000) and are not displayed. We did not perform formal null-hypothesis testing, so no prespecified level for statistical significance was set. Additional details appear in the eMethods in the Supplement.

Results

Across the UCHealth population of 2 252 219 individuals, with 1 225 533 women (54.4%) and a mean (SD) age of 42.9 (22.3) years, we identified 28 036 patients (1.2%) with 6-month incident AF (Table 1). Cardiac disease and known risk factors were more common among patients who developed incident AF than among those with no AF, including older age (mean [SD] age, 71.7 [16.5] years vs 42.9 [22.3] years) and male sex (12 919 [46.1%] women vs 1 212 586 [54.5%] women) (Table 1).

We first examined the role of undersampling and oversampling methods to identify the optimal approach to manage the imbalance between cases and controls that we identified in this data set. Using a 7-layer deep neural network algorithm with hyperbolic tangent activation and 20% dropout, we found that random oversampling provided the best classification F1 score (0.101) compared with other methods (eg, synthetic minority oversampling technique, 0.090; random undersampling, 0.099), including no resampling, which provided substantially poorer classification (0.002) than other methods (Table 2).

Using the random oversampling strategy, we examined several classification algorithms to identify a potential overall best model. Among the approaches examined, we found that a single-layer shallow neural network using the 200 most common EHR features, including age and sex, was superior to other methods (F1 score, 0.110; AUC, 0.800), which included regularized regression (F1, 0.088; AUC, 0.806), gradient boosted descent (F1, 0.108; AUC, 0.762), random forest classification (F1, 0.076; AUC, 0.792), and a deep neural network (F1, 0.101; AUC, 0.800) (Table 3). The single-layer neural network had a specificity of 84.9%, a sensitivity (ie, recall) of 75.2%, negative predictive value of 99.6%, and positive predictive value of 5.9% at a probability (ie, decision) cutoff of 0.5, with relatively poor calibration across predicted probabilities (Figure; eTable 3A and eFigure in the Supplement).

Finally, we compared the optimal model with an unregularized logistic regression model based on known AF risk factors. The logistic regression model had an F1 score of 0.079 and AUC of 0.794 (eTable 3B and eTable 4 in the Supplement).

Discussion

In this investigation of the application of an ML model using harmonized EHR data for identifying patients at risk of 6-month incident AF, we found that a shallow neural network using random oversampling provided the optimal classification of risk, although this model was not substantially better than a simpler model based on known AF risk factors. These results are important because, in addition to motivating future investigations to apply ML methods to EHR data to identify patients at risk of AF, they also incorporated harmonized data. This means that the optimal model can be directly applied not only in our institution’s EHR but in data from the EHR of any other medical institution participating in OMOP or the Observational Health Data Sciences and Informatics collaboration. In clinical application, our model could thus be inserted directly in the user interface to guide targeted screening of patients at risk of AF, including the development of prospective follow-up studies to use the prediction for targeted screening for AF, such as routine electrocardiograms, implantable devices, or wearable devices.

However, there are several reasons for hesitancy before taking these results directly to the bedside to guide clinical management without additional investigation. First, the model that we identified was not extremely accurate, with an F1 score under 20% and a sensitivity of approximately 75% based on a cutoff probability of 0.5 for risk. Although the probability threshold could be lowered to improve sensitivity of classification, the decreased specificity and number of false-positive results with such an approach would result in a large number of patients undergoing unfruitful screening. Second, as mentioned, we found that a simple logistic regression model based on known and clinical risk factors60-63 performed nearly as well as the more sophisticated ML model. This limitation has practical importance because many EHR systems allow for built-in regression modeling that can be plugged directly into the system in real time.

In the process of developing a 6-month risk prediction model for AF, we made several important observations about the application of ML to EHR data. First, we found that, for a rare condition like 6-month incident AF (which occurred in 1.2% of our sample), oversampling to rebalance the data was superior to using the imbalanced data set. The limitations of applying classification algorithms to imbalanced data sets are well described,64-66 and yet many investigations continue to use imbalanced data to draw conclusions from statistical modeling.

Second, we found that a shallow neural network provided the optimal classification algorithm, with better classification than a deep neural network approach. This finding demonstrates that, although deep learning approaches may be superior for classification of structured data sets such as in image67 or voice68 recognition, they do not always perform better than other standard ML algorithms. This finding also highlights the importance of examining all approaches for each classification problem, rather than assuming a given approach is optimal.

There are several additional strengths of this investigation. First, all models were created using a harmonization scheme (ie, OMOP-CDM) that could allow for direct application and validation to data mapped from a separate EHR. Such harmonization allows for exploring transfer-learning69 approaches, which could provide additional insights into similar and divergent AF risk factors across populations. Second, we conducted a systematic approach to identify the best resampling and classification algorithm for this outcome. Further work in other outcomes is needed to determine if the combination we identified for predicting 6-month incident AF would also be optimal for prevalent or longer-term AF predictions as well as for outcomes that are more or less common than AF. Finally, we examined a data set with more than 2 million individuals, which provided a large enough sample size from our single institution to conduct cross-validation and out-of-sample validation. This power from the use of big data is possible because of the unique circumstances of our relationship with the Google Cloud Platform. However, many other EHRs are moving to the cloud, providing further opportunities for development and testing.

Limitations

There are several limitations in this study, many of which are the subject of future, more targeted investigations. First, our study included a very simple method for managing the temporal relationships between features in our data set, which did not account for time-varying effects or censoring. An AF event that occurred the day after an encounter was modeled the same as an AF event occurring the day before a subsequent encounter, and a diagnosis or medication that was given a month before the AF diagnosis was weighted the same as that given 4 years prior. While we suggest that the approach we used for this investigation was reasonable based on the typical 6-month follow-up schedule for patients seen in cardiology clinics, we realize that additional information about temporal risk will be needed for more accurate prediction approaches. More sophisticated methods such as recurrent neural networks70,71 or parametric survival functions72 could provide more accurate predictions in future investigations. A second limitation is that we excluded some data elements, such as laboratory values and diagnostic test results, that may have had prognostic value for predicting 6-month AF.73,74 Some of these values have been difficult to harmonize across data sets via OMOP-CDM, and others, such as echo measures of diastolic function, have high variability in interinstitutional measurement.62,75 Nonetheless, there are many additional biomarkers73,74 likely to have a more biological association with risk of AF than a diagnostic association, and future applications that include this information would be expected to provide both predictive and inferential knowledge about the risk of AF. Third, despite having a reasonably good AUC for the classification of 6-month risk of AF (ie, 0.80), the actual increase in risk based on the model is not particularly elevated (5.9% for positive predicted risk, and 0.4% for negative prediction) because of the overall low incidence of AF across the sample. Clinically, we believe that this limitation indicates that our model approach would be insufficient for guiding treatments, such as initiating anticoagulation medication, but we suggest that that it could be used to identify individuals who may or may not be candidates for enhanced screening, eg, through the use of wearable monitors. Further prospective studies would be needed to identify the actual benefit that could be obtained from integration of this approach into a screening program for AF. However, we believe that because our model was constructed directly from EHR data, integration into a systemwide screening program would be more practical than if the model were created using data that needed to be collected outside the EHR. Also, although the systematic harmonized approach we used in this study holds potential for cross-institutional validation, much work is needed in terms of data sharing before actual testing can be performed. We did not have data on race/ethnicity or socioeconomic status readily available in our data set, but we are certain that our population is likely to be different from those cared for in many other health systems, and we believe that it is only through external validation in other EHRs that an unbiased risk prediction model could be developed. Our group and others are working in this direction, and the hope is that, sometime soon, all EHRs will incorporate a standard risk prediction model for AF and many other conditions.

Conclusions

We studied the development of an ML model for predicting 6-month risk of AF using harmonized EHR data and found that the combination of random oversampling and single-layer neural network classification provided superior prediction than other ML models. Further work is needed to explore the technical and clinical applications of this model to improving outcomes.

Back to top
Article Information

Accepted for Publication: November 18, 2019.

Published: January 17, 2020. doi:10.1001/jamanetworkopen.2019.19396

Open Access: This is an open access article distributed under the terms of the CC-BY License. © 2020 Tiwari P et al. JAMA Network Open.

Corresponding Author: Michael A. Rosenberg, MD, Division of Cardiology and Cardiac Electrophysiology, University of Colorado School of Medicine, 12631 E 17th Ave, Mail Stop B130, Aurora, CO 80045 (michael.a.rosenberg@ucdenver.edu).

Author Contributions: Mr Tiwari and Dr Rosenberg had full access to all of the data in the study and take responsibility for the integrity of the data and the accuracy of the data analysis.

Concept and design: Tiwari, Xing, Rosenberg.

Acquisition, analysis, or interpretation of data: Tiwari, Colborn, Smith, Ghosh, Rosenberg.

Drafting of the manuscript: Smith, Rosenberg.

Critical revision of the manuscript for important intellectual content: All authors.

Statistical analysis: Tiwari, Colborn, Smith, Rosenberg.

Obtained funding: Rosenberg.

Administrative, technical, or material support: Tiwari, Rosenberg.

Supervision: Rosenberg.

Conflict of Interest Disclosures: None reported.

Funding/Support: This work was funded by grant 5K23 HL127296 from the National Heart, Lung, and Blood Institute.

Role of the Funder/Sponsor: The funder had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; preparation, review, or approval of the manuscript; and decision to submit the manuscript for publication.

Meeting Presentation: This work was presented at the 2019 Scientific Sessions of the American Heart Association; November 16, 2019; Philadelphia, Pennsylvania.

Previous Posting: This manuscript was posted as a preprint on bioRxiv.org.

References
1.
Majeed  A, Moser  K, Carroll  K.  Trends in the prevalence and management of atrial fibrillation in general practice in England and Wales, 1994-1998: analysis of data from the general practice research database.  Heart. 2001;86(3):284-288. doi:10.1136/heart.86.3.284PubMedGoogle ScholarCrossref
2.
Benjamin  EJ, Blaha  MJ, Chiuve  SE,  et al; American Heart Association Statistics Committee and Stroke Statistics Subcommittee.  Heart disease and stroke statistics, 2017 update: a report from the American Heart Association.  Circulation. 2017;135(10):e146-e603. doi:10.1161/CIR.0000000000000485PubMedGoogle ScholarCrossref
3.
Benjamin  EJ, Wolf  PA, D’Agostino  RB, Silbershatz  H, Kannel  WB, Levy  D.  Impact of atrial fibrillation on the risk of death: the Framingham Heart Study.  Circulation. 1998;98(10):946-952. doi:10.1161/01.CIR.98.10.946PubMedGoogle ScholarCrossref
4.
Benjamin  EJ, Levy  D, Vaziri  SM, D’Agostino  RB, Belanger  AJ, Wolf  PA.  Independent risk factors for atrial fibrillation in a population-based cohort: the Framingham Heart Study.  JAMA. 1994;271(11):840-844. doi:10.1001/jama.1994.03510350050036PubMedGoogle ScholarCrossref
5.
Singer  DE, Hughes  RA, Gress  DR,  et al; Boston Area Anticoagulation Trial for Atrial Fibrillation Investigators.  The effect of low-dose warfarin on the risk of stroke in patients with nonrheumatic atrial fibrillation.  N Engl J Med. 1990;323(22):1505-1511. doi:10.1056/NEJM199011293232201PubMedGoogle ScholarCrossref
6.
 Stroke Prevention in Atrial Fibrillation Study: final results.  Circulation. 1991;84(2):527-539. doi:10.1161/01.CIR.84.2.527PubMedGoogle ScholarCrossref
7.
 Warfarin versus aspirin for prevention of thromboembolism in atrial fibrillation: Stroke Prevention in Atrial Fibrillation II Study.  Lancet. 1994;343(8899):687-691.PubMedGoogle Scholar
8.
Petersen  P, Boysen  G, Godtfredsen  J, Andersen  ED, Andersen  B.  Placebo-controlled, randomised trial of warfarin and aspirin for prevention of thromboembolic complications in chronic atrial fibrillation: the Copenhagen AFASAK study.  Lancet. 1989;1(8631):175-179. doi:10.1016/S0140-6736(89)91200-2PubMedGoogle ScholarCrossref
9.
Ezekowitz  MD, Bridgers  SL, James  KE,  et al; Veterans Affairs Stroke Prevention in Nonrheumatic Atrial Fibrillation Investigators.  Warfarin in the prevention of stroke associated with nonrheumatic atrial fibrillation.  N Engl J Med. 1992;327(20):1406-1412. doi:10.1056/NEJM199211123272002PubMedGoogle ScholarCrossref
10.
Connolly  SJ, Laupacis  A, Gent  M, Roberts  RS, Cairns  JA, Joyner  C.  Canadian Atrial Fibrillation Anticoagulation (CAFA) Study.  J Am Coll Cardiol. 1991;18(2):349-355. doi:10.1016/0735-1097(91)90585-WPubMedGoogle ScholarCrossref
11.
 Risk factors for stroke and efficacy of antithrombotic therapy in atrial fibrillation: analysis of pooled data from five randomized controlled trials.  Arch Intern Med. 1994;154(13):1449-1457. doi:10.1001/archinte.1994.00420130036007PubMedGoogle ScholarCrossref
12.
Hart  RG, Pearce  LA, Aguilar  MI.  Meta-analysis: antithrombotic therapy to prevent stroke in patients who have nonvalvular atrial fibrillation.  Ann Intern Med. 2007;146(12):857-867. doi:10.7326/0003-4819-146-12-200706190-00007PubMedGoogle ScholarCrossref
13.
van Walraven  C, Hart  RG, Singer  DE,  et al.  Oral anticoagulants vs aspirin in nonvalvular atrial fibrillation: an individual patient meta-analysis.  JAMA. 2002;288(19):2441-2448. doi:10.1001/jama.288.19.2441PubMedGoogle ScholarCrossref
14.
Cooper  NJ, Sutton  AJ, Lu  G, Khunti  K.  Mixed comparison of stroke prevention treatments in individuals with nonrheumatic atrial fibrillation.  Arch Intern Med. 2006;166(12):1269-1275. doi:10.1001/archinte.166.12.1269PubMedGoogle ScholarCrossref
15.
Connolly  SJ, Ezekowitz  MD, Yusuf  S,  et al; RE-LY Steering Committee and Investigators.  Dabigatran versus warfarin in patients with atrial fibrillation.  N Engl J Med. 2009;361(12):1139-1151. doi:10.1056/NEJMoa0905561PubMedGoogle ScholarCrossref
16.
Patel  MR, Mahaffey  KW, Garg  J,  et al; ROCKET AF Investigators.  Rivaroxaban versus warfarin in nonvalvular atrial fibrillation.  N Engl J Med. 2011;365(10):883-891. doi:10.1056/NEJMoa1009638PubMedGoogle ScholarCrossref
17.
Connolly  SJ, Eikelboom  J, Joyner  C,  et al; AVERROES Steering Committee and Investigators.  Apixaban in patients with atrial fibrillation.  N Engl J Med. 2011;364(9):806-817. doi:10.1056/NEJMoa1007432PubMedGoogle ScholarCrossref
18.
Jaakkola  J, Mustonen  P, Kiviniemi  T,  et al.  Stroke as the first manifestation of atrial fibrillation.  PLoS One. 2016;11(12):e0168010. doi:10.1371/journal.pone.0168010PubMedGoogle Scholar
19.
Curry  SJ, Krist  AH, Owens  DK,  et al; US Preventive Services Task Force.  Screening for atrial fibrillation with electrocardiography: US Preventive Services Task Force recommendation statement.  JAMA. 2018;320(5):478-484. doi:10.1001/jama.2018.10321PubMedGoogle ScholarCrossref
20.
Jonas  DE, Kahwati  LC, Yun  JDY, Middleton  JC, Coker-Schwimmer  M, Asher  GN.  Screening for atrial fibrillation with electrocardiography: evidence report and systematic review for the US Preventive Services Task Force.  JAMA. 2018;320(5):485-498. doi:10.1001/jama.2018.4190PubMedGoogle ScholarCrossref
21.
Freedman  B, Camm  J, Calkins  H,  et al; AF-Screen Collaborators.  Screening for atrial fibrillation: a report of the AF-SCREEN international collaboration.  Circulation. 2017;135(19):1851-1867. doi:10.1161/CIRCULATIONAHA.116.026693PubMedGoogle ScholarCrossref
22.
Kocher  R, Emanuel  EJ, DeParle  NA.  The Affordable Care Act and the future of clinical medicine: the opportunities and challenges.  Ann Intern Med. 2010;153(8):536-539. doi:10.7326/0003-4819-153-8-201010190-00274PubMedGoogle ScholarCrossref
23.
Dexheimer  JW, Talbot  TR, Sanders  DL, Rosenbloom  ST, Aronsky  D.  Prompting clinicians about preventive care measures: a systematic review of randomized controlled trials.  J Am Med Inform Assoc. 2008;15(3):311-320. doi:10.1197/jamia.M2555PubMedGoogle ScholarCrossref
24.
Sadoughi  F, Erfannia  L.  Health information system in a cloud computing context.  Stud Health Technol Inform. 2017;236:290-297.PubMedGoogle Scholar
25.
Jensen  RE, Rothrock  NE, DeWitt  EM,  et al.  The role of technical advances in the adoption and integration of patient-reported outcomes in clinical care.  Med Care. 2015;53(2):153-159. doi:10.1097/MLR.0000000000000289PubMedGoogle ScholarCrossref
26.
Casey  JA, Schwartz  BS, Stewart  WF, Adler  NE.  Using electronic health records for population health research: a review of methods and applications.  Annu Rev Public Health. 2016;37:61-81. doi:10.1146/annurev-publhealth-032315-021353PubMedGoogle ScholarCrossref
27.
Huang  Y, Lee  J, Wang  S, Sun  J, Liu  H, Jiang  X.  Privacy-preserving predictive modeling: harmonization of contextual embeddings from different sources.  JMIR Med Inform. 2018;6(2):e33. doi:10.2196/medinform.9455PubMedGoogle Scholar
28.
Jiang  G, Kiefer  RC, Sharma  DK, Prud’hommeaux  E, Solbrig  HR.  A consensus-based approach for harmonizing the OHDSI Common Data Model with HL7 FHIR.  Stud Health Technol Inform. 2017;245:887-891.PubMedGoogle Scholar
29.
Jiang  G, Evans  J, Oniki  TA,  et al.  Harmonization of detailed clinical models with clinical study data standards.  Methods Inf Med. 2015;54(1):65-74. doi:10.3414/ME13-02-0019PubMedGoogle ScholarCrossref
30.
Ocaña  K, de Oliveira  D.  Parallel computing in genomic research: advances and applications.  Adv Appl Bioinform Chem. 2015;8:23-35.PubMedGoogle Scholar
31.
Korb  O, Finn  PW, Jones  G.  The cloud and other new computational methods to improve molecular modelling.  Expert Opin Drug Discov. 2014;9(10):1121-1131. doi:10.1517/17460441.2014.941800PubMedGoogle ScholarCrossref
32.
Rosenberg  MA, Samuel  M, Thosani  A, Zimetbaum  PJ.  Use of a noninvasive continuous monitoring device in the management of atrial fibrillation: a pilot study.  Pacing Clin Electrophysiol. 2013;36(3):328-333. doi:10.1111/pace.12053PubMedGoogle ScholarCrossref
33.
Diederichsen  SZ, Haugan  KJ, Køber  L,  et al.  Atrial fibrillation detected by continuous electrocardiographic monitoring using implantable loop recorder to prevent stroke in individuals at risk (the LOOP study): rationale and design of a large randomized controlled trial.  Am Heart J. 2017;187:122-132. doi:10.1016/j.ahj.2017.02.017PubMedGoogle ScholarCrossref
34.
LeCun  Y, Bengio  Y, Hinton  G.  Deep learning.  Nature. 2015;521(7553):436-444. doi:10.1038/nature14539PubMedGoogle ScholarCrossref
35.
Stallkamp  J, Schlipsing  M, Salmen  J, Igel  C.  Man vs. computer: benchmarking machine learning algorithms for traffic sign recognition.  Neural Netw. 2012;32:323-332.PubMedGoogle ScholarCrossref
36.
Kooi  T, Litjens  G, van Ginneken  B,  et al.  Large scale deep learning for computer aided detection of mammographic lesions.  Med Image Anal. 2017;35:303-312. doi:10.1016/j.media.2016.07.007PubMedGoogle ScholarCrossref
37.
Minarro-Giménez  JA, Marín-Alonso  O, Samwald  M.  Exploring the application of deep learning techniques on medical text corpora.  Stud Health Technol Inform. 2014;205:584-588.PubMedGoogle Scholar
38.
Zhu  Q, Li  X, Conesa  A, Pereira  C.  GRAM-CNN: a deep learning approach with local context for named entity recognition in biomedical text.  Bioinformatics. 2018;34(9):1547-1554. doi:10.1093/bioinformatics/btx815PubMedGoogle ScholarCrossref
39.
Moravčík  M, Schmid  M, Burch  N,  et al.  DeepStack: expert-level artificial intelligence in heads-up no-limit poker.  Science. 2017;356(6337):508-513. doi:10.1126/science.aam6960PubMedGoogle ScholarCrossref
40.
Schaeffer  J, Burch  N, Björnsson  Y,  et al.  Checkers is solved.  Science. 2007;317(5844):1518-1522. doi:10.1126/science.1144079PubMedGoogle ScholarCrossref
41.
Silver  D, Huang  A, Maddison  CJ,  et al.  Mastering the game of Go with deep neural networks and tree search.  Nature. 2016;529(7587):484-489. doi:10.1038/nature16961PubMedGoogle ScholarCrossref
42.
Ramírez  J, Monasterio  V, Mincholé  A,  et al.  Automatic SVM classification of sudden cardiac death and pump failure death from autonomic and repolarization ECG markers.  J Electrocardiol. 2015;48(4):551-557. doi:10.1016/j.jelectrocard.2015.04.002PubMedGoogle ScholarCrossref
43.
Beaulieu-Jones  BK, Greene  CS; Pooled Resource Open-Access ALS Clinical Trials Consortium.  Semi-supervised learning of the electronic health record for phenotype stratification.  J Biomed Inform. 2016;64:168-178. doi:10.1016/j.jbi.2016.10.007PubMedGoogle ScholarCrossref
44.
Luo  Y, Li  Z, Guo  H,  et al.  Predicting congenital heart defects: a comparison of three data mining methods.  PLoS One. 2017;12(5):e0177811. doi:10.1371/journal.pone.0177811PubMedGoogle Scholar
45.
Amini  P, Ahmadinia  H, Poorolajal  J, Moqaddasi Amiri  M.  Evaluating the high risk groups for suicide: a comparison of logistic regression, support vector machine, decision tree and artificial neural network.  Iran J Public Health. 2016;45(9):1179-1187.PubMedGoogle Scholar
46.
Churpek  MM, Yuen  TC, Winslow  C, Meltzer  DO, Kattan  MW, Edelson  DP.  Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards.  Crit Care Med. 2016;44(2):368-374. doi:10.1097/CCM.0000000000001571PubMedGoogle ScholarCrossref
47.
Johnson  KW, Torres Soto  J, Glicksberg  BS,  et al.  Artificial intelligence in cardiology.  J Am Coll Cardiol. 2018;71(23):2668-2679. doi:10.1016/j.jacc.2018.03.521PubMedGoogle ScholarCrossref
48.
Collins  GS, Reitsma  JB, Altman  DG, Moons  KG.  Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement.  BMJ. 2015;350:g7594. doi:10.1136/bmj.g7594PubMedGoogle ScholarCrossref
49.
Makadia  R, Ryan  PB.  Transforming the Premier Perspective hospital database into the Observational Medical Outcomes Partnership (OMOP) Common Data Model.  EGEMS (Wash DC). 2014;2(1):1110.PubMedGoogle Scholar
50.
Chawla  NV, Bowyer  KW, Hall  LO, Kegelmeyer  WP.  SMOTE: synthetic minority over-sampling technique.  J Artif Intell Res. 2002;16:321-357. doi:10.1613/jair.953Google ScholarCrossref
51.
Chai  KE, Anthony  S, Coiera  E, Magrabi  F.  Using statistical text classification to identify health information technology incidents.  J Am Med Inform Assoc. 2013;20(5):980-985. doi:10.1136/amiajnl-2012-001409PubMedGoogle ScholarCrossref
52.
Davis  J, Goadrich  M.  The relationship between precision-recall and ROC curves.  In:  ICML ’06 Proceedings of the 23rd International Conference on Machine Learning. New York, NY: Association for Computing Machinery; 2006:233-240. doi:10.1145/1143844.1143874Google Scholar
53.
Hripcsak  G, Rothschild  AS.  Agreement, the f-measure, and reliability in information retrieval.  J Am Med Inform Assoc. 2005;12(3):296-298. doi:10.1197/jamia.M1733PubMedGoogle ScholarCrossref
54.
Wang  Q, Garrity  GM, Tiedje  JM, Cole  JR.  Naive Bayesian classifier for rapid assignment of rRNA sequences into the new bacterial taxonomy.  Appl Environ Microbiol. 2007;73(16):5261-5267. doi:10.1128/AEM.00062-07PubMedGoogle ScholarCrossref
55.
Mansoor  H, Elgendy  IY, Segal  R, Bavry  AA, Bian  J.  Risk prediction model for in-hospital mortality in women with ST-elevation myocardial infarction: A machine learning approach.  Heart Lung. 2017;46(6):405-411.PubMedGoogle ScholarCrossref
56.
Qu  Y, Adam  BL, Yasui  Y,  et al.  Boosted decision tree analysis of surface-enhanced laser desorption/ionization mass spectral serum profiles discriminates prostate cancer from noncancer patients.  Clin Chem. 2002;48(10):1835-1843.PubMedGoogle Scholar
57.
Penny  W, Frost  D.  Neural networks in clinical medicine.  Med Decis Making. 1996;16(4):386-398. doi:10.1177/0272989X9601600409PubMedGoogle ScholarCrossref
58.
Agresti  A, Coull  BA.  Approximate is better than “exact” for interval estimation of binomial proportions.  Am Stat. 1998;52(2):119-126. http://users.stat.ufl.edu/~aa/articles/agresti_coull_1998.pdf. Accessed December 2, 2019.Google Scholar
59.
Lakhani  P, Langlotz  CP.  Automated detection of radiology reports that document non-routine communication of critical or significant results.  J Digit Imaging. 2010;23(6):647-657. doi:10.1007/s10278-009-9237-1PubMedGoogle ScholarCrossref
60.
Nattel  S, Shiroshita-Takeshita  A, Brundel  BJ, Rivard  L.  Mechanisms of atrial fibrillation: lessons from animal models.  Prog Cardiovasc Dis. 2005;48(1):9-28. doi:10.1016/j.pcad.2005.06.002PubMedGoogle ScholarCrossref
61.
Nattel  S, Harada  M.  Atrial remodeling and atrial fibrillation: recent advances and translational perspectives.  J Am Coll Cardiol. 2014;63(22):2335-2345. doi:10.1016/j.jacc.2014.02.555PubMedGoogle ScholarCrossref
62.
Rosenberg  MA, Manning  WJ.  Diastolic dysfunction and risk of atrial fibrillation: a mechanistic appraisal.  Circulation. 2012;126(19):2353-2362. doi:10.1161/CIRCULATIONAHA.112.113233PubMedGoogle ScholarCrossref
63.
Alonso  A, Krijthe  BP, Aspelund  T,  et al.  Simple risk model predicts incidence of atrial fibrillation in a racially and geographically diverse population: the CHARGE-AF consortium.  J Am Heart Assoc. 2013;2(2):e000102. doi:10.1161/JAHA.112.000102PubMedGoogle Scholar
64.
Li  DC, Hu  SC, Lin  LS, Yeh  CW.  Detecting representative data and generating synthetic samples to improve learning accuracy with imbalanced data sets.  PLoS One. 2017;12(8):e0181853. doi:10.1371/journal.pone.0181853PubMedGoogle Scholar
65.
Colborn  KL, Bronsert  M, Amioka  E, Hammermeister  K, Henderson  WG, Meguid  R.  Identification of surgical site infections using electronic health record data.  Am J Infect Control. 2018;46(11):1230-1235. doi:10.1016/j.ajic.2018.05.011PubMedGoogle ScholarCrossref
66.
Blagus  R, Lusa  L.  SMOTE for high-dimensional class-imbalanced data.  BMC Bioinformatics. 2013;14:106. doi:10.1186/1471-2105-14-106PubMedGoogle ScholarCrossref
67.
Lee  JG, Jun  S, Cho  YW,  et al.  Deep learning in medical imaging: general overview.  Korean J Radiol. 2017;18(4):570-584. doi:10.3348/kjr.2017.18.4.570PubMedGoogle ScholarCrossref
68.
Agarwalla  S, Sarma  KK.  Machine learning based sample extraction for automatic speech recognition using dialectal Assamese speech.  Neural Netw. 2016;78:97-111.PubMedGoogle ScholarCrossref
69.
Chen  Y, Ghosh  J, Bejan  CA,  et al.  Building bridges across electronic health record systems through inferred phenotypic topics.  J Biomed Inform. 2015;55:82-93. doi:10.1016/j.jbi.2015.03.011PubMedGoogle ScholarCrossref
70.
Pham  T, Tran  T, Phung  D, Venkatesh  S.  Predicting healthcare trajectories from medical records: a deep learning approach.  J Biomed Inform. 2017;69:218-229. doi:10.1016/j.jbi.2017.04.001PubMedGoogle ScholarCrossref
71.
Yang  H, Pan  Z, Tao  Q.  Robust and adaptive online time series prediction with long short-term memory.  Comput Intell Neurosci. 2017;2017:9478952. doi:10.1155/2017/9478952PubMedGoogle Scholar
72.
Zhang  Z.  Parametric regression model for survival data: Weibull regression model as an example.  Ann Transl Med. 2016;4(24):484. doi:10.21037/atm.2016.08.45PubMedGoogle ScholarCrossref
73.
Rosenberg  MA, Maziarz  M, Tan  AY,  et al.  Circulating fibrosis biomarkers and risk of atrial fibrillation: the Cardiovascular Health Study (CHS).  Am Heart J. 2014;167(5):723-8.e2.PubMedGoogle ScholarCrossref
74.
Patton  KK, Ellinor  PT, Heckbert  SR,  et al.  N-terminal pro-B-type natriuretic peptide is a major predictor of the development of atrial fibrillation: the Cardiovascular Health Study.  Circulation. 2009;120(18):1768-1774. doi:10.1161/CIRCULATIONAHA.109.873265PubMedGoogle ScholarCrossref
75.
Rosenberg  MA, Gottdiener  JS, Heckbert  SR, Mukamal  KJ.  Echocardiographic diastolic parameters and risk of atrial fibrillation: the Cardiovascular Health Study.  Eur Heart J. 2012;33(7):904-912. doi:10.1093/eurheartj/ehr378PubMedGoogle ScholarCrossref
×