Machine Learning Prediction in Cardiovascular Diseases: A Meta Analysis
Machine Learning Prediction in Cardiovascular Diseases: A Meta Analysis
Machine Learning Prediction in Cardiovascular Diseases: A Meta Analysis
com/scientificreports
Several machine learning (ML) algorithms have been increasingly utilized for cardiovascular disease
prediction. We aim to assess and summarize the overall predictive ability of ML algorithms in
cardiovascular diseases. A comprehensive search strategy was designed and executed within the
MEDLINE, Embase, and Scopus databases from database inception through March 15, 2019. The
primary outcome was a composite of the predictive ability of ML algorithms of coronary artery
disease, heart failure, stroke, and cardiac arrhythmias. Of 344 total studies identified, 103 cohorts,
with a total of 3,377,318 individuals, met our inclusion criteria. For the prediction of coronary artery
disease, boosting algorithms had a pooled area under the curve (AUC) of 0.88 (95% CI 0.84–0.91), and
custom-built algorithms had a pooled AUC of 0.93 (95% CI 0.85–0.97). For the prediction of stroke,
support vector machine (SVM) algorithms had a pooled AUC of 0.92 (95% CI 0.81–0.97), boosting
algorithms had a pooled AUC of 0.91 (95% CI 0.81–0.96), and convolutional neural network (CNN)
algorithms had a pooled AUC of 0.90 (95% CI 0.83–0.95). Although inadequate studies for each
algorithm for meta-analytic methodology for both heart failure and cardiac arrhythmias because the
confidence intervals overlap between different methods, showing no difference, SVM may outperform
other algorithms in these areas. The predictive ability of ML algorithms in cardiovascular diseases is
promising, particularly SVM and boosting algorithms. However, there is heterogeneity among ML
algorithms in terms of multiple parameters. This information may assist clinicians in how to interpret
data and implement optimal algorithms for their dataset.
Machine learning (ML) is a branch of artificial intelligence (AI) that is increasingly utilized within the field
of cardiovascular medicine. It is essentially how computers make sense of data and decide or classify a task with
or without human supervision. The conceptual framework of ML is based on models that receive input data
(e.g., images or text) and through a combination of mathematical optimization and statistical analysis predict
outcomes (e.g., favorable, unfavorable, or neutral). Several ML algorithms have been applied to daily activities.
As an example, a common ML algorithm designated as SVM can recognize non-linear patterns for use in facial
recognition, handwriting interpretation, or detection of fraudulent credit card t ransactions1,2. So-called boost-
ing algorithms used for prediction and classification have been applied to the identification and processing of
spam email. Another algorithm, denoted random forest (RF), can facilitate decisions by averaging several nodes.
While convolutional neural network (CNN) processing, combines several layers and apples to image classification
1
Section of Cardiology, Baylor College of Medicine, Houston, TX, USA. 2Harrington Heart & Vascular Institute, Case
Western Reserve University, University Hospitals Cleveland Medical Center, Cleveland, OH, USA. 3Department
of Cardiovascular Diseases, New York University School of Medicine, New York, NY, USA. 4Robert D. and Patricia
E. Kern Center for the Science of Health Care Delivery, Rochester, MN, USA. 5Division of Health Care Policy
and Research, Department of Health Sciences Research, Mayo Clinic, Rochester, MN, USA. 6Department of
Genetics and Genomic Sciences, Institute for Next Generation Healthcare, Icahn School of Medicine at Mount
Sinai, New York, NY, USA. 7Levy Library, Icahn School of Medicine at Mount Sinai, New York, NY, USA. 8Division
of Cardiovascular Diseases, Mayo Clinic, Rochester, MN, USA. 9Department of Cardiovascular Diseases, Icahn
School of Medicine at Mount Sinai, Mount Sinai Hospital, Mount Sinai Heart, New York, NY, USA. 10Department
of Cardiovascular Medicine, Heart and Vascular Institute, Cleveland Clinic, Cleveland, OH, USA. *email:
Chayakrit.Krittanawong@bcm.edu
Vol.:(0123456789)
www.nature.com/scientificreports/
and segmentation3–5. We have previously described technical details of each of these a lgorithms6–8, but no con-
sensus has emerged to guide the selection of specific algorithms for clinical application within the field of car-
diovascular medicine. Although selecting optimal algorithms for research questions and reproducing algorithms
in different clinical datasets is feasible, the clinical interpretation and judgement for implementing algorithms
are very challenging. A deep understanding of statistical and clinical knowledge in ML practitioners is also a
challenge. Most ML studies reported a discrimination measure such as the area under an ROC curve (AUC),
instead of p values. Most importantly, an acceptable cutoff for AUC to be used in clinical practice, interpretation
of the cutoff, and the appropriate/best algorithms to be applied in cardiovascular datasets remain to be evalu-
ated. We previously proposed the methodology to conduct ML research in m edicine6. Systematic review and
meta-analysis, the foundation of modern evidence-based medicine, have to be performed in order to evaluate
the existing ML algorithm in cardiovascular disease prediction. Here, we performed the first systematic review
and meta-analysis of ML research over a million patients in cardiovascular diseases.
Methods
This study is reported in accordance with the Preferred Reporting Information for Systematic Reviews and Meta-
Analysis (PRISMA) recommendations. Ethical approval was not required for this study.
Search strategy. A comprehensive search strategy was designed and executed within the MEDLINE,
Embase, and Scopus databases from database inception through March 15, 2019. One investigator (R.P.)
designed and conducted the search strategy using input from the study’s principal investigator (C.K.). Con-
trolled vocabulary, supplemented with keywords, was used to search for studies of ML algorithms and coronary
heart disease, stroke, heart failure, and cardiac arrhythmias. The detailed strategy is available from the reprint
author. The full search strategies can be found in the supplementary documentation.
Study selection. Search results were exported from all databases and imported into Covidence9, an online
systematic review tool, by one investigator (R.P.). Duplicates were identified and removed using Covidence’s
automated de-duplication functionality. The de-duplicated set of results was screened independently by two
reviewers (C.K. and H.V.) in two successive rounds to identify studies that met the pre-specified eligibility cri-
teria. In the initial screening, two investigators (C.K. and H.V.) independently examined the titles and abstracts
of the records retrieved from the search via the Covidence portal and used a standard extraction form. Conflicts
were resolved through consensus and reviewed by other investigators. We included abstracts with sufficient eval-
uation data, including methodology, the definition of outcomes, and an appropriate evaluation matrix. Studies
without any kind of validation (external validation or internal validation) were excluded. We excluded reviews,
editorials, non-human studies, letters without sufficient data.
Data extraction. We extracted the following information, if possible, from each study: authors, year of
publication, study name, test types, testing indications, analytic models, number of patients, endpoints (CAD,
AMI, stroke, heart failure, and cardiac arrhythmias), and performance measures ((AUC, sensitivity, specificity,
positive cases (the number of patients who used the AI and were positively diagnosed with the disease), negative
cases (the number of patients who used the AI and were negative with the AI test), true positives, false positives,
true negatives, and false negatives)). CAD was defined as coronary artery stenosis > 70% using angiography or
FFR-based significance. Cardiac arrhythmias included studies involving bradyarrhythmias, tachyarrhythmias,
atrial, and ventricular arrhythmias. Data extraction was conducted independently by at least two investigators
for each paper. Extracted data were compared and reconciled through consensus. In case studies which did not
report positive and negative cases, we manually calculated by standard formulae using statistics available in the
manuscripts or provided by the authors. We contacted the authors if the data of interest were not reported in the
manuscripts or abstracts. The order of contact originated with the corresponding author, followed by the first
author, and then the last author. If we were unable to contact the authors as specified above, the associated stud-
ies were excluded from the meta-analysis (but still included it in the systematic review). We also excluded manu-
scripts or abstracts without sufficient evaluation data after contacting the authors.
Quality assessment. We created the proposed guidance quality assessment of clinical ML research based
on our previous recommendation (Table 1)6. Two investigators (C.K. and H.V.) independently assessed the
quality of each ML study by using our proposed guideline to report ML in medical literature (Supplementary
Table S1). We resolved disagreements through discussion amongst the primary investigators or by involving
additional investigators to adjudicate and establish a consensus. We scored study quality as low (0–2), moderate
(2.5–5), and high quality (5.5–8) as clinical ML research.
µSeni = logit(Seni )
µspci = logit(Spc i )
Vol:.(1234567890)
www.nature.com/scientificreports/
Algorithms
Clarity of algorithms
Propose new algorithms
Select the proper algorithms
Compare alternative algorithms
Resources
Reliable database/center
Number of database/centers
Number of samples (patients/images)
Type and diversity of data
Sufficient reported data
Manuscript with sufficient supplementary information
Letter or editor, short article, abstract
Report baseline characteristics of patients
Ground truth
Comparison to expert clinicians
Comparison to validated clinical risk models
Outcome
Assessment of outcome based on standard medical taxonomy
External validation cohort
Interpretation
Report both discrimination and calibration metrics
Report one or more of the following: sensitivity, specificity, positive, negative cases, balanced accuracy
2
µSeni µSen σSen σSenSpc
∼N , 2
µspci µSpc σSenSpc σSpc
πi1 = Seni and πi0 =1- Spc i . Xij = − 21 when no disease and Xij = 21 for those with disease. And θi and αi follow
normal distribution.
We conducted subgroup analyses stratified by ML algorithms. We assessed the performances of a subgroup-
specific and statistical test of interaction among subgroups. We performed all statistical analyses using Open-
MetaAnalyst for 64-bit (Brown University), R version 3.2.3 (Metafor and Phia packages), and Stata version 16.1
(Stata Corp, College Station, Texas). The meta-analysis has been reported in accordance with the Meta-analysis
of Observational Studies in Epidemiology guidelines (MOOSE)11.
Results
Study search. The database searches between 1966 and March 15, 2019, yielded 15,025 results. 3,716 dupli-
cates were removed by algorithms. After the screening process, we selected 344 articles for full-text review. After
full text and supplementary review, we excluded 289 studies due to insufficient data to perform meta-analytic
approaches despite contacting corresponding authors. Overall, 103 cohorts (55 studies) met our inclusion crite-
ria. The disposition of studies excluded after the full-text review is shown in Fig. 1.
Study characteristics. Table 2 shows the basic characteristics of the included studies. In total, our meta-
analysis of ML and cardiovascular diseases included 103 cohorts (55 studies) with a total of 3,377,318 indi-
viduals. In total, 12 cohorts assessed cardiac arrhythmias (3,144,799 individuals), 45 cohorts are CAD-related
(117,200 individuals), 34 cohorts are stroke-related (5,577 individuals), and 12 cohorts are HF-related (109,742
individuals). The characteristics of the included studies are listed in Table 2. We performed post hoc sensitivity
analysis, excluding each study, and found no difference among the results.
ML algorithms and prediction of CAD. For the CAD, 45 cohorts reported a total of 116,227 individu-
als. 10 cohorts used CNN algorithms, 7 cohorts used SVM, 13 cohorts used boosting algorithm, 9 cohorts used
custom-built algorithms, and 2 cohorts used RF. The prediction in CAD was associated with pooled AUC of
0.88 (95% CI 0.84–0.91), sensitivity of 0.86 (95% CI 0.77–0.92), and specificity of 0.70 (95% CI 0.51–0.84), for
boosting algorithms and pooled of AUC 0.93 (95% CI 0.85–0.97), sensitivity of 0.87 (95% CI 0.74–0.94), and
specificity of 0.86 (95% CI 0.73–0.93) for custom-built algorithms (Fig. 2).
ML algorithms and prediction of stroke. For the stroke, 34 cohorts reported a total of 7,027 individu-
als. 14 cohorts used CNN algorithms, 4 cohorts used SVM, 5 cohorts used boosting algorithm, 2 cohorts used
decision tree, 2 cohorts used custom-built algorithms, and 1 cohort used random forest (RF). For prediction
of stroke, SVM algorithms had a pooled AUC of 0.92 (95% CI 0.81–0.97), sensitivity 0.57 (95% CI 0.26–0.96),
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 1. Study design. This flow chart illustrates the selection process for published reports.
and specificity 0.93 (95% CI 0.71–0.99); boosting algorithms had a pooled AUC of 0.91 (95% CI 0.81–0.96),
sensitivity 0.85 (95% CI 0.66–0.94), and specificity 0.85 (95% CI 0.67–0.94); and CNN algorithms had a pooled
AUC of 0.90 (95% CI 0.83–0.95), sensitivity of 0.80 (95% CI 0.70–0.87), and specificity of 0.91 (95% CI 0.77–
0.97) (Fig. 3).
ML algorithms and prediction of HF. For the HF, 12 cohorts reported a total of 51,612 individuals. 3
cohorts used CNN algorithms, 4 cohorts used logistic regression, 2 cohorts used boosting algorithm, 1 cohort
used SVM, 1 cohort used in-house algorithm, and 1 cohort used RF. We could not perform analyses because we
had too few studies (≤ 5) for each model.
ML algorithms and prediction of cardiac arrhythmias. For the cardiac arrhythmias, 12 cohorts
reported a total of 3,204,837 individuals. 2 cohorts used CNN algorithms, 2 cohorts used logistic regression,
3 cohorts used SVM, 1 cohort used k-NN algorithm, and 4 cohorts used RF. We could not perform analyses
because we had too few studies (≤ 5) for each model.
Discussion
To the best of our knowledge, this is the first and largest novel meta-analytic approach in ML research to date,
which drew from an extensive number of studies that included over one million participants, reporting ML
algorithms prediction in cardiovascular diseases. Risk assessment is crucial for the reduction of the world-
wide burden of CVD. Traditional prediction models, such as the Framingham risk s core12, the PCE m odel13,
14 15
SCORE , and QRISK have been derived based on multiple predictive factors. These prediction models have
been implemented in guidelines; specifically, the 2010 American College of Cardiology/American Heart Associa-
tion (ACC/AHA) guideline16 recommended the Framingham Risk Score, the United Kingdom National Institute
for Health and Care Excellence (NICE) guidelines recommend the QRISK3 s core17, and the 2016 European
Society of Cardiology (ESC) guidelines recommended the SCORE model18. These traditional CVD risk scores
have several limitations, including variations among validation cohorts, particularly in specific populations such
as patients with rheumatoid a rthritis19,20. Under some circumstances, the Framingham score overestimates CVD
vertreatment20. In general, these risk scores encompass a limited number of predic-
risk, potentially leading to o
tors and omit several important variables. Given the limitations of the most widely accepted risk models, more
robust prediction tools are needed to more accurately predict CVD burden. Advances in computational power to
process large amounts of data has accelerated interest in ML-based risk prediction, but clinicians typically have
limited understanding of this methodology. Accordingly, we have taken a meta-analytic approach to clarify the
insights that ML modeling can provide for CVD research.
Unfortunately, we do not know how or why the authors of the analyzed studies selected the chosen algorithms
from the large array of options available. Researchers/authors may have selected potential models for their
Vol:.(1234567890)
www.nature.com/scientificreports/
Vol.:(0123456789)
www.nature.com/scientificreports/
Vol:.(1234567890)
www.nature.com/scientificreports/
Table 2. Characteristics of the included studies. SVM support vector machine, RF random forest, CNN
convolutional neural network, kNN k-nearest neighbors, PCA principal component analysis, GBRT gradient
boosted regression trees, MLP multilayer perceptron, HER electronic health record, HF heart failure, HFpEF
heart failure with preserved ejection fraction, ADHF acute decompensated heart failure, SAH subarachnoid
hemorrhage, SDH subdural hematoma, ICH intracerebral hemorrhage, CAD coronary artery disease, ACS
acute coronary syndrome, CCTAcoronary computed tomography angiography, FFR fractional flow reserve,
IVUS intravascular ultrasound, ICD implantable cardioverter-defibrillator, AF atrial fibrillation, ECG
electrocardiogram.
databases and performed several models (e.g., running parallel, hyperparameter tuning) while only reporting the
best model, resulting in overfitting to their data. Therefore, we assume the AUC of each study is based upon the
best possible algorithm available to the associated researchers. Most importantly, pooled analyses indicate that,
in general, ML algorithms are accurate (AUC 0.8–0.9 s) in overall cardiovascular disease prediction. In subgroup
analyses of each ML algorithms, ML algorithms are accurate (AUC 0.8–0.9 s) in CAD and stroke prediction. To
date, only one other meta-analysis of the ML literature has been reported, and the underlying concept was similar
to ours. The investigators compared the diagnostic performance of various deep learning models and clinicians
based on medical imaging (2 studies pertained to cardiology)21. The investigators concluded that deep learning
algorithms were promising but identified several methodological barriers to matching clinician-level accuracy21.
Although our work suggests that boosting models and support vector machine (SVM) models are promising for
predicting CAD and stroke risk, further study comparing human expert and ML models are needed.
First, the results showed that custom-built algorithms tend to perform better than boosting algorithm for
CAD prediction in terms of AUC comparison. However, there is significant heterogeneity among custom-built
algorithms that do not disclose their details. The boosting algorithm has been increasingly utilized in mod-
ern biomedicine22,23. In order to implement in clinical practice, the essential stages of designing a model and
interpretation need to be u niform24. For implementation in clinical practice, custom-built algorithms must be
transparent and replicated in multiple studies using the same set of independent variables.
Second, the result showed that boosting algorithms and SVM provides similar pooled AUC for stroke pre-
diction. SVMs and boosting shared a common margin to address the clinical question. SVM seems to perform
better than boosting algorithms in patients with stroke perhaps due to discrete, linear data or a proper non-linear
kernel that fits the data better with improved generalization. SVM is an algorithm designed for maximizing a
particular mathematical function with respect to a given collection of data. Compared to the other ML meth-
ods, SVM is more powerful at recognizing hidden patterns in complicated clinical d atasets2,25. Both boosting
and SVM algorithms have been widely used in biomedicine and prior studies showed mixed results26–30. SVM
seems to outperform boosting in image recognition tasks28, while boosting seems to be superior in omic tasks27.
However, in subgroup analysis, using research questions or types of protocols or images showed no difference
in algorithm predictions.
Third, for heart failure and cardiac arrhythmias, we could not perform meta-analytic approaches due to
the small number of studies for each model. However, based on our observation in our systematic review, SVM
seems to outperform other predictive algorithms in detecting cardiac arrhythmias, especially in one large s tudy31.
Interestingly, in HF, the results are inconclusive. One small study showed promising results from S VM32. CNN
seems to outperform others, but the results are suboptimal33. Although we assumed all reported algorithms have
optimal variables, technical heterogeneity exists in ML algorithms (e.g., number of folds for cross-validation,
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 2. ROC curves comparing different machine learning models for CAD prediction. The prediction in
CAD was associated with pooled AUC of 0.87 (95% CI 0.76–0.93) for CNN, pooled AUC of 0.88 (95% CI 0.84–
0.91) for boosting algorithms, and pooled of AUC 0.93 (95% CI 0.85–0.97) for others (custom-built algorithms).
bootstrapping techniques, how many run time [epochs], multiple parameters adjustments). In addition, optimal
cut off for AUC remained unclear in clinical practice. For example, high or low sensitivity/specificity for each test
depends on clinical judgement based on clinically correlated. In general, very high AUCs (0.95 or higher) are
recommended, and it is known that AUC 0.50 is not able to distinguish between true and false. In some fields such
as applied p sychology34, with several influential variables, AUC values of 0.70 and higher would be considered
strong effects. Moreover, standard practice for ML practitioners recommended reporting certain measures (e.g.,
AUC, c-statistics) without optimal sensitivity and specificity or model calibration, while interpretation in clinical
practice is challenging. For example, the difference in BNP cut off for HF patients could result in a difference in
volume management between diuresis and IV fluid in pneumonia with septic shock.
Compared to conventional risk scores, most ML models shared a common set of independent demographic
variables (e.g., age, sex, smoking status) and include laboratory values. Although those variables are not well-
validated individually in clinical studies, they may add predictive value in certain circumstances. Head-to-head
studies comparing ML algorithms and conventional risk models are needed. If these studies demonstrate an
advantage of ML-based prediction, the optimal algorithms could be implemented through electronic health
records (EHR) to facilitate application in clinical practice. The EHR implementation is well poised for ML based
prediction since the data are readily accessible, mitigating dependency on a large number of variables, such as
Vol:.(1234567890)
www.nature.com/scientificreports/
Figure 3. ROC curves comparing different machine learning models for stroke prediction. The prediction
in stroke was associated with pooled AUC of 0.90 (95% CI 0.83–0.95) for CNN, pooled AUC of 0.92 (95% CI
0.81–0.97) for SVM algorithms, and pooled AUC of 0.91 (95% CI 0.81–0.96) for boosting algorithms.
discrete laboratory values. While it may be difficult for physicians in resource-constrained practice settings to
access the input data necessary for ML algorithms, it is readily implemented in more highly developed clinical
environments.
To this end, the selection of ML algorithm should base on the research question and the structure of the
dataset (how large the population is, how many cases exist, how balanced the dataset is, how many available
variables there are, whether the data is longitudinal or not, if the clinical outcome is binary or time to event,
etc.) For example, CNN is particularly powerful in dealing with image data, while SVM can reduce the high
dimensionality of the dataset if the kernel is correctly chosen. While when the sample size is not large enough,
deep learning methods will likely overfit the data. Most importantly, this study’s intent is not to identify one
algorithm that is superior to others.
Vol.:(0123456789)
www.nature.com/scientificreports/
Limitations
Although the performance of ML-based algorithms seems satisfactory, it is far from optimal. Several methodo-
logical barriers can confound results and increase heterogeneity. First, technical parameters such as hyperpa-
rameter tuning in algorithms are usually not disclosed to the public, leading to high statistical heterogeneity.
Indeed, heterogeneity measures the difference in effect size between studies. Therefore, in the present study,
heterogeneity is inevitable as several factors can lead to this (e.g., fine-tuning models, hyperparameter selection,
epochs). It is also a not good indicator to use as, in our HSROC model, we largely controlled the heterogeneity.
Second, the data partition is also arbitrary because of no standard guidelines for utilization. In the present study,
most included studies use 80/20 or 70/30 for training and validation sets. In addition, since the sample size for
each type of CVD is small, the pooled results could potentially be biased. Third, feature selection methodolo-
gies, and techniques are arbitrary and heterogeneous. Fourth, due to the ambiguity of custom-built algorithms,
we could not classify the type of those algorithms. Fifth, studies report different evaluation matrices (e.g., some
did not report positive or negative cases, sensitivity/specificity, F-score, etc.). We did not report the confusion
matrix for this meta-analytic approach as it required aggregation of raw numbers from studies without adjusting
for difference between studies, which could result in bias. Instead, we presented pooled sensitivity and specificity
using the HSROC model. Although ML algorithms are robust, several studies did not report complete evaluation
metrics such as positive or negative cases, Beyes, bias accuracy, or analysis in the validation cohort since there
are many ways to interpret the data depending on the clinical context. Most importantly, some analyses did not
correlate with the clinical context, which made it more difficult to interpret. The efficacy of meta-analysis is to
increase the power of the study by using the same algorithms. In addition, clinical data are heterogeneous and
usually imbalanced. Most ML research did not report balanced accuracy, which could mislead the readers. Sixth,
we did not register the analysis in PROSPERO. Finally, some studies reported only the technical aspect without
clinical aspects, likely due to a lack of clinician supervision.
Conclusion
Although there are several limitations to overcome to be able to implement ML algorithms in clinical practice,
overall ML algorithms showed promising results. SVM and boosting algorithms are widely used in cardiovascular
medicine with good results. However, selecting the proper algorithms for the appropriate research questions,
comparison to human experts, validation cohorts, and reporting of all possible evaluation matrices are needed
for study interpretation in the correct clinical context. Most importantly, prospective studies comparing ML
algorithms to conventional risk models are needed. Once validated in that way, ML algorithms could be inte-
grated with electronic health record systems and applied in clinical practice, particularly in high resources areas.
References
1. Noble, W. S. Support vector machine applications in computational biology. Kernel Methods Comput. Biol. 71, 92 (2004).
2. Aruna, S. & Rajagopalan, S. A novel SVM based CSSFFS feature selection algorithm for detecting breast cancer. Int. J. Comput.
Appl. 31, 20 (2011).
3. Lakhani, P. & Sundaram, B. Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using
convolutional neural networks. Radiology 284, 574–582 (2017).
4. Yasaka, K. & Akai, H. Deep learning with convolutional neural network for differentiation of liver masses at dynamic contrast-
enhanced CT: A preliminary study. Radiology 286, 887–896 (2018).
5. Christ, P. F. et al. Automatic Liver and Lesion Segmentation in CT Using Cascaded Fully Convolutional Neural Networks and 3D
Conditional Random Fields. International Conference on Medical Image Computing and Computer-Assisted Intervention 415–423
(Springer, Berlin, 2016).
6. Krittanawong, C. et al. Deep learning for cardiovascular medicine: A practical primer. Eur. Heart J. 40, 2058–2073 (2019).
7. Krittanawong, C., Zhang, H., Wang, Z., Aydar, M. & Kitai, T. Artificial intelligence in precision cardiovascular medicine. J. Am.
Coll. Cardiol. 69, 2657–2664 (2017).
8. Krittanawong, C. et al. Future direction for using artificial intelligence to predict and manage hypertension. Curr. Hypertens. Rep.
20, 75 (2018).
9. Covidence systematic review software. Melbourne AVHIAawcoAD.
10. Rutter, C. M. & Gatsonis, C. A. A hierarchical regression approach to meta-analysis of diagnostic test accuracy evaluations. Stat.
Med. 20, 2865–2884 (2001).
11. Stroup, D. F. et al. Meta-analysis of observational studies in epidemiology: A proposal for reporting. Meta-analysis Of Observational
Studies in Epidemiology (MOOSE) group. JAMA 283, 2008–2012 (2000).
12. Wilson, P. W. et al. Prediction of coronary heart disease using risk factor categories. Circulation 97, 1837–1847 (1998).
13. Goff, D. C. Jr. et al. 2013 ACC/AHA guideline on the assessment of cardiovascular risk: A report of the American College of
Cardiology/American Heart Association Task Force on Practice Guidelines. J. Am. Coll. Cardiol. 63, 2935–2959 (2014).
14. Conroy, R. M. et al. Estimation of ten-year risk of fatal cardiovascular disease in Europe: The SCORE project. Eur. Heart J. 24,
987–1003 (2003).
15. Hippisley-Cox, J. et al. Predicting cardiovascular risk in England and Wales: Prospective derivation and validation of QRISK2.
BMJ (Clinical research ed) 336, 1475–1482 (2008).
16. Greenland, P. et al. 2010 ACCF/AHA guideline for assessment of cardiovascular risk in asymptomatic adults: A report of the
American College of Cardiology Foundation/American Heart Association Task Force on Practice Guidelines. Circulation 122,
e584-636 (2010).
17. Hippisley-Cox, J., Coupland, C. & Brindle, P. Development and validation of QRISK3 risk prediction algorithms to estimate future
risk of cardiovascular disease: Prospective cohort study. BMJ (Clinical research ed) 357, j2099 (2017).
18. Piepoli, M. F. et al. 2016 European Guidelines on cardiovascular disease prevention in clinical practice: The Sixth Joint Task Force
of the European Society of Cardiology and Other Societies on Cardiovascular Disease Prevention in Clinical Practice (constituted
Vol:.(1234567890)
www.nature.com/scientificreports/
by representatives of 10 societies and by invited experts) Developed with the special contribution of the European Association for
Cardiovascular Prevention & Rehabilitation (EACPR). Eur. Heart J. 37, 2315–2381 (2016).
19. Kremers, H. M., Crowson, C. S., Therneau, T. M., Roger, V. L. & Gabriel, S. E. High ten-year risk of cardiovascular disease in newly
diagnosed rheumatoid arthritis patients: A population-based cohort study. Arthritis Rheum. 58, 2268–2274 (2008).
20. Damen, J. A. et al. Performance of the Framingham risk models and pooled cohort equations for predicting 10-year risk of car-
diovascular disease: A systematic review and meta-analysis. BMC Med. 17, 109 (2019).
21. Liu, X. et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical
imaging: A systematic review and meta-analysis. Lancet Digit. Health 1, e271–e297 (2019).
22. Mayr, A., Binder, H., Gefeller, O. & Schmid, M. The evolution of boosting algorithms. From machine learning to statistical model-
ling. Methods Inf. Med. 53, 419–427 (2014).
23. Buhlmann, P. et al. Discussion of “the evolution of boosting algorithms” and “extending statistical boosting”. Methods Inf. Med.
53, 436–445 (2014).
24. Natekin, A. & Knoll, A. Gradient boosting machines, a tutorial. Front. Neurorobot. 7, 21–21 (2013).
25. Noble, W. S. What is a support vector machine?. Nat. Biotechnol. 24, 1565–1567 (2006).
26. Zhang H, & Gu C. Support vector machines versus Boosting.
27. Ogutu, J. O., Piepho, H. P. & Schulz-Streeck, T. A comparison of random forests, boosting and support vector machines for genomic
selection. BMC Proc. 5(Suppl 3), S11 (2011).
28. Sun, T. et al. Comparative evaluation of support vector machines for computer aided diagnosis of lung cancer in CT based on a
multi-dimensional data set. Comput. Methods Programs Biomed. 111, 519–524 (2013).
29. Huang, M.-W., Chen, C.-W., Lin, W.-C., Ke, S.-W. & Tsai, C.-F. SVM and SVM ensembles in breast cancer prediction. PLoS One
12, e0161501–e0161501 (2017).
30. Caruana R, Karampatziakis N, & Yessenalina A. An empirical evaluation of supervised learning in high dimensions. In Proceedings
of the 25th International Conference on Machine Learning: ACM, 2008, 96–103.
31. Hill, N. R. et al. Machine learning to detect and diagnose atrial fibrillation and atrial flutter (AF/F) using routine clinical data.
Value Health 21, S213 (2018).
32. Rossing, K. et al. Urinary proteomics pilot study for biomarker discovery and diagnosis in heart failure with reduced ejection
fraction. PLoS One 11, e0157167 (2016).
33. Golas, S. B. et al. A machine learning model to predict the risk of 30-day readmissions in patients with heart failure: A retrospective
analysis of electronic medical records data. BMC Med. Inform. Decis. Mak. 18, 44 (2018).
34. Rice, M. E. & Harris, G. T. Comparing effect sizes in follow-up studies: ROC Area, Cohen’s d, and r. Law Hum Behav. 29, 615–620
(2005).
Author contributions
C.K., H.H., S.B., Z.W., K.W.J., R.P., H.Z., S.K., B.N., T.K., U.B., J.L.H., W.T. had full access to all of the data in the
study and take responsibility for the integrity of the data and the accuracy of the data analysis. Study concept and
design: C.K., H.H., K.W.J., Z.W. Acquisition of data: C.K., H.H., R.P., H.J., T.K. Analysis and interpretation of
data: B.N., Z.W. Drafting of the manuscript: C.K., H.H., S.B., U.B., J.L.H., T.W. Critical revision of the manuscript
for important intellectual content: T.W., Z.W. Study supervision: C.K., T.W.
Funding
There was no funding for this work.
Competing interests
The authors declare no competing interests.
Additional information
Supplementary information is available for this paper at https://doi.org/10.1038/s41598-020-72685-1.
Correspondence and requests for materials should be addressed to C.K.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Vol.:(0123456789)