Healthcare 09 00971 v2
Healthcare 09 00971 v2
Healthcare 09 00971 v2
Article
Multimodal Early Alzheimer’s Detection, a Genetic Algorithm
Approach with Support Vector Machines
Ana G. Sánchez-Reyna 1,† , José M. Celaya-Padilla 1,† , Carlos E. Galván-Tejada 1 , Huizilopoztli Luna-García 1 ,
Hamurabi Gamboa-Rosales 1 , Andres Ramirez-Morales 2 , Jorge I. Galván-Tejada 1, *
and on behalf of the Alzheimer’s Disease Neuroimaging Initiative ‡
1 Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Jardín Juárez 147,
Centro Historico, Zacatecas 98000, Mexico; ing.agsreyna19@gmail.com (A.G.S.-R.);
jose.celaya@uaz.edu.mx (J.M.C.-P.); ericgalvan@uaz.edu.mx (C.E.G.-T.); hlugar@uaz.edu.mx (H.L.-G.);
hamurabigr@uaz.edu.mx (H.G.-R.)
2 Department of Physics, Kyungpook National University, 80 Daehak-ro, Daegu 41566, Korea;
andres@knu.ac.kr
* Correspondence: gatejo@uaz.edu.mx
† These authors contributed equally to this work.
‡ Membership of the Alzheimer’s Disease Neuroimaging Initiative is provided in the acknowledgments.
Abstract: Alzheimer’s disease (AD) is a neurodegenerative disease that mainly affects older adults.
Citation: Sánchez-Reyna, A.G.; Currently, AD is associated with certain hypometabolic biomarkers, beta-amyloid peptides, hy-
Celaya-Padilla, J.M.; Galvan-Tejada, perphosphorylated tau protein, and changes in brain morphology. Accurate diagnosis of AD, as
C.E.; Luna-García, H.; Gamboa- well as mild cognitive impairment (MCI) (prodromal stage of AD), is essential for early care of the
Rosales, H.; Ramirez-Morales, A.; disease. As a result, machine learning techniques have been used in recent years for the diagnosis
Galvan-Tejada, J.I.; Alzheimer’s of AD. In this research, we propose a novel methodology to generate a multivariate model that
Disease Neuroimaging Initiative. combines different types of features for the detection of AD. In order to obtain a robust biomarker,
Multimodal Early Alzheimer’s ADNI baseline data, clinical and neuropsychological assessments (1024 features) of 106 patients
Detection, a Genetic Algorithm
were used. The data were normalized, and a genetic algorithm was implemented for the selection of
Approach with Support Vector
the most significant features. Subsequently, for the development and validation of the multivariate
Machines. Healthcare 2021, 9, 971.
classification model, a support vector machine model was created, and a five-fold cross-validation
https://doi.org/10.3390/
with an AUC of 87.63% was used to measure model performance. Lastly, an independent blind test
healthcare9080971
of our final model, using 20 patients not considered during the model construction, yielded an AUC
Academic Editors: Yimin Zhou and of 100%.
Zuguo Chen
Keywords: Alzheimer’s disease; support vector machine; genetic algorithm
Received: 1 July 2021
Accepted: 26 July 2021
Published: 31 July 2021
1. Introduction
Publisher’s Note: MDPI stays neutral Alzheimer’s disease is one of the most common neurodegenerative diseases, mainly
with regard to jurisdictional claims in
affecting older adults. According to the World Health Organization [1] and Alzheimer’s
published maps and institutional affil-
Disease International [2], in 2018, dementia affected approximately 50 million people
iations.
worldwide, with an estimate of 75 million by 2030 and approximately 150 million by 2050.
Moreover, there is a co-occurrence of AD with several chronic diseases, such as diabetes
mellitus, which aggravate the treatment and outcome [3].
Although Alzheimer’s disease has no cure, there are pharmacological treatments to
Copyright: © 2021 by the authors. control the symptoms. Diagnosing AD in the mild stages allows the use of treatments that
Licensee MDPI, Basel, Switzerland. might delay the progression of the disease. Late AD detection, however, may lower the
This article is an open access article effectiveness of a given treatment. Hence, early detection is imperative for maximum effi-
distributed under the terms and
ciency [4]. There are efforts devoted to study Alzheimer’s disease, such as the Alzheimer’s
conditions of the Creative Commons
Disease Neuroimaging Initiative (ADNI) [5], which has documented a database of medical
Attribution (CC BY) license (https://
images, data from biological markers (biomarkers), and clinical and neuropsychological
creativecommons.org/licenses/by/
assessments of patients since 2004; these data are publicly available for scientific research.
4.0/).
Table 1. Examples of clinical and neuropsychological assessments and laboratory analysis for the
diagnosis of AD considered in this work.
3. Methodology
The proposed methodology for this study consists of six stages, as shown in Figure 1.
In the first stage, the used datasets are described (Figure 1A). In the second stage, the dataset
of interest is created by selecting the subjects according to a given inclusion criteria
(Figure 1B). In the third stage, data preprocessing is applied, and verification and treat-
ment of the empty fields and data transformation are performed (Figure 1C). In the fourth
stage, feature selection is implemented by means of a genetic algorithm (Figure 1D). Then,
a representative set of biomarkers are studied using support vector machine classifiers
(Figure 1E). Finally, a validation test is done considering different metrics (accuracy, sensi-
tivity, specificity and AUC) to determine the performance of our model (Figure 1F).
Figure 1. Flowchart of the proposed methodology. The green squares refer to the data processing
methodology, while the white squares detail the task involved in each step. (A) The different datasets
(gene indexes and clinical and neuropsychological assessments) are obtained from the ADNI database.
(B) Each dataset is analyzed and new data sets are created by selecting subjects according to the
criteria described in Table 2. (C) A preprocessing of the data is applied: handle the empty fields and
perform data transformations. (D) The use of genetic algorithms is implemented to extract the main
data features. (E) Using the main features for Alzheimer’s detection in patients, several models are
generated using the support vector machines. (F) The validation of our results is carried out using
different metrics (accuracy, sensitivity, specificity and AUC) to determine which of the models has
the best performance.
MCI and early AD. For up-to-date information, see www.adni-info.org; accessed on 7
September 2020.
Inclusion Criteria
1. Patients should have visit codes of baseline (bl), 12 months (m12), 24 months (m24) or 36 months (m36).
2. Verify and check the participant roster ID (RID) to ensure that the measurements were from the same
patient in the different datasets.
3. In case of examinations and evaluation scales, only the final score was taken, avoiding
redundant information.
4. The age of the patients should be between 53 and 95 years at the enrolment date.
5. No distinction of gender, education, ethnicity, race, marital status was performed.
6. Patients should have biological, clinical and neuropsychological assessments.
7. Patients with duplicated records were merged using the 1st non-empty record.
The resulting filtered dataset (FDS), after applying the above inclusion criteria of Table 2,
contains information corresponding to 106 patients (42 Female/64 Male), age (75.95 ± 6.02),
clinical and neuropsychological evaluations and diagnoses (CN = 36, MCI = 52, AD = 18).
Data Preprocessing
The FDS dataset, fulfilling the visit code “bl” requirement, comprises 103 observations
(42 Female/61 Male) and excludes 3 patients lacking this visit code. For the qualitative
features, a nominal scale was made; thus, the feature used for the diagnosis of the patient
(DX) remained as a binary variable. When the patient’s diagnosis is CN or MCI, a “0”
label was assigned, otherwise, if the patient’s diagnosis is AD, a “1” was assigned. Once
the dataset was composed of only numerical variables, a filter was performed to drop
features missing more than 6.8% of the values. Thus, the final dataset size consists of
103 observations and 927 features. From this point, two versions of FDS were ccreated
under the following criteria:
• Dataset 1 (D1): The missing values in a given feature were substituted with the mean
value of this feature to complete the records (103 observations and 927 features).
• Dataset 2 (D2): From the above dataset D1, the neuropsychological features are
eliminated, leaving a grand total of (103 observations and 904 features).
In D2, neuropsychological features were removed in an attempt to find new features
that would aid in the diagnosis of AD, given these neuropsychological features were found
to have a high correlation with AD during the experimentation described in Section 3.3.1.
Subsequently, the datasets were scaled to transform their features. In this case, a
z-score transformation was applied, that is, the mean and the standard deviation of each
feature are transformed to zero and one, respectively. The transformed values, zi , are
expressed as [31],
x − x̄
zi = i , (1)
σ
where xi are the raw values, x̄ is their mean and σ is their standard deviation of each feature
in the dataset.
Healthcare 2021, 9, 971 6 of 19
Finally, for both D1 and D2, 80% of the data was used for training and testing, and the
remaining (20%) was saved for an independent blind test.
20%
Train Blind
DATA
Feature Selection
Cross-validation Test
Best Model
GA
80%
Full train
Figure 2. Flowchart of the proposed methodology for the model generation and validation.
The GA was used to select the best subset of features from the FDS. The GA fitness of
the chromosomes is calculated, employing support vector machines as binary classifiers.
The support vector machine (SVM) model, introduced by Vladimir Vapnik [25], was
chosen since it is robust and could be used to solve binary classification ML problems.
The SVM model uses the theory of Structural Risk Minimization to maximize its prediction
accuracy and procures avoiding data overfitting [35]. The SVM classification is carried
out by mapping the original feature space, with kernel functions, to a hyperspace where a
hyperplane is constructed, which separates the data of one class from the other [25,34].
For this study, an SVM [36] with a radial kernel was used as a classification method
in the feature genetic search; the specific parameters of the model are shown in Table 3.
The top-50 most frequent features are obtained. This ranking is then used in the next step to
build the final model. The features appearing more frequently in this selection suggest that
they are of importance to the classification of AD patients, see Figure 3 (see Figure A1 for a
zoomed view). Subsequently, a model refinement was carried out by means of forward
selection and backwards elimination to select the most compact and accurate model (see
Table 4) [37].
Table 3. GA input parameters. The genetic selection parameters are as follows: The chromosome size
is set according to the recommendation found in Reference [34]. The number of solutions is defined
to avoid bias. The number of generations is set to allow most of the models to converge (see Figure 4
or Figure A2 for a zoomed view). The goal fitness is defined to obtain a minimum performance
required. The SVM hyper-parameters are as follows: The cost C is set to control the trade-off between
decision and classification error and to avoid overfitting. A small γ value restricts the curvature of
the decision boundary. A radial basis function is selected as the SVM kernel since it yields good
out-of-box performance [38].
Parameter Value
Genetic selection Classifier SVM
Chromosome size 5
Max solutions 300
Max generations 200
Goal fitness 0.9
SVM Cost C 1
Gamma γ 0.2
Kernel Radial
Figure 3. Gene frequency and rank in the models determined by implementing GA using the
parameters in Table 3 for the selection of the top features in the dataset. (A) Gene frequency shows
the number of times that a feature has been present in the models. (B) Gene rank shows the stability
and frequency of each feature within the models, ordered by rank. For a zoomed view see Figure A1.
Healthcare 2021, 9, 971 8 of 19
Table 4. Most important features for classification of patients with Alzheimer’s obtained through the GA.
Multivariate
Dataset Version Model Name Final Model Length Features
Model Type
D1 ADvsMCI/CN-m1 SVM 2 MMSE, CDRSB
CDGLOBAL,
D2 ADvsMCI/CN-m2 SVM 2
CDRSB
The forward selection algorithm creates models by adding one feature at a time and
keeps this feature in the model only if it contributes to the overall model accuracy. The for-
ward selection process generates models that allow obtaining a high level of classification
accuracy; however, this process can add a large number of features, which could overfit the
data. To avoid the latter, a backwards elimination process was carried out; in this process,
one feature is removed one at a time if the performance does not drop considerably. On the
other hand, if the performance of the model decreases, the feature is kept. Figure 5 (see
Figure A3 for a zoomed view) shows the performance of the models obtained during the
forward selection process; for more details, please refer to the results in Section 4.
Figure 4. Evolution of the maximum fitness score across generations. The horizontal axis represents
a given generation, whilst the vertical axis represents the fitness score. The average fitness, plotted
with a blue solid line, considers all models. The average unfinished fitness, plotted with a cyan solid
line, considers all searches that failed for a given generation and represents the average worst case
expectation. The established GA goal fitness is plotted with the red dotted line. For a zoomed view
see Figure A2.
Figure 5. Performance of the most compact and accurate models after using the forward selection
methodology. The horizontal axis represents the features ordered by rank. The vertical axis shows
the classification accuracy. For a zoomed view see Figure A3.
Healthcare 2021, 9, 971 9 of 19
AD vs. MCI/CN
Process Metrics Average Error
Training AUC 0.9079 0.0437
Specificity 0.9882 0.0156
Sensitivity 0.8276 0.0890
Accuracy 0.9631 0.0185
Testing AUC 0.8763 0.1024
Specificity 0.9811 0.0307
Sensitivity 0.7715 0.1957
Accuracy 0.9433 0.0444
Tp
Sensitivity = (2)
Tp + Fn
Tn
Speci f icity = (3)
Tn + Fp
Tp + Tn
Accuracy(1 − Error ) = (4)
Tp + Tn + Fp + Fn
where
Tp = True positive, number of subjects with dementia correctly classified.
Fp = False positive, number of healthy subjects incorrectly classified.
Tn = True negative, number of healthy subjects correctly classified.
Healthcare 2021, 9, 971 10 of 19
4. Results
The obtained models and the classification metrics are presented in Tables 4–6. It is
observed that the ADvsMCI/CN-m1 model and the ADvsMCI/CN-m2 model performed
equally in the blind test. This test reproduces the conditions in a real-life scenario to
diagnose AD in new unseen patients. Consequently, to choose the best model, an addi-
tional comparison of the length of the models, their features and the method of calculat-
ing their scores (MMSE, CDRSB, CDGLOBAL), was performed: The ADvsMCI/CN-m1
model contains only two features, MMSE and CDRSB, and has scores that are easier to
calculate than the CDGLOBAL assessment. Furthermore, in clinical and research areas,
the MMSE and CDRSB are more widely used to stage the severity of dementia. Therefore,
the ADvsMCI/CN-m1 model was established as the best performing model for classifying
AD patients.
Figures 3–5 (see Appendix A for a zoomed view), show the results obtained from the
application of the GA considering the GA parameters in Table 3. The selected top features,
for the development of the most representative model, are found by this GA configuration.
Figure 3 (see Figure A1 for a zoomed view) shows the results of the feature occurrences
in the models. The horizontal axis in Figure 3A shows the features. The left-vertical axis
shows the gene frequency, that is, the number of times a feature has been present in the
models. The right-vertical axis shows the corresponding percentage in relation to the total
number of models. Figure 3B shows the GA outcome rank stabilization, and this graph
shows the frequency (vertical axis) of the best features found by the GA algorithm in a
rank-descent fashion (horizontal axis), where the solid colors represent stable features
that always aid in the classification. For a zoomed view of the figures, please refer to
Appendix A and feature selection in Section 3.3.1 for more details. The inclusion procedure,
applied to the FDS dataset, included 103 patients and 927 features. The feature selection
was implemented using a GA, which evolved a total of 200 generations, and it was repeated
300 times. Figure 3 shows the stability ranking of the first 50 features found through this
GA. The features are ordered from the most to the least frequent appearance.
Figure 4 (see Figure A2 for a zoomed view) shows the fitness of the evolved models,
where the blue line represents the mean fitness considering all models and the red line
represents the generation in which the average fitness reaches the goal fitness. Analyzing
this figure, it was determined that the GA parameters in (Table 3) are appropriate, since
the number of generations needed to find an optimal model is less than 50 generations
on average.
With the ranked features (Figure 3), a forward selection procedure was used to create
a representative model to classify AD vs. MCI/CN. Figure 5 (see Figure A3 for a zoomed
view) demonstrates how the performance increased as features were added; the model was
then reduced by a backward elimination process to select the most compact model with
the highest classification accuracy and the lowest number of features.
Subsequently, multivariate SVM classification models were created with a linear
kernel, using the features obtained from the feature selection process by the GA, and refined
by forward selection and backward elimination (see Table 4). The final SVM models have
only two features each. To evaluate their performance and choose the most optimal model,
they were subjected to cross-validation and blind tests.
The models in Table 4 were subjected to a five-fold cross-validation. Eighty percent of
the FDS data were used for this process, which was separated to train and test for each of
the models. The results obtained from training and testing the five-fold cross-validation
Healthcare 2021, 9, 971 11 of 19
of the ADvsMCI/CN-m1 model are shown in Table 5. This table reports the mean of the
classification metrics for the five folds and the error that refers to the standard deviation of
the obtained results.
For measuring the performance of the model in a new environment, the model was
trained using the whole training dataset (80%) and subsequently validated by its perfor-
mance on the blind test dataset (20%). The results of this blind test validation are presented
in Table 6.
According to the results obtained in the blind test (Table 6), each of the classification
metrics used to measure the performance of the models had a value of 1. To validate these
results, the data were plotted using only the two features of each model to observe the
correlation of the data (see Figure 6). The plots show that the data are linearly separable.
This suggests that the use of a linear kernel in the SVM models for this study is appropriate.
Figure 6. Correlation of the features of each model during training (80% training data) and blind
test (20% blind subset). (A) Shows the correlation of ADvsMCI/CN-m1 model features in training;
(B) shows the correlation of ADvsMCI/CN-m1 model features in the blind test; (C) shows the correla-
tion of ADvsMCI/CN-m2 model features in training; (D) shows the correlation of ADvsMCI/CN-m2
model features in the blind test. In these plots, the model support vectors are represented with “X”;
the points represented with “O” are the remaining data. The red color classifies the data where
Alzheimer’s disease is present, while the black color classifies MCI/CN.
5. Discussion
The proposed methodology demonstrates the effectiveness of using genetic algorithms
and support vector machines systems for the classification of AD vs MCI/CN using multi-
source information.
Healthcare 2021, 9, 971 12 of 19
The methodology combined data (gene indexes and clinical and neuropsychological
assessments) from the ADNI1 study in its baseline stage of 103 patients.
Subsequently, in the normalization stage, the features were scaled (z-score transforma-
tion) for use in patient classification. Using the features, the genetic algorithms generated
200 generations for 300 solutions in order to find the best performing multivariate model.
As the models evolved, the average accuracy was plotted as depicted in Figure 4, the models
reached their best performance within the first fifteen generations. Hence, 200 generations
were defined as an optimal parameter, since no more generations were needed. The final
model was refined using forward selection and backward elimination. The SVM models
were constructed with the features obtained in Table 4.
The performance of the final model was evaluated using a cross-validation and a blind
test to simulate a real-world scenario. The cross-validation model was trained and tested
using 80% of the FDS dataset (see Table 4), while for the blind test, the model was trained
using 80% of the FDS dataset and validated using the remaining 20% of the unseen data.
The final SVM model to classify AD vs. MCI / CN was ADvsMCI/CN-m1, which obtained
a sensitivity of 100% and a specificity of 100% in the blind test. The value of both validation
metrics suggests that the model is robust. From over 900 features, the ADvsMCI/CN-m1
model included the following two features: MMSE and CDRSB.
The features included in the final model have been previously used as individual
diagnostic features, such as the MMSE proposed by [18] and the CDRSB proposed by [24].
Our model combines individual prediction performance into a multivariate model capable
of improving early diagnosis of AD. It was also observed that the CDRSB feature, which
represents one of the scores of the CDR assessment, appears in both models, proving to be
an important assessment in the classification of patients with some degree of dementia and
healthy patients.
The model proposed by Zhang et al. [6] has a good performance for the classification
of AD vs. CN patients. Nevertheless, in this model, most of the used features are extracted
from medical images of patients (MRI and PET). The proposed model obtained in the
present work avoids the use of features from medical images and obtains a performance
as good as the one proposed by Zhang et al.; the latter empowers the proposed model, in
this paper, with the advantage to be available in places where the access to medical images
studies is limited.
Additionally, the proposed models in this study avoid the use of features obtained
from laboratory tests to diagnose/classify patients between CN and MCI (or AD). This
leads to a natural reduction of the required features. Our models show similar performance
to the models proposed by Hassan et al. [26] and Stamate et al. [28], where the number of
features is higher than fifteen. Using fewer features could be advantageous for patients
who are vulnerable to laboratory tests or biopsies. It is hoped that the models proposed in
this study are a viable alternative for this type of patient.
6. Conclusions
The proposed methodology in this study selects the most relevant features of AD data
(gene indexes and clinical and neuropsychological assessment) through the use of genetic
algorithms. These features were used to generate supervised classification algorithms
with an SVM architecture. The efficiency of the generated models was evaluated by a
cross-validation and a blind test, selecting the model with the highest sensitivity, specificity,
and whose features exhibited a good performance during the blind test, for early detection
of AD, between subjects with AD and MCI or CN subjects.
The novelty of this study is that it uses only non-imaging biomarkers, and yet a
similar performance to those derived from medical images is reached. The obtained models
integrated features that were previously individually validated by the research community.
Therefore, the proposed multivariate study combines individual predictions into a more
robust biomarker to detect early Alzheimer’s disease.
Healthcare 2021, 9, 971 13 of 19
7. Future Work
For future work we, propose to combine features extracted directly from MRI and use
them with the biomarkers obtained in this study to predict the likelihood of a CN patient
evolving into an AD patient. We will also investigate the possibility of replacing those
features that come from CSF analyses and blood-based metabolomics tests (since these
analyses are considered invasive techniques) with features obtained from MRI and develop
more robust ML models for the classification of patients with AD.
Author Contributions: A.G.S.-R. and J.M.C.-P. performed the study. A.G.S.-R. and J.M.C.-P. per-
formed the study design and data analysis. A.G.S.-R., J.M.C.-P., C.E.G.-T., J.I.G.-T., A.R.-M., H.G.-R.
and H.L.-G. contributed to materials and methods used in this study. J.I.G.-T., C.E.G.-T., J.M.C.-P.
and A.R.-M. performed statistical analysis with critical feedback to the authors. H.L.-G. and A.R.-M.
contributed with critical feedback on the methodology and manuscript writing. A.R.-M., H.G.-R.,
J.I.G.-T. and C.E.G.-T. provided technical feedback from the results. All authors interpreted findings
from the analysis and drafted the manuscript. All authors have read and agreed to the published
version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The data presented in this study are available on request from corre-
sponding author.
Acknowledgments: Data collection and sharing for this study was funded by the Alzheimer’s Dis-
ease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U01 AG024904) and DOD
ADNI (Department of Defense award number W81XWH-12-2-0012). ADNI is funded by the National
Institute on Aging, the National Institute of Biomedical Imaging and Bioengineering, and through
generous contributions from the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug
Discovery Foundation; Araclon Biotech; BioClinica, Inc.; Biogen; Bristol-Myers Squibb Company;
CereSpir, Inc.; Cogstate; Eisai Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F.
Hoffmann-La Roche Ltd and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO
Ltd.; Janssen Alzheimer Immunotherapy Research and Development, LLC.; Johnson & Johnson
Pharmaceutical Research and Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso
Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals
Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition
Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI
clinical sites in Canada. Private sector contributions are facilitated by the Foundation for the Na-
tional Institutes of Health (www.fnih.org; accessed on 7 August 2020). The grantee organization is
the Northern California Institute for Research and Education, and the study is coordinated by the
Alzheimer’s Therapeutic Research Institute at the University of Southern California. ADNI data are
disseminated by the Laboratory for Neuro Imaging at the University of Southern California. Andres
Ramirez-Morales acknowledges support from the National Research Foundation (NRF) of Korea,
Grants 2018R1A6A1A06024970, 2019R1I1A3A01058933, 2020R1I1A1A01066423. (Data used in prepa-
ration of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI)
database (adni.loni.usc.edu; accessed on 7 September 2020). As such, the investigators within the
ADNI contributed to the design and implementation of ADNI and/or provided data but did not par-
ticipate in analysis or writing of this report. A complete listing of ADNI investigators can be found at:
http://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf;
(accessed on 7 September 2020).
Conflicts of Interest: The authors declare no conflict of interest.
Healthcare 2021, 9, 971 14 of 19
Abbreviations
The following abbreviations are used in this manuscript:
AD Alzheimer’s disease
ADAS Alzheimer’s Disease Assessment Scale
ADAS-Cog 11 ADAS-Cognitive Subscale with 11 items
ADAS-Cog 13 ADAS-Cognitive Subscale with 13 items
ADNI Alzheimer’s Disease Neuroimaging Initiative
AI Artificial Intelligence
AUC Area Under the Receiver Operating Characteristic Curve
bl Baseline
CDGLOBAL Clinical Dementia Rating Global Score
CDR Clinical Dementia Rating
CDRSB Clinical Dementia Rating sum-of-boxes score
CN Cognitive Normal
CSF Cerebrospinal Fluid
D1 Dataset 1
D2 Dataset 2
Fn False Negative
Fp False Positive
FAQ Functional Activities Questionnaire
FDS Filtered Dataset
GA GALGO Genetic Algorithm
GDS Geriatric Depression Scale
m12 12 Months
m24 24 Months
m36 36 Months
MCI Mild Cognitive Impairment
ML Machine Learning
MMSE Mini-Mental State Examination
MRI Magnetic Resonance Imaging
PET Positron Emission Tomography
RID Participant Roster ID
SVM Support Vector Machine
Tn True Negative
Tp True Positive
Healthcare 2021, 9, 971 15 of 19
Appendix A
Figure A1. Gene frequency and rank in the models determined by implementing GA using the
parameters in Table 3 for the selection of the top features in the dataset. (A) Gene frequency shows
the number of times that a feature has been present in the models. (B) Gene rank shows the stability
and frequency of each feature within the models, ordered by rank.
Healthcare 2021, 9, 971 16 of 19
Figure A2. Evolution of the maximum fitness score across generations. The horizontal axis represents
a given generation, whilst the vertical axis represents the fitness score. The average fitness, plotted
with a blue solid line, considers all models. The average unfinished fitness, plotted with a cyan solid
line, considers all searches that failed for a given generation and represents the average worst-case
expectation. The established GA goal fitness is plotted with the red dotted line.
Healthcare 2021, 9, 971 17 of 19
Figure A3. Performance of the most compact and accurate models after using the forward selection
methodology. The horizontal axis represents the features ordered by rank. The vertical axis shows
the classification accuracy.
Healthcare 2021, 9, 971 18 of 19
References
1. World Health Organization. Global Action Plan on the Public Health Response to Dementia 2017–2025; World Health Organization:
Geneva, Switzerland, 2017
2. Patterson, C. World Alzheimer Report 2018—The State of the Art of Dementia Research: New Frontiers; Technical Report; Alzheimers
Disease International (ADI): London, UK, 2018.
3. Surguchov, A. Caveolin: A new link between diabetes and ad. Cell. Mol. Neurobiol. 2020, 1–8. [CrossRef]
4. Frozza, R.L.; Lourenco, M.V.; De Felice, F.G. Challenges for Alzheimer’s disease therapy: Insights from novel mechanisms beyond
memory defects. Front. Neurosci. 2018, 12, 37. [CrossRef]
5. ADNI | Alzheimer’s Disease Neuroimaging Initiative. 2003. Available online: http://adni.loni.usc.edu/ (accessed on
7 September 2020).
6. Zhang, D.; Wang, Y.; Zhou, L.; Yuan, H.; Shen, D.; Alzheimer’s Disease Neuroimaging Initiative. Multimodal classification of
Alzheimer’s disease and mild cognitive impairment. Neuroimage 2011, 55, 856–867. [CrossRef]
7. Falahati, F.; Westman, E.; Simmons, A. Multivariate data analysis and machine learning in Alzheimer’s disease with a focus on
structural magnetic resonance imaging. J. Alzheimer’s Dis. 2014, 41, 685–708. [CrossRef]
8. Varma, V.R.; Oommen, A.M.; Varma, S.; Casanova, R.; An, Y.; Andrews, R.M.; O’Brien, R.; Pletnikova, O.; Troncoso, J.C.; Toledo,
J.; et al. Brain and blood metabolite signatures of pathology and progression in Alzheimer disease: A targeted metabolomics
study. PLoS Med. 2018, 15, e1002482. [CrossRef]
9. Liu, L.; Zhao, S.; Chen, H.; Wang, A. A new machine learning method for identifying Alzheimer’s disease. Simul. Model. Pract.
Theory 2020, 99, 102023. [CrossRef]
10. Grassi, M.; Rouleaux, N.; Caldirola, D.; Loewenstein, D.; Schruers, K.; Perna, G.; Dumontier, M.; Alzheimer’s Disease Neuroimag-
ing Initiative. A novel ensemble-based machine learning algorithm to predict the conversion from mild cognitive impairment to
Alzheimer’s disease using socio-demographic characteristics, clinical information, and neuropsychological measures. Front.
Neurol. 2019, 10, 756. [CrossRef]
11. Pozueta, A.; Rodríguez-Rodríguez, E.; Vazquez-Higuera, J.L.; Mateo, I.; Sánchez-Juan, P.; González-Perez, S.; Berciano, J.;
Combarros, O. Detection of early Alzheimer’s disease in MCI patients by the combination of MMSE and an episodic memory
test. BMC Neurol. 2011, 11, 1–5. [CrossRef]
12. Bondi, M.W.; Edmonds, E.C.; Jak, A.J.; Clark, L.R.; Delano-Wood, L.; McDonald, C.R.; Nation, D.A.; Libon, D.J.; Au, R.; Galasko,
D.; et al. Neuropsychological criteria for mild cognitive impairment improves diagnostic precision, biomarker associations, and
progression rates. J. Alzheimer’s Dis. 2014, 42, 275–289. [CrossRef]
13. Cao, C.; Liu, F.; Tan, H.; Song, D.; Shu, W.; Li, W.; Zhou, Y.; Bo, X.; Xie, Z. Deep learning and its applications in biomedicine.
Genom. Proteom. Bioinform. 2018, 16, 17–32. [CrossRef]
14. Ting, F.F.; Tan, Y.J.; Sim, K.S. Convolutional neural network improvement for breast cancer classification. Expert Syst. Appl. 2019,
120, 103–115. [CrossRef]
15. Uddin, S.; Khan, A.; Hossain, M.E.; Moni, M.A. Comparing different supervised machine learning algorithms for disease
prediction. BMC Med. Inform. Decis. Mak. 2019, 19, 1–16. [CrossRef]
16. Mitchell, M. An Introduction to Genetic Algorithms; MIT Press: Cambridge, MA, USA, 1998
17. Nalepa, J.; Kawulok, M. Selecting training sets for support vector machines: A review. Artif. Intell. Rev. 2019, 52, 857–900.
[CrossRef]
18. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. Mini-mental state: A practical method for grading the cognitive state of patients for
the clinician. J. Psychiatry Res. 1975, 12, 189–198. [CrossRef]
19. Rosen, W.G.; Mohs, R.C.; Davis, K.L. A new rating scale for Alzheimer’s disease. Am. J. Psychiatry 1984. [CrossRef]
20. Mohs, R.C.; Knopman, D.; Petersen, R.C.; Ferris, S.H.; Ernesto, C.; Grundman, M.; Sano, M.; Bieliauskas, L.; Geldmacher, D.; Clark,
C.; et al. Development of cognitive instruments for use in clinical trials of antidementia drugs: Additions to the Alzheimer’s
Disease Assessment Scale that broaden its scope. Alzheimer Dis. Assoc. Disord. 1997, 11, S13–S21. [CrossRef]
21. Yesavage, J.A. Geriatric depression scale. Psychopharmacol. Bull. 1988, 24, 709–711.
22. Pfeffer, R.I.; Kurosaki, T.T.; Harrah, C., Jr.; Chance, J.M.; Filos, S. Measurement of functional activities in older adults in the
community. J. Gerontol. 1982, 37, 323–329. [CrossRef]
23. Reisberg, B.; Ferris, S.H.; De Leon, M.; Crook, T. Global deterioration scale (GDS). Psychopharmacol. Bull. 1988, 24, 661–663.
24. Morris, J.C. The clinical dementia rating (cdr): Current version and. Young 1991, 41, 1588–1592.
25. Cortes, C.; Vapnik, V. Support vector machine. Mach. Learn. 1995, 20, 273–297. [CrossRef]
26. Hassan, S.A.; Khan, T. A machine learning model to predict the onset of alzheimer disease using potential cerebrospinal fluid
(csf) biomarkers. Int. J. Adv. Comput. Sci. Appl. 2017, 8, 124–131.
27. Quinlan, J.R. C4. 5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014.
28. Stamate, D.; Kim, M.; Proitsi, P.; Westwood, S.; Baird, A.; Nevado-Holgado, A.; Hye, A.; Bos, I.; Vos, S.J.; Vandenberghe, R.; et al. A
metabolite-based machine learning approach to diagnose Alzheimer-type dementia in blood: Results from the European Medical
Information Framework for Alzheimer disease biomarker discovery cohort. Alzheimer’s Dement. Transl. Res. Clin. Interv. 2019,
5, 933–938. [CrossRef] [PubMed]
29. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H. Xgboost: Extreme Gradient Boosting. 2015. Available online:
https://CRAN.R-project.org/package=xgboost (accessed on 7 September 2020).
Healthcare 2021, 9, 971 19 of 19
30. Shaw, L.M.; Vanderstichele, H.; Knapik-Czajka, M.; Clark, C.M.; Aisen, P.S.; Petersen, R.C.; Blennow, K.; Soares, H.; Simon, A.;
Lewczuk, P.; et al. Cerebrospinal fluid biomarker signature in Alzheimer’s disease neuroimaging initiative subjects. Ann. Neurol.
2009, 65, 403–413. [CrossRef]
31. Han, J.; Kamber, M.; Pei, J. Data mining concepts and techniques third edition. Morgan Kaufmann Ser. Data Manag. Syst. 2011,
5, 83–124.
32. Alpaydin, E. Introduction to Machine Learning; MIT Press: Cambridge, MA, USA, 2020.
33. Moreno-Torres, J.G.; Sáez, J.A.; Herrera, F. Study on the impact of partition-induced dataset shift on k-fold cross-validation. IEEE
Trans. Neural Netw. Learn. Syst. 2012, 23, 1304–1312. [CrossRef]
34. Trevino, V.; Falciani, F. GALGO: An R package for multivariate variable selection using genetic algorithms. Bioinformatics 2006,
22, 1154–1156. [CrossRef]
35. Aruna, S.; Rajagopalan, S. A novel SVM based CSSFFS feature selection algorithm for detecting breast cancer. Int. J. Comput.
Appl. 2011, 31, 1154–1156.
36. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2011, 2, 1–27.
[CrossRef]
37. Celaya-Padilla, J.M.; Galván-Tejada, C.E.; López-Monteagudo, F.E.; Alonso-González, O.; Moreno-Báez, A.; Martínez-Torteya,
A.; Galván-Tejada, J.I.; Arceo-Olague, J.G.; Luna-García, H.; Gamboa-Rosales, H. Speed bump detection using accelerometric
features: A genetic algorithm approach. Sensors 2018, 18, 443. [CrossRef]
38. Meyer, D. An Interface Libsvm Package E1071; FH Technikum Wien: Vienna, Austria, 2015.
39. Chang, Y.W.; Hsieh, C.J.; Chang, K.W.; Ringgaard, M.; Lin, C.J. Training and testing low-degree polynomial data mappings via
linear SVM. J. Mach. Learn. Res. 2010, 11, 1471–1490.
40. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997,
30, 1145–1159. [CrossRef]