Asd 6
Asd 6
Asd 6
ABSTRACT Due to the complexity of the etiology of autism spectrum disorders, the existing autism
diagnosis method is still based on scales. With the continuous development of artificial intelligence, image-
aided diagnosis of brain diseases has been widely studied and concerned. However, many doctors and
researchers still doubt the diagnosis basis of the neural network and think that the neural network belongs to
a limited interpretable black-box function approximator. They are not sure whether the neural network has
learned some interpretive image features like humans. In order to solve this problem, three new models (2D
CAM, 3D CAM and 3D Grad-CAM) are proposed for structural Magnetic Resonance Imaging (sMRI) data.
The Regions Of Interest (ROI) of subcortical tissues among models and between groups are analyzed based
on the heat maps of the three models. The experimental results show that these models mainly distinguish
the autism group and the control group according to the voxel value of these ROIs. There are significant
differences in mean voxel value and standard deviation of voxel value between the autism group and the
control group, such as in the left amygdala, optic chiasm and right hippocampus. According to medical
references, these ROIs are closely related to people’s speech, cognition and behavior. This can partly explain
why autistic patients have unusual symptoms such as speech communication disorder, stereotyped repetitive
behavior and so on. The proposed visualization models can provide a good bridge for doctors to understand
the brain features learned by the neural network. The research method of this paper may provide a new
way for doctors and researchers to find the diagnostic biomarkers of autism, which can greatly speed up the
process of modern medical diagnosis and treatment strategies, and liberate doctors from the traditional trial
and error.
INDEX TERMS Autism spectrum disorders, class activation mapping, sMRI, biomarker.
In addition, doctors usually need the information provided by layer includes multiple convolution layers, pooling layers and
the family members of autistic patients, as well as the behav- fully connected layers, which is responsible for extracting the
ioral characteristics of autistic patients observed by doctors. abstract features of the image; the output layer is also called
After all the information has been collected, the doctor can classification layer, which classifies the image according to
make the final diagnosis [5]–[7]. However, this method is the abstract features. Although deep learning models have
not objective and is hard to distinguish autism from other demonstrated outstanding capabilities in disease prediction,
diseases (such as depression, mental retardation and language in order to be applied to more advanced artificial intelligence
development). systems, such as large-scale surgery and large-scale control,
With the rapid development of computer technology, these models must be interpretable (to ensure that decisions
Computer Aided Diagnosis (CAD) technology has been are clearly explained) and unbiased (without showing pref-
developed rapidly in some medical developed countries, erence for some certain behaviors). This is the main reason
especially in the field of medical imaging. Statistical data why neural networks have not been fully applied to practical
show that CAD plays an active role in improving the diag- engineering applications in recent years. They are still limited
nostic accuracy, reducing missed diagnosis and improving interpretable black box function approximators [23], [24].
work efficiency [8], [9]. At present, CAD research is mostly The use of machine learning model for accurate diag-
limited to the segmental lesions of breast and chest [10]. The nosis of disease has been unable to meet the needs of
research on CT-based diagnosis of liver diseases and MRI- researchers. How to find out the biomarkers of diseases
based diagnosis of brain tumors is not mature [11], [12]. for diagnosis and treatment has become a research hotspot,
Therefore, CAD research of breast and pulmonary nodules for example, [25] proposed a multiview feature learning
can basically represent the highest level of CAD in medical method with multiatlas based functional connectivity net-
imaging. works to improve mild cognitive improvement diagnosis, [26]
Since autism is essentially a highly heterogeneous neu- identified the disease subtypes by analyzing, via unsuper-
rodevelopmental disorder, structural Magnetic Resonance vised and supervised machine learning, the power-envelope-
Imaging (sMRI) can be used to detect brain lesions. MRI based connectivity of signals reconstructed from high-density
is a new medical imaging technology, which has been used resting-state electroencephalography, [27] designed a latent-
in clinic since 1982. It uses static magnetic field and radio space machine-learning algorithm tailored for resting-state
frequency magnetic field to make human tissue image. In electroencephalography (EEG).
the process of imaging, it can obtain clear image with high In addition, the interpretation of deep neural network is
contrast without electron radiation or contrast agent. It can also helpful to find disease-related biomarkers. For the inter-
reflect the disorder and early pathological changes of human pretability of neural network, many experts have done a lot of
organs from the inside of human molecules. relevant research and analysis, such as deconvolution, occlu-
Using deep learning to develop CAD technology has sion, attention model [28], guided back-propagation, Class
become a trend. It has been widely used in the diagnosis Activation Mapping (CAM) and so on.
of diseases based on Magnetic Resonance Imaging (MRI) The concept of deconvolution was first proposed by Zeiler
data, such as liver tumor [13], breast cancer [14], [15], in [29]. In this paper, he mainly explained the relationship
brain tumor [16], [17], Alzheimer’s disease [18], [19] and between convolution layer and deconvolution layer. Decon-
AHDH [20]–[22]. In this paper, we use deep learning to volution is not actually the inverse operation of convolution,
classify autistic patients, and find out the corresponding diag- it is just the transposition of convolution. He processed the
nostic biomarkers according to the output of neural network, high-dimensional abstract features extracted by the trained
so as to provide objective diagnosis basis for doctors. CNN through depooling, deactivation, deconvolution and
other operations. Finally, the abstract features are expanded
II. RELATED WORK to the size of the original image, so as to observe the features
Some research statistics show that in the medical field, more learned by CNN after training. Occlusion is mainly derived
than 70% of clinical diagnosis technology needs medical from [30]. The basic idea of occlusion is that when training
image data. How to make full use of and correctly analyze a neural network for image classification, we want to know
these data to assist doctors in diagnosis and treatment has whether this model can locate the position of the main target
become a more and more popular research direction. Convo- in the image for classification, or only classify through some
lution Neural Network (CNN) in deep learning has become surrounding context. Through partial occlusion of the original
an indispensable means. It is the most common classification image, we can observe the changes of the features extracted
model, mainly composed of three parts. The first part is the by the middle layer and the final predicted value after modi-
first layer of the neural network, which is called the input fying the image input. The guided back-propagation method
layer. The second part is the second layer to the penultimate comes from [31]. Guided back-propagation firstly calculates
layer of the neural network, which is called the feature extrac- the gradient of output with respect to input through back-
tion layer. The third part is the last layer of the neural network, propagation, and then uses the ReLU function to only keep
which is called the output layer. For image classification, the regions that have positive influence on the output. Com-
the input layer is the original image; the feature extraction pared with the original back-propagation method, this method
can not only remove the noise, but also greatly improve the prevent over-fitting problem, but also retain the spatial infor-
visualization effect of neural network. mation of features. After training the neural network, there is
Although deconvolution, occlusion and guided back- a one-dimensional vector with the same dimension as the last
propagation can reflect the sensitivity of neural network to convolution channel for each class of objects. By weighting
some features or images, the biggest drawback of these meth- each piece of the last convolution layer, the interpretation
ods is that they are greatly affected by noise, the original image can be obtained. If yc is the final classification result
image area is still not so obvious, and the interpretation of neural network, then the output of CAM model can be
images of different types of images are not very different. represented by
Another kind of neural network interpretation method q u v
is represented by CAM-based models, such as CAM [32], X 1 XX k
yc = wck Aij (1)
Gradient-weighted Class Activation Mapping (Grad-CAM) Z
k i j
[33], Grad-CAM++ [34], etc. This kind of models mainly
uses the last convolution layer of CNN for mathemati- where q is the total number of the feature map Ak , wck is the
cal transformation. Compared with the above algorithm, weight of class c for feature map Ak , the width and height
the interpretation images given by these models are clearer of feature map Ak are u and v respectively, the total number
and more explanatory. of pixels is Z , and Akij is the pixel value at (i, j) of the k th
The CAM-based models have been successfully applied to feature map. In order to obtain the interpretative region of
the classification of muscular dystrophies [35], tumor diag- the specified class in the original image, the final heat map is
nosis [36], EEG signal interpretation [37] and other medical obtained as follows
research fields [38]. The biggest advantage of these models q
is that it can accurately locate the abnormal positions of c
X
LCAM = wck Ak (2)
features (such as electrical signals, cerebral cortex, patho-
k
logical images, etc.) that are difficult for human eyes to
recognize. However, most of these articles only deal with Different from the method of CAM model, Grad-CAM
two-dimensional data, and only analyze a single CAM model. does not need to change the fully connection layer of the
MRI is a kind of three-dimensional image data. Both two- model into GAP. Grad-CAM model calculates the weights of
dimensional and three-dimensional features have rich med- feature map by using gradients. If the last convolution layer
ical significance, which need to be analyzed from different of CNN is weighted by this gradient information, the ROI of
perspectives. On this basis, we change the original CAM neural network in the original image can be finally obtained.
model into three different types of neural network model. Grad-CAM calculates weights from the global average of
In addition to the above models, the innovation of this gradients, as follows
manuscript is mainly reflected in the exploration method of u v
autism diagnostic biomarkers. (1) In the application of med- 1 X X ∂yc
αkc = (3)
ical image, the subcortical tissue that CAM model focuses Z ∂Akij
i j
on is generally given by doctors or researchers. Unlike in
this paper, the analysis is based on the accurate subcortical The preliminary heat map is obtained by weighting αkc and
tissue segmentation data provided by FreeSurfer software. (2) the feature map Ak . In order to get clearer results, only the
According to the characteristics of different models focusing positive value in the feature map are used, the final heat map
on different features of the brain, this paper analyzes the is obtained as follows
similarities and differences of autism diagnostic biomarkers X q
!
between models and groups, so as to provide a more compre- c
LGrad−CAM = ReLU αk A
c k
(4)
hensive analysis method. k
and then we use (8) and (4) to get the heat map of 3D
Grad-CAM.
Compared with the traditional CAM models used in sMRI,
FIGURE 2. Flow chart of 3D CAM model for sMRI data.
most CAM models only analyze several slices of the sMRI
data. For 2D CAM, the independent models of each slice are
fused at the end of the model. For 3D CAM and 3D Grad-
CAM, this paper makes a complete three-dimensional feature
extraction for sMRI data. Therefore, the experimental results
of these three models are more convincing.
FIGURE 6. (a) The histogram of different indexes in different parts, including volume, mean voxel value, standard deviation of voxel value,
maximum voxel value and minimum voxel value. (b) the violin diagram of indexes that have significant differences between the autism group and
the control group. (c) 3D schematic of ROIs in subcortical tissue of 2D CAM.
In order to make our experimental conclusions (or diagnostic of left amygdala and optic chiasm. There are significant
biomarkers) more reliable and convincing, we analyzed the differences in the mean voxel value and the minimum voxel
experimental results between models and between groups value of left amygdala, and the mean voxel value and standard
respectively in Table 2, instead of just analyzing a certain deviation of the voxel value of optic chiasm. For the data of
group of a certain model. Therefore, we will analyze the fol- the above indexes in the autism group and the control group,
lowing five situations: (1) For 2D CAM, analyze the common we respectively fitted the corresponding probability distribu-
ROIs between autism and control group; (2) For 3D CAM, tion model, as shown in Fig. 6 (b). Specifically, as shown
analyze the common ROIs between autism and control group; in Fig. 6 (b), we can see that the mean value of the proba-
(3) For 3D Grad-CAM, analyze the common ROIs between bility distribution model of the mean voxel value of the left
autism and control group; (4) For the autism group, analyze amygdala and optic chiasm in autism group is significantly
the common ROIs among the three CAM-based models; (5) higher than that of the control group. Besides, the minimum
For the control group, analyze the common ROIs among the voxel value of the left amygdala and the standard deviation
three CAM-based models; of voxel value of optic chiasm in the autism group are more
concentrated, that is to say, the variance of the corresponding
TABLE 2. The common ROIs for each model and group.
probability distribution model is smaller than that of the
control group.
(2) For 3D CAM, the ROIs shared by autism and control
groups are in the third row and fourth column of Table 2. The
subcortical tissues are 18 (Left Amygdala), 44 (Right Inferior
Lateral Ventricle) and 53 (Right Hippocampus). As shown
in Fig. 7 (a), the dotted line indicates that the p-value between
the autism group and the control group is equal to 0.05.
For 3D CAM model, it mainly determines whether the
sample has autism by identifying the voxel values of left
amygdala and right hippocampus, and the volume of right
hippocampus. There are significant differences in the mean
voxel value and the minimum voxel value of left amygdala,
and the volume and mean voxel value of right hippocampus.
For the data of the above indexes in the autism group and the
control group, we respectively fitted the corresponding proba-
bility distribution model, as shown in Fig. 7 (b). Specifically,
as shown in Fig. 7 (b), we can see that the mean value of
(1) For 2D CAM, the ROIs shared by autism and control the probability distribution model of the mean voxel value of
groups are in the second row and fourth column of Table 2. the left amygdala and right hippocampus in autism group is
The subcortical tissues are 18 (Left Amygdala) and 85 (Optic significantly higher than that of the control group. Similarly,
Chiasm). As shown in Fig. 6 (a), the dotted line indicates that the volume of the right hippocampus in autism group is also
the p-value between the autism group and the control group higher than that of the control group. Besides, the standard
is equal to 0.05. For 2D CAM model, it mainly determines deviation of the probability distribution model of the mean
whether the sample has autism by identifying the voxel values voxel value of the right hippocampus in the autism group
FIGURE 7. (a) The histogram of different indexes in different parts, including volume, mean voxel
value, standard deviation of voxel value, maximum voxel value and minimum voxel value. (b) the
violin diagram of indexes that have significant differences between the autism group and the control
group. (c) 3D schematic of ROIs in subcortical tissue of 3D CAM.
FIGURE 8. (a) The histogram of different indexes in different parts, including volume, mean voxel value, standard deviation of voxel value, maximum
voxel value and minimum voxel value. (b) the violin diagram of indexes that have significant differences between the autism group and the control group.
(c) 3D schematic of ROIs in subcortical tissue of 3D Grad-CAM.
are more concentrated, that is to say, the variance of the of left cerebellum cortex and 4th-ventricle in autism group is
corresponding probability distribution model is smaller than significantly different than that of the control group. Besides,
that of the control group. the standard deviation of the probability distribution model
(3) For 3D Grad-CAM, the ROIs shared by autism and of standard deviation of the voxel value of left cerebellum
control groups are in the second row and fourth column of cortex and 4th-ventricle in the autism group are also very
Table 2. The subcortical tissues are 7 (Left Cerebellum White different, that is to say, the variance of the corresponding
Matter), 8 (Left Cerebellum Cortex), 15 (4th-Ventricle) and probability distribution model is bigger (or smaller) than that
46 (Right Cerebellum White Matter). of the control group. Similarly, the standard deviation of the
As shown in Fig. 8 (a), the dotted line indicates that the probability distribution model of the maximum voxel value
p-value between the autism group and the control group is of 4th-ventricle is significantly higher in autism group than
equal to 0.05. For 3D Grad-CAM model, it mainly deter- that of the control group.
mines whether the sample has autism by identifying the voxel (4) For autism group, the ROIs shared by the three
values of left cerebellum cortex and 4th-ventricle. There are CAM-based models are in the second row and fourth column
significant differences in the mean voxel value and standard of Table 2. The subcortical tissues are 7 (Left Cerebellum
deviation of the voxel value of left cerebellum cortex, and the White Matter) and 15 (4th-Ventricle). As shown in Fig. 9 (a),
mean voxel value, the maximum voxel value and standard the dotted line indicates that the p-value between the autism
deviation of the voxel value of 4th-ventricle. For the data group and the control group is equal to 0.05. The three
of the above indexes in the autism group and the control models mainly determine whether the sample has autism
group, we respectively fitted the corresponding probabil- by identifying the voxel values of 4th-ventricle. There are
ity distribution model, as shown in Fig. 8 (b). Specifically, significant differences in the mean voxel value, the maxi-
as shown in Fig. 8 (b), we can see that the mean value of mum voxel value and standard deviation of the voxel value
the probability distribution model of the mean voxel value of 4th-ventricle. For the data of the above indexes in the
FIGURE 9. (a) The histogram of different indexes in different parts, including volume, mean voxel value, standard deviation of voxel value, maximum
voxel value and minimum voxel value. (b) the violin diagram of indexes that have significant differences between the autism group and the control
group. (c) 3D schematic of ROIs in subcortical tissue between three CAM-based models for autism group.
autism group and the control group, we respectively fitted Grad-CAM models. As mentioned above, CAM-based mod-
the corresponding probability distribution model, as shown els can explain the classification basis of neural network and
in Fig. 9 (b). Specifically, as shown in Fig. 9 (b), we can see reflect the features extracted by neural network. Based on
that the mean value of the probability distribution model of this, according to the heat maps of these three CAM-based
the mean voxel value, the maximum voxel value and standard models, this paper analyzes the ROIs among models and
deviation the voxel value of 4th-ventricle in autism group between groups. The results show that these models diagnose
are all more concentrated, that is to say, the variance of the autism mainly through the voxel values of subcortical tissues,
corresponding probability distribution model is smaller than such as left amygdala, chiasma and right hippocampus. Sta-
that of the control group. tistical analysis show that there are significant differences in
(5) For control group, the ROIs shared by the three the mean or standard deviation of voxel values between the
CAM-based models are in the second row and fourth column autism group and the control group.
of Table 2. The subcortical tissues are 18 (Left Amygdala) However, there are some differences in MRI acquisition
and 85 (Optic Chiasm). The three models mainly determine methods, acquisition equipment and autism diagnosis meth-
whether the sample is normal by identifying the voxel values ods in different sites. If all data sets are used, the classification
of the above two parts. These two parts are the same as those accuracy of neural network is not high. Therefore, this paper
used in previous 2D CAM model. Therefore, the analysis and only analyzes the data set of a single site to obtain the ideal
the corresponding figure of these two parts are the same as (1) accuracy. In this way, the diagnostic biomarkers obtained
and they will not be repeated here. from the analysis of three CAM-based models are more
To sum up, according to the experimental results of the reliable. In the future research, we will study how to analyze
above five cases, most of them distinguish autism from the multi site data by dimensionality reduction and clustering
control group according to the voxel value of subcortical algorithms, so as to get more accurate diagnostic biomarkers
tissue. Although the minimum and maximum voxels in some and more generalized classification algorithm.
ROIs do show significant differences, which can be used as Besides, in the training process, our proposed models only
a reference for the diagnosis of autism. However, due to the depends on the cross entropy loss function, which leads
different data collection methods of different sites and differ- to some meaningless ROI regions in some heat maps. The
ent devices, these indicators are easy to be disturbed by noise recently proposed CAM model also add the segmentation
and may only be applicable to the data of single site. For data loss function provided by the heat map, which can show the
collected from multiple sites or different devices, the standard heat region of the model more accurately, such as Guided
deviation and mean voxel values of these ROIs are relatively Attention Information Network (GAIN) [43], Score-CAM
reliable. In addition, from the above analysis, most of the [44] and so on. In addition, we will use the multi-task learn-
ROIs are related to self-learning, communication, behavioral ing algorithm to further improve the above models. While
decision-making, and some cognitive functions. Thus, there completing the previous training tasks, we will use the sMRI
is a possibility that these brain structural abnormalities of segmentation label data provided by FreeSurfer software to
autistic patients lead to behavioral abnormalities. It can partly learn how to segment each subcortical tissue part, so that the
explain why autistic patients have unusual symptoms such models can learn more medical features.
as stereotyped repetitive behavior, speech communication
disorder and so on. ADDITIONAL INFORMATION
The open source data sets come from the website ABIDE.
V. CONCLUSION The usage agreement clearly states, ‘‘Consistent with the
In this paper, we propose 2D CAM, 3D CAM and 3D policies of the 1000 functional connectors project, data usage
Grad-CAM for sMRI data based on existing CAM and is unrestricted for non-commercial research purposes.’’
This paper only uses a part of the data sets, that is, [16] V. P. G. P. Rathi and S. Palani, ‘‘Brain tumor detection and classification
New York University Langone Medical Center. The research using deep learning classifier on MRI images,’’ Res. J. Appl. Sci., Eng.
Technol., vol. 10, no. 2, pp. 177–187, May 2015.
in this paper is completely for non-commercial purposes. All [17] N. Noreen, S. Palaniappan, A. Qayyum, I. Ahmad, M. Imran, and
the authors of this paper are very grateful for all the funding M. Shoaib, ‘‘A deep learning model based on concatenation approach for
for supporting open source data sets. Because open source the diagnosis of brain tumor,’’ IEEE Access, vol. 8, pp. 55135–55144,
2020.
data sets have received a lot of funding, they will not be [18] H. Guo and Y. Zhang, ‘‘Resting state fMRI and improved deep learning
listed here one by one. Finally, the research in this paper is algorithm for earlier detection of Alzheimer’s disease,’’ IEEE Access,
completely in line with the usage agreement of ABIDE, and vol. 8, pp. 115383–115392, 2020.
[19] S. Al-Shoukry, T. H. Rassem, and N. M. Makbol, ‘‘Alzheimer’s diseases
the detailed rules can be seen in [40]. detection by using deep learning algorithms: A mini-review,’’ IEEE Access,
vol. 8, pp. 77131–77141, 2020.
REFERENCES [20] S. Abdolmaleki and M. S. Abadeh, ‘‘Brain MR image classifica-
tion for ADHD diagnosis using deep neural networks,’’ in Proc. Int.
[1] J. G. Williams, J. P. T. Higgins, and C. E. G. Brayne, ‘‘Systematic review Conf. Mach. Vis. Image Process. (MVIP), Feb. 2020, pp. 1–5, doi: 10.
of prevalence studies of autism spectrum disorders,’’ Arch. Disease Child- 1109/MVIP49855.2020.9116877.
hood, vol. 91, no. 1, pp. 8–15, Apr. 2005. [21] L. Shao, D. Zhang, H. Du, and D. Fu, ‘‘Deep forest in ADHD data
[2] E. Simonoff, A. Pickles, T. Charman, S. Chandler, T. Loucas, and G. Baird, classification,’’ IEEE Access, vol. 7, pp. 137913–137919, 2019.
‘‘Psychiatric disorders in children with autism spectrum disorders: Preva- [22] L. Zou, J. Zheng, C. Miao, M. J. Mckeown, and Z. J. Wang, ‘‘3D CNN
lence, comorbidity, and associated factors in a population-derived sample,’’ based automatic diagnosis of attention deficit hyperactivity disorder using
J. Amer. Acad. Child Adolescent Psychiatry, vol. 47, no. 8, pp. 921–929, functional and structural MRI,’’ IEEE Access, vol. 5, pp. 23626–23636,
Aug. 2008. 2017.
[3] M. W. Kuzniewicz, S. Wi, Y. Qian, E. M. Walsh, M. A. Armstrong, and [23] J. Zhang and J. Zhang, ‘‘An analysis of CNN feature extractor based
L. A. Croen, ‘‘Prevalence and neonatal factors associated with autism on KL divergence,’’ Int. J. Image Graph., vol. 18, no. 3, Jul. 2018,
spectrum disorders in preterm infants,’’ J. Pediatrics, vol. 164, no. 1, Art. no. 1850017.
pp. 20–25, Jan. 2014. [24] I. Rio-Torto, K. Fernandes, and L. F. Teixeira, ‘‘Understanding the deci-
[4] W. M. State and N. Šestan, ‘‘Neuroscience. The emerging biology of sions of CNNs: An in-model approach,’’ Pattern Recognit. Lett., vol. 133,
autism spectrum disorders,’’ Science, vol. 337, no. 6100, pp. 1301–1303, pp. 373–380, May 2020.
Sep. 2012. [25] Y. Zhang, H. Zhang, E. Adeli, X. Chen, M. Liu, and D. Shen, ‘‘Multiview
[5] K. Gotham, A. Pickles, and C. Lord, ‘‘Standardizing ADOS scores for feature learning with multiatlas-based functional connectivity networks for
a measure of severity in autism spectrum disorders,’’ J. Autism Develop. MCI diagnosis,’’ IEEE Trans. Cybern., early access, Dec. 14, 2020, doi:
Disorders, vol. 39, no. 5, pp. 693–705, Dec. 2008. 10.1109/TCYB.2020.3016953.
[6] M. Reyes, P. Ponce, D. Grammatikou, and A. Molina, ‘‘Methodology [26] Y. Zhang et al., ‘‘Identification of psychiatric disorder subtypes from
to weight evaluation areas from autism spectrum disorder ADOS-G test functional connectivity patterns in resting-state electroencephalography,’’
with artificial neural networks and taguchi method,’’ Revista Mexicana de Nature Biomed. Eng., vol. 5, pp. 309–323, 2021.
Ingenier Biomédica, vol. 35, no. 3, pp. 223–240, Nov. 2014. [27] W. Wu et al., ‘‘An electroencephalographic signature predicts antidepres-
[7] T. Overton, C. Fielding, and R. G. de Alba, ‘‘Brief report: Exploratory sant response in major depression,’’ Nature Biotechnol., vol. 38, no. 4,
analysis of the ADOS revised algorithm: Specificity and predictive value pp. 439–447, Apr. 2020, doi: 10.1038/s41587-019-0397-3.
with hispanic children referred for autism spectrum disorders,’’ J. Autism [28] V. Mnih, N. Heess, and A. Graves, ‘‘Recurrent models of visual attention,’’
Develop. Disorders, vol. 38, no. 6, pp. 1166–1169, Nov. 2007. in Proc. Adv. Neural Inf. Process. Syst., vol. 27, 2014, pp. 2204–2212.
[8] A. R. Jamieson, M. L. Giger, K. Drukker, H. Li, Y. Yuan, and N. Bhooshan, [29] M. D. Zeiler, D. Krishnan, G. W. Taylor, and R. Fergus, ‘‘Deconvolutional
‘‘Exploring nonlinear feature space dimension reduction and data repre- networks,’’ in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recog-
sentation in breast CADx with Laplacian eigenmaps and t-SNE: Nonlinear nit., Jun. 2010, pp. 2528–2535.
dimension reduction and representation in breast CADx,’’ Med. Phys., [30] D. M. Zeiler and R. Fergus, ‘‘Visualizing and understanding convolu-
vol. 37, no. 1, pp. 339–351, Dec. 2009. tional networks,’’ in Proc. Eur. Conf. Comput. Vis., Zürich, Switzerland,
[9] S. J. Kim, W. K. Moon, S.-Y. Kim, J. M. Chang, S. M. Kim, and vol. 8689, 2014, pp. 818–833.
N. Cho, ‘‘Comparison of two software versions of a commercially avail- [31] J. T. Springenberg, A. Dosovitskiy, T. Brox, and M. Riedmiller, ‘‘Striv-
able computer-aided detection (CAD) system for detecting breast cancer,’’ ing for simplicity: The all convolutional net,’’ 2014, arXiv:1412.6806.
Acta Radiologica, vol. 51, no. 5, pp. 482–490, Jun. 2010. [Online]. Available: http://arxiv.org/abs/1412.6806
[10] F. Li, R. Engelmann, K. Doi, and H. Macmahon, ‘‘True detection versus [32] B. Zhou, A. Khosla, A. Lapedriza, A. Oliva, and A. Torralba, ‘‘Learning
‘accidental’ detection of small lung cancer by a computer-aided detection deep features for discriminative localization,’’ 2015, arXiv:1512.04150.
(CAD) program on chest radiographs,’’ J. Digit. Imag., vol. 23, no. 1, [Online]. Available: http://arxiv.org/abs/1512.04150
pp. 66–72, 2010. [33] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and
[11] D. Chyzhyk, A. Savio, and M. Graña, ‘‘Evolutionary ELM wrapper fea- D. Batra, ‘‘Grad-CAM: Visual explanations from deep networks via
ture selection for Alzheimer’s disease CAD on anatomical brain MRI,’’ gradient-based localization,’’ Int. J. Comput. Vis., vol. 128, no. 2,
Neurocomputing, vol. 128, pp. 73–80, Mar. 2014. pp. 336–359, Feb. 2020.
[12] J. Virmani, V. Vinod, N. Kalra, and N. Khandelwa, ‘‘PCA-SVM based [34] A. Chattopadhyay, A. Sarkar, P. Howlader, and N. V. Balasubrama-
CAD system for focal liver lesions using B-mode ultrasound images,’’ nian, ‘‘Grad-CAM++: Improved visual explanations for deep convo-
Defence Sci. J., vol. 63, no. 5, pp. 478–486, Oct. 2013. lutional networks,’’ in Proc. IEEE Winter Conf. Appl. Comput. Vis.
[13] E. Trivizakis, G. C. Manikis, K. Nikiforaki, K. Drevelegas, (WACV), Mar. 2017, pp. 839–847. [Online]. Available: https://arxiv.org/
M. Constantinides, A. Drevelegas, and K. Marias, ‘‘Extending 2-D abs/1710.11063
convolutional neural networks to 3-D for advancing deep learning cancer [35] J. Cai, F. Xing, A. Batra, F. Liu, G. A. Walter, K. Vandenborne, and
classification with application to MRI liver tumor differentiation,’’ IEEE L. Yang, ‘‘Texture analysis for muscular dystrophy classification in MRI
J. Biomed. Health Informat., vol. 23, no. 3, pp. 923–930, May 2019. with improved class activation mapping,’’ Pattern Recognit., vol. 86,
[14] G. Amit, R. Ben-Ari, O. Hadad, E. Monovich, N. Granot, and S. Hashoul, pp. 368–375, Feb. 2019.
‘‘Classification of breast MRI lesions using small-size training sets: Com- [36] S. Lee, J. Lee, J. Lee, C.-K. Park, and S. Yoon, ‘‘Robust tumor localiza-
parison of deep learning approaches,’’ Proc. SPIE, vol. 10134, Mar. 2017, tion with pyramid grad-CAM,’’ in Proc. Comput. Vis. Pattern Recognit.,
Art. no. 101341H. May 2018. [Online]. Available: https://arxiv.org/abs/11393
[15] S. Marrone, G. Piantadosi, R. Fusco, A. Petrillo, M. Sansone, and [37] S. Jonas, A. O. Rossetti, M. Oddo, S. Jenni, P. Favaro, and F. Zubler, ‘‘EEG-
C. Sansone, ‘‘An investigation of deep learning for lesions malignancy based outcome prediction after cardiac arrest with convolutional neural
classification in breast DCE-MRI,’’ in Proc. Int. Conf. Image Anal. Pro- networks: Performance and visualization of discriminative features,’’ Hum.
cess., Catania, Italy, 2017, pp. 479–489. Brain Mapping, vol. 40, no. 16, pp. 4606–4617, Jul. 2019.
[38] Y. Sato, Y. Takegami, T. Asamoto, Y. Ono, T. Hidetoshi, R. Goto, HUANPING LIU was born in Hubei, China,
A. Kitamura, and S. Honda, ‘‘A computer-aided diagnosis system using in 1996. He received the bachelor’s degree from
artificial intelligence for hip fractures -multi-institutional joint develop- the Hubei University of Technology, Wuhan,
ment research,’’ 2020, arXiv:2003.12443. [Online]. Available: http://arxiv. China, in 2019. He is currently pursuing the
org/abs/2003.12443 master’s degree with the School of Mechanical
[39] M. Lin, Q. Chen, and S. Yan, ‘‘Network in network,’’ 2013, Engineering. His current research interests include
arXiv:1312.4400. [Online]. Available: http://arxiv.org/abs/1312.4400 medical image processing, machine learning, and
[40] C. Cameron et al., ‘‘The neuro bureau preprocessing initiative: Open
statistical analysis of the autistic brain.
sharing of preprocessed neuroimaging data and derivatives,’’ Frontiers
Neuroinform., vol. 7, 2013.
[41] B. Alexander, W. Y. Loh, L. G. Matthews, A. L. Murray,
C. Adamson, R. Beare, J. Chen, C. E. Kelly, P. J. Anderson, L. W. Doyle,
A. J. Spittle, J. L. Y. Cheong, M. L. Seal, and D. K. Thompson, ‘‘Desikan-
killiany-tourville atlas compatible version of M-CRIB neonatal parcellated
whole brain atlas: The M-CRIB 2.0,’’ Frontiers Neurosci., vol. 13, p. 34,
Feb. 2019.
[42] O. Potvin, L. Dieumegarde, and S. Duchesne, ‘‘Freesurfer cortical norma-
tive data for adults using Desikan-Killiany-Tourville and ex vivo proto-
cols,’’ NeuroImage, vol. 156, pp. 43–46, Aug. 2017.
[43] K. Li, Z. Wu, K.-C. Peng, J. Ernst, and Y. Fu, ‘‘Tell me where to look:
Guided attention inference network,’’ in Proc. IEEE Conf. Comput. Vis. MINGCHENG ZHOU was born in Anhui, China,
Pattern Recognit. (CVPR), Feb. 2018, pp. 9215–9223. [Online]. Available: in 1995. He received the bachelor’s degree from
https://arxiv.org/abs/1802.10171v1 Chuzhou University, in 2020. He is currently
[44] H. Wang, Z. Wang, M. Du, F. Yang, Z. Zhang, S. Ding, P. Mardziel, pursuing the master’s degree with the School
and X. Hu, ‘‘Score-CAM: Score-weighted visual explanations for con- of Mechanical Engineering, Hubei University of
volutional neural network,’’ in Proc. IEEE/CVF Conf. Comput. Vis.
Technology. His current research interests include
Pattern Recognit. (CVPR), Apr. 2020, pp. 24–25. [Online]. Available:
medical image processing, machine learning, and
https://arxiv.org/abs/1910.01279
statistical analysis of the autistic brain.