Current Ophthalmology Reports (2019) 7:204–211
https://doi.org/10.1007/s40135-019-00218-9
CORNEA (T YAMAGUCHI, SECTION EDITOR)
Artificial Intelligence in Corneal Diagnosis: Where Are we?
Bernardo T. Lopes 1,2,3 & Ashkan Eliasy 1 & Renato Ambrosio Jr 2,3,4
Published online: 9 July 2019
# The Author(s) 2019
Abstract
Purpose of Review In this paper, we prospectively review some artificial intelligence (AI) techniques and models used to enhance
clinical decisions for patients with corneal diseases and conditions.
Recent Findings Cornea subspeciality was a pioneer in aggregating technology to clinical practice. It provided the ability to early
diagnose diseases and improve treatments. Currently, we face the challenge of dealing with a tremendous amount of information
from complementary multimodal imaging devices. The analysis of such data for enhancing clinical decisions is perfectly suitable
for AI. While AI models are rapidly growing in various fields, some are already available and in use by clinicians, for instance to
help in refractive surgery screening.
Summary AI models represent a boundless method for helping to deal with and avoid the overload of an extraordinary amount of
information provided by advances in complementary diagnosis. It is currently ready to use in refractive surgery screening. The
challenge is to coordinate multicentre collaborations in order to build good quality and large data collection to train and improve
AI models. AI is an instrument to upturn clinical decision power with many possible applications for ophthalmologists.
Keywords Artificial intelligence . Corneal imaging . Diagnosis . Refractive surgery
Introduction
The cornea subspecialty has continuously been a pioneer on
the use of technology to aid treatment and diagnosis in ophthalmic diseases. The first corneal surface characterisation
dates back to the early XVII when Scheiner did his first experiments observing the image reflected on calibrated glass
spheres and the cornea [1]. During the 1900s, the keratometer
and the keratoscope were developed and used separately until
we had the computational power to effectively combine the
objective and subjective approach of both methods in the
1980s [2]. This made possible the diagnosis of early forms
This article is part of the Topical Collection on Cornea
* Bernardo T. Lopes
blopes@liverpool.ac.uk
1
School of Engineering, University of Liverpool, Liverpool L69 3GH,
UK
2
Rio de Janeiro Corneal Tomography and Biomechanics Study
Group, Rio de Janeiro, Brazil
3
Federal University of São Paulo, São Paulo, Brazil
4
Federal University of the State of Rio de Janeiro, Rio de
Janeiro, Brazil
of progressive corneal diseases even before the vision was
affected. The slow advances of the last four centuries have
given way to a rapid transformation. In a few decades, several
new devices and technologies arouse. We can now evaluate
the corneal curvature and elevation of both surfaces along
with the pachymetric map [3]. It is possible to study the corneal tissue anatomy by layers, to evaluate its histology and
biomechanical properties in vivo [4–8].
All these new technologies provide us with an unprecedented amount of information, which in one hand is vital to
early disease detection, but on the other, its excess can disrupt
sensible decision-making. Information and knowledge are not
synonyms. In fact, since the 1960s, it is proposed that information overload could be a barrier to the formation of knowledge [9]. There are some strategies to deal with this information overload, but be it to keep up-to-date in the speciality or to
extract meaningful information from a complementary exam,
the use of machines seems to be the most efficient way of
doing it [10]. Artificial neural networks, deep learning, and
other machine learning (ML) techniques had become useful
tools in clinician’s arsenal to help to deliver the best quality of
care to their patients.
One of the earliest examples of its application in corneal
imaging was proposed by Maeda et al. in 1994 with the use of
a classification tree combined with a linear discriminant
Curr Ophthalmol Rep (2019) 7:204–211
function to distinguish between a keratoconus and a nonkeratoconic pattern [11]. Throughout the 1990s, some other
techniques such as neural networks have been proposed to
identify the keratoconus pattern based on corneal topography
[12–14]. But the greatest push to improve disease susceptibility detection came with the first reported cases of post-Lasik
ectasia [15–18]. Currently, refractive surgery screening is the
most prolific field for ML development in corneal disease, yet
research in other fields is growing, despite promising ML
techniques are full of dangerous pitfalls. A meticulous process
of developing the models should be followed in order to get
reliable information from it.
Overview of ML Techniques
In brief, we will discuss some of the ML techniques most
commonly used in corneal diagnosis, their benefits and limitations, and how to avoid unexceptional problems and misinterpretation inherent to it.
Artificial intelligence (AI), introduced in the 1950s and
1960s, represents algorithms that through learning and thinking ability has enabled computers to become intelligent.
Learning is defined as the ability to update parameters and
coefficients of an algorithm upon the availability of data.
Historically, machine learning is built on three fundamental
branches of symbolic learning [19], statistics [20], and neural
networks [21]. These led to the development of advanced
approaches that include pattern and statistical recognition (knearest neighbours, Bayesian classifiers, and Fisher’s linear
discriminants), symbolic learning (decision trees, logical
programmes, and decision rules) and artificial neural networks
(ANN) (deep learning, recurrent networks, and convolutional
neural networks) [22, 23].
There are various types of machine learning and they can
mainly be categorised into unsupervised (data are not labelled), supervised (data are labelled), and reinforcement
learning [24]. For both supervised and unsupervised learning,
if the data are continuous, it demands a regression analysis and
in case of discrete data, a classification method is required.
Regression analysis is a statistical method that reveals the
relationships between two or more variables. An example of
regression analysis can be the correlation of visual acuity improvement after a surgical intervention based on the preoperative status of the patient. On the other hand, an example of a
classification problem is disease diagnosis using complementary exams’ data. Reinforcement learning is a very new type of
machine learning algorithms that rewards are provided for the
actions taken by the machine. A good example is the software
packages that learn to play computer games. It has three components that are an agent (the decision maker that functions in
an environment), environment (leads to making a decision),
and actions (what agent can do based on the decision made).
205
Artificial neural networks are designed on the concept of
biological neurons. A single perceptron, that is a building
block for a neural network, can be mathematically described
as shown in Fig. 1. Features (individual measurable characteristic or property) from the input provide information about the
problem under consideration. A randomly allocated weight
will be multiplied by the input and passed into the activation
function. The activation function is a simple equation that
defines the output; for instance, it can be if the inputs are
positive, print 1 otherwise print 0. The problem will arise
when the input is zero in some cases, which would make the
output zero no matter what the activation function is. Hence,
to overcome this problem, scientist included a bias term into
each perceptron calculations.
This process can be described mathematically using Eq. 1:
n
∑ w i xi þ b
ð1Þ
i¼0
where n is the number of inputs, w is the weight of the input, x
is the input, and b is the bias term. Multiple perceptrons that
connect the input to the outputs are called middle layer or
hidden layer. Any neural network is constructed from three
layers of input, output, and hidden layer. If the number of
hidden layers is more than three, it is called deep learning.
Deep learning enables the network to gain a deeper understanding of the inputs and outputs. Hence, in a deep neural
network, the function above is applied through multiple layers
as below:
Hidden layer 1 : Z ð1Þ ¼ w0 xi þ b0
Hidden layer 2 : Z ð2Þ ¼ w1 Z 1 þ b1
Hidden layer 3 : Z ð3Þ ¼ w2 Z 2 þ b2
ð2Þ
Hidden layer n : Z ðnÞ ¼ wn−1 Z n−1 þ bn−1
where Z is the output of each layer. A neural network always
goes through a process, Fig. 2. The mathematical complexity
of this process is out of the scope of this paper and will not be
Fig. 1 A demonstration of the perceptron model that shows two inputs
and a bias term that multiplied by random weights and fed into an
activation function which numerically predicted the output
206
Curr Ophthalmol Rep (2019) 7:204–211
Fig. 2 Process for building a machine learning software starts with data
acquisition and cleaning. Then, the dataset should be split into training
and test set. This will be followed by training the model, evaluating the
results, and adjusting the parameters through an iterative process. Finally,
the model can be deployed to utilise new inputs for predictions
discussed here. However, readers are encouraged to read
about it at [25].
There are a few important points regarding the training of
any ANN that enables supervision of AI-related projects. This
include the following:
provide any explanation to the clinician of why a decision
has been made. Hence, abnormalities may be misd
iagnosed which is a major concern. This problem can be
overcome by refining the algorithm over time by comparing its decisions with the clinicians.
Other commonly used techniques that share some of
the same advantages and limitations of the abovedescribed artificial neural networks are the random forests
and the support vector machines [27, 28]. The random
forest uses the concept of the decision tree (DT) models.
In these models, a flowchart-like structure is built. In each
node, there is a test using one of the independent variables
that will split the data in two mutual subgroups (bran
ches). This process is repeated several times until the final
decision of class assignment (leaves) is reached.
The random forest combine several trees. Each DT in this
method is built based on a random subset of the data that is
generated by a bootstrap resampling technique and for each
split, the best variable is chosen from a subset composed of a
pre-defined number of randomly selected variables.
Accordingly, to its final decision, each tree gets a “vote” and
the mode is used in classification problems while the mean is
used in regression. Figure 3 exemplifies this process. Some
advantages of the method are that it can model nonlinear class
boundaries and can give variable importance. On the downside, it is a slow method and it is hard to get insights into the
decision rules.
In support vector machine (SVM), a previously classified
dataset (supervised learning) is used to train the model. The
algorithm will search for an n-dimensional hyperplane able to
separate the group with the largest margin. In many cases, in
ophthalmology, a linear solution (two-dimensional) is not possible, so finding the solution in a higher dimensional space is
needed as observed in Fig. 4. The SVM looks for this solutions with a relatively less computational cost using the kernel
trick. The kernel trick is a function used to obtain nonlinear
variants of a selected algorithm with an ability to be casted in
dot products’ formation [29]. However, choosing a suitable
kernel is rather easy and an inappropriate kernel could lead
to overfitting [30].
The applications of these methods will be discussed below
in clinical scenarios.
&
&
Cleaning of input data in an organised manner
Choosing one appropriate cross-validation method (in the
order of extrapolation accuracy and computational cost
from lower to higher)
–
Hold-out: Splitting the data into a randomly selected
training set (normally 70–90% of the data) and test set
K-fold: Splitting the data in k folds, normally 10, separating 9 of the folds for training and 1 for testing. Repeating
the process 10 times, we can get the average accuracy and
its standard deviation.
Leave-one-out: Use all data set but one case to train the
data and test the accuracy in this remaining case.
Repeating the process n times (size of the dataset) will
get a robust estimation of the generalisation ability of the
model.
–
–
&
&
&
The normalisation of training data to achieve consistency
Optimising the architecture of the network, pre-set parameters and selecting a suitable NN through trials and errors
Evaluating the performance of the network through the
cost function (minimising the error between predictions
and expected data)
The advantage of using ANN is that the algorithm is able to
deal with noisy and missed clinical data and understanding
complex patterns in the data in a way that is not possible with
linear and non-linear equations [26].
The problem with ANN is that they require extremely
large clinical datasets for training. For achieving a globally accepted performance, these clinical data should be
collected from different ethnicities. This is not always
easy to obtain considering patient data is often expensive,
highly regulated, and time-consuming to collect in the
desired manner. On the other hand, what is going on in
the network is more like a black box as it does not
Curr Ophthalmol Rep (2019) 7:204–211
207
Fig. 3 Example of a random forest model. a The patients included in the
train and test sets, for exemplification purposes, different characteristics
of height, age, and weight are graphically expressed with geometrical
shapes (triangle, square, and circle, respectively). b A very simple
random forest model composed of 3 trees trained with the three
characteristics to separate healthy (H) and diseased (D) and the
classification path (in green) of a new patient
Keratoconus Diagnosis and Refractive
Surgery Screening
identification of these very early forms of the disease with
high accuracy [33••, 34–36].
The pentacam random forest index (PRFI) is a random
forest model built using data from the tomographer
Pentacam HR (Oculus, Wetzlar, Germany). It was the only
model trained with the preoperative exam of patients that have
developed ectasia. It was trained using a large data set of
patients from 3 different continents to better assess the patient’s normal variability. While the index already available
on the device (BAD-D) presented 55.3% of sensitivity, the
PRFI was able to correctly identify 80% of the cases. In the
external validation set, 85% of accuracy was found in detecting the normal topographic eye of very asymmetric cases
(VAE-NT) maintaining specificity of 96.6% [33••].
A single decision tree method was proposed based on the
data of a different tomographer, the Galilei Dual Scheimpflug
Analyzer (Ziemer Ophthalmic Systems AG, Port,
Switzerland). This index held sensitivity of 90% with specificity of 86% to detect the early forms [34]. Analysing the data
from the tomographer Sirius (CSO, Firenze, Italy), the
The progressive character of the ecstatic corneal disease has
always stimulated the search for a means of early diagnosis.
With the introduction of corneal cross-linking, the disease
progression could be halted, and with an accurate early diagnosis of the disease, the vision of patients could be preserved.
But the increasing number of refractive surgeries and the report of iatrogenic keratectasia in cases without anterior surface
alterations pushed the need to screen for susceptible cases
[31]. It means to identify those cases that could experience a
biomechanical failure after the procedure even before any
mild alteration is present in the anterior corneal surface. To
accomplish this task, a series of methods including risk scores,
linear models, and more recently artificial intelligence models
have been proposed with a progressive accuracy increase evaluating data from different devices [32, 33••]. Data from 4
different tomographers were analysed with 4 different artificial intelligence techniques and it made possible the
Fig. 4 a Linear separable data
with good margins. b Linearly
impossible to separate data (2
dimensions) c Separation
obtained with a hyperplane in a
higher dimensional space (3
Dimensions)
208
identification of cases with first signs of the disease (a stage
slightly later than the VAE-NT) using an SVM model presented sensitivity of 92% with 97.7% of specificity [35].
Discriminant linear models were also successfully used to
analyse the Orbscan II data (Technolas, Munich, Germany)
with 92% and 96% of sensitivity and specificity, respectively,
in a first validation set [36] and 70.8% and 98.1% of sensitivity and specificity, respectively, in a different ethnical background population [37].
The in vivo assessment of corneal biomechanics also provided a torrent of new and hard to interpret information.
Neural networks have been used to evaluate the waveform
signal of the Ocular Response Analyzer (Reichert
Ophthalmic Instruments, Buffalo, USA) and resulted in high
accuracy on the study validation sample composed of early
forms of keratoconus (AUC 0.978) [38]. However, these results had not been validated in independent samples. The data
from the other commercially available device the Corvis ST
(Oculus Optikgeräte GmbH, Wetzlar, Germany) has also been
analysed individually with logistic regression and high accuracy was found in keratoconus detection with 98.8% of the
cases correctly classified [39]. The relatively low accuracy in
detecting initial cases for the device used independently was
overcome with the integration with tomographic data using
AI. The random forest model named the tomography and
biomechanical index (TBI) achieved 90.3% of sensitivity in
detecting the VAE-NT with 96% of specificity [40••]. The
tomographic and biomechanical combined parameter showed
to be superior to both methods used alone [41].
One of the main advantages of the AI-derived models is
their ability to eliminate the gap between research and clinical
application, by providing a simple output parameter that can
be directly used as a risk profiling tool. Different corneal imaging devices (tomographers and biomechanical analysers)
have already implemented on its software these AI-based indices and made them available in daily clinical practice as
objective screening parameters.
In Vivo Corneal Morphology Exams
The in vivo corneal morphology evaluation has also benefitted
from the AI models. Neural networks have been used to automatically identify the healthy corneal layers on the confocal
microscopy exam with significant improvement over the previous methods and eliminating the need of the image processing step of binarization, which is cumbersome and can often
lead to information loss [42]. Evaluating the ultra-highresolution OCT images with convolutional neural networks,
it was also possible to precisely identify the corneal layers in
keratoconic eyes. In these cases, a non-uniform thickness was
observed in at least one layer, and this result was also proposed to be used for diagnosis [43, 44].
Curr Ophthalmol Rep (2019) 7:204–211
Going further into the evaluation of more specific histological features of the corneal tissue is possible to study in vivo
the endothelial cell and the subbasal nerve plexus characteristics. The endothelial cell number and shape are calculated by
means of specular or corneal confocal microscopy exams. To
estimate cell density, pleomorphism, and polymegethism, the
images acquired are manually annotated. Although it seems a
trivial task to identify cell borders, due to frequent low image
quality with blurred regions, low contrast, and artefacts, this
automatization is often difficult. In order to fully automate the
process, increasing the speed and avoiding imprecisions of the
manual appraisal of the exam, several methods to delineate the
cell borders have been proposed. The machine learning
models have proven to be a fast and more accurate method
of characterising the endothelium [45, 46•].
The corneal subbasal nerve plexus is composed of short
nerve fibres that can be noninvasively studied with confocal
microscopy. It has gained a special interest in evaluating diabetic sensorimotor polyneuropathy (DSPN), a common longterm complication of the disease affecting up to 50% of the
patients [47]. The promising results in identifying the DSPN
in cross-sectional studies face considerable difficulties in longitudinal evaluation since the images are manually analysed in
an imprecise and time-consuming process [48]. This gap was
filled with the introduction of neural network and random
forest models to fully automate the nerve segmentation and
morphology study, allowing the development of an objective
and precise method to early characterise the disease [49•, 50].
Hyphae detection in corneal fungal infections and even the
image segmentation of corneal ulcers, difficult tasks for subjective characterisation, have also had their accuracies improved with the aid of artificial intelligence models [51, 52].
Corneal Surgery
Another growing field where machine learning techniques are
being used is corneal surgery. Automated objective quantification of haze and the demarcation line post-cross-linking
surgery is achieved with a support vector machine model.
This method provides the clinician’s haze statistics along with
visual demarcation on the OCT images of the shape and location of haze and the demarcation line [53, 54•].
In corneal posterior lamellar transplant surgery, graft detachment is one of the main complications, especially during
the initial part of the learning curve [55]. Intraoperative corneal OCT can be used to evaluate the fluid in the graft-host
interface [56]. The identification of a larger residual interface
fluid volume by the automated graph searching approach is
associated with early graft dislocation [57]. On the postoperative period, the graft dislocation can also be objectively quantified using a convolutional neural network [58].
Curr Ophthalmol Rep (2019) 7:204–211
Current Limitations and Future Perspectives
AI techniques have already shown its value for enhancing
clinical decisions for patients with corneal conditions, and its
application keeps growing. However, there are some considerations and important limitations that preclude an even more
spread use of it. One big limitation of deep learning is that it
requires large training sets, on the order of tens of thousands to
be accurately trained and to be possible to generalise the results. The very large datasets are also required to accurately
deal with the high amount of noise derived from biological
data. Successfully trained deep neural networks to classify
retinal funduscopic images utilised a dataset of more than
the 100,000 images [59, 60].
In corneal imagining, there are a few hurdles to build a very
large dataset making it an extremely hard task, if not impossible.
The high cost of devices such as the tomographers makes their
availability relatively lower than the retinographers. The technical challenges to acquire images with devices such as the confocal microscope, that requires highly trained operators, is also a
challenge. Another limitation is the differences between devices,
even those that use the same technology. For instance, with
Scheimpflug imaging, the scans obtained from the same patients
with different devices are not easily interchangeable, which restricts the datasets to usually a single device type. One change in
this scenario is the clever adaptations for smartphone cameras
that allow self-imaging of the cornea and the anterior segment,
and even high-quality imaging at a sub-cellular resolution [61,
62]. With the relative low cost and wide spread of these portable
devices, a bigger dataset of corneal imaging are more likely to be
acquired and the AI applications are numerous with a high potential to promote healthcare in remote areas [63].
With big datasets, there is also the need of high computational
power to evaluate the features in a reasonable time which compels the participation of big high tech companies to make them
possible to be applied in real life. AI models also need to be
continuously trained and exposed to new data to be able to identify more subtle variations of the patterns and normal ethnical
differences. Considering patient data confidentialities, an efficient data sharing system under strict privacy rules needs to be
implemented to facilitate AI advancements.
The participation of high tech companies to provide the
computational power needed, multicentric collaborations to
gather big datasets along with efficient data sharing systems
to constantly train the models is a vital step to improve the
accuracy of AI models applied to medicine in general and also
corneal diseases.
Conclusion
In conclusion, machines have been used to augment the clinical ability of ophthalmologies, either to reveal characteristics
209
initially imperceptible to our senses in the typical clinical exam or to aid in the interpretation of the amount of information
that they itself produce. AI models diagnostic indices are already available and widely used by clinicians in refractive
surgery screening, and several other application are under fast
development.
Compliance with Ethical Standards
Conflict of Interest Ashkan Eliasy declares no potential conflicts of
interest. Bernardo T. Lopes has received research funding from
OCULUS Optikgeräte GmbH. Renato Ambrosio Jr. is a consultant for
Oculus and Alcon.
Human and Animal Rights and Informed Consent This article does not
contain any studies with human or animal subjects performed by any of
the authors.
Open Access This article is distributed under the terms of the Creative
Commons Attribution 4.0 International License (http://
creativecommons.org/licenses/by/4.0/), which permits unrestricted use,
distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons license, and indicate if changes were made.
References
Papers of particular interest, published recently, have been
highlighted as:
• Of importance
•• Of major importance
1.
Daxecker F. Christoph Scheiner’s eye studies. In: Henkes HE, editor. History of Ophthalmology 5: Sub auspiciis Academiae
Ophthalmologicae Internationalis. Dordrecht: Springer
Netherlands; 1993. p. 27–35.
2. Belin MW, Khachikian SS. An introduction to understanding
elevation-based topography: how elevation data are displayed - a
review. Clin Exp Ophthalmol. 2009;37(1):14–29.
3. Liu Z, Huang AJ, Pflugfelder SC. Evaluation of corneal thickness
and topography in normal eyes using the Orbscan corneal topography system. Br J Ophthalmol. 1999;83(7):774–8.
4. Reinstein DZ, Gobbe M, Archer TJ, Silverman RH, Coleman J.
Epithelial thickness in the normal cornea: three-dimensional display
with Artemis very high-frequency digital ultrasound. J Refract
Surg. 2008;24(6):571–81.
5. Jalbert I, et al. In vivo confocal microscopy of the human cornea. Br
J Ophthalmol. 2003;87(2):225–36.
6. Luce DA. Determining in vivo biomechanical properties of the
cornea with an ocular response analyzer. J Cataract Refract Surg.
2005;31(1):156–62.
7. Valbon BF, Ambrósio R Jr, Fontes BM, Luz A, Roberts CJ, Alves
MR. Ocular biomechanical metrics by CorVis ST in healthy
Brazilian patients. J Refract Surg. 2014;30(7):468–73.
8. Eliasy A, Chen KJ, Vinciguerra R, Lopes BT, Abass A, Vinciguerra
P, et al. Determination of corneal biomechanical behavior in-vivo
for healthy eyes using CorVis ST tonometry: stress-strain index.
Front Bioeng Biotechnol. 2019;7.
210
9.
Gross BM. The managing of organizations: the administrative
struggle. New York: Free Press; 1964. p. 1964.
10. Smith R. Strategies for coping with information overload. BMJ.
2010;341:c7126.
11. Maeda N, Klyce SD, Smolek MK, Thompson HW. Automated
keratoconus screening with corneal topography analysis. Invest
Ophthalmol Vis Sci. 1994;35(6):2749–57.
12. Maeda N, Klyce SD, Smolek MK. Neural network classification of
corneal topography. Preliminary demonstration. Invest Ophthalmol
Vis Sci. 1995;36(7):1327–35.
13. Smolek MK, Klyce SD. Current keratoconus detection methods
compared with a neural network approach. Invest Ophthalmol Vis
Sci. 1997;38(11):2290–9.
14. Kalin NS, et al. Automated topographic screening for keratoconus
in refractive surgery candidates. CLAO J. 1996;22(3):164–7.
15. Smolek MK, Klyce SD. Screening of prior refractive surgery by a
wavelet-based neural network. J Cataract Refract Surg.
2001;27(12):1926–31.
16. Hafezi F, Kanellopoulos J, Wiltfang R, Seiler T. Corneal collagen
crosslinking with riboflavin and ultraviolet A to treat induced
keratectasia after laser in situ keratomileusis. J Cataract Refract
Surg. 2007;33(12):2035–40.
17. Seiler T, Koufala K, Richter G. Iatrogenic keratectasia after laser in
situ keratomileusis. J Refract Surg. 1998;14(3):312–7.
18. Seiler T, Quurke AW. Iatrogenic keratectasia after LASIK in a case
of forme fruste keratoconus. J Cataract Refract Surg. 1998;24(7):
1007–9.
19. Hunt EB, Marin J, Stone PJ. Experiments in induction. 1966.
20. Nilsson, N.J., Learning machines. 1965.
21. Rosenblatt F. Principles of neurodynamics. Perceptrons and the
theory of brain mechanisms. 1961. CORNELL
AERONAUTICAL LAB INC BUFFALO NY.
22. Mitchie D, Spiegelhalter DJ, Taylor CC. Machine learning, neural
and statistical classification. 1994.
23. Schmidhuber J. Deep learning in neural networks: an overview.
Neural Netw. 2015;61:85–117.
24. Ayodele TO. Types of machine learning algorithms, in New advances in machine learning: IntechOpen; 2010.
25. Priddy KL, Keller PE. Artificial neural networks: an introduction,
vol. 68: SPIE press; 2005.
26. Livingstone DJ, Manallack DT, Tetko IV. Data modelling with neural networks: advantages and limitations. J Comput Aided Mol Des.
1997;11(2):135–42.
27. Segal MR. Machine learning benchmarks and random forest regression. UCSF: Center for Bioinformatics and Molecular Biostatistics,
2004. Retrieved from https://escholarship.org/uc/item/35x3v9t4.
28. Scholkopf B, Smola AJ. Learning with kernels: support vector machines, regularization, optimization, and beyond. Cambridge: MIT
press; 2001.
29. Schölkopf B, Burges CJ, Smola AJ. Advances in kernel methods:
support vector learning. Cambridge: MIT press; 1999.
30. Liu Y, Liao S. Preventing over-fitting of cross-validation with kernel stability. Berlin, Heidelberg: Springer Berlin Heidelberg; 2014.
31. Ambrosio R Jr, Randleman JB. Screening for ectasia risk: what are
we screening for and how should we screen for it? J Refract Surg.
2013;29(4):230–2.
32. Chan C, Saad A, Randleman JB, Harissi-Dagher M, Chua D, Qazi
M, et al. Analysis of cases and accuracy of 3 risk scoring systems in
predicting ectasia after laser in situ keratomileusis. J Cataract
Refract Surg. 2018;44(8):979–92.
33.•• Lopes BT, et al. Enhanced tomographic assessment to detect corneal ectasia based on artificial intelligence. Am J Ophthalmol.
2018;195:223–32 In this paper, some artificial intelligence
models were developed based on a multicentre dataset to improve preoperative screening for refractive surgery.
Curr Ophthalmol Rep (2019) 7:204–211
34.
35.
36.
37.
38.
39.
40.••
41.
42.
43.
44.
45.
46.•
47.
48.
49.•
50.
Smadja D, Touboul D, Cohen A, Doveh E, Santhiago MR, Mello
GR, et al. Detection of subclinical keratoconus using an automated
decision tree classification. Am J Ophthalmol. 2013;156(2):237–
246 e1.
Arbelaez MC, Versaci F, Vestri G, Barboni P, Savini G. Use of a
support vector machine for keratoconus and subclinical
keratoconus detection by topographic and tomographic data.
Ophthalmology. 2012;119(11):2231–8.
Saad A, Gatinel D, Barbara A. Validation of a new scoring system
for the detection of early forme of keratoconus. Int J Keratoconus
Ectatic Corneal Dis. 2012;1:100–8.
Chan C, Ang M, Saad A, Chua D, Mejia M, Lim L, et al. Validation
of an objective scoring system for forme fruste keratoconus detection and post-LASIK ectasia risk assessment in Asian eyes. Cornea.
2015;34(9):996–1004.
Ventura BV, Machado AP, Ambrósio R Jr, Ribeiro G, Araújo LN,
Luz A, et al. Analysis of waveform-derived ORA parameters in
early forms of keratoconus and normal corneas. J Refract Surg.
2013;29(9):637–43.
Vinciguerra R, Ambrósio R Jr, Elsheikh A, Roberts CJ, Lopes B,
Morenghi E, et al. Detection of keratoconus with a new biomechanical index. J Refract Surg. 2016;32(12):803–10.
Ambrosio R Jr, et al. Integration of Scheimpflug-based corneal
tomography and biomechanical assessments for enhancing ectasia
detection. J Refract Surg. 2017;33(7):434–43 This publication enlightens the role of AI on the multimodal preoperative screening for refractive surgery combining data from different
devices.
Ferreira-Mendes J, Lopes BT, Faria-Correia F, Salomão MQ,
Rodrigues-Barros S, Ambrósio R Jr. Enhanced ectasia detection
using corneal tomography and biomechanics. Am J Ophthalmol.
2019;197:7–16.
Ruggeri A, Pajaro S. Automatic recognition of cell layers in corneal
confocal microscopy images. Comput Methods Prog Biomed.
2002;68(1):25–35.
Dos Santos VA, et al. CorneaNet: fast segmentation of cornea OCT
scans of healthy and keratoconic eyes using deep learning. Biomed
Opt Express. 2019;10(2):622–41.
Sudharshan Mathai, T., K. Lathrop, and J. Galeotti Learning to
segment corneal tissue interfaces in OCT images. arXiv e-prints,
2018.
Kolluru C et al. Machine learning for segmenting cells in corneal
endothelium images. SPIE Medical Imaging. Vol. 10950. 2019:
SPIE.
Nurzynska K. Deep learning as a tool for automatic segmentation of
corneal endothelium images. Symmetry. 2018. 10(3). The conventional methods of automatic endothelial cell characterisation
were outperformed by a convolution neural network model.
Tesfaye S, Vileikyte L, Rayman G, Sindrup SH, Perkins BA,
Baconja M, et al. Painful diabetic peripheral neuropathy: consensus
recommendations on diagnosis, assessment and management.
Diabetes Metab Res Rev. 2011;27(7):629–38.
Dehghani C, Pritchard N, Edwards K, Vagenas D, Russell AW,
Malik RA, et al. Morphometric stability of the corneal subbasal
nerve plexus in healthy individuals: a 3-year longitudinal study
using corneal confocal microscopy. Invest Ophthalmol Vis Sci.
2014;55(5):3195–9.
Al-Fahdawi S, et al. A fully automatic nerve segmentation and
morphometric parameter quantification system for early diagnosis
of diabetic neuropathy in corneal images. Comput Methods
Programs Biomed. 2016;135:151–66 This publication highlights
the use of AI to automate the evaluation of corneal confocal
microscopy images to non-invasively assess an important complication of diabetes.
Chen X, Graham J, Dabbah MA, Petropoulos IN, Tavakoli M,
Malik RA. An automatic tool for quantification of nerve fibers in
Curr Ophthalmol Rep (2019) 7:204–211
corneal confocal microscopy images. IEEE Trans Biomed Eng.
2017;64(4):786–94.
51. Wu X, Qiu Q, Liu Z, Zhao Y, Zhang B, Zhang Y, et al. Hyphae
detection in fungal keratitis images with adaptive robust binary
pattern. IEEE Access. 2018;6:13449–60.
52. Deng L et al. Automatic segmentation of corneal ulcer area based
on ocular staining images. SPIE Medical Imaging. Vol. 10578.
2018: SPIE.
53. Dhaini AR, Chokr M, el-Oud SM, Fattah MA, Awwad S.
Automated detection and measurement of corneal haze and demarcation line in spectral-domain optical coherence tomography images. IEEE Access. 2018;6:3977–91.
54.• Awwad ST, Fattah MA, Shokr M, Dhaini AR. Automated detection
of the stromal demarcation line using optical coherence tomography in keratoconus eyes after corneal cross-linking. Am J
Ophthalmol. 2019;199:177–83 In this publication, a new ultrafast software to detect the demarcation line after cross-linking is
able to optimise and standardise the indicator for treatment
success.
55. Pereira NC, Gomes JÁP, Moriyama AS, Chaves LF, Forseto AS.
Descemet membrane endothelial keratoplasty outcomes during the
initial learning curve of cornea fellows. Cornea. 2019;38:806–11.
56. Xu D, Dupps WJ Jr, Srivastava SK, Ehlers JP. Automated volumetric analysis of interface fluid in descemet stripping automated endothelial keratoplasty using intraoperative optical coherence tomography. Invest Ophthalmol Vis Sci. 2014;55(9):5610–5.
57. Hallahan KM, Cost B, Goshe JM, Dupps WJ Jr, Srivastava SK,
Ehlers JP. Intraoperative interface fluid dynamics and clinical outcomes for intraoperative optical coherence tomography-assisted
211
Descemet stripping automated endothelial keratoplasty from the
PIONEER study. Am J Ophthalmol. 2017;173:16–22.
58. Treder M, et al. Using deep learning in automated detection of graft
detachment in Descemet membrane endothelial keratoplasty: a pilot
study. Cornea. 2019;38(2):157–61.
59. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D,
Narayanaswamy A, et al. Development and validation of a deep
learning algorithm for detection of diabetic retinopathy in retinal
fundus photographs. JAMA. 2016;316(22):2402–10.
60. Tufail A, Rudisill C, Egan C, Kapetanakis VV, Salas-Vega S, Owen
CG, et al. Automated diabetic retinopathy image assessment software: diagnostic accuracy and cost-effectiveness compared with
human graders. Ophthalmology. 2017;124(3):343–51.
61. Kaya A. Ophthoselfie: detailed self-imaging of cornea and anterior
segment by smartphone. Turk J Ophthalmol. 2017;47(3):130–2.
62. Toslak D, Thapa D, Erol MK, Chen Y, Yao X. Smartphone-based
imaging of the corneal endothelium at sub-cellular resolution. J
Mod Opt. 2017;64(12):1229–32.
63. Rono HK, Bastawrous A, Macleod D, Wanjala E, di Tanna GL,
Weiss HA, et al. Smartphone-based screening for visual impairment
in Kenyan school children: a cluster randomised controlled trial.
Lancet Glob Health. 2018;6(8):e924–32.
Publisher’s Note Springer Nature remains neutral with regard to
jurisdictional claims in published maps and institutional affiliations.