0% found this document useful (0 votes)
0 views20 pages

remotesensing-15-00482 (1)

Download as pdf or txt
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 20

remote sensing

Article
Comparative Analysis of Machine Learning Algorithms for Soil
Erosion Modelling Based on Remotely Sensed Data
Daniel Fernández 1 , Eromanga Adermann 1 , Marco Pizzolato 1 , Roman Pechenkin 1 , Christina G. Rodríguez 1
and Alireza Taravat 2, *

1 Fléttan, Kópavogsbraut 85, 200 Kópavogur, Iceland


2 Deimos Space UK Ltd., AirSpeed 1, 151 Eighth St., Harwell Oxford, Didcot OX11 0QR, UK
* Correspondence: art23130@gmail.com

Abstract: Recent years have seen an increase in the use of remote-sensing based methods to assess
soil erosion, mainly due to the availability of freely accessible satellite data, with successful results on
a consistent basis. There would be valuable benefits from applying these techniques to the Arctic
areas, where ground local studies are typically difficult to perform due to hardly accessible roads
and lands. At the same time, however, the application of remote-sensing methods comes with its
own set of challenges when it comes to the peculiar features of the Arctic: short growing periods,
winter storms, wind, and frequent cloud and snow cover. In this study we perform a comparative
analysis of three commonly used classification algorithms: Support Vector Machine (SVM), Random
Forest (RF) and Multilayer Perceptron (MLP), in combination with ground truth samples from regions
all over Iceland, provided by Iceland’s Soil Conservation Service department. The process can be
automated to predict soil erosion risk for larger, less accessible areas from Sentinel-2 images. The
analysis performed on validation data sets supports the effectiveness of both approaches for modeling
soil erosion, albeit differences are highlighted.

Keywords: soil erosion; Sentinel-2; remote sensing; machine learning; support vector machine;
random forest; multilayer perceptron; image classification; arctic

Citation: Fernández, D.; Adermann,


E.; Pizzolato, M.; Pechenkin, R.;
Rodríguez, C.G.; Taravat, A.
1. Introduction
Comparative Analysis of Machine
Learning Algorithms for Soil Erosion
Soil erosion is becoming a major land deterioration hazard worldwide. Making strides
Modelling Based on Remotely Sensed
in our understanding of the likely future rates of soil erosion, quickened by human activity
Data. Remote Sens. 2023, 15, 482. and climate alterations, is one of the foremost conclusive components when it comes to
https://doi.org/10.3390/rs15020482 making choices on preservation arrangements and for Earth ecosystem modelers looking
to diminish the unreliability of global expectations [1]. Despite this, detailed information is
Academic Editor: Dino Ienco
typically based on limited field campaigns and is typically only known locally [2].
Received: 4 December 2022 This is the situation in Iceland, where Landgræðslan, the country’s Soil Conservation
Revised: 9 January 2023 Service, is the sole source of the majority of information about soil degradation in the
Accepted: 11 January 2023 land [3]. The deterioration of Iceland’s ecosystem can be considered as a form of desertifi-
Published: 13 January 2023 cation. Due to their lack of vegetation, its badlands have prominent similitudes to desolate
zones in dry weathered nations. Soil erosion forecast plays a key part in relieving such
progression [4].
From a historical perspective, pioneers include Björn Johannesson [5], who first pub-
Copyright: © 2023 by the authors.
lished a soil erosion map in his book on the soils of Iceland. The current FAO classification
Licensee MDPI, Basel, Switzerland.
for Icelandic soils was attempted decades later [6]. The Agricultural Research Institute of
This article is an open access article
Iceland (Rala), which became a part of the Agricultural University of Iceland (AUI) in 2005,
distributed under the terms and
has done the majority of the work on Icelandic soil science. The joint European COST-622
conditions of the Creative Commons
Action [7,8] provides a wealth of information regarding the chemical and physical properties
Attribution (CC BY) license (https://
of Iceland’s soils. Numerous Icelandic and international research efforts have contributed to
creativecommons.org/licenses/by/
4.0/).
our understanding of the effects of both man-caused and natural degradation.

Remote Sens. 2023, 15, 482. https://doi.org/10.3390/rs15020482 https://www.mdpi.com/journal/remotesensing


Remote Sens. 2023, 15, 482 2 of 20

The government of Iceland aspires to control soil erosion and achieve sustainable
land use as soon as possible. The interaction of past and present grazing effects with
delicate soils and vegetation is the primary cause of desertification. In order to better
manage and monitor grazing practices in protected areas that are at risk of erosion and to
restore denuded land, the soil conservation authorities—primarily the soil conservation
service—were given stronger capabilities [9].
Measurements taken on the ground and aerial photographs taken at various points
in time are two of the methods used to evaluate the progression of soil erosion. With
aerial photographs, there are two methods: scanning and image analysis on one hand,
and digitizing on the other. There is a certain amount of room for error when using them,
however. On-site measurement is difficult because these are expensive tasks, especially in
some parts of the country where access is difficult. As a result, not all areas of interest can
currently be investigated.
Soil erosion in Arctic regions such as Iceland may also have an impact on the global
climate, in addition to the effects that climate change may already be having on the ecosys-
tems of those regions. Up to half of the carbon in the soil of Earth is stored in northern
latitudes; approximately twice as much carbon as is stored in the atmosphere. The signifi-
cance of this carbon sink cannot be overstated. Northern circumpolar soils are a significant
carbon sink because of extensive peatlands and permanently frozen ground that keeps
this organic carbon locked up in the soil [10]. The Northern Circumpolar Soil Carbon
Database’s most recent estimates indicate that the northern permafrost region contains
approximately 1672 Pg of organic carbon, with approximately 1466 Pg, or 88%, occurring
in soils and deposits that are perpetually frozen [11]. Thermal collapse and erosion of these
carbon-rich deposits may accelerate with Arctic amplification of climate warming [12].
Through the development of satellite remote sensing technology, one extremely useful
tool has been made available to overcome the limitations outlined above [13]. Its application
to evaluate soil erosion, in any event, when in view of uninhibitedly available satellite
information, has resulted in numerous breakthroughs as of late [14,15]. Some multi-
resolution approaches to soil erosion risk mapping have been used in previous studies, but
they only used indicators such as rainfall and vegetation cover to apply them to Iceland.
In this work, ground truth samples are used in our proposed solution to analyse
pre-processed remote images over a specific area, to then automate the analysis for larger,
more difficult-to-access areas. This solution is being developed specifically for Icelandic
soil erosion studies, using Sentinel-2 satellite data combined with Landgræðslan’s local
assessment data. Given that they are the oldest soil erosion department in the world, their
historical data is more extensive than usual.
This is made possible by training a machine learning (ML) algorithm that can produce
appropriate weights that represent erosion-prone areas using a set of geo-environmental
parameters. In operation research, the development of mathematical and computational
tools to support decision makers’ intellectual evaluation of criteria and alternatives has been
incorporating an increasing number of ML models [16]. Training the algorithm requires
parameters of bare ground cover, which can be calculated solely from satellite images by
using information from observationally correct areas devoid of vegetation as a calibration;
Icelandic soil profiles, which will be examined to determine how the profile relates to the
intensity of soil erosion; as well as data on arable land, including cultivated land plant
species, and the parameters of agriculture use. These are all included in the available data.
One of the areas of ML where inputs can be given a class label is classification. Image
classification has been the focus of several remote sensing applications using Random
Forest (RF), Support Vector Machine (SVM), and Multilayer Perceptron (MLP) among the
available ML algorithms. One of the hot topics in multiple research fields that use images
of different resolutions taken by different platforms, while including several limitations
and priorities, is how to choose the best image classification algorithm.
Remote Sens. 2023, 15, 482 3 of 20

Based on the authors’ knowledge, there are very few studies that have focused on
applying these methods to the study of soil erosion in Arctic areas, despite machine
learning models potentially being very competitive in terms of accuracy and predictive
power. Similar image processing classification techniques have been used to study the arctic
tundra vegetation [17], the ionic composition of arctic lakes [18] and land cover types [19].
Deep learning has been employed to monitor polar sea ice [20].
Regarding soil erosion, early remote sensing studies showed an increase in erosion rate
during the last decades in areas such as Alaska [21]. Contemporary techniques are throwing
interesting conclusions that link coastal erosion to sea surface temperature fluctuations [22]
and complex processes related to coastal permafrost bluff erosion [23]. In fact, deep learning
techniques were recently applied to extract the erosion rates of Arctic permafrost [24]. As
such, this is an exciting area in which many discoveries are yet to be made. Particularly,
there is a lack of detailed analyses of the influence of slope steepness and height on erosion,
since these are factors that do not play an important role in other areas of the world. The
present study aims at providing one of the first steps toward establishing the most reliable,
fast and accurate classical ML algorithm to predict soil erosion quantitatively in the Arctic,
which may in turn allow us to locate the main process driving it.

2. Materials and Methods


2.1. Data and Preprocessing
2.1.1. Icelandic Soil Erosion Data
Iceland is an island of size 103,000 km2 located within the Arctic Circle that presents
a unique combination of erosion factors not seen in any other region on the planet. The
region experiences high levels of volcanic and seismic activity due to its position over a
hotspot on the Mid-Atlantic Ridge (a divergent tectonic plate boundary), as well as a harsh
climate of cold winters and cool summers due to the moderating effects of the Atlantic Gulf
Stream [25,26]. Its unique geography and weather include glaciers (which cover 11% of
its surface and are melting), very active volcanoes, carbon rich volcanic soils, sand dunes,
freeze-thaw cycles and extreme winds. The landscape is sparsely vegetated (<50% of land
is vegetation) and there is little to no tree cover, which is a relatively recent change that
resulted from livestock overgrazing and human activity paired with the harsh climates,
winds and volcanic activity [27,28], the effects of which are still felt today. Vegetation loss
exposes top layers of soil, which are easily disturbed by wind erosion [29,30]. Moreover,
volcanic eruptions can produce large dust storms that further threaten vegetation, and
melting glaciers from volcanic activity can impact soil through water erosion [26,27,29].
With the continued threat of erosion and degradation, the Soil Conservation Service
of Iceland, known as Landgræðslan, monitors and assesses the level of soil erosion across
Iceland as part of its efforts to protect, restore and improve the island’s vegetation and soil
resources. Some of this data on soil erosion levels was provided to us by the organisation,
consisting of a database of 748 regions, which are visualised in Figure 1. Each point consists
of a value reflecting the measured level of soil erosion within the region, which was used
as ground truth values for our training data. This data was obtained through various
programmes of assessment aimed at land restoration, since the scale of soil erosion and
desertification in Iceland has long been a cause of ecological concern in the country [29].

2.1.2. Earth Observations Data


The Copernicus Sentinel-2 data, characterized by its free access policy, has a global
coverage of Earth’s land surfaces from 56º S to 84º N. It offers improved data compared
to other low spatial resolution satellite images, especially in temporal and spatial reso-
lution [31]. The 13 bands for Sentinel-2 images have spatial resolutions ranging from
10 to 60 m [32]. The other valuable characteristic of Sentinel-2 data is its high temporal
resolution of 5 days [33]. Its sensor has low radiometric calibration uncertainty that makes
the radiance of the images to produce reliable results. Thus, Sentinel-2 images have the
potential to produce highly accurate information and to support different applications [13].
Remote Sens. 2023, 15, 482 4 of 20

Figure 1. Map of Iceland. The regions where soil erosion measurements were taken by Landgræðslan
are marked by black circles. Altitude is indicated by the color ramp.

The data we used to train our machine learning models was obtained through WEkEO,
the European Union Copernicus DIAS reference service, which provides pre-processed L2A
Sentinel-2 data. The data consists of geo-referenced and atmospherically corrected images
of Iceland between 23 June 2015 and 26 July 2022. Scene classification layers are included
for each picture, which can be projected into the original in order to figure out which pixels
depict snow, clouds or non-information areas. After that, an automated procedure is used
to create a combined image without these undesirable effects. This was performed with
images taken over a 30-day summer period for the present results. Depending on the
satellite pass, the data are accessible every 5 to 10 days [34].
We generated images of all over Iceland for each band between 2 July 2019 and
24 September 2021. We decided to focus on bands with resolutions between 10 and 20 m.
After atmospheric correction, pixels either with clouds or shadows of clouds, or with no
data or overly saturated (surface reflectance greater than 1) were considered as bad pixels.
The latter may happen due to sub glint or snow reflectance. Snow and ice should also be
removed, as well as water bodies, which are of no interest if they are constantly full. After
cropping clouds and bad pixels out, these images are then used to create indices [35]. Indices
that represent or highlight particular characteristics, such as vegetation, soil crusting, bare
soil, and red edge indices, were created by combining data from various spectral bands.
The raw Sentinel-2 images may appear changed between different dates and thus
require appropriate data filtering. This is a challenging task, since our atmosphere is
quite dynamic and can change for many reasons. Even after cloudy observations have
been filtered out by the preprocessing system [36] of Sentinel-2, some cloud shadows are
misclassified by default and therefore artifacts are still present. What is left is a mixture
of valid observations and a set of undetected anomalous observations, which need to
be identified. Due to the Arctic atmospheric conditions, this problem of cloud coverage
Remote Sens. 2023, 15, 482 5 of 20

becomes exceptionally significant for our studies. Most images are covered by clouds or by
snow being carried over by wind into cloudless areas.
In order to solve this, we used an iterative script that takes multiple images of the
same area for a selected time period and delivers one image (per band) that contains the
least amount of bad data. Sentinel-2 data is downloaded by tiles, which represent UTM
zones. We used the scene classification layer to identify the unusable bad pixels mentioned
above on the training tile location, as well as on the finally classified tile. After all tiles are
processed, the tiles are stitched together to produce an image of the entire area of interest.
The composite was created by scanning the images until one was found with a usable
pixel in every location. An automation of this process ensures the production of a dataset
including images from each summer (the timeframe with least snow ranges from mid July
to the end of September) with no clouds, shadows or snow under the datapoints.
To ensure resolution consistency across our training dataset, we included only the
data from Bands 2, 3, 4, 8 and 8a. Some bands, like 8a, have 20 m resolution. These needed
to be resampled to 10 m by increasing the number of pixels artificially, so that we have
the same resolution across the spectrum for the training. We also adopted a normalisation
procedure for the Sentinel-2A image arrays in these bands, which consisted of scaling down
the elements in each image array by 10,000 and applying a ceiling of 1, in order to lessen
the impact of any images with too many saturated pixels.
We also included in our training set the Normalized Difference Vegetation Index,
(NDVI). Vegetation reflects near-infrared wavelengths strongly while absorbing red light,
in such a way that this index is useful to measure the amount of vegetation in a given
location, and is known to be correlated with the level of land degradation [37].

2.1.3. Arctic Digital Elevation Models


In addition to satellite imagery, we included height data and slope data to train our
machine learning models. We expected some dependence of erosion risk on the height and
slope of a region, since vegetation growth is generally reduced at greater elevations due to
longer periods of snow cover, shorter growth periods and stronger winds primarily coming
from one side. A larger gradient would also be associated with an increased likelihood of
landslides or water erosion from eroding material falling down the slopes.
The height and slope data used for this study were originally produced by the National
Land Survey of Iceland and the Polar Geospatial Center of the University of Minnesota.
It derives from the Arctic Digital Elevation Models (DEMs), which is elevation data for
the Arctic region (latitude north of 60° N, which covers Iceland) that is publicly available
through the ArcticDEM project [38]. The DEMs were repeatedly acquired over a period
from 2012 to present day, with some DEMs dating back to 2008, through satellite sub-meter
stereo imagery (WorldView 1-3 and GeoEye-1) and processed using the open source digital
photogrammetric software, SETSM [39–41].
We obtained a modified, processed version of the original DEM data from the Google
Earth Engine community [42], selecting a dataset in the EPSG:3057 (ISN93) coordinate
system with 10 × 10 m resolution (resampled from the original 2 × 2 m resolution) as it
more closely matches the resolution of the Sentinel-2 bands.

2.1.4. Creating the Training Data


The Sentinel-2 data, Landgræðslan-provided soil erosion data and DEM data were
combined to form a training dataset to train and evaluate the ML models. The Sentinel-2
data and DEM data were used as inputs for the model, while the soil erosion data was used
as ground truth values.
To produce the inputs, we matched the Sentinel-2 data with the soil erosion data by
cropping square tiles of length 11 pixels centred around the coordinate points where the
soil erosion was measured by Landgræðslan, such that every pixel within the square was
labelled according to the erosion class of the central point, as illustrated in Figure 2. This
process was followed for each of the bands of interest, so that we had a set of 11 by 11 pixel
Remote Sens. 2023, 15, 482 6 of 20

tiles per band for each location. The DEM slopes and height data were also cropped to the
same 11 by 11 pixels shape and centred on each location with a soil erosion measurement,
where each pixel had a resolution of 10 m, matching the spatial resolution of some of the
bands of interest.

B2, B3, B4, B8, B8a, height, slope

B3 B8a

B4
height

B8
slope

Figure 2. Reconstruction of a data point with erosion value 0 and coordinates (398,402.355,
439,987.376) and the 11 × 11 surrounding pixels that were classified with that erosion value. Bands 2,
3, 4 are the RGB layers, other bands are also shown, as well as height and slope maps.

Each tile was then checked to remove those with blank data (black regions of null
pixel value), those with the same pixel value across the entire tile and those that are not of
size 11 by 11 pixels. The coordinate systems were all converted into EPSG:3057 coordinates
for consistency.
This resulted in a dataset that consisted of:
• 10,040 Sentinel-2A tiles of 11 by 11 pixels size across all 10 bands, spanning data
collected between 2 July 2019 and 24 September 2021, and with a resolution of 10 m
(resampled from a resolution of 20 m to 10 m as required),
• 742 DEM-based 11 by 11 pixel tiles representing height values in metres, with a
resolution of 10 m per pixel, equivalent to 110 by 110 m pixels,
• 742 DEM-derived 11 by 11 pixel tiles representing slope values in degrees, with a
resolution of 10 m per pixel, equivalent to 110 by 110 m regions,
• 553 datapoints in geopackage format with soil erosion classification between 0 (not
eroded) and 5 (highly eroded),
from which we created our training set. The reason for the greater number of Sentinel-2A
tiles per band than the number of DEM height and slope tiles is due to the overlapping of
satellite images, such that multiple images cover the same region.
Lastly, we removed the duplicate Sentinel-2 observations of the same objects (regions)
from the dataset to prevent bias from entering the model, resulting in 487 distinct objects.
In any case, given the noisy data we had to work with, this was insufficient to create a
strong, accurate model, so we split the pictures into single pixels and regarded the pixels
as little pictures representing 10 by 10 m regions for the classifier to learn to classify. This
resulted in a training set of 58,927 samples.
Remote Sens. 2023, 15, 482 7 of 20

The resulting dataset was somewhat imbalanced across the 6 classes of soil erosion, as
can be seen in Table 1. Class 0 has fewer samples than any other class, which we take into
account with stratification during the training process. Since the imbalance is small (i.e., the
largest ratio between classes is approximately 2.4:1), we do not apply sampling techniques
to create an evenly distributed dataset. The imbalance in the dataset is a reflection of the
current distribution in erosion level of Iceland. There are fewer regions in Iceland with
minimal risk/level of erosion compared to regions that are somewhat to severely eroded.

Table 1. Soil erosion classes in the training set.

Soil Erosion Class Description No. of Samples % of Dataset


0 No to low erosion risk 5687 9.7
1 Low erosion risk 12,342 20.9
2 Lower to moderate erosion risk 12,463 21.1
3 Moderate erosion risk 10,406 17.7
4 Moderate to higher erosion risk 8228 14.0
5 High erosion risk 9801 16.6

2.2. Algorithms
We trained three different machine learning algorithms to recognise the 6 different
levels of soil erosion identified in the previous section: an SVM model, a RF model and an
MLP model.
We applied a stratified k-fold approach for cross-validation to ensure that the models
were trained with a balanced training sample despite the imbalance present in our full
dataset. The training, test and validation sets were randomly sampled from the full dataset
provided as input to the models, following an 60:20:20 split, respectively. Each model was
trained to predict the soil erosion severity of specific regions of Iceland based on Sentinel-2
satellite images in several height, slope and the calculated NDVI of each region in multiple
wavebands. The training data used for each model was randomly sub-sampled from the
training data we detailed in Section 2.1.4.
For both models, we applied k-fold cross-validation with k = 5 to detect signs of
overfitting. As the training set was not well-balanced, we adopted a stratified approach
to splitting the training and testing data, and used a stratified k-fold approach for cross-
validation, which ensures the training and testing distributions reflect the distribution of
the complete dataset. We set aside 20 % of the original dataset for testing purposes, which
amounted to 47,141 training samples and 11,786 samples.

2.2.1. Support Vector Machine Model


The Support Vector Machine classification algorithm is a supervised machine learning
technique that works by determining the hyperplanes—also known as decision boundaries—
that best discern among the data categories. The hyperplanes that maximize the distances
between them and the closest data points within each category they separate are the best-
fitting ones. The SVM algorithm will find n-dimensional hyperplanes for a training set
with n features. For our classification problem, we chose to start with the SVM algorithm
because it is faster and performs better with smaller training sets than neural networks do.
Using the SVM algorithm that is included in the SCIKIT- LEARN Version 1.1.1 li-
brary [43], the SVM model was constructed in Python 3. A radial basis function kernel was
used as it produced the best results. We tested model performance across a wide range
of values for the parameters C and γ, which control the margin of error for the decision
boundaries and the similarity radius for data points in the same class, respectively. We
tested values between 0.101 and 99.999 for C, and between 0.001 and 9.999 for γ. Within
these intervals, C = 10.00 and γ = 0.50 were the values that produced the highest accuracy
Remote Sens. 2023, 15, 482 8 of 20

without being either too small or too big (as to avoid overfitting and maximise performance
at the same time).

2.2.2. Random Forest Model


The Random Forest algorithm is an ensemble learning algorithm that works by fitting
independent decision tree classifiers to sub-samples of the training dataset, then combines
the output of the decision trees to arrive at a single classification. The RF approach is suited
to a training set of noisy data, such as satellite image data, as it tends to reduce variance in
the output classification by combining the result of multiple decision tree classifiers.
The RF model was built in Python 3, using the RF algorithm available within the
SCIKIT- LEARN Version 1.1.1 library [43]. We used 125 estimators (decision tree classifiers)
for our model, a figure we arrived at after experimenting with the number of estimators,
ranging from 25 estimators to 200, and finding that the highest accuracy was produced
with 125 estimators.

2.2.3. Multilayer Perceptron Model


An MLP is a class of feedforward artificial neural networks composed of multiple
layers of perceptrons or nodes. The simplest form of an MLP consists of three layers of
nodes: an input layer, a hidden layer and an output layer, where each node is a neuron
with a non-linear activation function, and information is fed only in the forward direction
from the input layer through to the output layer.
We ran numerous experiments on different architectures for the model before settling
on a 4-layered model, which consists of an input layer with 256 nodes, 2 hidden layers with
128 and 64 nodes respectively, and an output layer with 6 nodes to match the number of
classes in the problem. MLPs are suited for classification problems and are very flexible,
and their relative simplicity compared to other neural network architectures allows them
to be used as a baseline point of comparison with other models.
In contrast to the normalisation approach we followed to produce the training data
for the SVM and RF models, we chose to use a min–max scaling technique to normalise the
satellite data, the height data and the slope data before feeding them in to the MLP model
for training. For each value in a given feature column of the training data, we used the
following simple formula to perform the min–max scaling,

X − Xmin
Xscaled = , (1)
Xmax − Xmin

where Xscaled is the scaled value, X is the original value, and Xmin and Xmax are the
minimum and maximum values of the original feature column, respectively. This method
of scaling places the data into a fixed range between 0 and 1, and was done to prevent the
occurrence of different scales for different features as well as large input values, which
negatively impact performance for deep learning models. We performed the scaling on
each of the satellite band values, the height values and the slope values. The NDVI values
were not re-scaled as they are by definition normalised to the range [−1, 1]. This scaling
was tested for SVM as well; however, it led to a loss of accuracy for that model.
For our loss function, we used categorical cross entropy as it is specifically suited to our
multi-class classification problem. For activation functions, we used the ReLU (Rectified
Linear Unit) function for both the input and hidden layers, and the SoftMax function for
the output layer. The optimiser chosen was the RMSprop optimiser with a learning rate of
0.001. We ran the training over 500 epochs, with 5 cross-validation folds, and with a batch
size of 64.
The MLP model was trained in Python 3, using the K ERAS software library. The
code was executed on Google Colaboratory, which provides free access to GPUs that were
required for training the model.
Remote Sens. 2023, 15, 482 9 of 20

3. Results
3.1. Model Performance Metrics
The overall accuracy achieved by the SVM model with C = 10.00 and γ = 0.50 was
0.92. In all of our cross-validation runs, there was little variation in the overall performance
of the model. The F1-scores and precision for each soil erosion class ranged from 0.89 to
0.95, while the recall ranged from 0.89 to 0.96, indicating that the vast majority of each class
is correctly being identified by the model and when the model predicts that a sample is of a
particular class, it is highly likely to be correct. The classes with the lowest F1-score and
precision were classes 2 and 3, while class 3 had the lowest recall. The best scoring class
across all three metrics was class 5.
The overall accuracy achieved by the RF model was 0.93, and again, there was little
variation in performance across all cross-validation runs. The F1-scores and precision for
each soil erosion class all ranged from 0.90 to 0.98, while the recall ranged from 0.89 to 0.97.
The lowest F1-scores occurred for classes 2 and 3, class 2 had the lowest precision and class
3 had the lowest recall. The best scoring class across all 3 metrics was again class 5.
The performance of the MLP model is significantly lower than that of the SVM and RF
models. It achieved an accuracy of 0.74, which exhibited only small variation across all the
k-folds. The model performed significantly worse in precision, recall and F1-score for class
3 than for other classes, while class 5 exhibited the highest precision, recall and F1-score,
consistent with the performance of the SVM and RF models. Unlike the other two models,
the MLP model did not show worse performance in identifying class 3 samples compared
to other classes. It achieved higher precision, recall and F1-scores for class 2 compared to
classes 0, 1, 3 and 4.
We trialled a range of different architectures for the MLP model and evaluated their
performance in order to identify the best performing MLP configuration. Specifically, we
experimented with different layer numbers, different node numbers in each layer and the
use of dropout layers to minimise overfitting.
The training and testing accuracies obtained with different model depths are shown in
Table 2. Our results demonstrate that despite increasing the depth of the MLP architecture
and the number of nodes in each layer, the accuracy does not increase much. Where
training accuracy is under 0.90, we can assume the model was underfit. Where the training
accuracies were high, the testing accuracies were not able to exceed 0.80. Interestingly,
extending the model to five layers (three hidden layers) only marginally improved the
testing accuracy, despite the number of extra nodes available. Deeper models would likely
be affected by overfitting.

Table 2. The training and testing accuracies for different MLP architectures.

No. of Nodes in No. of Nodes in No. of Nodes in No. of Nodes in


Training Accuracy Testing Accuracy
Input Layer 1st Hidden Layer 2nd Hidden Layer 3rd Hidden Layer
128 64 - - 0.85 0.65
256 128 - - 0.92 0.71
512 256 - - 0.96 0.74
8 4 2 - 0.49 0.44
16 8 4 - 0.48 0.48
32 16 8 - 0.69 0.58
64 32 16 - 0.84 0.66
128 64 32 - 0.91 0.72
256 128 64 - 0.96 0.74
256 128 64 32 0.97 0.75
512 256 128 64 0.98 0.77
Remote Sens. 2023, 15, 482 10 of 20

Table 3 shows the impact of including drop out layers to a three or four layer model
architecture, where the input layer has 256 nodes, the first hidden layer has 128 nodes and
the second hidden layer (if applicable) has 64 nodes. Dropout layers prune preceding layers
by removing nodes randomly to reduce the overfitting to the data, which we hypothesised
may improve the testing accuracy compared to the training accuracy. The results shown in
Table 3 indicate that the inclusion of drop out layers did not improve the testing accuracies
measured for each MLP configuration, and largely resulted in a decline in testing accuracy
if more than 10% of nodes were affected in each layer. However, for the instances where
10% drop out layers were applied, the training accuracy was reduced as the number of
drop out layers increased, which led to a reduction in the discrepancy between the training
and testing accuracies. For drop outs that exceeded 10%, the testing accuracies declined
along with the training accuracy. These accuracies and standard deviations were calculated
across 10 cross-validation folds.

Table 3. Training and testing accuracies for the 3- and 4-layer MLP architecture with dropout layers
(1 and 2 hidden layers, respectively).

1st Hidden 2nd Hidden


Input Layer Dropout Training Accuracy Testing Accuracy
Layer Dropout Layer Dropout
3-layer architecture
None None - 0.92 0.74
None 0.1 - 0.90 0.72
None 0.5 - 0.73 0.65
0.1 0.1 - 0.86 0.73
0.1 0.5 - 0.73 0.66
0.5 0.5 - 0.59 0.58
4-layer architecture
None None None 0.96 0.74
0.1 None None 0.95 0.74
0 0.1 None 0.93 0.74
0.1 0.1 None 0.90 0.74
0 0.1 0.1 0.90 0.72
0.1 0.1 0.1 0.85 0.72
0.5 0.5 0.5 0.54 0.53

The above results suggest that the MLP architecture is not well-suited to this particular
training set. We selected the simplest architecture that produced the highest testing accuracy
for comparison with SVM and RF, resulting in the mentioned 4 layer arrangement.
Since there is a small imbalance in the training data used to develop the models, we
also calculated Cohen’s κ statistic, defined by the formula

p0 − p e
κ= , (2)
1 − pe

to validate the other performance metrics we obtained. p0 is the overall model accuracy
and pe is the hypothetical probability of an agreement with ground truth values by random
chance, which we can calculate by summing the probabilities of predictions randomly
agreeing with ground truth values for each class, over all classes.
Remote Sens. 2023, 15, 482 11 of 20

This is an alternative metric used to determine the agreement between model predic-
tions and ground truth values that takes into account class distribution imbalances in the
training data [44]. Specifically, it compares the measured accuracy with the probability of
agreement with ground truth if predictions were done by random chance. κ has a range
between −1 and 1, where 0 represents the amount of agreement that can be expected from
random chance, and 1 represents perfect agreement between the ground truth and the
model predictions.
We find that Cohen’s κ statistic reaches κ = 0.90 for the SVM model, κ = 0.92 for
the RF model and κ = 0.68 for the MLP model. This means that the SVM and RF models’
predictions have reached very strong agreement with the ground truth classes, while the
MLP model has relatively weak agreement.
The key performance results for the SVM, RF and MLP models are displayed in Table 4.
The accuracy reported is the overall accuracy achieved when using the models to predict
on the test set, while the k-fold accuracy is the average overall accuracy (with standard
deviation) obtained during cross-validation across the 5 folds reserved for validation during
training. The macro recall, precision and F1-score are the unweighted averages of the recall,
precision and F1-scores achieved for each class.

Table 4. Performance metrics for the SVM, RF and MLP models.

k-Fold Macro
Method Accuracy Precision Recall
Accuracy F1-Score
SVM 0.92 0.919 ± 0.002 0.92 0.92 0.92
RF 0.93 0.932 ± 0.002 0.93 0.93 0.93
MLP 0.74 0.739 ± 0.005 0.73 0.74 0.73

3.2. Visual Representations


For illustrative purposes, we selected a small region to visualise the predictions for soil
erosion levels made by the models, shown in Figure 3. The predictions made by each model
are displayed in Figure 4. The erosion risk level is color-coded so that dark green represents
Class 0 (very low erosion risk) and dark red represents Class 5 (high erosion risk).

Figure 3. Test site location, with coordinates geolocated in EPSG:3057 ISN93.


Remote Sens. 2023, 15, 482 12 of 20

Figure 4. Predictions for soil erosion level made by the ML models: (a) MLP predictions (b) RF
predictions (c) SVM predictions. The erosion risk level is color-coded from Class 0 (very low erosion
risk) to Class 5 (high erosion risk).

There is significant disagreement in the predictions made by each model. The SVM
predictions are only partially consistent with what we expect of the underlying geography
of the test region. For example, parts of the vegetation areas marked by green in the test
region image were assigned an erosion risk level of 4, which is not expected for vegetative
land. We also find that the predictions for the elevated areas at the bottom and mid-right of
the test region seem to bleed across to the vegetative regions surrounding them, despite the
fact that these land types are quite different in nature. This is in contrast to the RF and MLP
predictions, which appear to assign different soil erosion risk levels to the distinct regions.
The SVM predictions are also quite varied within each distinct region. The vegetation
section in the middle of the test region is classed into three different levels of erosion risk.
The boundaries between these different classifications match small changes in the land
that are observable in the test region, which suggests that these small changes in land
characteristics were enough to cause the SVM model to assign different classifications to
them, while the RF and MLP models were more robust to these small changes. Furthermore,
Remote Sens. 2023, 15, 482 13 of 20

these variations within distinct regions are not always continuous variations. For example,
in the elevated land on the mid-right side of the test region, the SVM model shows variation
from Class 1 to 2, and then Class 2 to 4, while within the vegetation region in the middle of
the test region, we observe Class 1 transitioning directly into Class 4.
The RF model predictions exhibit far better matching with the underlying test region
geography, and display a more continuous variation in classification of erosion risk wher-
ever variations occur. However, the predictions are somewhat grainy in nature, which may
be due to noisiness in the training dataset. The model predicts a low level of soil erosion
risk for the vegetative patches and a medium to high risk of erosion for the elevated land,
as expected.
The MLP model predicts a highly uniform level of soil erosion risk across each distinct
sub-region. Although it predicts a low risk of erosion for the vegetation patches, the level
of risk it assigns is lower than what the RF model assigns. Due to the lack of subtlety and
variation in its predictions, we can only say that it generally agrees with the RF predictions
in distinguishing between high risk and low risk of erosion depending on land type. The
model fails to distinguish between more subtly different levels of erosion risk. The model
we trained does not appear to be as sensitive to these subtle changes in features that the RF
model is sensitive to, while the SVM model could potentially be too sensitive.
This is not surprising as the testing accuracy we obtained for the MLP model was only
0.74; it is clear that the MLP model was underfit. In contrast, the SVM model achieved a
very high testing accuracy, but when applied to an unseen patch of land, its predictions did
not match well with expectations. This is indicative of a model that was overfit to the data
we trained it on. Out of the three models, the RF model exhibits the best performance in
predicting level of soil erosion risk.

3.3. Dependence of Erosion Risk on Height, Slope and NDVI


To test whether the risk of erosion within a region is correlated with its height, slope
or NDVI, we calculated and plotted box-and-whisker distributions for each of the above
characteristics, separated by erosion risk class and using all samples in the dataset described
in Section 2.1.4. Our results are displayed in Figures 5–7.

Figure 5. The height distribution across all regions in the training dataset, separated by soil erosion
risk class.
Remote Sens. 2023, 15, 482 14 of 20

Figure 6. The slope distribution across all regions in the training dataset, separated by soil erosion
risk class.

Figure 7. The distribution of NDVI values across all regions in the training dataset, separated by soil
erosion risk class.

The box-and-whisker plots can be understood as follows: the median is indicated by the
central orange line within the box, the box marks the values within the interquartile range
(IQR = Q3 − Q1), the whiskers mark the values for Q1 − 1.5 × IQR and Q3 + 1.5 × IQR, and
the crosses represent the values that lie outside the range [ Q1 − 1.5 × IQR, Q3 + 1.5 × IQR],
which may be considered outliers depending on the distribution.
Our results show that there is no clear correlation between height or slope and level of
erosion. Within each erosion class, the distributions are very similar. The regions of greatest
height (>800 m) all have the highest level of erosion, but otherwise there is no preferred soil
erosion class at any height range. There is a small indication that a steeper area is more likely
to be severely eroded than flatter areas; we observe that Class 5 contains many more regions
that have a slope of >30 degrees than any of the other classes (including classes 1 and 2), while
Remote Sens. 2023, 15, 482 15 of 20

all regions within Class 0 have slopes under 17 degrees. However, classes 1 and 2 contain
some of the steepest regions within the dataset, and neither the whiskers nor the medians of
the slope distributions shift towards higher slope values with increasing erosion level.
These results indicate that while there is a theoretical motivation for a correlation to
exist between soil erosion and height/slope, any such correlation is not clearly visible in
the data, perhaps due to other contributing factors that more strongly impact soil quality.
On the other hand, Figure 7 shows that NDVI value is negatively correlated with
level of erosion risk, as expected. The risk of erosion rises with decreasing median NDVI
(NDV I = 0.71 for Class 0 and NDV I = 0.08 for Class 5. In the classes representing highest
risk of erosion, the majority of samples have NDVIs closer to 0. Similarly, within the classes
representing lowest erosion risk, the bulk of samples have NDVIs above 0.50. Out of all
of the classes, the NDVI seems to be best at differentiating regions of severe erosion risk
(Class 5), since the distribution within this class is fairly narrow and the vast majority
of NDVIs occur below 0.25. As the samples within all other classes are more broadly
distributed and there is considerable overlap in NDVI values between two or more these
classes, the NDVI is less effective at distinguishing between them, particularly if its value
falls between 0.25 and 0.75.
Interestingly, there are more NDVI values within Class 1 than Class 0 that fall between
0.75 and 1, so an NDVI in this range has greatest probability of being part of Class 1, rather
than Class 0. There is even a small chance that a region with NDVI close to 1 is actually
severely eroded, as class 5 contains some possible outliers that go against the general trend
of low NDVIs.
Since vegetation is expected to reflect NIR wavelengths while absorbing red wave-
lengths of light, and therefore produce higher values of NDVI, it is not surprising that
the higher values of NDVI are more likely to be classed as very low to moderate erosion
risk. The fact that regions of moderate to high erosion risk can also exhibit higher values
of NDVI suggests that there are other geographies other than vegetation that reflect NIR
while absorbing red light, or that some vegetative regions are at risk of erosion.

4. Discussion
The results presented in this paper demonstrate that machine learning models are
capable of accurately predicting level of soil erosion risk in Iceland based on satellite
observations of the regions of interest and height and slope information. We find that
the SVM and RF models achieve very high accuracy, precision and recall when tested on
samples from the original training dataset, and the level of performance per class is quite
consistent across all classes. In contrast, the MLP model achieves a relatively low accuracy,
precision and recall.
The visual maps of the model predictions in Section 3.3 show that despite the high
performance of the SVM model as measured on the testing set, the quality of its predictions
on the test region is far lower than that of the RF model. This suggests that the SVM model
does not generalise well to unseen data, and is certainly overfit to the training data. This
is consistent with expectations, as the SVM algorithm tends to be sensitive to noise and
outliers. As we observed in Figure 4, subtle and small differences in the land characteristics
can cause the model to assign different risks of erosion to otherwise similar areas. Based
on our observation, this is the main reason of over or underfitting the models. On the
other hands, from soil science point of view, relying on just two topographic layers for soil
erosion modelling does not make sense and thus layer significance in this case will not
affect the model development and performance improvement. The RF algorithm does not
have the same weakness, as it is an ensemble technique that relies on the averaging out of
many independent classification models. This leads to both higher prediction accuracy and
smaller likelihood of overfitting to meaningless noise. The SVM algorithm does employ
regularisation principles to prevent overfitting, but it remains less robust to noise than
RF. This is supported by other studies that have shown that RF outperforms SVM [45],
Remote Sens. 2023, 15, 482 16 of 20

particularly as it is fairly insensitive to overfitting where there is high data dimensionality


and potential multicollinearity [46].
The predictions of the MLP model on the test region are also poor, but this is primarily
due to underfitting by the MLP model rather than overfitting. However, there is another
possible cause relating to how the min–max scaling was performed. For the test region, we
used Equation (1) to perform our scaling, which is consistent with how the training data
was scaled prior to training the MLP model. However, the scaling factors used to scale the
test region data and the original training data may be significantly different, due to the
fact that the training data consists of pixels from hundreds of regions while the test region
data consists of pixels from only the test region. The minimum and maximum values of
both datasets may differ significantly, and result in a less accurate representation of the
MLP model’s predictions on the test region. Identifying a suitable global minimum and
maximum independent of the input data, by which to perform the min–max scaling, may
result in a better set of predictions, though these predictions would still be affected by the
low predictive power of the model.
The poorer predictive power of the MLP model could indicate that the general MLP
architecture is not well-suited to the training data we used to train it, especially as changing
the specific configuration of the MLP model did not significantly increase the testing
accuracy. If a deep learning approach is preferred for this problem, then a Convolutional
Neural Network (CNN) may be a better choice. Alternatively, the MLP model may train
better if the data is kept as full 11 by 11 pixel tiles rather than split into individual pixels.
Another likely factor contributing to the lower accuracy and generalisability of the
SVM model and the underfitting of the MLP model could simply be an excess of noise in
the original training set, in the form of inaccurately classified pixels. The approach we used
to create the training set required splitting the original 11 by 11 pixel tiles, each of which
was labelled under a single class. Splitting the tiles into individual pixels and treating each
pixel as a single data sample may have introduced a significant proportion of inaccurately
labelled pixels into training, since the original 11 by 11 pixel tiles are very unlikely to all
have the same ground truth level of soil erosion risk. The most accurately labelled pixels
would be the ones closest to the centre of the 11 by 11 pixel tiles, while the pixels at the
edge may be off by one (or more) class(es).
The fact that the RF model predictions (and to an extent, the SVM model predictions)
are quite grainy may also be due to the inaccurately labelled pixels in the training set,
particularly as the ‘grains’ seem to appear within areas where the predicted class differs
by one. However, since RF is expected to be quite robust to noise, this may not be the
primary contributing factor. A more likely explanation may be the model’s sensitivity to
the DEM data inputs; there is a chance that the grainy pattern is correlated with the DEM
data. Applying a smoothing filter to the predictions can address the issue of graininess to
improve the classification.
Despite the issues highlighted above, our results indicate that there is strong potential
for using remote sensing combined with machine learning to identify land areas that are
at risk of soil erosion. Our results support the recommendation of RF over SVM and
MLP for predicting the level of soil erosion risk based on satellite observations. Although
CNNs have been shown to exhibit better performance than SVM and RF in remote sensing
applications [47,48], our results demonstrate that RF is still a good choice of model for
such applications. The RF algorithm is capable of producing highly accurate predictions
and has the advantage of being much more easily implementable and interpretable than
deep learning algorithms. For certain problems in remote sensing, such as where training
samples or computing resources are limited, or where simplicity and training speed are
important, SVM and RF models may be better options than deep learning models.
To further improve the models, particularly the performance of SVM, it is most
important to reduce the number of inaccurately labelled pixels in the training set, perhaps by
changing our pixel by pixel approach to use smaller tiles, or by implementing a weighting
that varies with distance from the centre of the tile. Moreover, in our present study, we
Remote Sens. 2023, 15, 482 17 of 20

have only utilised one spectral index (the NDVI) out of many indices that may provide
additional predictive power to our models. It is worth experimenting with other indices in
future to enhance their performances.
The extension of our most viable algorithm to a nationwide prediction study would
result in the first soil erosion map of the country since the map from RALA of 1997,
published in [4] and then later in [49]. A comparison would be of great interest.
Regarding conducting an ecosystem assessment of our results, this is challenging task
since it would demand a more extensive analysis of historical data. Historic classification
is limited by the quality and availability of historic imagery, and long term monitoring of
ecosystem properties throughout the Arctic from satellite information is still largely absent
to this date [50], particularly in Iceland. Quantitative analyses such as the ones described
in this study can open a new avenue for the understanding of landscape change dynamics.
Furthermore, the ecosystem function can differ greatly between land cover types,
suggesting that further input parameters may need to be added in order to draw conclusions
about the interdependence of the processes involved in long term observations. As a case
in point, our study has focused on the NDVI index, which is susceptible to trends in
precipitation-evapotranspiration. The presence of surface water absorbs light in the near
infrared, thereby reducing this index while landscape drying increases it [51].
Changes in ecosystem type also alter interactions between climate and permafrost,
which may evolve in response to climate change. Accurate prediction of the permafrost
carbon climate feedback will require detailed understanding of changes in terrestrial
ecosystem distribution and function, which depend on the net effects of multiple feedback
processes operating across scales in space and time [52].
Further improvement could also come from testing topographic derivatives other than
slope and purely elevation values, such as aspect (due to wind erosion), texture, terrain
ruggedness, wetness indices, etc.

5. Conclusions
The aim of this study was to investigate the suitability of using ML algorithms with
satellite imagery as inputs to assess the risk of soil erosion in Iceland. Specifically, we
evaluated the performance and potential of an SVM model, a RF model and an MLP model
in classifying soil into one of six classes of erosion risk.
To conduct the study, we created a training set composed of cropped geo-referenced
and atmospherically corrected Copernicus Sentinel-2 images of Iceland (normalised satellite
image arrays in the Bands 2, 3, 4, 8 and 8a) and Arctic DEM-derived heights and slopes.
We also calculated the NDVI index as an additional input to the models, and showed that
it is correlated with soil erosion risk, while height and slope are not. For our ground truth
values, we used measurements of soil erosion in Iceland provided by the Icelandic Soil
Conservation Services.
We found that both the SVM and RF models achieved very high testing accuracies,
F1-scores, precisions and recalls, based on evaluations completed on a reserved portion
of the training set, while the MLP model performed comparatively worse on the same
measures. The SVM model achieved an average accuracy of 0.917 ± 0.001, the RF model
achieved an average accuracy of 0.931 ± 0.002 and the MLP model achieved an accuracy of
0.739 ± 0.005, across the five folds. However, when testing the predictions of the models on
an unseen region in Iceland, we found that the neither the SVM model nor the MLP model
performs as well as the RF model. The SVM model predictions do not match the underlying
geography as well as the RF model predictions, and the model generally exhibits too much
sensitivity to minor and irrelevant changes in the land characteristics. In contrast, the MLP
model predictions match the underlying geography reasonably well, but the model was
not sensitive enough to changes in land features that were linked to soil erosion.
We conclude that the SVM model is overfit and does not generalise well to unseen
data, while the MLP model is underfit, suggesting that the underlying model architecture
is not suited to this problem. Out of all the models, RF exhibited the best performance and
Remote Sens. 2023, 15, 482 18 of 20

appears to generalise well to new data. The performance of the RF model shows that simpler
machine learning algorithms can still produce accurate classifications for soil erosion and
may be viable alternatives to more complex and resource-intensive implementations such
as deep learning models.
The appropriateness of erosion concepts commonly employed in model structures
is still questionable. Despite the fact that physical processes of detachment, transport,
and deposition in overland flow are well recognized and have been widely incorporated
within erosion models, the experimental procedures to test conditions when processes are
occurring concurrently have only recently been developed. Our initiative of using machine
learning approaches therefore proves to be a promising alternative for erosion prediction
in which it overcomes obstacles in parametrization, calibration, and validation processes
that are often considered to be the main difficulties while applying conventional models.
The goal of the research described in this work intends to produce a reliable, widely
applicable, and cost-effective method to classify Icelandic soils into various categories
of erosion risk. A proof of concept that, once engineered, could be easily expanded and
applied to other Arctic regions, such as Greenland and Canada.

Author Contributions: All authors acknowledge equal contribution to this study as part of the Fléttan
collaboration. D.F. was responsible for leading the manuscript efforts, and writing the introduction,
methodology and conclusion section of the manuscript. E.A. was responsible for developing and
evaluating the SVM, RF and MLP models and writing the methodology, results, discussion and
conclusion section of the manuscript. M.P. was responsible for creating the training dataset out of the
satellite data, the soil erosion data provided by Landgræðslan and the DEM data, and developing the
MLP model. R.P. was responsible for obtaining and preprocessing the satellite data from Sentinel-2.
C.G.R. was the originator of the project and was responsible for leading the project efforts. A.T. was
responsible for guiding and advising the team during this study, especially with model development
and evaluation. All authors have read and agreed to the published version of the manuscript.
Funding: This research was funded by the Icelandic Technology Development Fund (Rannís) grant
number 2215789-0611.
Data Availability Statement: Not applicable.
Acknowledgments: Authors wish to thank Landgræðslan, the Soil Conservation Service of Iceland,
for their collaboration with the project and for providing us with the data about existing soil erosion
that is being used as a reference for our model. We also wish to thank the CASSINI program
for awarding us the first place at the 2nd CASSINI Hackathon Iceland and enrolling us in their
mentorship program by experts who are helping us develop our project both on the business and on
the technological side.
Conflicts of Interest: The authors declare no conflict of interest.

Abbreviations
The following abbreviations are used in this manuscript:

AUI Agricultural University of Iceland


CNN Convolutional Neural Network
DEM Digital Elevation Model
DIAS Data and Information Access Services
EPSG European Petroleum Survey Group
FAO Food and Agriculture Organization
GPU Graphics Processing Unit
ML Machine Learning
MLP Multilayer Perceptron
NDVI Normalized Difference Vegetation Index
NIR Near Infra-Red
ReLU Rectified Linear Unit
Remote Sens. 2023, 15, 482 19 of 20

RF Random Forest
SETSM Surface Extraction with TIN-based Search space Minimization
SVM Support Vector Machine
UTM Universal Transverse Mercator
WEkEO WE knowledge Earth Observation

References
1. FAO. Status of the World’s Soil Resources: Main Report; Food and Agriculture Organization of the United Nations and Intergovern-
mental Technical Panel on Soils: Rome, Italy, 2015.
2. Verheijen, F.; Jones, R.; Rickson, R.; Smith, C. Tolerable versus actual soil erosion rates in Europe. Earth-Sci. Rev. 2009, 94, 23–38.
[CrossRef]
3. Runólfsson, S. Landgræðslan í 80 ár 1907–1987 [Eighty Years of Soil Conservation Service of Iceland in 1907–1987]; Árbók Landgræðs-
lunnar: Græðum Ísland, 1988; pp. 99–124.
4. Arnalds, O.; Thorarinsdottir, E.; Metúsalemsson, S.; Jonsson, A.; Gretarsson, E.; Arnason, A. Soil Erosion in Iceland; Soil
Conservation Service and Agricultural Research Institute: Reykjavik, Iceland, 2001.
5. Johannesson, B. The Soils of Iceland. University Research Institute, Reykjavik 1960. 140 sider. Et kort. Geografisk Tidsskrift 1961,
60. Available online: https://tidsskrift.dk/geografisktidsskrift/article/view/46690 (accessed on 21 November 2022).
6. Gudmundsson, T. The FAO Classification System Adapted to Icelandic Conditions; Rala Report; Agricultural Research Institute:
Reykjavik, Iceland, 1994; Volume 167.
7. Bartoli, F.; Buurman, P.; Delvaux, B.; Madeira, M. Volcanic soils: Properties and processes as a function of soil genesis and land
use. Geoderma 2003, 117, 183–184. [CrossRef]
8. Arnalds, O.; Stahr, K. Volcanic soil resources: Occurrence, development, and properties. Catena 2004, 56, 1–2. [CrossRef]
9. UNEP. Iceland Country Profile. In Country Profiles Series; UNEP: Washington, DC, USA, 2002.
10. Jones, A.; Stolbovoy, V.; Tarnocai, C.; Broll, G.; Spaargaren, O.; Montanarella, L. Soil Atlas of the Northern Circumpolar Region;
European Commission: Brussels, Belgium, 2009. [CrossRef]
11. Tarnocai, C.; Canadell, J.G.; Schuur, E.A.G.; Kuhry, P.; Mazhitova, G.; Zimov, S. Soil organic carbon pools in the northern
circumpolar permafrost region. Glob. Biogeochem. Cycles 2009, 23. [CrossRef]
12. Vonk, J.E.; Sánchez-García, L.; van Dongen, B.E.; Alling, V.; Kosmach, D.; Charkin, A.; Semiletov, I.P.; Dudarev, O.V.; Shakhova, N.;
Roos, P.; et al. Activation of old carbon by erosion of coastal and subsea permafrost in Arctic Siberia. Nature 2012, 489, 137–140.
[CrossRef] [PubMed]
13. Phiri, D.; Simwanda, M.; Salekin, S.; Nyirenda, V.R.; Murayama, Y.; Ranagalage, M. Sentinel-2 Data for Land Cover/Use Mapping:
A Review. Remote Sens. 2020, 12, 2291. [CrossRef]
14. Žížala, D.; Juřicová, A.; Zádorová, T.; Zelenková, K.; Minařík, R. Mapping soil degradation using remote sensing data and
ancillary data: South-East Moravia, Czech Republic. Eur. J. Remote Sens. 2019, 52, 108–122. [CrossRef]
15. Huffman, E.; Eilers, R.; Padbury, G.; Wall, G.; MacDonald, K. Canadian agri-environmental indicators related to land quality:
integrating census and biophysical data to estimate soil cover, wind erosion and soil salinity. Agric. Ecosyst. Environ. 2000,
81, 113–123. [CrossRef]
16. Arabameri, A.; Pradhan, B.; Pourghasemi, H.; Rezaei, K. Identification of erosion-prone areas using different multi-criteria
decision-making techniques and GIS. Geomat. Nat. Hazards Risk 2018, 9, 1129–1155. [CrossRef]
17. Davidson, S.J.; Santos, M.J.; Sloan, V.L.; Watts, J.D.; Phoenix, G.K.; Oechel, W.C.; Zona, D. Mapping Arctic Tundra Vegetation
Communities Using Field Spectroscopy and Multispectral Satellite Data in North Alaska, USA. Remote Sens. 2016, 8, 978.
[CrossRef]
18. Pathak, P.; Whalen, S. Using Geospatial Techniques to Analyze Landscape Factors Controlling Ionic Composition of Arctic Lakes,
Toolik Lake Region, Alaska. Int. J. Appl. Geospat. Res. 2012, 3, 37–57. [CrossRef]
19. Lin, D.H.; Johnson, D.R.; Andresen, C.; Tweedie, C.E. High spatial resolution decade-time scale land cover change at multiple
locations in the Beringian Arctic (1948–2000s). Environ. Res. Lett. 2012, 7, 025502. [CrossRef]
20. Boulze, H.; Korosov, A.; Brajard, J. Classification of Sea Ice Types in Sentinel-1 SAR Data Using Convolutional Neural Networks.
Remote Sens. 2020, 12, 2165. [CrossRef]
21. Mars, J.; Houseknecht, D. Quantitative remote sensing study indicates doubling of coastal erosion rate in past 50 yr along a
segment of the Arctic coast of Alaska. Geology 2007, 35, 583–586. [CrossRef]
22. Wang, J.; Li, D.; Cao, W.; Lou, X.; Shi, A.; Zhang, H. Remote Sensing Analysis of Erosion in Arctic Coastal Areas of Alaska and
Eastern Siberia. Remote Sens. 2022, 14, 589. [CrossRef]
23. Jones, B.M.; Farquharson, L.M.; Baughman, C.A.; Buzard, R.M.; Arp, C.D.; Grosse, G.; Bull, D.L.; Günther, F.; Nitze, I.;
Urban, F.; et al. A decade of remotely sensed observations highlight complex processes linked to coastal permafrost bluff erosion
in the Arctic. Environ. Res. Lett. 2018, 13, 115001. [CrossRef]
24. Philipp, M.; Dietz, A.; Ullmann, T.; Kuenzer, C. Automated Extraction of Annual Erosion Rates for Arctic Permafrost Coasts
Using Sentinel-1, Deep Learning, and Change Vector Analysis. Remote Sens. 2022, 14, 3656. [CrossRef]
25. Arnalds, O.; Barkarson, B. Soil erosion and land use policy in Iceland in relation to sheep grazing and government subsidies.
Environ. Sci. Policy 2003, 6, 105–113. [CrossRef]
Remote Sens. 2023, 15, 482 20 of 20

26. Arnalds, O. Volcanic soils of Iceland. Catena 2004, 56, 3–20. [CrossRef]
27. Arnalds, A. Ecosystem Disturbance in Iceland. Arct. Alp. Res. 1987, 19, 508–513. [CrossRef]
28. Sigurdsson, B.D.; Snorrason, A. Carbon sequestration by afforestation and revegetation as a means of limiting net-CO2 emissions
in Iceland. Biotechnol. Agron. Soc. Environ. 2000, 4, 303–307.
29. Ása L. Aradóttir.; Petursdóttir, T.; Halldorsson, G.; Svavarsdóttir, K.; Arnalds, Ó. Drivers of Ecological Restoration: Lessons from
a Century of Restoration in Iceland. Ecol. Soc. 2013, 18, 33.
30. Runólfsson, S. Land Reclamation in Iceland. Arct. Alp. Res. 1987, 19, 514–517. [CrossRef]
31. Novelli, A.; Aguilar, M.A.; Nemmaoui, A.; Aguilar, F.J.; Tarantino, E. Performance evaluation of object based greenhouse detection
from Sentinel-2 MSI and Landsat 8 OLI data: A case study from Almería (Spain). Int. J. Appl. Earth Obs. Geoinf. 2016, 52, 403–411.
[CrossRef]
32. Immitzer, M.; Vuolo, F.; Atzberger, C. First Experience with Sentinel-2 Data for Crop and Tree Species Classifications in Central
Europe. Remote Sens. 2016, 8, 166. [CrossRef]
33. Vuolo, F.; Neuwirth, M.; Immitzer, M.; Atzberger, C.; Ng, W.T. How much does multi-temporal Sentinel-2 data improve crop type
classification? Int. J. Appl. Earth Obs. Geoinf. 2018, 72, 122–130. [CrossRef]
34. Main-Knorn, M.; Pflug, B.; Louis, J.; Debaecker, V.; Müller-Wilm, U.; Gascon, F. Sen2Cor for Sentinel-2. In Proceedings of the
Image and Signal Processing for Remote Sensing XXIII, Warsaw, Poland, 11–13 September 2017; p. 3.
35. Heinrich, V.; Krauss, G.; Goetze, C.; Sandow, C. IDB—Index DataBase. 2012. Available online: https://www.indexdatabase.de/
db/is.php?sensor_id=96 (accessed on 5 September 2022).
36. Sentinel-2 SR Cloudless Browser. Available online: https://showcase.earthengine.app/view/s2-sr-browser-s2cloudless-nb
(accessed on 13 November 2022).
37. Yengoh, G.T.; Dent, D.; Olsson, L.; Tengberg, A.E.; Tucker, C.J. Applications of NDVI for Land Degradation Assessment. In Use of
the Normalized Difference Vegetation Index (NDVI) to Assess Land Degradation at Multiple Scales: Current Status, Future Trends, and
Practical Considerations; Springer International Publishing: Cham, Switzerland, 2016; pp. 17–25. [CrossRef]
38. Porter, C.; Morin, P.; Howat, I.; Noh, M.J.; Bates, B.; Peterman, K.; Keesey, S.; Schlenk, M.; Gardiner, J.; Tomko, K.; et al. ArcticDEM.
2018. Available online: https://doi.org/10.7910/DVN/OHHUKH (accessed on 21 November 2022).
39. Noh, M.J.; Howat, I.M. Automated stereo-photogrammetric DEM generation at high latitudes: Surface Extraction with TIN-based
Search-space Minimization (SETSM) validation and demonstration over glaciated regions. GISci. Remote Sens. 2015, 52, 198–217.
[CrossRef]
40. Noh, M.J.; Howat, I.M. The Surface Extraction from TIN based Search-space Minimization (SETSM) algorithm. ISPRS J.
Photogramm. Remote Sens. 2017, 129, 55–76. [CrossRef]
41. Noh, M.J.; Howat, I.M. Automatic relative RPC image model bias compensation through hierarchical image matching for
improving DEM quality. ISPRS J. Photogramm. Remote Sens. 2018, 136, 120–133. [CrossRef]
42. Roy, S. Awesome Gee Community Datasets. 2022. Available online: https://samapriya.github.io/awesome-gee-community-
datasets/projects/iceland_dem (accessed on 21 November 2022).
43. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.;
Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830.
44. Kohen, J. A coefficient of agreement for nominal scale. Educ. Psychol. Meas. 1960, 20, 37–46. [CrossRef]
45. Fernández-Delgado, M.; Cernadas, E.; Barro, S.; Amorim, D. Do we need hundreds of classifiers to solve real world classification
problems? J. Mach. Learn. Res. 2014, 15, 3133–3181.
46. Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm.
Remote Sens. 2016, 114, 24–31. [CrossRef]
47. Heydari, S.S.; Mountrakis, G. Effect of classifier selection, reference sample size, reference class distribution and scene hetero-
geneity in per-pixel classification accuracy using 26 Landsat sites. Remote Sens. Environ. 2018, 204, 648–658. [CrossRef]
48. Heydari, S.S.; Mountrakis, G. Meta-analysis of deep neural networks in remote sensing: A comparative study of mono-temporal
classification to support vector machines. ISPRS J. Photogramm. Remote Sens. 2019, 152, 192–210. [CrossRef]
49. Petursdóttir, T.; Arnalds, O.; Baker, S.; Montanarella, L.; Aradóttir, Á. A Social-Ecological System Approach to Analyze
Stakeholders’ Interactions within a Large-Scale Rangeland Restoration Program. Ecol. Soc. 2013, 18. [CrossRef]
50. Callaghan, T.V.; Tweedie, C.E.; Åkerman, J.; Andrews, C.; Bergstedt, J.; Butler, M.G.; Christensen, T.R.; Cooley, D.; Dahlberg, U.;
Danby, R.K.; et al. Multi-decadal changes in tundra environments and ecosystems: synthesis of the International Polar Year-Back
to the Future Project (IPY-BTF). Ambio 2011, 40, 705–716. [CrossRef]
51. Goswami, S.; Gamon, J.A.; Tweedie, C.E. Surface hydrology of an arctic ecosystem: Multiscale analysis of a flooding and draining
experiment using spectral reflectance. J. Geophys. Res. Biogeosci. 2011, 116. [CrossRef]
52. Loranty, M.M.; Abbott, B.W.; Blok, D.; Douglas, T.A.; Epstein, H.E.; Forbes, B.C.; Jones, B.M.; Kholodov, A.L.; Kropp, H.;
Malhotra, A.; et al. Reviews and syntheses: Changing ecosystem influences on soil thermal regimes in northern high-latitude
permafrost regions. Biogeosciences 2018, 15, 5287–5313. [CrossRef]

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.

You might also like