mathematics-12-00026-v2
mathematics-12-00026-v2
mathematics-12-00026-v2
Article
Optimal Tilt-Wing eVTOL Takeoff Trajectory Prediction Using
Regression Generative Adversarial Networks
Shuan-Tai Yeh 1 and Xiaosong Du 2, *
Abstract: Electric vertical takeoff and landing (eVTOL) aircraft have attracted tremendous attention
nowadays due to their flexible maneuverability, precise control, cost efficiency, and low noise. The
optimal takeoff trajectory design is a key component of cost-effective and passenger-friendly eVTOL
systems. However, conventional design optimization is typically computationally prohibitive due
to the adoption of high-fidelity simulation models in an iterative manner. Machine learning (ML)
allows rapid decision making; however, new ML surrogate modeling architectures and strategies are
still desired to address large-scale problems. Therefore, we showcase a novel regression generative
adversarial network (regGAN) surrogate for fast interactive optimal takeoff trajectory predictions of
eVTOL aircraft. The regGAN leverages generative adversarial network architectures for regression
tasks with a combined loss function of a mean squared error (MSE) loss and an adversarial binary
cross-entropy (BC) loss. Moreover, we introduce a surrogate-based inverse mapping concept into
eVTOL optimal trajectory designs for the first time. In particular, an inverse-mapping surrogate
takes design requirements (including design constraints and flight condition parameters) as input
and directly predicts optimal trajectory designs, with no need to run design optimizations once
trained. We demonstrated the regGAN on optimal takeoff trajectory designs for the Airbus A3
Vahana. The results revealed that regGAN outperformed reference surrogate strategies, including
multi-output Gaussian processes and conditional generative adversarial network surrogates, by
matching simulation-based ground truth with 99.6% relative testing accuracy using 1000 training
Citation: Yeh, S.-T.; Du, X. Optimal
Tilt-Wing eVTOL Takeoff Trajectory
samples. A parametric study showed that a regGAN surrogate with an MSE weight of one and a
Prediction Using Regression BC weight of 0.01 consistently achieved over 99.5% accuracy (denoting negligible predictive errors)
Generative Adversarial Networks. using 400 training samples, while other regGAN models require at least 800 samples.
Mathematics 2024, 12, 26. https://
doi.org/10.3390/math12010026 Keywords: eVTOL; optimal takeoff trajectory design; machine learning; surrogate modeling; genera-
tive adversarial networks; regGAN; inverse mapping
Academic Editor: Daniel-Ioan
Curiac
MSC: 93C85
Received: 29 November 2023
Revised: 15 December 2023
Accepted: 19 December 2023
Published: 21 December 2023 1. Introduction
The attention on electric vertical takeoff and landing (eVTOL) aircraft has grown
significantly because eVTOL is suitable for missions that require flexible maneuverability
Copyright: © 2023 by the authors.
and precise control, such as urban air mobility (UAM) [1,2]. UAM represents a safe
Licensee MDPI, Basel, Switzerland. and efficient air transportation system where everything from small package delivery
This article is an open access article drones using unmanned aerial vehicles (UAVs) to passenger-carrying air taxis through
distributed under the terms and eVTOL aircraft can operate above populated areas [3,4]. UAVs have broad applications,
conditions of the Creative Commons including the sixth-generation communication using UAVs and an aerial base station for
Attribution (CC BY) license (https:// supporting the Internet of Things deployment in remote and disaster areas [5]. Hua et al. [6]
creativecommons.org/licenses/by/ proposed 3D non-stationary modeling for UAV-to-ground communications and managed
4.0/). to demonstrate its validity and practicality against measured results. eVTOL aircraft’s
unique characteristics of precise delivery, lower cost, and reduced noise have motivated
significant developments [7,8]. Conceptions under development can be categorized into a
few major aircraft types, including lift+cruise (such as Aurora Flight Sciences eVTOL [9])
and tilt-wing (such as Airbus A3 Vahana [10]).
Among the aforementioned eVTOL types, the tilt-wing configuration enables the
aircraft to combine the flexibility of a helicopter for vertical takeoff and landing with the
efficiency of airplanes during cruising. Thus, the transition optimization on tilt-wing
eVTOL aircraft attracts special interest since transition has major effects on the success of
flight tasks [11]. Pradeep and Wei [12,13] put an emphasis on the formulation of multiphase
optimal control problems with energy consumption index for tilt-wing and multirotor
eVTOL vehicles. The proposed multiphase optimal control problem formulation and the
numerical solution allowed an eVTOL air taxi to fulfill the specified required arrival time
while attaining the most energy-efficient trajectory for arrival. This capability played a vital
role in enabling safe and efficient future operations of eVTOL aircraft, facilitating passenger
transportation and cargo delivery. Chauhan and Martins [14] constructed an Airbus
A3 Vahana [10] model and optimized its takeoff-to-cruise trajectory with the objective
of minimizing energy consumption. They concluded that the optimal takeoff trajectory
involved stalling the wings or flying near the stall angle of attack. Moreover, in the
absence of acceleration constraints, the optimized trajectories involved a rapid transition to
forward flight, followed by climbing at a relatively constant speed, and then accelerating
to the desired cruising speed. When considering passenger comfort and incorporating
acceleration constraints, the transition, climb, and acceleration phases exhibited more
gradual and less distinct behavior, as expected. However, completed work on transition
optimization as well as conventional engineering design optimizations rely on iterative
simulation model evaluations, which prohibit real-time decision making.
Thus, surrogate models have emerged as an effective alternative for fast interactive
decision making in lieu of time-consuming simulation models [15,16]. Surrogate-based
design optimization is a methodology that leverages surrogate models to approximate the
behavior of design candidates. To elaborate, surrogate models are mathematical models
that are trained using a limited number of data points obtained from computationally
expensive simulations or physical experiments. These surrogate models are then used in
optimization algorithms to efficiently make predictions, explore design space, and find
optimal solutions. Surrogate-based design optimization is advantageous over conventional
simulation-based optimization for computational efficiency [17,18], the effective exploration
of high-dimensional design spaces [19,20], multi-objective optimization capabilities [21,22],
sensitivity analysis [23], and uncertainty analysis [24–26].
The Gaussian process (GP) [27] is one of the most commonly used surrogate models
due to its capability and flexibility for modeling and predicting unknown functions based
on observed data [28]. The concept of a multi-output GP (MOGP) has emerged to further
handle problems with multiple outputs [29,30]. By extending the concept of a GP, an
MOGP is a surrogate modeling technique that allows for the joint modeling of multiple
correlated outputs and the simultaneous capture of their dependencies. In particular,
by jointly modeling the outputs by estimating the covariance matrix (also known as the
correlation function), an MOGP leverages the shared information among the outputs,
which reduces the overall modeling complexity. For more mathematical and surrogate
modeling details, please refer to Section 2.3. Meanwhile, deep learning surrogates are going
through revolutionary developments and pushing forward cutting-edge research in the
design optimization community. Thelen et al. [31] computed design variable derivatives
by analytically linking objective definition, mesh, and geometry, and they demonstrated
the derivatives through the multi-fidelity Broyden–Fletcher–Goldfarb–Shannon algorithm
for high-dimensional aerodynamic and aeroelastic design optimizations. Tao and Sun [32]
developed a multi-fidelity surrogate-based optimization framework based on deep belief
networks. The results for airfoil and wing designs under uncertainty indicated that the
multi-fidelity surrogate model performed well by significantly improving optimization
Mathematics 2024, 12, 26 3 of 25
lack of research study in this field. Third, we introduce the inverse mapping concept (from
design requirements, including design constraints and flight condition parameters, directly
into optimal designs) to eVTOL trajectory design for the first time.
We organize the rest of the paper as follows. Section 2 introduces the optimization
framework and simulation models used in this work, followed by MOGP, cGAN, and
regGAN setups for surrogate modeling. We demonstrate the regGAN as well as other
surrogates on eVTOL optimal takeoff trajectory predictions in Section 3. We end this paper
with conclusions in Section 4.
2. Methodology
In this section, we introduce the open-source optimization toolbox Dymos (https://gith
ub.com/OpenMDAO/RevHack2020/tree/master/problems/evtol_trajectory/evtol_dymo
s, accessed on 1 August 2022) [42] within OpenMDAO (https://github.com/OpenMDAO/Op
enMDAO, accessed on 1 August 2022) [43] and simulation models. Additionally, we detail
an MOGP toolkit (MOGPTK (https://github.com/GAMES-UChile/mogptk, accessed on 1
August 2022)) [44] followed by DNN and GAN series models (namely, GAN, cGAN, and
regGAN) within TensorFlow [45] for surrogate modeling.
in the following, such as design variables (initial time t0 , duration t p ), state variables at
the state discretization nodes (x̄d ), dynamic control variables at the control discretization
¯ Herman and Conway proposed the high-order
nodes (ūd ), and the design parameters (d).
LGL method, which is the primary method used in OTIS [46] and is a general extension
of the Hermite-Simpson collocation to higher orders [47]. In a phase, the segments are
discretized at the LGL nodes so that, for each segment, the total number of nodes is an
odd value. The state discretization nodes are the nodes with an even index, where the
index starts at zero for the first node. Dynamic controls are provided at designated control
discretization nodes. In LGL collocation, all nodes within the phase are involved in the
control discretization subset. The evaluation of an LGL phase in Dymos progresses with the
¯ are given
following steps: Initially, the values of the design variables (t0 , t p , x̄d , ūd , and d)
by the optimizer and user. Secondly, the OpenMDAO system offering the state dynamics is
assessed at the state discretization nodes, contributing the state time derivatives at the state
discretization nodes:
tseg
x̄c = [ Ai ] xd + [ Bi ] x̄˙ d , (2)
2
2
x̄c′ = [ A ] x + [ Bd ] x̄˙ d , (3)
seg d d
where tseg is the duration of the segment where each nodes originates. At the collocation
nodes, the state system provided by the system at this point is evaluated and is given by
t̄seg
¯ = x̄c′ −
∆ x̄˙ . (5)
2 col
2.2.1. Aerodynamics
The aerodynamics model assumes that the forward and rear wings are equivalent and
have the same reference area for simplicity, and there is no flow interaction between the
two wings. The rotations of the two wings are assumed to be the same, so that the angles
of attack are equivalent as well as the lift and drag generated by the two wings. During the
transition from vertical to horizontal flight, separated-flow conditions are considered. A
model developed by Tangler and Ostowari [49] for wing aerodynamics is implemented to
predict the lift and drag beyond the linear-lift region. The poststall lift coefficient is
cos2 α
CL = A1 sin2α + A2 , (6)
sinα
Mathematics 2024, 12, 26 6 of 25
where
C1
A1 = , (7)
2
sinαs
A2 = (CLs − C1 sinαs cosαs ) , (8)
cos2 αs
and
C1 = 1.1 + 0.018AR, (9)
where α is the wing angle of attack (in radians), αs is the angle of attack at stall (in radians),
CLs is the lift coefficient at stall, and AR is the wing aspect ratio.
The drag coefficient at a wing angle of attack between 27.5 and 90 degrees is given by
where
B1 = CDmax , (11)
CDs − CDmax sinαs
B2 = , (12)
cosαs
and
1.0 + 0.065AR
CDmax = , (13)
0.9 + t/c
where CDs is the drag coefficient at stall, and t/c is the airfoil thickness-to-chord ratio. For
the poststall drag coefficient at a wing angle of attack below 27.5 degrees, the equation is
where α is in radians.
The well-known finite wing corrections from lifting-line theory for unswept wings in
incompressible flow are utilized to modify the lift and drag of the airfoil prior to stalling,
αairfoil
αwing = , (15)
1 + (αairfoil /(π · AR · e))
where αwing is the finite-wing lift–curve slope, αairfoil is the airfoil lift–curve slope, and e is
the span efficiency factor. The prestall lift curve is assumed to be linear, and the wing stall
angle of attack is 15 deg.
In order to obtain the total drag of the wing before stall, induced drag has been
added using the well-known formula based on the lifting-line theory for prestall parasite
drag coefficients,
CL2
C Di = , (16)
π · AR · e
where CDi is the induced drag coefficient, CL is the wing’s lift coefficient, and AR = 8 for
each wing of the configuration.
Additional drag on an assumed drag area for the fixed landing gear and the fuselage
is assumed to be independent of the freestream angle of attack. The induced drag for the
configuration can be expressed as
L2wing L2wing
! !
Dinduced = 2 1.4 =2 , (17)
πqb2 · 0.95 πqb2 · 0.68
where Lwing is lift per wing, q is freestream dynamic pressure, b is wing span.
Mathematics 2024, 12, 26 7 of 25
2.2.2. Propulsion
Momentum theory [50] has been utilized to compute thrust from the propellers as a
function of power.
s
2
V∞⊥
V T
Pdisk = TV∞⊥ + κT − ∞⊥ + + , (18)
2 4 2ρAdisk
where Pdisk is the power supplied to the propeller disk excluding profile power, T is the
thrust, V∞⊥ is the freestream velocity component normal to the propeller disk, ρ is the
air density, Adisk is the disk area of the propeller, and κ is the correction factor utilized to
incorporate induced power losses associated with non-uniform flow, tip effects, and other
simplifications made in momentum theory (κ = 1 for ideal power). Power is used as a
design variable in the optimization problems, and the Newton–Raphson method [51] has
been chosen to solve the nonlinear equation for thrust with power as an input.
Applying blade-element theory to a rotor operating in nonaxial forward flight [52,53]
estimates the profile power coefficient CPp and profile power Pp
σCd0p
CPp = (1 + 4.6µ2 ), (19)
8
and
Pp
CPp = , (20)
ρAdisk R3 Ω3
where R is the radius of the propeller, Ω is the angular speed, σ is the solidity, Cd0p is a
representative constant profile drag coefficient, and µ is defined as
V∞∥
µ= , (21)
ΩR
where V∞∥ is the freestream velocity component parallel to the disk. For our cases, we
assume that Ω = 181 rad/s for R = 0.75 m, σ = 0.13, and Cd0p = 0.012.
With electrical power as an input, we use a factor kelec to account for electrical and me-
chanical losses related to batteries, electrical systems, and motors. Pdisk can be expressed as
where q⊥ is the dynamic pressure based on the freestream velocity component normal to
the propeller disk, β is the blade pitch angle at 0.75R and is assumed to change linearly
from 10 deg at a flight speed of 0 m/s to 35 deg at a cruise speed of 67 m/s, Adisk is the
propeller area, and αin is the incidence angle. The remaining terms are calculated in the
following manner. The effective solidity σe is
2Bcb
σe = , (24)
3πR
where B is the number of blades per propeller, cb is the average chord length of the blades,
and R is the propeller radius. The thruster factor f is
Mathematics 2024, 12, 26 8 of 25
√
1 + Tc − 1 Tc
f = 1+ + , (25)
2 4(2 + Tc )
T
Tc = , (26)
q⊥ Adisk
2.2.4. Dynamics
A two degrees of freedom (DOF) representation and the forward Euler method [54] are
used for simulating the trajectory of the aircraft. The horizontal and vertical components of
the aircraft velocity are solved as functions of time, considering the control variables, which
include the wing-tilt angle and electrical power. The horizontal component of velocity at
each time step is calculated as follows:
T sin θ − Dfuse sin(θ + α∞ ) − Dwings sin(θ + αEFS ) − Lwings cos(θ + αEFS ) − N cos θ
v x i +1 = v x i + ∆t, (28)
m
where i is index of a time step, ∆t is a time step length, θ is a wing angle relative to the
vertical, α∞ is a freestream angle of attack, αEFS is the effective freestream angle of attack
experienced by the wings due to the propeller influence, m is the mass of the aircraft, T is
the total thrust, Dfuse is the drag of the fuselage, Dwings is the total drag of the two wings,
Lwings is the total lift of the two wings, and N is the total normal force generated by the
propellers. Figure 1 displays the forces and the angles on the aircraft. Likewise, the vertical
component of velocity at each time step is calculated as follows:
T cos θ − Dfuse cos(θ + α∞ ) − Dwings cos(θ + αEFS ) + Lwings sin(θ + αEFS ) + N sin θ − mg
v y i +1 = v y i + ∆t, (29)
m
where g is the gravitational acceleration. To determine the time step length, ∆t, the total
takeoff time, which is a design variable in optimization problem formulation, is divided by
500. A convergence study concluded the optimized takeoff flight time would be typically
below 50 s, resulting in adequately small time steps of less than 0.1 s [14].
Mathematics 2024, 12, 26 9 of 25
Lrear y
Nrear
Trear
kinvi x
Ve ective freestream
Lfwd
Drear V∞
Nfwd
V∞ θ Tfwd
α∞
Dfuse αEFS Ve ective freestream
V∞
Dfwd
kinvi
ff mg
In other words, the mean function represents the average function value at input x across
all functions in the distribution. To simplify posterior computations and rely only on the
covariance function for inference, the prior mean function is frequently assumed to be a
constant. In this work, we apply the constant mean function (m(x) = b1 ) and the linear
mean function (m(x) = ax + b2 ), where a, b1 , and b2 are arbitrary constant coefficients.
The covariance function k(x, x′ ) captures the relationship between the function values at
different input points x and x′ :
k x, x′ = E ( f (x) − m(x)) f x′ − m x′ .
(32)
The function k is commonly called the kernel of the Gaussian process. The kernel defines
the shape and characteristics of the GP model. We implement the Matérn kernel and the
square exponential kernel function in our study, and they are described as follows.
1. Matérn kernel
The Matérn covariance between two points with the distance τ = |x − x′ | is
1− ν √ √
22 τ τ
k ν (τ ) = σ 2ν Kν 2ν , (33)
Γ(ν) ρ ρ
where Γ is the gamma function, Kν is the modified Bessel function of the second kind,
and ρ and ν are non-negative covariance parameters. The Matérn kernel is stationary
since the covariance only depends on distances between points.
2. Squared exponential kernel
|x − x′ |
2 1
k SE = s exp − , (34)
2 λ
Mathematics 2024, 12, 26 10 of 25
where s is the scale factor. In the case where we use a single length scale parameter,
the squared exponential (SE) kernel is an example of a radial basis function.
The GP has been extended to multiple series (or channels), which are referred to
as MOGP. The key difference lies in their implementation and capabilities. The GP is
designed for single-output regression, predicting a single target variable given inputs,
while an MOGP extends the GP to handle multiple outputs simultaneously. An MOGP
is useful when dealing with correlated outputs, enabling the modeling of dependencies
between different target variables. The complexity of a GP relates to the number of data
points, while an MOGP introduces additional complexity due to modeling correlations
between multiple outputs, but the MOGP remains efficient and more capable compared
with training individual GPs for each output. Thus, a key feature of the MOGP is to harness
applicable information across outputs to provide more accurate predictions than separately
modeling correlated outputs [28].
In this study, the multi-output spectral mixture [56] kernel is used as the multivariate
kernel for the MOGP model, and can be expressed as
Q
(q) ⊤ (q)
(q) ⊤ (q)
1
∑ ij
(q) (q) (q)
k ij (τ ) = α exp − τ + θ ij Σ ij τ + θ ij cos τ + θ ij µ ij + ϕij , (35)
q =1
2
o = σ ( W · h + b ), (36)
where h is a data vector from previous layers, σ (·) represents an activation function used
in the current layer, and o is an output vector of the current layer. During training, W
and b are adjusted to optimize the model’s performance. We utilize two commonly used
activation functions in this work to introduce nonlinearity into the DNN. The rectified
Mathematics 2024, 12, 26 11 of 25
linear unit (relu) [59] activation function follows the equation σrelu (h) = max(0, h),while
the sigmoid [60] activation function has an expression of σsigmoid (h) = 1/ 1 + e−h .
In this work, we harness two types of loss functions for training DNN surrogates to
minimize the difference between predicted responses and corresponding actual values.
The MSE [61] loss function is formulated as
N
1
LMSE =
N ∑ (|vi − v̂i |2 )2 , (37)
i =1
where N is the number of training data points in the data set, | · |2 is L2 norm, vi and v̂i
are actual values and predicted values for the ith training sample. Binary crossentropy
(BC) [62] is formulated as
N
1
LBC = −
N ∑ (vi · log v̂i + (1 − vi ) · log(1 − v̂i )), (38)
i =1
where vi is the ith actual binary label (0 or 1), and v̂i is the predicted probability (between 0
and 1). The loss function penalizes the model more severely when it predicts the opposite
class with high confidence. In this work, the DNN unknown parameters are tuned by
minimizing loss functions through the gradient-based Adam optimizer (see the following
section) enabled by backpropagation within Tensorflow [45].
where
∂L
mt = βmt−1 + (1 − β) , (40)
∂wt
where the subscripts t − 1, t, and t + 1 are the indices of previous, current, and next
optimization steps, respectively. Initially, m0 = 0, w is a weight to be determined, α
∂L
is learning rate, ∂w t
is the derivative of loss function with respect to weight at current
optimization step, and β is a constant moving average parameter.
RMSP [64] is an adaptive learning algorithm with a mathematical formulation as
∂L 2
α
w t +1 = w t − , (41)
(vt + ϵ)1/2 ∂wt
where
∂L
vt = βvt−1 + (1 − β) . (42)
∂wt
A small positive constant (ϵ = 10−7 ) is utilized. The remaining variables remain the same
as in Equation (40).
The Adam optimizer integrates the moment and RMSP algorithms as follows.
α
wt+1 = wt − m̂t √ , (43)
v̂t + ϵ
where
mt vt
m̂t = , v̂t = , (44)
1 − βt1 1 − βt2
where mt and vt follow the updating process as described above.
Mathematics 2024, 12, 26 12 of 25
0:z 1:e
Figure 2. GAN contains the discriminator to compete with the generator. The generator generates
shapes (g) based on random variables (z), typically following user-defined distributions (uniform
distribution in this work). The discriminator distinguishes between the existing data (e) and the
generated data (g). During training, the discriminator adjusts its weights (wd ) to make the probability
p g of g approach zero’s while increasing the probability pe of e towards one’s. Conversely, during
training, the generator adjusts its weights (w g ) to increase the probability of p g towards one’s. The
generator, after training, generates shapes similar to the existing data.
min max LGAN ( D, G ) = Ex∼ Pdata [log D (x)] + Ez∼ Pz [log(1 − D ( G (z)))], (45)
G D
where x is sampled from the existing data distribution Pdata , z is sampled from the noise
variable distribution Pz , and G and D are the generator and discriminator. Thus, a trained
GAN model generates reasonable designs with ample shape variability within prior noise
variable distributions.
forms, such as text descriptions and class labels, depending on the utilization. The condi-
tioning is achieved by providing y as an additional input layer to both the discriminator
and generator. Within the generator, the prior input noise Pz (z) and y are combined in a
joint hidden representation, benefiting from the adversarial training framework’s flexibility
in composing this hidden representation. In addition, the discriminator takes as inputs
x and y pairs as well as shapes generated by the generator to complete the adversarial
competition with the generator. Similarly, as a GAN loss function (Equation (45)), the
training on loss functions could also be expressed as a minimax problem,
min max LcGAN ( D, G ) = Ex∼ Pdata (x) [log D (x|y)] + Ez∼ Pz (z) [log(1 − D ( G (z|y)))]. (46)
G D
In addition, cGANs can be adapted for regression tasks by encoding the regression
labels (i.e., model observations) as additional conditioning information (Figure 3). The
cGAN generator takes random noise variables as the original GAN, together with arbitrary
model observations as an additional input group. The goal is to generate data samples
that correspond to the model observations. The discriminator, which tries to distinguish
between real and generated data samples, will now take both real data samples x and
generated data samples g as input. During training, the cGAN generator learns to map
the random noise variables and model observations to corresponding data samples, which
helps in regression tasks by generating samples consistent with the given target values.
This approach offers a way to tackle problems with complex relationships between inputs
and targets, and incorporate the adversarial feature of GANs into regression processes. In
this work, x represents the eVTOL optimal takeoff trajectory control points to be predicted,
y includes design requirements, and we set z with only one element and assign a Uniform
(0, 1) distribution to z. During regression tasks, an average of predictions over a fixed set
of y and 100 randomly sampled z is used as the predicted model response [39], which in
this work, is the optimal takeoff trajectory (x). The network architecture settings for the
generator and discriminator within the cGAN follow the same architecture as the regGAN,
which will be discussed in Section 2.5.3.
0 : z, y 1 : x, y
Figure 3. cGAN has similar structures to GAN (shown in Figure 2). cGANs take design requirements
(y) and random noise variables (z) as input for the generator network. The time-sequence optimal
takeoff trajectories (x) and generated data g are considered inputs for the discriminator.
where wBC and wMSE are constant weights on BC and MSE loss functions; we conduct a
parametric study in Section 3 for the best predictive performance.
The MSE loss (also known as context loss) guides the regGAN predictions to match
model observations in a similar way as traditional surrogates, while the BC loss (also known
as adversarial loss due to the competitive training on GAN models) on the discriminator
engenders similar patterns between predicted results and observations. In this work, we
leverage the regGAN for eVTOL takeoff trajectory inverse mapping, i.e., directly from
design requirements to optimal takeoff trajectories. Thus, the regGAN generator reads
the design requirements (y) and predicts the corresponding optimal takeoff trajectories
(x). Meanwhile, the discriminator attempts to distinguish between predicted/generated
trajectories and actual optimal takeoff trajectories. Thus, the nature of the eVTOL optimal
takeoff trajectory prediction makes it necessary to develop and introduce the regGAN
surrogate. First, the regGAN makes use of the predictive capability of the DNN via
the generator. Second, the optimal takeoff trajectory profiles typically follow realistic
patterns. For instance, optimal power and wing angle profiles follow ascending trends in
general, meaning that the battery provides more and more power to balance the weight
and gradually moves forward by turning the wing from vertical to horizontal positions. In
this way, the BC loss based on the GAN architecture facilitates the training by automatically
filtering out unrealistic trajectory profiles.
TrainingGen,MSE Time-sequence
LMSE
design variables
TrainingGen,BC TrainingDis,BC
Figure 4. regGAN includes a generator and a discriminator as the original GAN and cGAN. The
generator generates time-sequence design variables based on corresponding design requirements
through DNN. The discriminator takes generated optimal designs and true time-sequence optimal
trajectories in order to distinguish the differences. This results in adversarial competition, which
causes an adversarial loss function (using binary crossentropy, LBC ) to train the generator and the
discriminator. Additionally, we minimize the mean square error between generated and true design
variables (LMSE ) to explicitly train the generator for regression tasks.
The regGAN generator architecture setting is as follows (Table 1). The generator
consists of one input layer, two hidden layers, and one output layer. Five neurons in the
input layer correspond to five different design requirements. Both hidden layers have
100 neurons, followed by relu activation functions. The output layer has twenty-one
Mathematics 2024, 12, 26 15 of 25
Table 1. Neural architectures and training setups of the generator and the discriminator within
regGAN.
Generator Discriminator
Input layer 5 neurons, no activation 21 neurons, no activation
Hidden layer 1 100 neurons, relu activation 100 neurons, relu activation
Hidden layer 2 100 neurons, relu activation 100 neurons, relu activation
Output layer 21 neurons/1 neuron, sigmoid activation 1 neurons, sigmoid activation
Training algorithm Adam optimizer Adam optimizer
Training parameters β 1 = 0.9, β 2 = 0.999 β 1 = 0.9, β 2 = 0.999
Learning rate 0.001 0.001
Batch size 20 20
Epochs 1000 1000
where Ntest is the number of testing samples, ytrue,i and ypred,i are the true observations and
predicted values of the ith set of design requirements. The relative accuracy is calculated as
ACCL1 = 1 − ϵ̄ L1 . (52)
The L1 relative accuracy measures the average relative match between predicted
values and true observations, expressed in percentage. It gives an indication of to what
extend surrogate predictions agree with true observations relative to true observations.
The percentage exhibits deeper insights through the errors relative to model response
magnitudes.
1.5 300
2.75
power (105)
1.0 200
theta
2.50
y
0.5 100
2.25
0
0 5 10 15 20 25 0 5 10 15 20 25 0 200 400 600 800
time time x
6 0.2 8
energy (106)
thrust (103)
4 0.1
aoa
6
2 0.0
0 4
0 5 10 15 20 25 0 5 10 15 20 25 30 0 5 10 15 20 25
time time time
0.30 1.0
0.25
0.5 0.05
acc
cd
0.20
cl
0.15 0.0
0.00
0 5 10 15 20 25 30 0 5 10 15 20 25 0 5 10 15 20 25
time time time
Dymos simulation Verify
Figure 5. We verified results with Dymos package. The blue line represents the direct result of the
outputs from the Dymos optimization data by NASA, and the black dot line presents the results by
Chauhan and Martins [14] as verification data. The units for each y axis are given by the following:
power (W), theta (rad), y (m), energy (J), aoa (rad), thrust (N), acc (g), cl (-), cd (-).
Table 3. MOGP model conditions and mean ± standard deviation of testing accuracy on testing data
set. SE stands for SE kernel function.
We chose an arbitrary case to further reveal the predictive performance of each MOGP
surrogate. Table 4 shows the design requirements for the visualization case. Table 5 shows
the testing accuracy on design variables for MOGP models 1–4. We compare the predicted
optimal trajectory profiles by MOGP models 1–4 in Figure 6, where all models achieve simi-
lar predictive accuracies. The testing accuracies, shown in Table 5, are all over 90%; however,
visualization exhibits obvious discrepancies between surrogate-based and simulation-based
optimal trajectories. The results indicate that MOGP surrogates intend to use higher power
early but not the maximum power. The early power usage results in higher thrust and
greater acceleration in the early takeoff phase by MOGP surrogates than by the simulation-
based optimal design. The acceleration by the surrogate-based optimal design in the early
phase even violates the maximum acceleration constraint. So, we conclude that the MOGP
cannot realize a sufficient accuracy level using 1000 training samples.
Mathematics 2024, 12, 26 18 of 25
1.5
3.0
power (105)
1.0 200
theta
2.8
y
2.6 0.5
0
0 5 10 15 20 25 0 5 10 15 20 25 0 200 400 600 800
time time x
7.5 0.2 10.0
energy (106)
thrust (103)
5.0
7.5
aoa
0.1
2.5
5.0
0.0 0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
0.6 1.0 0.06
0.4 0.04
0.5
acc
cd
cl
0.2 0.02
0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
True value, Dymos SE, const SE, linear Matern, const Matern, linear
Figure 6. Optimized takeoff trajectory profile for MOGP models 1–4 verified against simulation-based
ground truth.
as well as the addition of the discriminator to drive prediction matching training data set
shapes. However, there are still some unexpected wiggles when we look at the acceleration
profile. In addition, the cGAN approximates ground truth based on an average of a number
of predictions (100 predictions in this work) over the same set of input parameters (design
requirements in this work), where the prediction may vary slightly due to Monte Carlo
properties. Hence, we develop and introduce the regGAN surrogate for further predictive
improvement as follows.
Table 6. cGAN models exhibit better predictive performance over MOGP in terms of mean ± standard
deviation in testing accuracy of design variables.
ϵ̄L1 , tflight
Model Loss Function ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
(%)
cGANBC BC 98.4 ± 0.858 99.0 ± 0.653 98.6 ± 0.424
cGANMSE MSE 98.3 ± 0.843 98.8 ± 0.745 98.7 ± 0.426
Table 7. Testing accuracy of visualized case for cGAN model BC and MSE.
Model Loss Function ϵ̄L1 , tflight (%) ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
cGANBC BC 98.8 99.8 98.7
cGANMSE MSE 98.9 99.7 98.7
1.5
3.0
power (105)
y
2.6 0.5
0
0 5 10 15 20 25 0 5 10 15 20 25 0 200 400 600 800
time time x
7.5 0.2 10.0
energy (106)
thrust (103)
5.0
7.5
aoa
0.1
2.5
5.0
0.0 0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
0.6 1.0 0.06
0.4 0.04
0.5
acc
cd
cl
0.2 0.02
0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
True value, Dymos MOGP, Model 2 cGAN, Model BC cGAN, Model MSE
Figure 7. Optimal takeoff trajectory profile comparison on cGAN model BC and MSE and MOGP
model 2 with simulation-based ground truth.
Table 8. Mean ± standard deviation of testing accuracy for regGAN model CL1.
ϵ̄L1 , tflight
Model wMSE wBC ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
(%)
CL1 1 0 99.5 ± 0.246 99.7 ± 0.196 99.7 ± 0.117
Table 9. Testing accuracy of visualization case for regGAN model BC and MSE.
ϵ̄L1 , tflight
Model wMSE wBC ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
(%)
CL1 1 0 99.4 99.8 99.8
1.5
3.0
power (105)
y
2.6 0.5
0
0 5 10 15 20 25 0 5 10 15 20 25 0 200 400 600 800
time time x
7.5 0.2 10.0
energy (106)
thrust (103)
5.0
7.5
aoa
0.1
2.5
5.0
0.0 0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
0.6 1.0 0.06
0.4 0.04
0.5
acc
cd
cl
0.2 0.02
0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
True value, Dymos MOGP, Model 2 cGAN, Model BC cGAN, Model MSE regGAN, Model CL1
Figure 8. Optimal takeoff trajectory profile for regGAN model CL1 matches the ground truth and
outperforms cGAN models BC and MSE and MOGP model 2.
Table 10. regGAN model conditions and mean ± standard deviation of testing accuracy for design
variables.
ϵ̄L1 , tflight
Model wMSE wBC ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
(%)
CL2 1 0.01 99.7 ± 2.388 99.7 ± 0.118 99.7 ± 0.123
CL3 1 0.001 99.7 ± 0.193 99.7 ± 0.160 99.7 ± 0.124
CL4 1 0.0001 99.6 ± 0.216 99.7 ± 0.145 99.7 ± 0.128
Mathematics 2024, 12, 26 21 of 25
Table 11. Testing accuracy for visualization case of regGAN model CL2–4.
ϵ̄L1 , tflight
Model wMSE wBC ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
(%)
CL2 1 0.01 99.6 99.9 99.8
CL3 1 0.001 99.9 99.8 99.8
CL4 1 0.0001 99.9 99.9 99.5
1.5 300
3.0
power (105)
1.0 200
2.8 theta
y
2.6 100
0.5
0
0 5 10 15 20 25 0 5 10 15 20 25 0 100 200 300 400 500 600 700 800
time time x
7.5 0.2
8
energy (106)
thrust (103)
5.0
aoa
0.1 6
2.5
4
0.0 0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
0.4 1.0
0.06
0.3 0.04
0.5
acc
cd
cl
0.2
0.02
0.1 0.0
0 5 10 15 20 25 0 5 10 15 20 25 0 5 10 15 20 25
time time time
True value, Dymos regGAN, Model CL1 regGAN, Model CL2 regGAN, Model CL3 regGAN, Model CL4
Figure 9. Optimal takeoff trajectory profiles predicted by regGAN models CL1–4 match well with
simulation-based ground truth.
We also conducted a parametric study for different combined loss weights to further
investigate regGAN performance. We use a different number of training samples, ranging
from 50, 100, 200, 400, 600, 800, to 1000. Table 12 indicates that using 100 and 200 training
samples is able to obtain the above with an overall 97% accuracy. Note that when provided
with 50 training samples, CL1 has the lowest mean testing accuracy and highest standard
deviation time tflight prediction, while CL2, CL3, and CL4 have better overall predictive
performance. Moreover, regGAN model CL2 (wMSE = 1 and wBC = 0.01) achieves over
99.5% accuracy starting with 400 training samples, while all the other regGAN surrogates
require at least 800 training samples. In addition, regGAN model CL2 consistently shows
lower predictive standard deviations (such as 0.251%, 0.257%, and 0.145% on tflight , P,
and θ) on the testing data set, further confirming a better and more robust predictive
performance. Based on the visualization case, a mean testing accuracy of over 99.5% can be
considered sufficiently accurate with negligible differences.
In sum, the MOGP surrogates could not match ground truths or capture the general
trend of optimal takeoff trajectories well using 1000 training samples, which may be due
to the Gaussian assumption of GP series models. The cGAN surrogates achieve better
performance over MOGP using the 1000 training samples, with a closer match towards the
general trend of ground truth labels, mainly because of the guidance of the discriminator
for generating similar data patterns as the training data set. The regGAN surrogates
outperform MOGP and cGAN since regGAN is trained with a combined loss function of
MSE and BC adversarial losses. The BC adversarial loss leads regGAN to handle the general
trend of observations, while the MSE loss directly drives the match between predictions
and observations.
Mathematics 2024, 12, 26 22 of 25
Table 12. Parametric study for regGAN models exhibits deeper insights on mean ± standard
deviation of testing accuracy with respect to the number of training samples for each design variable
group. Note that we only vary wBC while keeping wMSE as 1.
Model wBC Samples ϵ̄L1 , tflight (%) ϵ̄L1 , P (%) ϵ̄L1 , θ (%)
CL1 0 50 95.2 ± 0.279 97.4 ± 1.58 98.3 ± 0.758
100 98.6 ± 0.906 98.5 ± 1.05 98.9 ± 0.414
200 99.3 ± 0.368 99.2 ± 0.461 99.4 ± 0.212
400 99.5 ± 0.314 99.5 ± 0.389 99.6 ± 0.167
600 99.2 ± 0.477 99.4 ± 0.294 99.6 ± 0.148
800 99.7 ± 0.281 99.7 ± 0.146 99.7 ± 0.119
1000 99.5 ± 0.246 99.7 ± 0.196 99.7 ± 0.117
CL2 0.01 50 97.4 ± 1.51 97.7 ± 0.852 97.6 ± 0.955
100 97.4 ± 0.732 98.6 ± 0.829 98.9 ± 0.434
200 99.3 ± 0.443 97.8 ± 0.980 98.9 ± 0.427
400 99.6 ± 0.251 99.6 ± 0.257 99.6 ± 0.145
600 99.6 ± 0.204 99.7 ± 0.215 99.6 ± 0.153
800 99.5 ± 0.314 99.7 ± 0.158 99.6 ± 0.147
1000 99.7 ± 0.239 99.7 ± 0.118 99.7 ± 0.123
CL3 0.001 50 95.3 ± 2.77 97.7 ± 1.29 98.0 ± 0.759
100 97.2 ± 1.72 96.4 ± 2.08 96.6 ± 1.31
200 99.3 ± 0.380 99.0 ± 0.557 99.1 ± 0.308
400 99.6 ± 0.343 99.4 ± 0.378 99.4 ± 0.230
600 99.7 ± 0.228 98.9 ± 0.767 99.4 ± 0.239
800 99.5 ± 0.228 99.7 ± 0.194 99.7 ± 0.127
1000 99.7 ± 0.193 99.7 ± 0.160 99.7 ± 0.124
CL4 0.0001 50 97.1 ± 1.64 98.5 ± 1.11 96.4 ± 1.78
100 98.5 ± 0.757 98.9 ± 0.648 97.7 ± 0.847
200 99.2 ± 0.384 98.8 ± 0.706 98.9 ± 0.494
400 99.4 ± 0.346 99.4 ± 0.358 99.3 ± 0.249
600 99.2 ± 0.611 99.3 ± 0.509 99.4 ± 0.195
800 99.6 ± 0.287 99.5 ± 0.363 99.5 ± 0.182
1000 99.6 ± 0.216 99.7 ± 0.145 99.7 ± 0.128
4. Conclusions
In this paper, we investigated surrogate-based optimal takeoff trajectory predictions
for electric vertical takeoff and landing (eVTOL) drones within the scope of urban air
mobility. We developed the regression generative adversarial network (regGAN), which
outperformed the Gaussian process (MOGP) and the conditional generative adversarial
network (cGAN) by achieving over 99.5% accuracy. We summarize the main contribution
of this work as follows.
First, we implemented the surrogate-based inverse mapping concept into eVTOL
optimal trajectory design for the first time. Specifically, surrogate models took design
requirements as input and predicted optimal trajectories. We realized fast interactive eVTOL
takeoff trajectory design without running any optimizations since the trained surrogates
directly predicted optimal trajectories. However, reducing training costs is essential since
each training sample requires a simulation-based trajectory design optimization.
Second, we introduced the MOGP, a representative traditional surrogate model, into
the eVTOL takeoff trajectory design for rapid predictions. The results showed that the
MOGP with a square exponential kernel function could accurately capture the inverse
mapping using 1000 training samples. We then implemented a cGAN for eVTOL inverse
mapping since cGAN also makes use of the GAN architecture for regression tasks. The
results revealed that the cGAN achieved over 98% generalization accuracy in predicting
optimal designs using the same 1000 training samples as the MOGP, which means the
cGAN outperformed the MOGP on predictive performance (around 92% accuracy). In
addition, the visualization case verified that the cGAN could match the general trend of
optimal designs well with actual observations from Dymos but missed detailed features.
Mathematics 2024, 12, 26 23 of 25
Third, we introduced the regGAN into takeoff trajectory design for the first time
and achieved over 99.6% accuracy in predicting optimal design variables with the same
1000 training samples as the MOGP and cGAN. By varying the weights of different loss
functions, the regGAN could achieve over 99.6% accuracy. Moreover, results indicated
that the best regGAN surrogate architecture consistently achieved over 99.5% accuracy
if provided with 400 or more random training samples. This confirmed the outstanding
predictive performance and potential generality of the regGAN.
In future work, we are planning to explore and develop other novel deep learning
architectures in regGAN. In addition, the simulation models used in this work are not
high-fidelity models but effective for describing the physics; we will increase the fidelity of
simulation models in future work, which may lead to a higher training cost for surrogate
modeling. Moreover, we will consider takeoff time as another constraint to make sure the
total takeoff will not take unreasonable time.
References
1. Wikipedia. EHang. 2023. Available online: https://en.wikipedia.org/wiki/EHang (accessed on 27 June 2023).
2. Wisk Aero LLC. Advanced Air Mobility Presents Opportunity to Bring Economic, Social, and Environmental Benefits to South
East Queensland. Tech. Rep. 2023. Available online: https://wisk.aero/news/press-release/advanced-air-mobility-presents-op
portunity-to-bring-economic-social-and-environmental-benefits-to-south-east-queensland/ (accessed on 15 January 2022).
3. Johnson, W.; Silva, C.; Solis, E. Concept Vehicles for VTOL Air Taxi Operations. In Proceedings of the AHS Technical Conference
on Aeromechanics Design for Transformative Vertical Flight, San Francisco, CA, USA, 16–18 January 2018.
4. Bacchini, A.; Cestino, E. Electric VTOL Configurations Comparison. Aerospace 2019, 6, 26. [CrossRef]
5. Na, Z.; Liu, Y.; Shi, J.; Liu, C.; Gao, Z. UAV-Supported Clustered NOMA for 6G-Enabled Internet of Things: Trajectory Planning
and Resource Allocation. IEEE Internet Things J. 2021, 8, 15041–15048. [CrossRef]
6. Hua, B.; Ni, H.; Zhu, Q.; Wang, C.-X.; Zhou, T.; Mao, K.; Bao, J.; Zhang, X. Channel Modeling for UAV-to-Ground Communications
With Posture Variation and Fuselage Scattering Effect. IEEE Trans. Commun. 2023, 71, 3103–3116. [CrossRef]
7. Boelens, J.-H. Pioneering the Urban Air Taxi Revolution. Tech. Rep. 2019. Available online: https://www.volocopter.com/urban
-air-mobility/ (accessed on 15 January 2022).
8. Electric VTOL News. Joby Aviation S4 2.0 (prototype). 2023 Available online: https://evtol.news/joby-s4 (accessed on 27
June 2023).
9. Electric VTOL News. Aurora Flight Sciences Pegasus PAV. 2019. Available online: https://evtol.news/aurora/ (accessed on 27
June 2023).
10. Wikipedia. Airbus A³ Vahana. 2022. Available online: https://en.wikipedia.org/wiki/Airbus_A%C2%B3_Vahana (accessed on
27 June 2023).
11. Yeh, S.-T.; Yan, G.; Du, X. Inverse Machine Learning Prediction for Optimal Tilt-Wing eVTOL Takeoff Trajectory. Aiaa Aviat. 2023
Forum 2023, 2023, 3593. [CrossRef]
12. Pradeep, P.; Wei, P. Energy Optimal Speed Profile for Arrival of Tandem Tilt-Wing eVTOL Aircraft with RTA Constraint. In
Proceedings of the IEEE/CSAA Guidance, Navigation and Control Conference (GNCC), Xiamen, China, 10–12 August 2018.
13. Pradeep, P.; Wei, P. Energy Efficient Arrival with RTA Constraint for Urban eVTOL Operations. In Proceedings of the 2018 AIAA
Aerospace Sciences Meeting, Kissimmee, FL, USA, 8–12 January 2018.
14. Chauhan, S.S.; Martins, J.R.R.A. Tilt-wing eVTOL takeoff trajectory optimization. J. Aircr. 2020, 57, 93–112. [CrossRef]
15. Li, J.; Du, X.; Martins, J.R. Machine learning in aerodynamic shape optimization. Prog. Aerosp. Sci. 2022, 134, 100849. [CrossRef]
16. Koziel, S.; Leifsson, L. Surrogate-Based Aerodynamic Shape Optimization by Variable-Resolution Models. AIAA J. 2013, 51,
94–106. [CrossRef]
Mathematics 2024, 12, 26 24 of 25
17. Nagawkar, J.; Leifsson, L. Applications of Polynomial Chaos-Based Cokriging to Simulation-Based Analysis and Design Under
Uncertainty. In Proceedings of the International Design Engineering Technical Conferences and Computers and Information in
Engineering Conference, Online, 17–19 August 2020.
18. Lobo do Vale, J.; Sohst, M.; Crawford, C.; Suleman, A.; Potter, G.; Banerjee, S. On the multi-fidelity approach in surrogate-based
multidisciplinary design optimisation of high-aspect-ratio wing aircraft. Aeronaut. J. 2023, 127, 2–23. [CrossRef]
19. Iuliano, E.; Quagliarella, D. Aerodynamic shape optimization via non-intrusive POD-based surrogate modelling. In Proceedings
of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; pp. 1467–1474. [CrossRef]
20. Li, S.; Trevelyan, J.; Wu, Z.; Lian, H.; Wang, D.; Zhang, W. An adaptive SVD–Krylov reduced order model for surrogate based
structural shape optimization through isogeometric boundary element method. Comput. Methods Appl. Mech. Eng. 2019, 349,
312–338. [CrossRef]
21. Singh, P.; Couckuyt, I.; Ferranti, F.; Dhaene, T. A constrained multi-objective surrogate-based optimization algorithm. In
Proceedings of the 2014 IEEE Congress on Evolutionary Computation (CEC), Beijing, China, 6–11 July 2014; pp. 3080–3087.
[CrossRef]
22. Shen, Y.; Huang, W.; Yan, L.; Zhang, T.-T. Constraint-based parameterization using FFD and multi-objective design optimization
of a hypersonic vehicle. Aerosp. Sci. Technol. 2020, 100, 105788. [CrossRef]
23. Li, M.; Wang, Z. Surrogate model uncertainty quantification for reliability-based design optimization. Reliab. Eng. System Saf.
2019, 192, 106432. [CrossRef]
24. Du, X.; Leifsson, L.; Koziel, S.; Bekasiewicz, A. Airfoil Design Under Uncertainty Using Non-Intrusive Polynomial Chaos Theory
and Utility Functions. Procedia Comput. Sci. 2017, 108, 1493–1499. [CrossRef]
25. Shao, J.; Shi, L.; Cheng, Y.; Li, T. Asynchronous Tracking Control of Leader–Follower Multiagent Systems With Input Uncertainties
Over Switching Signed Digraphs. IEEE Trans. Cybern. 2022, 52, 6379–6390. [CrossRef] [PubMed]
26. Li, W.; Qin, K.; Li, G.; Shi, M.; Zhang, X. Robust bipartite tracking consensus of multi-agent systems via neural network combined
with extended high-gain observer. ISA Trans. 2023, 136, 31–45. [CrossRef] [PubMed]
27. Rasmussen, C.E.; Williams, C.K.I. Gaussian Processes for Machine Learning; Adaptive Computation and Machine Learning; MIT
Press: Cambridge, MA, USA, 2006.
28. Liu, M.; Chowdhary, G.; Castra da Silva, B.; Liu, S.-Y.; How, J.P. Gaussian Processes for Learning and Control: A Tutorial with
Examples. IEEE Control Syst. Mag. 2018, 38, 53–86. [CrossRef]
29. Ver Hoef, J.M.; Barry, R.P. Constructing and fitting models for cokriging and multivariable spatial prediction. J. Stat. Plan. Inference
1998, 69, 275–294. [CrossRef]
30. Chiles, J.-P.; Delfiner, P. Geostatistics: Modeling Spatial Uncertainty; John Wiley & Sons: Hoboken, NJ, USA, 2012; Volume 713.
31. Thelen, A.S.; Bryson, D.E.; Stanford, B.K.; Beran, P.S. Multi-Fidelity Gradient-Based Optimization for High-Dimensional Aeroelas-
tic Configurations. Algorithms 2022, 15, 131. [CrossRef]
32. Tao, J.; Sun, G. Application of Deep Learning Based Multi-Fidelity Surrogate Model to Robust Aerodynamic Design Optimization.
Aerosp. Sci. Technol. 2019, 92, 722–737. [CrossRef]
33. Renganathan, S.A.; Maulik, R.; Ahuja, J. Enhanced data efficiency using deep neural networks and Gaussian processes for
aerodynamic design optimization. Aerosp. Sci. Technol. 2021, 111, 106522. [CrossRef]
34. O’Leary-Roseberry, T.; Du, X.; Chaudhuri, A.; Martins, J. R.; Willcox, K.; Ghattas, O. Learning high-dimensional parametric maps
via reduced basis adaptive residual networks. Comput. Methods Appl. Mech. Eng. 2022, 402, 115730. [CrossRef]
35. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial
Nets. In Advances in Neural Information Processing Systems 27; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger,
K.Q., Eds.; MIT Press: Cambridge, MA, USA, 2014; pp. 2672–2680.
36. Chen, W.; Chiu, K.; Fuge, M.D. Airfoil design parameterization and optimization using bézier generative adversarial networks.
AIAA J. 2020, 58, 4723–4735. [CrossRef]
37. Du, X.; He, P.; Martins, J.R.R.A. A B-Spline-based Generative Adversarial Network Model for Fast Interactive Airfoil Aerodynamic
Optimization. In AIAA SciTech Forum; AIAA: Orlando, FL, USA, 2020. [CrossRef]
38. Du, X.; Martins, J.R. Super Resolution Generative Adversarial Networks for Multi-Fidelity Pressure Distribution Prediction. In
AIAA SCITECH 2023 Forum; AIAA: Orlando, FL, USA, 2023; p. 0533.
39. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. 2014.
40. Aggarwal, K.; Kirchmeyer, M.; Yadav, P.; Keerthi, S.S.; Gallinari, P. Regression with conditional gan. arXiv 2019,
arXiv:1905.12868v1.
41. Ye, K.; Wang, Z.; Chen, P.; Piao, Y.; Zhang, K.; Wang, S.; Jiang, X.; Cui, X. A novel GAN-based regression model for predicting
frying oil deterioration. Sci. Rep. 2022, 12, 10424. [CrossRef]
42. Falck, R.; Gray, J.S.; Ponnapalli, K.; Wright, T. dymos: A Python package for optimal control of multidisciplinary systems. J. Open
Source Softw. 2021, 6, 2809. [CrossRef]
43. Gray, J.S.; Hwang, J.T.; Martins, J.R.R.A.; Moore, K.T.; Naylor, B.A. OpenMDAO: An open-source framework for multidisciplinary
design, analysis, and optimization. Struct. Multidiscip. 2019, 59, 1075–1104. [CrossRef]
44. de Wolff, T. Cuevas, A.; Tobar, F. MOGPTK: The Multi-Output Gaussian Process Toolkit. Neurocomputing 2020, 424, 49–53.
[CrossRef]
Mathematics 2024, 12, 26 25 of 25
45. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow:
Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467.
46. Paris, S.; Riehl, J.; Sjauw, W. Enhanced procedures for direct trajectory optimization using nonlinear programming and implicit
integration. In Proceedings of the AIAA/AAS Astrodynamics Specialist Conference and Exhibit, Keystone, CO, USA, 21–24
August 2006; p. 6309.
47. Hargraves, C.R.; Paris, S.W. Direct trajectory optimization using nonlinear programming and collocation. J. Guid. Dyn. 1987, 10,
338–342. [CrossRef]
48. Schubert, G.R. Algorithm 211: Hermite Interpolation. Commun. ACM 1963, 6, 617. [CrossRef]
49. Tangler, J.L.; Ostowari, C. Horizontal axis wind turbine post stall airfoil characteristics synthesization. In Proceedings of the
DOE/NASA Wind Turbine Technology Workshop, Cleveland, OH, USA, 8–10 May 1984.
50. Glauert, H. Airplane propellers. Aerodyn. Theory 1935, 169–360. [CrossRef]
51. Ypma, T.J. Historical development of the Newton–Raphson method. SIAM Rev. 1995, 37, 531–551. [CrossRef]
52. McCormick, B.W. Aerodynamics of V/STOL Flight, 1st ed.; Academic Press: Cambridge, MA, USA, 1967.
53. Leishman, J.G. Principles of Helicopter Aerodynamics, 1st ed.; The Press Syndicate of the University of Cambridge: New York, NY,
USA, 2000.
54. Biswas, B.; Chatterjee, S.; Mukherjee, S.; Pal, S. A discussion on Euler method: A review. Electron. J. Math. Anal. Appl. 2013, 1,
2090–2792.
55. Schulz, E.; Speekenbrink, M.; Krause, A. A tutorial on Gaussian process regression: Modelling, exploring, and exploiting functions.
J. Math. Psychol. 2018, 85, 1–16. [CrossRef]
56. Parra, G.; Tobar, F. Spectral Mixture Kernels for Multi-Output Gaussian Processes. In Advances in Neural Information Processing
Systems; Guyon, I., Luxburg, U.V.,Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.:
Red Hook, NY, USA, 2017; Volume 30. Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/333cb763
facc6ce398ff83845f224d62-Paper.pdf (accessed on 10 January 2022).
57. de G. Matthews, A.G.; van der Wilk, M.; Nickson, T.; Fujii, K.; Boukouvalas, A.; León-Villagrá, P.; Ghahramani, Z.; Hensman, J.
GPflow: A Gaussian Process Library using TensorFlow. J. Mach. Learn. 2017, 18, 1–6.
58. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [CrossRef]
59. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375.
60. Han, J.; Moraga, C. The influence of the sigmoid function parameters on the speed of backpropagation learning. In International
Workshop on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 1995; pp. 195–201.
61. Allen, D.M. Mean square error of prediction as a criterion for selecting variables. Technometrics 1971, 13, 469–475. [CrossRef]
62. Ho, Y.; Wookey, S. The real-world-weight cross-entropy loss function: Modeling the costs of mislabeling. IEEE Access 2019, 8,
4806–4813. [CrossRef]
63. Qian, N. On the Momentum Term in Gradient Descent Learning Algorithms. Neural Netw. 1999, 12, 145–151. [CrossRef] [PubMed]
64. Tieleman, S.; Hinton, G. Lecture 6.5—RMSProp: Neural Networks for Machine Learning. Coursera Tech. Rep. 2012, 6. Available
online: https://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf (accessed on 10 June 2022).
65. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.