Advancements in Deep Learning Techniques For Time
Advancements in Deep Learning Techniques For Time
Review
Advancements in Deep Learning Techniques for Time Series
Forecasting in Maritime Applications: A Comprehensive Review
Meng Wang 1 , Xinyan Guo 1 , Yanling She 2, *, Yang Zhou 3 , Maohan Liang 4 and Zhong Shuo Chen 1, *
1 School of Intelligent Finance and Business, Xi’an Jiaotong-Liverpool University, Suzhou 215400, China;
[email protected] (M.W.); [email protected] (X.G.)
2 Faculty of International Tourism and Management, City University of Macau, Avenida Padre Tom’as Pereira
Taipa, Macau 999078, China
3 Zhejiang “Eight Eight Strategy” Innovation and Development Research Institute, The Party School of Zhejiang
Provincial Committee of the Communist Party of China, Hangzhou 311121, China; [email protected]
4 Department of Civil and Environmental Engineering, National University of Singapore,
Singapore 117576, Singapore; [email protected]
* Correspondence: [email protected] (Y.S.); [email protected] (Z.S.C.)
Abstract: The maritime industry is integral to global trade and heavily depends on precise forecasting
to maintain efficiency, safety, and economic sustainability. Adopting deep learning for predictive
analysis has markedly improved operational accuracy, cost efficiency, and decision-making. This
technology facilitates advanced time series analysis, vital for optimizing maritime operations. This
paper reviews deep learning applications in time series analysis within the maritime industry, focus-
ing on three areas: ship operation-related, port operation-related, and shipping market-related topics.
It provides a detailed overview of the existing literature on applications such as ship trajectory predic-
tion, ship fuel consumption prediction, port throughput prediction, and shipping market prediction.
The paper comprehensively examines the primary deep learning architectures used for time series
forecasting in the maritime industry, categorizing them into four principal types. It systematically
analyzes the advantages of deep learning architectures across different application scenarios and
explores methodologies for selecting models based on specific requirements. Additionally, it analyzes
Citation: Wang, M.; Guo, X.; She, Y.;
data sources from the existing literature and suggests future research directions.
Zhou, Y.; Liang, M.; Chen, Z.S.
Advancements in Deep Learning
Keywords: deep learning; time series forecasting; maritime; shipping market; port operation
Techniques for Time Series
Forecasting in Maritime Applications:
A Comprehensive Review. Information
2024, 15, 507. https://doi.org/
10.3390/info15080507
1. Introduction
The maritime industry is navigating a complex environment, grappling with increas-
Academic Editor: Francesco
ingly stringent trade policies, heightened geopolitical tensions, and evolving globalization
Camastra
patterns [1]. Adding to these difficulties and challenges, the climate is intricate and the
Received: 30 July 2024 volume of sea traffic is immense, further straining the industry [2]. Nevertheless, maritime
Revised: 16 August 2024 trade remains the cornerstone of global commerce, driving economic growth and under-
Accepted: 17 August 2024 pinning the global economy [3,4]. Amid these complexities, the rapid development of
Published: 21 August 2024 satellite communications, large-scale wireless networks, and data science technologies has
revolutionized the shipping industry. These advancements have made the comprehensive
tracking of global ship movements and trends feasible [5]. The richness and diversity
of collected maritime data have significantly enhanced the accuracy and effectiveness of
Copyright: © 2024 by the authors.
maritime applications, making them indispensable in modern maritime operations [6–8].
Licensee MDPI, Basel, Switzerland.
Deep learning technologies for time series forecasting in maritime applications have gar-
This article is an open access article
nered significant interest, particularly in facing these challenges. These technologies offer
distributed under the terms and
conditions of the Creative Commons
promising solutions, helping the maritime industry manage safety, optimize operations,
Attribution (CC BY) license (https://
and address future challenges more effectively.
creativecommons.org/licenses/by/
However, there are several notable gaps in the existing literature. Existing studies
4.0/). primarily focus on applying and optimizing individual deep learning models, lacking
systematic and comprehensive comparative analysis. For instance, different studies employ
various models, datasets, and experimental setups, making it challenging to conduct
effective horizontal comparisons. This hinders a complete understanding of the strengths
and weaknesses of different deep learning methods in maritime time series analysis and
their appropriate application scenarios. Moreover, the performance variations of different
models in diverse application environments have not been thoroughly explored. While
some studies indicate that certain models excel in specific tasks, these results are often
based on particular datasets and experimental conditions, lacking generalizability. More
comprehensive research is needed to reveal which models have advantages across various
maritime applications and how to select and optimize models based on specific needs.
To overcome these research gaps, this review examines the applications of deep
learning in time series analysis within the maritime industry, focusing on three key areas:
ship operation-related, port operation-related, and shipping market-related topics. For
the purpose of obtaining relevant literature, extensive searches were conducted in major
international databases such as the Web of Science and Scopus to gather bibliometric data
suitable for analysis. Our strategy used specific keywords to identify studies employing
deep learning architectures for time series data in the maritime domain. After screening
and removing duplicates, we collected 89 papers on these topics. Most of the collected
literature, 79%, pertains to ship operations, highlighting the interest in forecasting ship
operation processes using deep learning. This comprehensive data collection underscores
the increasing reliance on advanced computational techniques to enhance maritime research
and operations.
This review explores the diverse applications of deep learning in the maritime do-
main, emphasizing its role in enhancing ship operation-related tasks, port operations, and
shipping market forecasting. Specifically, it delves into applications such as ship trajectory
prediction, anomaly detection, intelligent navigation, and fuel consumption prediction.
These applications leverage the unique advantages of deep learning models such as the
Artificial Neural Network (ANN), Multilayer Perceptron (MLP), Deep Neural Network
(DNN), Long Short-Term Memory (LSTM) network, Temporal Convolutional Network
(TCN), Recurrent Neural Network (RNN), Gated Recurrent Unit (GRU), Convolutional
Neural Network (CNN), Random Vector Functional Link (RVFL) network, and Transformer
models to address specific maritime challenges. Through comparative analysis, we aim
to uncover the performance and suitability of different deep learning models in various
maritime applications, thereby guiding future research and practical applications. Future
research directions emphasize improving data processing and feature extraction, optimiz-
ing deep learning models, and exploring specific maritime applications, including ship
trajectory prediction, port forecasts, ocean environment modeling, fault diagnosis, and
cross-domain applications.
The remainder of this work is organized as follows: Section 2 introduces the literature
review methods. In Section 3, deep learning methods are presented. The applications
of deep learning in time series forecasting within the maritime domain are discussed in
Section 4. An overall analysis is provided in Section 5. Finally, Section 6 concludes the
paper and outlines the limitations of the research.
Figure
Figure 1. The flow chart of1.data
The collection.
flow chart of Source:
data collection. Source: authors.
authors.
Figure 2. The
Figure 2. The foundational foundationalofarchitecture
architecture the ANN. of the ANN.
Source: Source:
authors authorsbased
redrawn redrawn
onbased
[11]. on [11].
3.1.1.
3.1.1. Multilayer Multilayer
Perceptron Perceptron (MLP)/Deep
(MLP)/Deep Neural Networks
Neural Networks (DNN) (DNN)
The term ANNThe termas
serves ANN serves
a broad as a broadfor
descriptor descriptor
variousfor variousincluding
models, models, including
the MLP,the a MLP,
fundamental archetype. The MLP’s architecture, as previously illustrated, relies on error error
a fundamental archetype. The MLP’s architecture, as previously illustrated, relies on
backpropagation and gradient descent algorithms to optimize network weights [12].
backpropagation and gradient descent algorithms to optimize network weights [12]. These
These methods work together to reduce prediction errors in the model’s output [13]. Equa-
methods work together to reduce prediction errors in the model’s output [13]. Equation (1)
tion (1) explains the formulation of the MLP’s output.
explains the formulation of the MLP’s output.
𝑍 𝑄
𝑜 𝐻
𝑦𝑝Z = 𝜑o 𝑜h{∑ Q𝜔𝑗𝑝 H 𝐻 (∑
[𝜑 𝜔𝑖𝑗 𝑥𝑖 )]} (1)
n io
y p = φo ∑ j=1 ω jp φ H 𝑗=1 ∑i=0 ij i
ω x 𝑖=0 (1)
𝐻 𝑜
where 𝜔𝑖𝑗 signifies the weights associated with the hidden layer and 𝜔𝑗𝑝 indicates the
where ωijH signifies theofweights
weights associated
the output with𝜑the
layer. 𝜑𝐻 and hidden layer and ω ojp indicates thehidden
𝑜 symbolize the activation functions for the
weights of the output layer. φ H and φo symbolize the activation functions for the hidden
and output layers, respectively. Commonly utilized activation functions include ReLU and
Sigmoid, with the selection primarily influenced by the particular application context and
the type of output required.
Conversely, the Deep Feedforward Network or Deep Neural Network (DNN), the
most fundamental model in deep learning, typically features more hidden layers [14]. The
increased architectural depth enhances the network’s ability to learn intricate features and
patterns from the input data, thereby improving its prediction capabilities.
Within the corpus of literature compiled for this paper, two distinct studies employed
the frameworks of ANN and MLP, respectively, to conduct time series forecasting. These
neural network architectures were specifically utilized for predicting ship Response Ampli-
tude Operators (RAOs) and ship classification purposes, respectively [15,16].
Information 2024, 15, 507 5 of 33
3.1.2. WaveNet
Recent advancements in ANN research have led to the development of high-performance
models, notably the WaveNet, a generative model based on convolutional neural networks
designed for audio waveform generation [17]. This model incorporates several advanced
deep learning technologies, including dilated causal convolutions, residual modules, and
gated activation units. The hallmark of WaveNet is its dilated convolution structure, which
enables the model to effectively learn long-term dependencies within audio data, demon-
strating robust learning capabilities. As a probabilistic autoregressive model, WaveNet
directly learns the mappings between sequential sample values, revealing its potential
across a variety of application scenarios. Figure 3 illustrates the foundational architecture of
WaveNet. Furthermore, for a sequence of x = { x1 , x2 . . . , xt−1 }, its joint probability p( x )
Information 2024,can be PEER
15, x FOR represented
REVIEW by Equation (2). WaveNet also utilizes gated activation units,5 of 36 whose
Figure4.4.Visualization
Figure Visualization of
of aa stack
stack of
of causal
causal convolutional layers.
convolutional layers.
Figure5.5.Visualization
Figure Visualization of
of aa stack
stack of
of dilated
dilated causal convolutional layers.
layers.
3.1.3.
3.1.3.Randomized
Randomized Neural
Neural Network
Network
Random
Random neural networks, commonly
neural networks, commonly usedused in
in shallow
shallow neural
neuralarchitectures,
architectures,initialize
initialize
the weights and biases of hidden layers randomly and keep them fixed
the weights and biases of hidden layers randomly and keep them fixed during training, during training,
enhancing training efficiency [20]. This approach, integral to the construction
enhancing training efficiency [20]. This approach, integral to the construction of Extreme of Extreme
Learning Machines (ELMs) and the RVFL network, is particularly effective
Learning Machines (ELMs) and the RVFL network, is particularly effective for rapid learn- for rapid
learning and streamlining model training
ing and streamlining model training [21,22]. [21,22].
ELM
ELMrepresents
represents aa type
type of
of random
random neural network, particularly
neural network, particularlyemphasizing
emphasizingthe theuse
use
of a single hidden layer, where the weights and biases of the hidden layer
of a single hidden layer, where the weights and biases of the hidden layer remain fixed remain fixed
after
afterinitialization.
initialization.Within
Withinthe
theELM
ELMframework,
framework,thetheonly
onlyparameters
parameters that need
that needoptimization
optimiza-
are the output weights, which are the weights connecting the hidden
tion are the output weights, which are the weights connecting the hidden nodes nodes to the output
to the
output nodes. These are typically computed efficiently via matrix inversion operations.
Figure 6 illustrates a simplified ELM architecture, and 𝛽 represents the weights associ-
ated with the connections in the network.
enhancing training efficiency [20]. This approach, integral to the construction of Extreme
Learning Machines (ELMs) and the RVFL network, is particularly effective for rapid learn-
ing and streamlining model training [21,22].
ELM represents a type of random neural network, particularly emphasizing the use
of a single hidden layer, where the weights and biases of the hidden layer remain fixed
Information 2024, 15, 507
after initialization. Within the ELM framework, the only parameters that need optimiza- 7 of 33
tion are the output weights, which are the weights connecting the hidden nodes to the
output nodes. These are typically computed efficiently via matrix inversion operations.
nodes.6 These
Figure are typically
illustrates computed
a simplified efficiently and
ELM architecture, via matrix inversion
𝛽 represents the operations. Figure 6
weights associ-
illustrates
ated with the a simplified
connectionsELM architecture,
in the network. and β represents the weights associated with the
connections in the network.
Figure
Figure 7. 7.
TheThe simplified
simplified architecture
architecture of theof the RVFL.
RVFL.
While traditional ELM and RVFL are not typically considered deep learning archi-
tectures, they have been applied to more complex data structure processing tasks by in-
troducing deep variants, such as deep ELM (DELM) and ensemble deep random vector
functional link (edRVFL), showing broad applicability across various application do-
mains. This is aimed at combining the benefits of deep learning, such as the ability to
Information 2024, 15, 507 8 of 33
While traditional ELM and RVFL are not typically considered deep learning archi-
tectures, they have been applied to more complex data structure processing tasks by
introducing deep variants, such as deep ELM (DELM) and ensemble deep random vector
functional link (edRVFL), showing broad applicability across various application domains.
This is aimed at combining the benefits of deep learning, such as the ability to capture more
complex data structures, with the efficient training processes of ELM/RVFL [25]. These
deep versions are generally achieved by increasing the number of hidden layers, allowing
the model to learn deeper data representations. Currently, DELM and edRVFL are widely
utilized in image processing or prediction tasks [26–28].
3.2. Convolutional
Information 2024, 15, x FOR PEER REVIEW Neural Network (CNN) 9 of 36
Lecun et al. [29] pioneered the CNN by integrating the backpropagation algorithm
with feedforward neural networks, laying the foundation for modern convolutional ar-
chitectures. and
interaction, Originally
parameter designed
sharingfor image
[14]. processing,
Although initiallythe CNN
used efficiently
primarily addresses
in machine vi-
sion and speech recognition, CNN has also begun to demonstrate its potential in inter-
complex nonlinear challenges through features like equivalent representation, sparse time
action, and parameter sharing [14]. Although initially used primarily in machine vision
series prediction [30,31]. These networks capture local features in temporal data and use
and speech recognition, CNN has also begun to demonstrate its potential in time series
their deep structures to extract time-dependent relationships, providing new perspectives
prediction [30,31]. These networks capture local features in temporal data and use their
and methods for time series analysis. Figure 8 illustrates the fundamental structure of the
deep structures to extract time-dependent relationships, providing new perspectives and
CNN.
methods for time series analysis. Figure 8 illustrates the fundamental structure of the CNN.
Figure 8.
Figure The foundational
8. The foundational architecture
architecture of
of the
the CNN.
CNN. Source:
Source: authors
authors redrawn
redrawn based
based on
on [11].
[11].
CNN has demonstrated superior performance in processing time series data. In time
CNN has demonstrated superior performance in processing time series data. In time
series prediction applications, CNN employs filters to process the input data, capturing
series prediction applications, CNN employs filters to process the input data, capturing
local patterns and features within the time series via convolutional operations, such as
local patterns and features within the time series via convolutional operations, such as
trends and periodic changes. These filters, utilizing weight sharing and regional connec-
trends and periodic changes. These filters, utilizing weight sharing and regional connec-
tions, reduce the complexity and computational load of the model while enabling the
tions, reduce the complexity and computational load of the model while enabling the ef-
effective identification of key structures in temporal data, such as spikes or drops [32]. The
fective identification of key structures in temporal data, such as spikes or drops [32]. The
mathematical expression for a convolutional layer is denoted as Equation (4).
mathematical expression for a convolutional layer is denoted as Equation (4).
x l
= φ
𝑥𝑗𝑙 j= 𝜑 (∑
∑i∈ Mj i𝑥 𝑙−1 ∗ij𝜔𝑙 j+ 𝑏𝑙 )
x l −1
∗ ω l
+ b l
(4)
(4)
𝑖 𝑖𝑗 𝑗
𝑖∈𝑀𝑗
In the equation of a convolutional layer, φ represents the activation function; ω𝑙ijl
In the equation of a convolutional layer, 𝜑 represents the activation function; 𝜔𝑖𝑗
denotes the weight values of the convolutional kernel; bl refers to the bias associated with
denotes the weight values of the convolutional kernel; 𝑏𝑗𝑙j refers to the bias associated with
the kernel; and M indicates the set of feature mappings.
the kernel; and 𝑀𝑗j indicates the set of feature mappings.
Following convolution, a pooling layer such as max or average pooling reduces the
Following convolution, a pooling layer such as max or average pooling reduces the
data’s temporal dimension, easing computational demands and helping prevent overfitting.
data’s temporal dimension, easing computational demands and helping prevent overfit-
The network’s later stages consist of fully connected layers that integrate and interpret
ting. The network’s later stages consist of fully connected layers that integrate and inter-
these features, although these often contain many parameters, posing computational chal-
pret these features, although these often contain many parameters, posing computational
lenges. Efficient design adjustments and regularization techniques are critical for enhancing
challenges. Efficient design adjustments and regularization techniques are critical for en-
CNN performance in time series prediction tasks, such as financial trend forecasting or
hancing CNN performance in time series prediction tasks, such as financial trend fore-
weather prediction [33–35]. Through these mechanisms, the CNN utilizes historical data to
casting or weather
effectively forecast prediction [33–35].
future trends, Through
showcasing itsthese
robustmechanisms,
capabilities inthetime
CNN utilizes
series his-
analysis.
torical data to effectively forecast future trends, showcasing its robust capabilities
Although CNN has demonstrated robust capabilities in time series analysis tasks, it in time
series analysis.
exhibits certain limitations in addressing long-term dependencies. To mitigate this chal-
Although
lenge, the TCNCNN has demonstrated
was introduced by Bai etrobust
al. [36],capabilities
emerging as in atime series
notable analysis tasks,
adaptation of CNNit
exhibits certain limitations in addressing long-term dependencies. To mitigate this chal-
lenge, the TCN was introduced by Bai et al. [36], emerging as a notable adaptation of CNN
specifically engineered for time series data processing. TCN processes time series data
exclusively through convolutional layers instead of recurrent layers, thereby effectively
capturing long-range dependencies within the time series. The architecture incorporates
causal convolutions, dilated convolutions, and residual blocks. These techniques collec-
Information 2024, 15, 507 9 of 33
specifically engineered for time series data processing. TCN processes time series data
exclusively through convolutional layers instead of recurrent layers, thereby effectively
capturing long-range dependencies within the time series. The architecture incorporates
Information 2024, 15, x FOR PEER REVIEW 10 of 36
causal convolutions, dilated convolutions, and residual blocks. These techniques collec-
tively address the challenge of extracting long-term information from series data. Figure 9
illustrates the fundamental architecture of the TCN.
Figure
Figure 9. 9.
TheThe foundational
foundational architecture
architecture ofof the
the TCN.
TCN. Source:
Source: authors
authors redrawn
redrawn based
based onon[37].
[37].
TCN
TCN processessequence
processes sequencedata
datausing
usinga aseries
seriesofofone-dimensional
one-dimensionalconvolutional
convolutionaloper-
opera-
tions, each comprising a convolutional layer and a nonlinear activation
ations, each comprising a convolutional layer and a nonlinear activation function. The function. The kernel
size, stride,
kernel and dilation
size, stride, rate can
and dilation ratebecan
adjusted in eachinconvolutional
be adjusted each convolutionaloperation as needed.
operation as
The technique of dilated convolution enables the kernel to skip over
needed. The technique of dilated convolution enables the kernel to skip over some aspectssome aspects in the
insequence, effectively
the sequence, expanding
effectively the receptive
expanding the receptivefield field
without adding
without moremore
adding parameters.
parame-By
stacking multiple layers of such convolutions, the TCN progressively abstracts higher-level
ters. By stacking multiple layers of such convolutions, the TCN progressively abstracts
features from the sequence data, enhancing the model’s expressive capacity. The equation
higher-level features from the sequence data, enhancing the model’s expressive capacity.
for operation F, representing dilated convolutions on elements p of a 1-D sequence, is
The equation for operation 𝐹, representing dilated convolutions on elements p of a 1-D
presented in Equation (5).
sequence, is presented in Equation (5). l −1
F ( p ) = ∑ i =0 f ( i ) x p − d ∗ i (5)
𝑙−1
where d represents the dilation𝐹(𝑝) factor;=f denotes
∑ 𝑓(𝑖)𝑥 (5)of
𝑝−𝑑∗𝑖 of filters; p refers to the size
the number
each filter; and d ∗ i specifies the direction of𝑖=0 the past.
where To avoid localthe
𝑑 represents optima during
dilation training
factor; and tothe
𝑓 denotes enhance
number performance,
of filters; 𝑝 TCN
refersemploys
to the
residual
size of eachconnections
filter; and between convolutional
𝑑 ∗ 𝑖 specifies layers
the direction of[38]. The fundamental concept of resid-
the past.
ualTo
connections
avoid local involves
optimaadding
duringthe output and
training of a current
to enhancelayerperformance,
to the output TCN
from employs
a previous
layer (or
residual several prior
connections layers),
between effectively creating
convolutional a “shortcut”
layers [38]. connection.
The fundamental This method
concept of re-
enhances the model’s efficiency in learning long-range dependencies within
sidual connections involves adding the output of a current layer to the output from a pre- the sequence.
Residual
vious layerconnections
(or several also help
prior to mitigate
layers), vanishing
effectively and aexploding
creating “shortcut”gradients and acceler-
connection. This
ate model convergence. The residual connections can be expressed
method enhances the model’s efficiency in learning long-range dependencies within in Equation (6). the
sequence. Residual connections also help to mitigate vanishing and exploding gradients
O = φ( F ( x ) + x ) (6)
and accelerate model convergence. The residual connections can be expressed in Equation
(6).
where φ represents the activation function. After all convolutional operations, a TCN
typically employs a global average 𝑂 pooling
= 𝜑(𝐹(𝑥)layer+to𝑥)aggregate all features and generate
(6)
a global feature vector. This vector is inputted into subsequent fully connected layers
where 𝜑 represents
designed the activation
for tasks such function.
as regression After all convolutional operations, a TCN
and classification.
typically employs a global average pooling layer
Finally, TCN employs a loss function to measure to aggregate all features
the discrepancy and generate
between a
predictions
global featureoutcomes,
and actual vector. This vector
using the is inputted into subsequent
backpropagation algorithm fully connected
to update layers de-
parameters. Dur-
signed for tasks such as regression and classification.
Finally, TCN employs a loss function to measure the discrepancy between predic-
tions and actual outcomes, using the backpropagation algorithm to update parameters.
During training, various optimization algorithms and regularization techniques enhance
the model’s generalization capabilities [39].
In the literature we collected, it is common to see CNN and TCN used in conjunction
Information 2024, 15, 507 10 of 33
Figure
Figure10.
10.The
Thefoundational
foundationalarchitecture
architectureof
ofthe
theRNN.
RNN.Source:
Source: authors
authorsredrawn
redrawn based
based on
on [44].
[44].
𝑥x represents
represents the
theinput layer;s denotes
inputlayer; 𝑠 denotes thethe
state of the
state of hidden W, U,𝑊,
layer;layer;
the hidden and𝑈, V and
signify
𝑉
the weights;
signify and o stands
the weights; for thefor
and 𝑜 stands output layer. layer.
the output “Unfold” typically
“Unfold” refers refers
typically to expanding
to expand-the
RNN
ing theacross
RNN the timethe
across series
timeinto a sequence
series of feedforward
into a sequence networks.
of feedforward The RNN,
networks. Thewith
RNN, its
unique memory capacity and flexibility, demonstrates a decisive advantage
with its unique memory capacity and flexibility, demonstrates a decisive advantage in the in the field of
series forecasting [45]. Whether it is predicting traffic flow, forecasting financial
field of series forecasting [45]. Whether it is predicting traffic flow, forecasting financial and stock
market
and stocktrends,
marketortrends,
anticipating weather and
or anticipating climate
weather andchanges, RNN plays
climate changes, RNN a crucial
plays role
a cru-in
producing
cial role in accurate
producing predictions
accurate [46,47]. The basic
predictions equations
[46,47]. The basic for equations
the RNN can for be
therepresented
RNN can
byrepresented
be Equations (7)byand (8).
Equations (7) and (8).
st = φ Uxt + Wst−1 + bs (7)
𝑠𝑡 = 𝜑(𝑈𝑥𝑡 + 𝑊𝑠𝑡−1 + 𝑏𝑠 ) (7)
ot = Vst + bo (8)
where bs and bo represent the biases and 𝑜𝑡φ=denotes
𝑉𝑠𝑡 + 𝑏an
𝑜 activation function. Certainly, while
(8)
RNN was designed to address long-term dependencies, its performance on long-duration
data in𝑏experimental
where 𝑠 and 𝑏𝑜 represent
settingsthe
hasbiases 𝜑 denotes
andentirely
not been an activation
optimal [48,49]. function. Certainly,
while RNN was designed to address long-term dependencies, its performance on long-
duration data in experimental settings has not been entirely optimal [48,49].
Information 2024, 15, x FOR PEER REVIEW 12 of 36
Information 2024, 15, 507 11 of 33
Figure 11. The foundational architecture of the LSTM. Source: authors redrawn based on [11].
ℎh represents
representsthe thehidden
hiddenstate, whilec denotes
state,while 𝑐 denotes thethe
long-term
long-term memory
memory cell.cell.
TheTheconnec-
con-
tion between h and c is finely controlled by three “gates”: the input
nection between ℎ and 𝑐 is finely controlled by three “gates”: the input gate, the forget gate, the forget gate,
and the
gate, andoutput gate. Ingate.
the output eachIn consecutive time step,
each consecutive time thestep,
LSTMthe utilizes
LSTMthe currentthe
utilizes input and
current
previous state to determine the extent of memory elimination (through
input and previous state to determine the extent of memory elimination (through the for- the forget gate), the
quantity
get gate),of thenew information
quantity of new to assimilateto
information into the memory
assimilate cellmemory
into the (throughcellthe(through
input gate),
the
and the amount of information transmission from the memory cell
input gate), and the amount of information transmission from the memory cell to the cur- to the current hidden
rent h (through
statehidden statethe output gate).
ℎ (through These gates
the output gate).primarily
These gates manage the flow
primarily manage of information,
the flow of
utilizing sigmoid and tanh activation functions. When
information, utilizing sigmoid and tanh activation functions. When error signalserror signals are relayedare fromre-
the output layer, the LSTM’s memory cell ccc captures vital error
layed from the output layer, the LSTM’s memory cell ccc captures vital error gradients, gradients, facilitating
prolonged information
facilitating retention. Compared
prolonged information retention. to RNN, LSTM
Compared alleviates
to RNN, LSTM thealleviates
problemsthe of
vanishing or exploding gradients to some extent through their
problems of vanishing or exploding gradients to some extent through their gating mech-gating mechanism, thereby
more effectively handling long-term dependencies. The internal details of the LSTM are
anism, thereby more effectively handling long-term dependencies. The internal details of
shown in Figure 12 and Equations (9) through (14) illustrate all the internal components of
the LSTM are shown in Figure 12 and Equations (9) through (14) illustrate all the internal
the LSTM.
components of the LSTM.
f t = σ W f [ h t −1 , x t ] + b f (9)
Figure
Figure12.
12.The
Theinternal
internal details ofthe
details of theLSTM.
LSTM.Source:
Source: authors
authors redrawn
redrawn based
based on [54].
on [54].
𝑐𝑡 = 𝑓𝑡 ⊙ 𝑐𝑡−1 + 𝑖𝑡 ⊙ 𝑔𝑡 (13)
∼
h t = (1 − z t ) h t −1 + z t h t (18)
∼
where h t and ht represent the candidate vector and output vector, respectively; σ denotes
an activation function; W refers to the weight matrices associated with the inputs to the
network’s activation functions; zt is identified as the update gate; and rt represents the
reset gate.
The papers we have gathered show that the RNN architecture is the most frequently
applied, with LSTM being the most extensively utilized variant. This widespread adoption
underscores the significant recognition that RNN architectures receive in the field of series
prediction, emphasizing their effectiveness in managing the complexities of series data
across various applications. Among these, they are widely used in tasks related to ship
operations, such as trajectory prediction using RNN-LSTM by Pan et al. [58]; the prediction
of Ship Encounter Situation Awareness using LSTM by Ma et al. [59]; and trajectory
prediction using GRU by Suo et al. [60]. This indicates the efficacy of RNN-based models
in enhancing time series predictions within the maritime domain.
QK T
A( Q, K, V ) = so f tmax √ V (19)
dk
In the existing
In time literature,with
series forecasting the AM is often usedmodel,
the Transformer with other models,
the process such
starts byas in the
convert-
GRU-AM for trajectory predictions by Zhang et al. [64]. The Transformer
ing input data into embedding vectors, which are then enriched with positional encodings model is par-
ticularly prevalent in prediction applications, with variations such as the
to add temporal context. The data pass through several layers of self-attention mecha- iTransformer
being
nisms employed for trajectory
and fully connected predictions
layers, with eachbylayer
Zhang et al. [65],
fortified and hybrid
by residual models that
connections and
combine
layer normalization to ensure stable training. The Transformer excels at analyzingtasks
Transformer and LSTM also being used for similar trajectory forecasting by
feature
Jiang et al. [66]. across various time intervals due to its parallel processing capability,
interrelationships
crucial for modeling long-term dependencies. The multi-head attention mechanism splits
3.5. Overview of Algorithms Usage
the input into subspaces, allowing for focused analysis of distinct patterns, which en-
Excluding
hances review andDuring
feature extraction. comparative studiesphase,
the decoding from the
the model
collected papers,
uses maskedFigure 15 il-
self-atten-
lustrates the proportionate usage of primary models in the documented research. In
tion to base predictions solely on previously known data, finishing with predictions gen-
this analytical review encompassing 89 papers, it is conspicuous that the LSTM model is
erated by a fully connected layer followed by a Softmax layer. This architecture makes the
prevalently utilized, appearing in over 40% of the examined studies. This predominance
Transformer particularly effective for complex time series forecasting involving long-term
underscores the model’s robustness in processing sequential data, affirming its continued
dependencies. Equations (19)–(21) demonstrate the self-attention mechanism applied to
relevance and centrality in the literature. While deep learning architectures are often used
the Query (Q), Key (K), and Value (V) components.
Information 2024, 15, 507 15 of 33
in combination to leverage their respective strengths, many of the collected papers exclu-
sively utilized LSTM for predicting time series data, likely due to its exceptional capability
in handling long-duration sequences [59,67–72]. Concurrently, the CNN model is also
featured and employed in approximately 23% of the papers. Unlike LSTM, RNN, and GRU,
the CNN model performs excellently in extracting features from image data, making them
particularly effective in this domain [73–76]. The Attention Mechanism-based AM and
Transformer is seldom used as a standalone predictive model; typically, it is integrated
with a conventional neural network model. Its advantage lies in employing a unique mech-
anism such as a querying process, which facilitates the allocation of weights across spatial
dimensions and temporal steps, thereby emphasizing crucial information [37,42,66,77–81].
Additionally, other models typically do not appear independently in tasks involving time
series prediction; they are usually employed in combinations, such as using certain models
for feature extraction and others for prediction. A CNN is frequently used for both feature
extraction and prediction tasks [54,82–84]. A GRU plays a significant role in prediction
tasks [60,64,82]. Notably, scholars often integrate various deep learning models, for in-
stance, by weighting to aggregate the results from each model [85]. Collectively, these
Information 2024, 15, x FOR PEER REVIEW
findings delineate prevailing technological inclinations within deep learning research18 of
and36
illuminate prospective trajectories for scholarly inquiry.
3% 3%
9%
9%
21%
8%
3%
4%
40%
Figure 15.
Figure 15. Percentages
Percentages of
of the
the deep
deep learning
learning algorithms
algorithms used
used in
in the
the collected
collected literature.
literature.
Forecasting in
4. Time Series Forecasting in Maritime
Maritime Applications
4.1.
4.1. Ship
Ship Operation-Related
Operation-Related Applications
Applications
This
This section
section explores
explores various applications of
various applications of deep
deep learning
learning inin ship
ship operations,
operations, includ-
includ-
ing enhancing navigation safety, ship anomaly detection, intelligent navigation
ing enhancing navigation safety, ship anomaly detection, intelligent navigation practices,practices,
meteorological
meteorological predictions,
predictions, and
and fuel
fuel consumption
consumption forecasting.
forecasting. ItIt highlights
highlights the
the use
use of
of
models
models such as LSTM, CNN, and Transformer to improve prediction accuracy andopera-
such as LSTM, CNN, and Transformer to improve prediction accuracy and oper-
tional
ationalefficiency.
efficiency.Additionally,
Additionally,applications in ship
applications type
in ship classification,
type traffic
classification, prediction,
traffic and
prediction,
collision risk assessment
and collision are discussed,
risk assessment demonstrating
are discussed, the broad
demonstrating theimpact
broad of deep learning
impact of deep
on ship operations.
learning on ship operations.
4.1.1. Ship Trajectory Prediction
4.1.1. Ship Trajectory Prediction
A. Navigation Safety Enhancement
A. Navigation Safety Enhancement
Accurate trajectory prediction is essential for maritime safety and navigation optimiza-
Accurate trajectory prediction is essential for maritime safety and navigation optimi-
tion. It involves estimating vessels’ future positions and states, such as longitude, latitude,
zation. It involves estimating vessels’ future positions and states, such as longitude, lati-
heading, and speed. Accurate predictions are vital for enhancing maritime traffic safety,
tude, heading, and speed. Accurate predictions are vital for enhancing maritime traffic
safety, optimizing navigation management, and aiding emergency response [86]. Li et al.
[86] introduced an innovative method based on AIS data, data encoding transformation,
the Attribute Correlation Attention (ACoAtt) module, and the LSTM network. This ap-
proach significantly enhances prediction accuracy and stability by capturing dynamic in-
teractions among ship attributes. The experimental results demonstrate that this method
Information 2024, 15, 507 16 of 33
optimizing navigation management, and aiding emergency response [86]. Li et al. [86]
introduced an innovative method based on AIS data, data encoding transformation, the
Attribute Correlation Attention (ACoAtt) module, and the LSTM network. This approach
significantly enhances prediction accuracy and stability by capturing dynamic interactions
among ship attributes. The experimental results demonstrate that this method outper-
forms traditional models. Similarly, Yu et al. [87] introduced an improved LSTM model
incorporating the ACoAtt module, significantly improving trajectory prediction accuracy.
This model utilizes a multi-head attention mechanism to capture intricate relationships
among ship speed, direction, and position, demonstrating excellent performance in han-
dling time series data. The experimental results indicate that the improved LSTM model
offers significant advantages over standard LSTM models regarding multi-step prediction
accuracy and stability. It has broad application prospects in maritime search and rescue
and navigation safety.
Enhancing ship trajectory prediction in complex environments is essential for effective
maritime operations. Xia et al. [88] proposed a specially designed pre-training model to
enhance ship trajectory prediction performance in complex environments. This model uses
a single-decoder architecture, pre-training on complete trajectory data, and fine-tuning
for specific prediction scenarios. It achieves high-precision predictions and demonstrates
broad adaptability. This method directly inputs trajectory data segments and immediately
initiates prediction, which is suitable for various downstream tasks and significantly
enhances model practicality. Cheng et al. [89] proposed a cross-scale SSE model based
on class-imbalanced vessel movement data to address data imbalance issues. This model
includes multi-scale feature learning modules, cross-scale feature learning modules, and
prototype classifier modules, aiming to directly learn imbalanced vessel movement data.
Compared to traditional softmax classifiers, the distance-based classifier significantly
improves classification performance. This model demonstrates superior scalability and
generality on multiple public datasets and excels in vessel movement data. Traditional
prediction methods are often limited by the irregularity of AIS data and the neglect of
intrinsic vessel information. To address these issues, Xia et al. [90] proposed a divide-and-
conquer model architecture incorporating vessel-specific data, TABTformer (Turtle and
Bunny-TCN-former), to reduce computational complexity and improve accuracy.
Improving the accuracy of ship trajectory prediction is highly significant. Pan et al. [58]
aimed to predict ship arrival times using AIS data to enhance container terminal opera-
tions. They employed a transitive closure method based on equivalence relations for the
fuzzy clustering of routes, achieving optimal route matching and constructing navigation
trajectory features. The proposed RNN-LSTM model combines deep learning with time
series analysis, demonstrating excellent predictive performance and providing technologi-
cal support for maritime intelligent transportation. Bin Syed and Ahmed [40] proposed
a 1D CNN-LSTM architecture for trajectory association. This architecture combines the
advantages of the CNN and the LSTM network, simulating AIS datasets’ spatial and tem-
poral features. By capturing spatial and temporal correlations in the data, this integrated
model surpasses other deep learning architectures in accuracy for trajectory prediction.
Experimental results demonstrate that the CNN-LSTM framework excels in processing
ship trajectory data, notably enhancing predictive performance.
B. Ship Anomaly Detection
Improving prediction accuracy and reducing training time are essential for effective
ship anomaly detection. Violos et al. [67] proposed an ANN architecture combined with
LSTM layers, utilizing genetic algorithms and transfer learning methods to predict the
future locations of moving objects. Genetic algorithms conduct intelligent searches within
the hypothesis space to approximate optimal ANN architectures, while transfer learning
utilizes pre-trained ANN models to expedite the training process. Evaluations based on
real data from various ship trajectories indicate that this model outperforms other advanced
methods in prediction accuracy and significantly reduces training time.
Information 2024, 15, 507 17 of 33
Predicting ship trajectories is crucial for vessel traffic service (VTS) supervision and
accident warning. Ran et al. [91] utilized AIS data to transform navigational dynamics
into time series, extracting navigation trajectory features and employing an attention
mechanism-based LSTM network for training and testing. The results indicate that this
model can effectively and accurately predict ship trajectories, providing valuable references
for VTS supervision and offering high practical application value in accident warnings.
This method presents a novel approach to ship navigation prediction, avoiding the need
to establish complex ship motion models. Yang et al. [92] proposed a ship trajectory
prediction method that combines data denoising and deep learning techniques. Their
process involves trajectory separation, data denoising, and standardization. By denoising
the original AIS data, the complexity and computational time of the input prediction
model are reduced, thereby improving prediction accuracy. Experimental results show
that this method outperforms ETS, ARIMA, SIR, RNN, and LSTM models in prediction
accuracy and computational efficiency, significantly reducing errors in ship trajectory
prediction. This contributes to improving maritime traffic efficiency and safety, preventing
maritime accidents.
Efficient maritime traffic monitoring and collision prevention are critical for ensuring
safe navigation. Sadeghi and Matwin [93] investigated anomaly detection in maritime
time series data using an unsupervised method based on autoencoder (AE) reconstruction
errors. They could identify varying degrees of anomalies by estimating the probability
density function of reconstruction errors. The results demonstrated that this method ef-
fectively pinpointed irregular patterns in ship movement trajectories, providing reliable
support for maritime traffic monitoring and collision prevention. Ma et al. [79] proposed
an AIS data-driven method to predict ship encounter risks by modeling inter-ship behavior
patterns. This method extracts multidimensional features from AIS tracking data to capture
spatial dependencies between encountering ships. It uses a sliding window technique to
create behavioral feature sequences, viewing future collision risk levels as sequence labels.
Utilizing the powerful temporal dependency and complex pattern modeling capabilities
of LSTM network, they established a relationship between inter-ship behavior and future
collision risk. This method efficiently and accurately predicts future collision risks, particu-
larly excelling when potential risks are high, making it promising for early prediction of
ship encounters and enhancing navigation safety.
Detecting ship behavior patterns and monitoring anomalies is essential for maritime
safety. Perumal et al. [94] proposed a novel hybrid model combining the CNN and
the LSTM network. By modifying the residual CNN network and integrating it with
LSTM/GRU, they introduced a new activation function to overcome the dead neuron
propagation issue of rectified linear unit (ReLU) activation functions, improving model
accuracy by over 1%. Furthermore, Hoque and Sharma [68] presented a solution for real-
time monitoring of ship anomalous behavior. They used spectral clustering and outlier
detection methods on historical AIS data to discern ship types and typical routes. LSTM
networks were then used for trajectory prediction and ship engine behavior to determine if
the current route was abnormal. This method applies the LSTM network for trajectory pre-
diction and multivariate time series anomaly detection, displaying real-time anomaly state
classification and considering route changes due to emergencies. It effectively monitors
and identifies ship anomalous behavior, enhancing maritime safety.
Accurate ship fault detection and prevention are vital for maintaining operational
efficiency and safety. Ji et al. [80] proposed a hybrid neural network (HNN) model (CNN-
BiLSTM-Attention) for predicting a ship’s diesel engine exhaust gas temperature (EGT)
based on deep learning. This model demonstrates significant advantages in prediction
accuracy, capable of extracting comprehensive spatial and temporal features of a ship’s
diesel engine EGT, aiding in improving engine performance, preventing faults, reducing
maintenance costs, and controlling emissions to protect the environment. Liu et al. [81]
constructed a ship diesel engine exhaust temperature prediction model combining the
feature extraction capabilities of the attention mechanism with the time series memory
Information 2024, 15, 507 18 of 33
for sea-state estimation. Sensitivity analysis assessed the impact of data preprocessing, and
real-time testing further demonstrated the model’s practicality.
Combining various neural networks and feature extraction techniques enhances sea-
state estimation accuracy. Wang et al. [75] proposed a deep neural network model named
DynamicSSE, which autonomously learns the dynamic correlations between sensors at
different locations, generating more valuable predictive features. DynamicSSE includes
feature extraction and dynamic graph structure construction modules, effectively capturing
graph structures and dependencies over long and short periods by combining a CNN
and the LSTM network. Their experimental results indicated that DynamicSSE outper-
formed baseline methods on two vessel motion datasets and demonstrated its practicality
and effectiveness through real-world testing. Additionally, Cheng et al. [74] introduced
a deep learning sea-state estimation model (SpectralNet) based on spectrograms. Com-
pared to methods directly applied to raw time series data, SpectralNet achieved higher
classification accuracy. This approach significantly enhances the interpretability and ac-
curacy of deep-learning-based sea-state estimation models by transforming time–domain
data into time–frequency spectrograms using vessel movement data from a commercial
simulation platform.
4.1.4. Others
These are other studies of deep learning work conducted in ship operation-related
applications. Ljunggren [99] demonstrated that deep neural networks could learn specific
ship features and outperform the 1-Nearest Neighbor (1NN) baseline method in ship-type
classification, mainly when the classification confidence was highest for 50% of the data.
Additionally, the study proposed an improved short-term traffic prediction model using
LSTM with RNN, chosen for its ability to remember long-term historical input data and
automatically determine the optimal time lag. The experimental results showed that this
model outperformed existing models in accuracy. Zhang et al. [84] combined CNN and
LSTM to construct a data-driven neural network model for forecasting the roll motion of
unmanned surface vehicles (USVs). CNN was employed to extract spatial correlations
and local time series features from USV sensor data. At the same time, the LSTM layer
captured the long-term motion processes of the USV and predicted the roll motion for
the next moment. A fully connected layer decoded the LSTM output and calculated the
final prediction results. Furthermore, Ma et al. [79] proposed a data-driven method for
predicting early collision risk by analyzing encountering vessels’ spatiotemporal motion
Information 2024, 15, 507 20 of 33
problem (DeepDLP). This framework aims to determine pricing strategies that adapt to
the dynamic nature of liner trade, characterized by large volumes, low costs, and high
capacity. DeepDLP features a dynamic price prediction module and an RL-based pricing
module, built by analyzing and studying data from real shipping scenarios. Experimental
simulations demonstrated that DeepDLP effectively improves liner trade revenue, veri-
fying its efficiency and effectiveness. Additionally, Alqatawna et al. [106] utilized time
series analysis techniques to predict resource demands for logistics companies, optimizing
order volume forecasts and determining staffing requirements. The study applied methods
like Autoregressive (AR), Autoregressive Integrated Moving Average (ARIMA), and Sea-
sonal Autoregressive Integrated Moving-Average with Exogenous Factors (SARIMAX) to
capture trends and seasonal components, providing interpretable results. This approach
helps companies accurately estimate resources needed for packaging based on predicted
order volumes.
Lim et al. [107] proposed an LSTM-based time series model to predict future values
of various liquid cargo transports in ship transport and port management. Traditional
models like ARIMA and Vector Autoregression (VAR) often fail to consider the linear
dependencies between different types of liquid cargo transport values. The proposed LSTM
model incorporates techniques for handling missing values, considers both short-term
and long-term dependencies, and uses auxiliary variables such as the inflation rate, the
USD exchange rate, the GDP value, and international oil prices to improve prediction
accuracy. Cheng et al. [108] introduced an integrated edRVFL algorithm for predicting ship
order dynamics. The algorithm enhances prediction performance and generalization by
combining deep feature extraction and ensemble learning with a minimal embedding strat-
egy. Accurate ship order predictions are critical for strategic planning in dynamic market
environments, and the edRVFL model demonstrated improved forecasting capabilities and
flexibility across different regions. Mo et al. [54] proposed a novel Convolutional Recurrent
Neural Network for time charter rate prediction under multivariate conditions, specifically
targeting monthly time charter rates for tankers and bulk carriers. The model showed
significant advantages in reflecting dynamic changes in the shipping market.
For ship market index prediction, Kamal et al. [85] proposed an integrated deep
learning method for short-term and long-term Baltic Dry Index (BDI) forecasting, assisting
stakeholders and shipowners in making informed business decisions and mitigating market
risks. The study employed Recurrent Neural Network models (RNN, LSTM, and GRU)
for BDI prediction and advanced sequential deep learning models for one-step and multi-
step forecasting. These models were combined into a Deep Ensemble Recurrent Network
(DERN) to enhance prediction accuracy. Their experimental results indicated that DERN
outperformed traditional methods such as ARIMA, MLP, RNN, LSTM, and GRU in both
short-term and long-term BDI predictions.
5. Overall Analysis
5.1. Literature Description
This section primarily explores the distribution of the 89 academic papers we have
collected. Through the analysis of extensively compiled scholarly literature, the content of
this section presents the development trends and the distribution of academic focus within
Information 2024, 15, x FOR PEER REVIEW 25 of 36
this research field.
Sensors
5.1.2. Literature Classification
Given the broad scope of the maritime field and the diverse types of time series data
Journal of Physics: Conference
available, Series
it is necessary to categorize the various applications found in the literature more
systematically. In conjunction with the data we have collected, we have referenced several
0 1 2 3 4 5 6 7
Figure 17. The number of papers published in journals or presented at conferences. Notes: The full
Figure 17. The number of papers published in journals or presented at conferences. Notes: The full
name of “Journal of Marine Science and…” is “Journal of Marine Science and Engineering”.
name of “Journal of Marine Science and. . .” is “Journal of Marine Science and Engineering”.
5.1.2. Literature Classification
Given the broad scope of the maritime field and the diverse types of time series data
available, it is necessary to categorize the various applications found in the literature more
systematically. In conjunction with the data we have collected, we have referenced several
Information 2024, 15, 507 23 of 33
5.1.2.
Information 2024, 15, x FOR PEER REVIEW Literature Classification 26 of 36
Given the broad scope of the maritime field and the diverse types of time series data
available, it is necessary to categorize the various applications found in the literature more
systematically.
review articles andIn conjunction
classified the 89with the data
primary paperswe have
into collected,
three we haveship
main categories: referenced
oper- several
review articles
ation-related, portand classified the 89and
operation-related, primary
shippingpapers into three areas
market-related main[108–112].
categories: ship operation-
Figure
18related,
displays a pie
port chart illustrating the
operation-related, andclassification of the collected literature.
shipping market-related Among Figure
areas [108–112]. the 18 dis-
data
playsweagathered,
pie chart71illustrating
papers related
thetoclassification
ship operations constitute
of the 79%literature.
collected of the total,Among
a pro- the data
portion visible in
we gathered, 71the chart.related
papers This indicates a strong scholarly
to ship operations interest
constitute 79%inofutilizing
the total,deep
a proportion
learning architectures and time series data to predict or classify tasks related to ship
visible in the chart. This indicates a strong scholarly interest in utilizing deep learning op-
erations, aiming to optimize ship operation processes.
architectures and time series data to predict or classify tasks related to ship operations,
aiming to optimize ship operation processes.
11%
9%
80%
21%
9%
54%
16%
AIS Data
High-Frequency Radar Data and Sensor Data
Container Throughput Data
Other Datasets
Figure 19.
Figure 19. Data
Datautilized
utilizedin in
maritime research.
maritime research.
5.3. Evaluation
5.3. Evaluation Parameters
Parameters
After analyzing
After analyzingthe assessment
the assessmentcriteria employed
criteria in 89 academic
employed papers, papers,
in 89 academic we discov-
we discov-
ered a wide range of measurement methods extensively used for assessingeffective-
ered a wide range of measurement methods extensively used for assessing the the effectiveness
ness of machine learning and deep learning models. Prominent metrics for regression
of machine learning and deep learning models. Prominent metrics for regression tasks in-
tasks include Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Abso-
clude Mean Squared Error (MSE), Root Mean Squared Error (RMSE), Mean Absolute Error
lute Error (MAE), and MAPE, which primarily focus on quantifying prediction errors.
(MAE),
Equationsand MAPE,
(22) throughwhich primarily
(25) present thefocus on quantifying
mathematical prediction
constructs for MSE, errors.
RMSE, Equations
MAE, (22)
through
and MAPE. (25) present the mathematical constructs for MSE, RMSE, MAE, and MAPE.
∑𝑛𝑖=1(𝑦 − 𝑦̂ )22
MSE = ∑in=1𝑟𝑒𝑎𝑙
yreal −𝑝𝑟𝑒𝑑
ŷ pred (22)
MSE = 𝑛 (22)
n
∑ 𝑛 − 𝑦̂𝑝𝑟𝑒𝑑 )22
𝑖=1(𝑦
v
n 𝑟𝑒𝑎𝑙
u
RMSE = √ u
t ∑i=1 yreal − ŷ pred (23)
RMSE = 𝑛 (23)
n
∑in=1 yreal − ŷ pred
MAE = (24)
n
Information 2024, 15, 507 25 of 33
TP + TN
Accuracy = (26)
TP + FP + FN + TN
TP
Precision = (27)
TP + FP
TP
Recall = (28)
TP + FN
Precision ∗ Recall
F1 score = 2 ∗ (29)
Precision + Recal
TP, FP, TN, and FN represent the four essential components of the Confusion Matrix
TP (True Positives), FN (False Negatives), FP (False Positives), and TN (True Negatives).
In this context, “True” and “False” denote the accuracy of the model’s classification; “True”
implies correct classification, while “False” indicates a classification error. “Positives” and
“Negatives” pertain to the predictions made by the classifier, where “Positives” suggests
a prediction of the presence of a condition and “Negatives” signifies the prediction of its
absence. These components are vital metrics for evaluating the efficacy of classification
models, enabling the computation of critical statistical indicators such as accuracy, recall,
precision, and the F1 score. These metrics are indispensable for assessing the performance
of predictive models in various classification scenarios. Some literature concerning novel
architectures or model constructions also considers model execution speed or computa-
tional time as evaluation metrics [115,118]. For different tasks, other evaluation metrics
such as maintenance score, cumulative reward, and distance are also used within the given
collection of papers [119,120]. This diverse array of metrics reflects the problem’s specific
needs, the data’s characteristics, and considerations in model design, all of which are crucial
for accurately assessing and further optimizing model performance.
more precise analytical tool for research in fields that require nuanced spatial and temporal
analysis [125].
6. Conclusions
This review comprehensively analyzes the various applications of deep learning in
maritime time series prediction. Through an extensive literature search in the Web of Sci-
ence and Scopus, 89 papers were collected, covering ship operations, port operations, and
shipping market forecasts. By examining various deep learning models, including ANN,
MLP, DNN, the LSTM network, TCN, RNN, GRU, CNN, RVFL, and Transformer models,
this review provides valuable insights into their potential and limitations in maritime
applications. The review investigates previous research on diverse applications such as pre-
dicting ship trajectories, weather forecasting, estimating ship fuel consumption, forecasting
port throughput, and predicting shipping market orders. It summarizes the distribution
and classification of the literature and analyzes the data utilized in maritime research,
including AIS data, high-frequency radar data and sensor data, container throughput
Information 2024, 15, 507 28 of 33
data, and other datasets. In summarizing future research directions, the review highlights
several key areas for further investigation, such as improving data handling techniques
to enhance accuracy, optimizing models for better performance, applying deep learning
models to specific scenarios, conducting long-term predictions, and exploring cross-domain
applications. By synthesizing the current state of deep learning applications in maritime
time series prediction, this review underscores both achievements and challenges, offering
a foundation for future studies to build upon.
The main contributions of this study are outlined as follows.
(1) This study fills the gap in the literature on advancements in deep learning techniques
for time series forecasting in maritime applications, focusing on three key areas: ship
operations, port operations, and shipping markets.
(2) Different types of deep learning models are compared to identify the most suitable
models for various applications. The differences between these models are discussed,
providing valuable insights for domain researchers.
(3) The study summarizes future research directions, helping to clarify future research
objectives and guide subsequent studies. These directions include enhancing data
processing and feature extraction, optimizing deep learning models, and exploring
specific application scenarios. Additionally, practical applications and long-term
predictions, environmental impacts, fault prediction, and cross-domain applications
are addressed to provide comprehensive guidance for future research efforts.
While significant progress has been made, addressing these gaps and leveraging
advanced deep learning techniques will enhance operational efficiency, safety, and sustain-
ability in maritime applications. However, this review has limitations, including the rapidly
evolving nature of the field, which may lead to the omission of the latest developments.
Additionally, the review may not fully capture the diversity of maritime applications and
the contextual nuances of different regions. Future reviews should continuously update
findings to reflect ongoing advancements and aim to include a broader range of sources
and perspectives to mitigate these limitations.
Author Contributions: Conceptualization, Z.S.C., Y.S. and M.L.; methodology, Y.S., Z.S.C., M.W.
and X.G.; software, M.W. and X.G.; validation, Z.S.C., Y.S. and Y.Z.; formal analysis, M.W. and
X.G.; investigation, M.W. and X.G.; resources, Z.S.C., Y.Z. and M.L.; data curation, M.W. and X.G.;
writing—original draft preparation, M.W. and X.G.; writing—review and editing, M.W., X.G., Z.S.C.
and M.L.; visualization, M.W. and X.G.; supervision, Z.S.C. and Y.S.; project administration, Y.S.
and Z.S.C.; funding acquisition, Y.S., Z.S.C., Y.Z. and M.L. All authors have read and agreed to the
published version of the manuscript.
Funding: This research was funded by Xi’an Jiaotong-Liverpool University, grant number RDF-22-02-026.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Data are contained within the article.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. UNCTAD. Review of Maritime Transport 2023; United Nations Conference on Trade and Development: Geneva, Switzerland, 2023;
Available online: https://www.un-ilibrary.org/content/books/9789213584569 (accessed on 1 April 2024).
2. Liang, M.; Liu, R.W.; Zhan, Y.; Li, H.; Zhu, F.; Wang, F.Y. Fine-Grained Vessel Traffic Flow Prediction With a Spatio-Temporal
Multigraph Convolutional Network. IEEE Trans. Intell. Transp. Syst. 2022, 23, 23694–23707. [CrossRef]
3. Liu, R.W.; Liang, M.; Nie, J.; Lim, W.Y.B.; Zhang, Y.; Guizani, M. Deep Learning-Powered Vessel Trajectory Prediction for
Improving Smart Traffic Services in Maritime Internet of Things. IEEE Trans. Netw. Sci. Eng. 2022, 9, 3080–3094. [CrossRef]
4. Dui, H.; Zheng, X.; Wu, S. Resilience analysis of maritime transportation systems based on importance measures. Reliab. Eng.
Syst. Saf. 2021, 209, 107461. [CrossRef]
5. Liang, M.; Li, H.; Liu, R.W.; Lam, J.S.L.; Yang, Z. PiracyAnalyzer: Spatial temporal patterns analysis of global piracy incidents.
Reliab. Eng. Syst. Saf. 2024, 243, 109877. [CrossRef]
Information 2024, 15, 507 29 of 33
6. Chen, Z.S.; Lam, J.S.L.; Xiao, Z. Prediction of harbour vessel emissions based on machine learning approach. Transp. Res. Part D
Transp. Environ. 2024, 131, 104214. [CrossRef]
7. Chen, Z.S.; Lam, J.S.L.; Xiao, Z. Prediction of harbour vessel fuel consumption based on machine learning approach. Ocean Eng.
2023, 278, 114483. [CrossRef]
8. Liang, M.; Weng, L.; Gao, R.; Li, Y.; Du, L. Unsupervised maritime anomaly detection for intelligent situational awareness using
AIS data. Knowl.-Based Syst. 2024, 284, 111313. [CrossRef]
9. Dave, V.S.; Dutta, K. Neural network based models for software effort estimation: A review. Artif. Intell. Rev. 2014, 42, 295–307.
[CrossRef]
10. Uslu, S.; Celik, M.B. Prediction of engine emissions and performance with artificial neural networks in a single cylinder diesel
engine using diethyl ether. Eng. Sci. Technol. Int. J. 2018, 21, 1194–1201. [CrossRef]
11. Chaudhary, L.; Sharma, S.; Sajwan, M. Systematic Literature Review of Various Neural Network Techniques for Sea Surface
Temperature Prediction Using Remote Sensing Data. Arch. Comput. Methods Eng. 2023, 30, 5071–5103. [CrossRef]
12. Dharia, A.; Adeli, H. Neural network model for rapid forecasting of freeway link travel time. Eng. Appl. Artif. Intell. 2003, 16,
607–613. [CrossRef]
13. Hecht-Nielsen, R. Applications of counterpropagation networks. Neural Netw. 1988, 1, 131–139. [CrossRef]
14. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016.
15. Veerappa, M.; Anneken, M.; Burkart, N. Evaluation of Interpretable Association Rule Mining Methods on Time-Series in the
Maritime Domain. Springer International Publishing: Cham, Switzerland, 2021; pp. 204–218.
16. Frizzell, J.; Furth, M. Prediction of Vessel RAOs: Applications of Deep Learning to Assist in Design. In Proceedings of the SNAME
27th Offshore Symposium, Houston, TX, USA, 22 February 2022. [CrossRef]
17. Van Den Oord, A.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K.
Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499.
18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
19. Ning, C.X.; Xie, Y.Z.; Sun, L.J. LSTM, WaveNet, and 2D CNN for nonlinear time history prediction of seismic responses. Eng.
Struct. 2023, 286, 116083. [CrossRef]
20. Schmidt, W.F.; Kraaijveld, M.A.; Duin, R.P. Feed forward neural networks with random weights. In International Conference on
Pattern Recognition; IEEE Computer Society Press: Washington, DC, USA, 1992; pp. 1–4.
21. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501.
[CrossRef]
22. Pao, Y.H.; Park, G.H.; Sobajic, D.J. Learning and generalization characteristics of the random vector functional-link net. Neurocom-
puting 1994, 6, 163–180. [CrossRef]
23. Zhang, L.; Suganthan, P.N. A comprehensive evaluation of random vector functional link networks. Inf. Sci. 2016, 367, 1094–1105.
[CrossRef]
24. Huang, G.; Huang, G.B.; Song, S.J.; You, K.Y. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48.
[CrossRef]
25. Shi, Q.S.; Katuwal, R.; Suganthan, P.N.; Tanveer, M. Random vector functional link neural network based ensemble deep learning.
Pattern Recognit. 2021, 117, 107978. [CrossRef]
26. Du, L.; Gao, R.B.; Suganthan, P.N.; Wang, D.Z.W. Graph ensemble deep random vector functional link network for traffic
forecasting. Appl. Soft Comput. 2022, 131, 109809. [CrossRef]
27. Rehman, A.; Xing, H.L.; Hussain, M.; Gulzar, N.; Khan, M.A.; Hussain, A.; Mahmood, S. HCDP-DELM: Heterogeneous chronic
disease prediction with temporal perspective enabled deep extreme learning machine. Knowl.-Based Syst. 2024, 284, 111316.
[CrossRef]
28. Gao, R.B.; Li, R.L.; Hu, M.H.; Suganthan, P.N.; Yuen, K.F. Online dynamic ensemble deep random vector functional link neural
network for forecasting. Neural Netw. 2023, 166, 51–69. [CrossRef]
29. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86,
2278–2324. [CrossRef]
30. Palaz, D.; Magimai-Doss, M.; Collobert, R. End-to-end acoustic modeling using convolutional neural networks for HMM-based
automatic speech recognition. Speech Commun. 2019, 108, 15–32. [CrossRef]
31. Fang, W.; Love, P.E.D.; Luo, H.; Ding, L. Computer vision for behaviour-based safety in construction: A review and future
directions. Adv. Eng. Inform. 2020, 43, 100980. [CrossRef]
32. Qin, L.; Yu, N.; Zhao, D. Applying the convolutional neural network deep learning technology to behavioural recognition in
intelligent video. Teh. Vjesn. 2018, 25, 528–535.
33. Hoseinzade, E.; Haratizadeh, S. CNNpred: CNN-based stock market prediction using a diverse set of variables. Expert Syst. Appl.
2019, 129, 273–285. [CrossRef]
34. Rasp, S.; Dueben, P.D.; Scher, S.; Weyn, J.A.; Mouatadid, S.; Thuerey, N. WeatherBench: A Benchmark Data Set for Data-Driven
Weather Forecasting. J. Adv. Model. Earth Syst. 2020, 12, e2020MS002203. [CrossRef]
35. Crivellari, A.; Beinat, E.; Caetano, S.; Seydoux, A.; Cardoso, T. Multi-target CNN-LSTM regressor for predicting urban distribution
of short-term food delivery demand. J. Bus. Res. 2022, 144, 844–853. [CrossRef]
Information 2024, 15, 507 30 of 33
36. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling.
arXiv 2018, arXiv:1803.01271.
37. Lin, Z.; Yue, W.; Huang, J.; Wan, J. Ship Trajectory Prediction Based on the TTCN-Attention-GRU Model. Electronics 2023, 12, 2556.
[CrossRef]
38. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity Mappings in Deep Residual Networks. In Computer Vision–ECCV 2016; Leibe, B.,
Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 630–645.
39. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks
from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958.
40. Bin Syed, M.A.; Ahmed, I. A CNN-LSTM Architecture for Marine Vessel Track Association Using Automatic Identification System
(AIS) Data. Sensors 2023, 23, 6400. [CrossRef] [PubMed]
41. Li, M.-W.; Xu, D.-Y.; Geng, J.; Hong, W.-C. A hybrid approach for forecasting ship motion using CNN–GRU–AM and GCWOA.
Appl. Soft Comput. 2022, 114, 108084. [CrossRef]
42. Zhang, B.; Wang, S.; Deng, L.; Jia, M.; Xu, J. Ship motion attitude prediction model based on IWOA-TCN-Attention. Ocean Eng.
2023, 272, 113911. [CrossRef]
43. Elman, J.L. Finding Structure in Time. Cogn. Sci. 1990, 14, 179–211. [CrossRef]
44. Shan, F.; He, X.; Armaghani, D.J.; Sheng, D. Effects of data smoothing and recurrent neural network (RNN) algorithms for
real-time forecasting of tunnel boring machine (TBM) performance. J. Rock Mech. Geotech. Eng. 2024, 16, 1538–1551. [CrossRef]
45. Apaydin, H.; Feizi, H.; Sattari, M.T.; Colak, M.S.; Shamshirband, S.; Chau, K.-W. Comparative Analysis of Recurrent Neural
Network Architectures for Reservoir Inflow Forecasting. Water 2020, 12, 1500. [CrossRef]
46. Ma, Z.; Zhang, H.; Liu, J. MM-RNN: A Multimodal RNN for Precipitation Nowcasting. IEEE Trans. Geosci. Remote Sens. 2023, 61,
1–14. [CrossRef]
47. Lu, M.; Xu, X. TRNN: An efficient time-series recurrent neural network for stock price prediction. Inf. Sci. 2024, 657, 119951.
[CrossRef]
48. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw.
1994, 5, 157–166. [CrossRef]
49. Kratzert, F.; Klotz, D.; Brenner, C.; Schulz, K.; Herrnegger, M. Rainfall–runoff modelling using Long Short-Term Memory (LSTM)
networks. Hydrol. Earth Syst. Sci. 2018, 22, 6005–6022. [CrossRef]
50. Hochreiter, S. Untersuchungen zu dynamischen neuronalen Netzen. Diploma Tech. Univ. München 1991, 91, 31.
51. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [CrossRef] [PubMed]
52. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 2000, 12, 2451–2471.
[CrossRef] [PubMed]
53. Witten, I.H.; Frank, E. Data mining: Practical machine learning tools and techniques with Java implementations. SIGMOD Rec.
2002, 31, 76–77. [CrossRef]
54. Mo, J.X.; Gao, R.B.; Liu, J.H.; Du, L.; Yuen, K.F. Annual dilated convolutional LSTM network for time charter rate forecasting.
Appl. Soft Comput. 2022, 126, 109259. [CrossRef]
55. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder
approaches. arXiv 2014, arXiv:1409.1259.
56. Yang, S.; Yu, X.; Zhou, Y. LSTM and GRU Neural Network Performance Comparison Study: Taking Yelp Review Dataset as an
Example. In Proceedings of the 2020 International Workshop on Electronic Communication and Artificial Intelligence (IWECAI),
Shanghai, China, 12–14 June 2020; pp. 98–101. [CrossRef]
57. Zhao, Z.N.; Yun, S.N.; Jia, L.Y.; Guo, J.X.; Meng, Y.; He, N.; Li, X.J.; Shi, J.R.; Yang, L. Hybrid VMD-CNN-GRU-based model for
short-term forecasting of wind power considering spatio-temporal features. Eng. Appl. Artif. Intell. 2023, 121, 105982. [CrossRef]
58. Pan, N.; Ding, Y.; Fu, J.; Wang, J.; Zheng, H. Research on Ship Arrival Law Based on Route Matching and Deep Learning. J. Phys.
Conf. Ser. 2021, 1952, 022023. [CrossRef]
59. Ma, J.; Li, W.K.; Jia, C.F.; Zhang, C.W.; Zhang, Y. Risk Prediction for Ship Encounter Situation Awareness Using Long Short-Term
Memory Based Deep Learning on Intership Behaviors. J. Adv. Transp. 2020, 2020, 8897700. [CrossRef]
60. Suo, Y.F.; Chen, W.K.; Claramunt, C.; Yang, S.H. A Ship Trajectory Prediction Framework Based on a Recurrent Neural Network.
Sensors 2020, 20, 5133. [CrossRef] [PubMed]
61. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473.
[CrossRef]
62. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv.
Neural Inf. Process. Syst. 2017, 30, 03762.
63. Nascimento, E.G.S.; de Melo, T.A.C.; Moreira, D.M. A transformer-based deep neural network with wavelet transform for
forecasting wind speed and wind energy. Energy 2023, 278, 127678. [CrossRef]
64. Zhang, L.; Zhang, J.; Niu, J.; Wu, Q.M.J.; Li, G. Track Prediction for HF Radar Vessels Submerged in Strong Clutter Based on
MSCNN Fusion with GRU-AM and AR Model. Remote Sens. 2021, 13, 2164. [CrossRef]
65. Zhang, X.; Fu, X.; Xiao, Z.; Xu, H.; Zhang, W.; Koh, J.; Qin, Z. A Dynamic Context-Aware Approach for Vessel Trajectory Prediction
Based on Multi-Stage Deep Learning. IEEE Trans. Intell. Veh. 2024, 1–16. [CrossRef]
Information 2024, 15, 507 31 of 33
66. Jiang, D.; Shi, G.; Li, N.; Ma, L.; Li, W.; Shi, J. TRFM-LS: Transformer-Based Deep Learning Method for Vessel Trajectory Prediction.
J. Mar. Sci. Eng. 2023, 11, 880. [CrossRef]
67. Violos, J.; Tsanakas, S.; Androutsopoulou, M.; Palaiokrassas, G.; Varvarigou, T. Next Position Prediction Using LSTM Neural
Networks. In Proceedings of the 11th Hellenic Conference on Artificial Intelligence, Athens, Greece, 2–4 September 2020;
pp. 232–240. [CrossRef]
68. Hoque, X.; Sharma, S.K. Ensembled Deep Learning Approach for Maritime Anomaly Detection System. In Proceedings of the 1st
International Conference on Emerging Trends in Information Technology (ICETIT), Inst Informat Technol & Management, New
Delhi, India, 21–22 June 2020; In Lecture Notes in Electrical Engineering; Volume 605, pp. 862–869.
69. Wang, Y.; Zhang, M.; Fu, H.; Wang, Q. Research on Prediction Method of Ship Rolling Motion Based on Deep Learning. In
Proceedings of the 2020 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; pp. 7182–7187. [CrossRef]
70. Choi, J. Predicting the Frequency of Marine Accidents by Navigators’ Watch Duty Time in South Korea Using LSTM. Appl. Sci.
2022, 12, 11724. [CrossRef]
71. Li, T.; Li, Y.B. Prediction of ship trajectory based on deep learning. J. Phys. Conf. Ser. 2023, 2613, 012023. [CrossRef]
72. Chondrodima, E.; Pelekis, N.; Pikrakis, A.; Theodoridis, Y. An Efficient LSTM Neural Network-Based Framework for Vessel
Location Forecasting. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4872–4888. [CrossRef]
73. Long, Z.; Suyuan, W.; Zhongma, C.; Jiaqi, F.; Xiaoting, Y.; Wei, D. Lira-YOLO: A lightweight model for ship detection in radar
images. J. Syst. Eng. Electron. 2020, 31, 950–956. [CrossRef]
74. Cheng, X.; Li, G.; Skulstad, R.; Zhang, H.; Chen, S. SpectralSeaNet: Spectrogram and Convolutional Network-based Sea State
Estimation. In Proceedings of the IECON 2020 the 46th Annual Conference of the IEEE Industrial Electronics Society, Singapore,
18–21 October 2020; pp. 5069–5074. [CrossRef]
75. Wang, K.; Cheng, X.; Shi, F. Learning Dynamic Graph Structures for Sea State Estimation with Deep Neural Networks. In
Proceedings of the 2023 6th International Conference on Intelligent Autonomous Systems (ICoIAS), Qinhuangdao, China, 22–24
September 2023; pp. 161–166.
76. Yu, J.; Huang, D.; Shi, X.; Li, W.; Wang, X. Real-Time Moving Ship Detection from Low-Resolution Large-Scale Remote Sensing
Image Sequence. Appl. Sci. 2023, 13, 2584. [CrossRef]
77. Ilias, L.; Kapsalis, P.; Mouzakitis, S.; Askounis, D. A Multitask Learning Framework for Predicting Ship Fuel Oil Consumption.
IEEE Access 2023, 11, 132576–132589. [CrossRef]
78. Selimovic, D.; Hrzic, F.; Prpic-Orsic, J.; Lerga, J. Estimation of sea state parameters from ship motion responses using attention-
based neural networks. Ocean Eng. 2023, 281, 114915. [CrossRef]
79. Ma, J.; Jia, C.; Yang, X.; Cheng, X.; Li, W.; Zhang, C. A Data-Driven Approach for Collision Risk Early Warning in Vessel Encounter
Situations Using Attention-BiLSTM. IEEE Access 2020, 8, 188771–188783. [CrossRef]
80. Ji, Z.; Gan, H.; Liu, B. A Deep Learning-Based Fault Warning Model for Exhaust Temperature Prediction and Fault Warning of
Marine Diesel Engine. J. Mar. Sci. Eng. 2023, 11, 1509. [CrossRef]
81. Liu, Y.; Gan, H.; Cong, Y.; Hu, G. Research on fault prediction of marine diesel engine based on attention-LSTM. Proc. Inst. Mech.
Eng. Part M J. Eng. Marit. Environ. 2023, 237, 508–519. [CrossRef]
82. Li, M.W.; Xu, D.Y.; Geng, J.; Hong, W.C. A ship motion forecasting approach based on empirical mode decomposition method
hybrid deep learning network and quantum butterfly optimization algorithm. Nonlinear Dyn. 2022, 107, 2447–2467. [CrossRef]
83. Yang, C.H.; Chang, P.Y. Forecasting the Demand for Container Throughput Using a Mixed-Precision Neural Architecture Based
on CNN–LSTM. Mathematics 2020, 8, 1784. [CrossRef]
84. Zhang, W.; Wu, P.; Peng, Y.; Liu, D. Roll Motion Prediction of Unmanned Surface Vehicle Based on Coupled CNN and LSTM.
Future Internet 2019, 11, 243. [CrossRef]
85. Kamal, I.M.; Bae, H.; Sunghyun, S.; Yun, H. DERN: Deep Ensemble Learning Model for Short- and Long-Term Prediction of Baltic
Dry Index. Appl. Sci. 2020, 10, 1504. [CrossRef]
86. Li, M.Z.; Li, B.; Qi, Z.G.; Li, J.S.; Wu, J.W. Enhancing Maritime Navigational Safety: Ship Trajectory Prediction Using ACoAtt–
LSTM and AIS Data. ISPRS Int. J. Geo-Inform. 2024, 13, 85. [CrossRef]
87. Yu, T.; Zhang, Y.; Zhao, S.; Yang, J.; Li, W.; Guo, W. Vessel trajectory prediction based on modified LSTM with attention mechanism.
In Proceedings of the 2024 4th International Conference on Neural Networks, Information and Communication Engineering,
NNICE, Guangzhou, China, 19–21 January 2024; pp. 912–918. [CrossRef]
88. Xia, C.; Peng, Y.; Qu, D. A pre-trained model specialized for ship trajectory prediction. In Proceedings of the IEEE Advanced
Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 15–17 March 2024;
pp. 1857–1860. [CrossRef]
89. Cheng, X.; Li, G.; Skulstad, R.; Chen, S.; Hildre, H.P.; Zhang, H. Modeling and Analysis of Motion Data from Dynamically
Positioned Vessels for Sea State Estimation. In Proceedings of the 2019 International Conference on Robotics and Automation
(ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 6644–6650. [CrossRef]
90. Xia, C.; Qu, D.; Zheng, Y. TATBformer: A Divide-and-Conquer Approach to Ship Trajectory Prediction Modeling. In Proceedings
of the 2023 IEEE 11th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), Chongqing,
China, 8–10 December 2023; pp. 335–339. [CrossRef]
Information 2024, 15, 507 32 of 33
91. Ran, Y.; Shi, G.; Li, W. Ship Track Prediction Model based on Automatic Identification System Data and Bidirectional Cyclic
Neural Network. In Proceedings of the 2021 4th International Symposium on Traffic Transportation and Civil Architecture,
ISTTCA, Suzhou, China, 12–14 November 2021; pp. 297–301. [CrossRef]
92. Yang, C.H.; Wu, C.H.; Shao, J.C.; Wang, Y.C.; Hsieh, C.M. AIS-Based Intelligent Vessel Trajectory Prediction Using Bi-LSTM. IEEE
Access 2022, 10, 24302–24315. [CrossRef]
93. Sadeghi, Z.; Matwin, S. Anomaly detection for maritime navigation based on probability density function of error of reconstruction.
J. Intell. Syst. 2023, 32, 20220270. [CrossRef]
94. Perumal, V.; Murugaiyan, S.; Ravichandran, P.; Venkatesan, R.; Sundar, R. Real time identification of anomalous events in coastal
regions using deep learning techniques. Concurr. Comput. Pract. Exp. 2021, 33, e6421. [CrossRef]
95. Xie, J.L.; Shi, W.F.; Shi, Y.Q. Research on Fault Diagnosis of Six-Phase Propulsion Motor Drive Inverter for Marine Electric
Propulsion System Based on Res-BiLSTM. Machines 2022, 10, 736. [CrossRef]
96. Han, P.; Li, G.; Skulstad, R.; Skjong, S.; Zhang, H. A Deep Learning Approach to Detect and Isolate Thruster Failures for
Dynamically Positioned Vessels Using Motion Data. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [CrossRef]
97. Cheng, X.; Wang, K.; Liu, X.; Yu, Q.; Shi, F.; Ren, Z.; Chen, S. A Novel Class-Imbalanced Ship Motion Data-Based Cross-Scale
Model for Sea State Estimation. IEEE Trans. Intell. Transp. Syst. 2023, 24, 15907–15919. [CrossRef]
98. Lei, L.; Wen, Z.; Peng, Z. Prediction of Main Engine Speed and Fuel Consumption of Inland Ships Based on Deep Learning.
J. Phys. Conf. Ser. 2021, 2025, 012012.
99. Ljunggren, H. Using Deep Learning for Classifying Ship Trajectories. In Proceedings of the 21st International Conference on
Information Fusion (FUSION), Cambridge, UK, 10–13 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 2158–2164.
100. Kulshrestha, A.; Yadav, A.; Sharma, H.; Suman, S. A deep learning-based multivariate decomposition and ensemble framework
for container throughput forecasting. J. Forecast. 2024, in press. [CrossRef]
101. Shankar, S.; Ilavarasan, P.V.; Punia, S.; Singh, S.P. Forecasting container throughput with long short-term memory networks. Ind.
Manag. Data Syst. 2020, 120, 425–441. [CrossRef]
102. Lee, E.; Kim, D.; Bae, H. Container Volume Prediction Using Time-Series Decomposition with a Long Short-Term Memory Models.
Appl. Sci. 2021, 11, 8995. [CrossRef]
103. Cuong, T.N.; You, S.-S.; Long, L.N.B.; Kim, H.-S. Seaport Resilience Analysis and Throughput Forecast Using a Deep Learning
Approach: A Case Study of Busan Port. Sustainability 2022, 14, 13985. [CrossRef]
104. Song, X.; Chen, Z.S. Shipping market time series forecasting via an Ensemble Deep Dual-Projection Echo State Network. Comput.
Electr. Eng. 2024, 117, 109218. [CrossRef]
105. Li, X.; Hu, Y.; Bai, Y.; Gao, X.; Chen, G. DeepDLP: Deep Reinforcement Learning based Framework for Dynamic Liner Trade
Pricing. In Proceedings of the Proceedings of the 2023 17th International Conference on Ubiquitous Information Management
and Communication, IMCOM, Seoul, Republic of Korea, 3–5 January 2023; pp. 1–8. [CrossRef]
106. Alqatawna, A.; Abu-Salih, B.; Obeid, N.; Almiani, M. Incorporating Time-Series Forecasting Techniques to Predict Logistics
Companies’ Staffing Needs and Order Volume. Computation 2023, 11, 141. [CrossRef]
107. Lim, S.; Kim, S.J.; Park, Y.; Kwon, N. A deep learning-based time series model with missing value handling techniques to predict
various types of liquid cargo traffic. Expert Syst. Appl. 2021, 184, 115532. [CrossRef]
108. Cheng, R.; Gao, R.; Yuen, K.F. Ship order book forecasting by an ensemble deep parsimonious random vector functional link
network. Eng. Appl. Artif. Intell. 2024, 133, 108139. [CrossRef]
109. Xiao, Z.; Fu, X.J.; Zhang, L.Y.; Goh, R.S.M. Traffic Pattern Mining and Forecasting Technologies in Maritime Traffic Service
Networks: A Comprehensive Survey. IEEE Trans. Intell. Transp. Syst. 2020, 21, 1796–1825. [CrossRef]
110. Yan, R.; Wang, S.A.; Psaraftis, H.N. Data analytics for fuel consumption management in maritime transportation: Status and
perspectives. Transp. Res. Part E Logist. Transp. Rev. 2021, 155, 102489. [CrossRef]
111. Filom, S.; Amiri, A.M.; Razavi, S. Applications of machine learning methods in port operations—A systematic literature review.
Transp. Res. Part E-Logist. Transp. Rev. 2022, 161, 102722. [CrossRef]
112. Ksciuk, J.; Kuhlemann, S.; Tierney, K.; Koberstein, A. Uncertainty in maritime ship routing and scheduling: A Literature review.
Eur. J. Oper. Res. 2023, 308, 499–524. [CrossRef]
113. Jia, H.; Prakash, V.; Smith, T. Estimating vessel payloads in bulk shipping using AIS data. Int. J. Shipp. Transp. Logist. 2019, 11,
25–40. [CrossRef]
114. Yang, D.; Wu, L.X.; Wang, S.A.; Jia, H.Y.; Li, K.X. How big data enriches maritime research—A critical review of Automatic
Identification System (AIS) data applications. Transp. Rev. 2019, 39, 755–773. [CrossRef]
115. Liu, M.; Zhao, Y.; Wang, J.; Liu, C.; Li, G. A Deep Learning Framework for Baltic Dry Index Forecasting. Procedia Comput. Sci.
2022, 199, 821–828. [CrossRef]
116. Wang, Y.C.; Wang, H.; Zou, D.X.; Fu, H.X. Ship roll prediction algorithm based on Bi-LSTM-TPA combined model. J. Mar. Sci.
Eng. 2021, 9, 387. [CrossRef]
117. Xie, H.T.; Jiang, X.Q.; Hu, X.; Wu, Z.T.; Wang, G.Q.; Xie, K. High-efficiency and low-energy ship recognition strategy based on
spiking neural network in SAR images. Front. Neurorobotics 2022, 16, 970832. [CrossRef]
Information 2024, 15, 507 33 of 33
118. Muñoz, D.U.; Ruiz-Aguilar, J.J.; González-Enrique, J.; Domínguez, I.J.T. A Deep Ensemble Neural Network Approach to
Improve Predictions of Container Inspection Volume. In Proceedings of the 15th International Work-Conference on Artificial
Neural Networks (IWANN), Gran Canaria, Spain, 12–14 June 2019; Springer: Berlin/Heidelberg, Germany, 2019; Volume 11506,
pp. 806–817. [CrossRef]
119. Velasco-Gallego, C.; Lazakis, I. Mar-RUL: A remaining useful life prediction approach for fault prognostics of marine machinery.
Appl. Ocean Res. 2023, 140, 103735. [CrossRef]
120. Zhang, X.; Zheng, K.; Wang, C.; Chen, J.; Qi, H. A novel deep reinforcement learning for POMDP-based autonomous ship
collision decision-making. Neural Comput. Appl. 2023, 1–15. [CrossRef]
121. Guo, X.X.; Zhang, X.T.; Lu, W.Y.; Tian, X.L.; Li, X. Real-time prediction of 6-DOF motions of a turret-moored FPSO in harsh sea
state. Ocean Eng. 2022, 265, 112500. [CrossRef]
122. Kim, D.; Kim, T.; An, M.; Cho, Y.; Baek, Y.; IEEE. Edge AI-based early anomaly detection of LNG Carrier Main Engine systems. In
Proceedings of the OCEANS Conference, Limerick, Ireland, 5–8 June 2023. [CrossRef]
123. Theodoropoulos, P.; Spandonidis, C.C.; Giannopoulos, F.; Fassois, S. A Deep Learning-Based Fault Detection Model for Optimiza-
tion of Shipping Operations and Enhancement of Maritime Safety. Sensors 2021, 21, 5658. [CrossRef] [PubMed]
124. Huang, B.; Wu, B.; Barry, M. Geographically and temporally weighted regression for modeling spatio-temporal variation in house
prices. Int. J. Geogr. Inf. Sci. 2010, 24, 383–401. [CrossRef]
125. Zhang, W.; Xu, Y.; Streets, D.G.; Wang, C. How does decarbonization of the central heating industry affect employment? A
spatiotemporal analysis from the perspective of urbanization. Energy Build. 2024, 306, 113912. [CrossRef]
126. Zhang, D.; Li, X.; Wan, C.; Man, J. A novel hybrid deep-learning framework for medium-term container throughput forecasting:
An application to China’s Guangzhou, Qingdao and Shanghai hub ports. Marit. Econ. Logist. 2024, 26, 44–73. [CrossRef]
127. Wang, Y.; Wang, H.; Zhou, B.; Fu, H. Multi-dimensional prediction method based on Bi-LSTMC for ship roll. Ocean Eng. 2021,
242, 110106. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.