Intelligent Tire Sensor-Based Real-Time Road Surface
Intelligent Tire Sensor-Based Real-Time Road Surface
Article
Intelligent Tire Sensor-Based Real-Time Road Surface
Classification Using an Artificial Neural Network
Dongwook Lee , Ji-Chul Kim, Mingeuk Kim and Hanmin Lee *
Department of Smart Industrial Machine Technologies, Korea Institute of Machinery and Materials,
156 Gajeongbuk-Ro, Yuseong-Gu, Daejeon 34103, Korea; [email protected] (D.L.); [email protected] (J.-C.K.);
[email protected] (M.K.)
* Correspondence: [email protected]
Abstract: Vehicles today have many advanced driver assistance control systems that improve vehicle
safety and comfort. With the development of more sophisticated vehicle electronic control and
autonomous driving technology, the need and effort to estimate road surface conditions is increasing.
In this paper, a real-time road surface classification algorithm, based on a deep neural network, is
developed using a database collected through an intelligent tire sensor system with a three-axis
accelerometer installed inside the tire. Two representative types of network, fully connected neural
network (FCNN) and convolutional neural network (CNN), are learned with each of the three-axis
acceleration sensor signals, and their performances were compared to obtain an optimal learning
network result. The learning results show that the road surface type can be classified in real-time
with sufficient accuracy when the longitudinal and vertical axis acceleration signals are trained
with the CNN. In order to improve classification accuracy, a CNN with multiple input that can
simultaneously learn 2-axis or 3-axis acceleration signals is suggested. In addition, by analyzing
how the accuracy of the network is affected by number of classes and length of input data, which is
related to delay of classification, the appropriate network can be selected according to the application.
Citation: Lee, D.; Kim, J.; Kim, M.; The proposed real-time road surface classification algorithm is expected to be utilized with various
Lee, H. Intelligent Tire Sensor-Based
vehicle electronic control systems and makes a contribution to improving vehicle performance.
Real-Time Road Surface Classification
Using an Artificial Neural Network.
Keywords: intelligent tire; road surface classification; deep neural network
Sensors 2021, 21, 3233. https://
doi.org/10.3390/s21093233
possible to improve the convenience by utilizing the roughness of the road to control an
active suspension system that changes the suspension properties.
Many previous studies have been conducted to estimate road surface conditions or
the road friction coefficient. The methods used in these studies can be divided into three
categories: (1) tire force-slip-based methods, (2) road property-based methods, and (3)
tire–road interaction-based methods.
The tire force-slip-based method has been applied most often in previous studies.
Based on a vehicle and tire dynamics model, the behavior of the vehicle and tires are
measured, the tire contact force and slip ratio (the speed ratio between the driving and
driven wheels) are estimated, and the resulting force-slip relationship is used to determine
the potential grip of the vehicle on the road [9–13]. This approach has the advantage of
making it possible to estimate the road surface friction coefficient with numerical value
from the vehicle dynamics model, and the number of experiments required is relatively
small because the experimental results are needed only for identifying the parameters of a
physical model. However, due to the difficulty in estimating the exact parameters of the
vehicle model, and in applying tire force-slip-based approaches in low excitation situations,
these approaches have limitations for application prospects in general driving situations.
A road property-based method classifies the road surface by extracting the physical
properties of the road surface by using noncontact sensors, such as cameras, light detection
and ranging (LiDAR), radar, ultrasonic sensors, or optical sensors [14–18]. This approach
has been applied mainly in autonomous vehicles that are basically installed with noncontact
sensors including cameras and LiDAR. These methods have the advantage of being able to
operate even when the vehicle is not driving; however, the performance of road property-
based methods can degrade significantly in adverse weather conditions such as snow, rain,
fog and sunglare. Moreover, the camera image-based approach requires a large database of
experiments that reflect a variety of environmental conditions, including changes in region,
weather conditions and the intensity of illumination.
In tire–road interaction-based methods, road surface conditions are classified by
measuring tire vibration or deformation caused by the friction force on the tire contact
patch [19–23]. Similar to road property-based methods, tire–road interaction-based meth-
ods have seen an upswing in research activity in recent years because these approaches
can obtain useful signals that reflect road surface properties even in low-excitation situ-
ations. Tire–road interaction-based methods also offer the additional benefit that most
of the sensors these methods require are inexpensive and are more robust to the external
environmental conditions than the sensors used in road property-based methods.
Among the various types of sensors that can measure tire–road interactions, intelligent
tire sensor that directly measures the acceleration or deformation of tire elements through
sensors installed inside the tire have been applied in many recent studies [24–28]. Current
studies that use intelligent tires for estimating road conditions have succeeded in finding
the characteristic features from the waveforms of acceleration sensor signals for different
types of road surfaces through experimental data, but no algorithm that can classify road
surface types from sensor signals in real time has yet been suggested. Therefore, the goal
of this study is to develop an algorithm that estimates road surface type in real time by
using a deep neural network (DNN) based on signals from an intelligent tire system with
an accelerometer attached inside the tires.
Section 2 describes the experimental setup, data acquisition experiment using intel-
ligent tire sensor on various road surfaces for an acceleration signal database, and the
data preprocessing and training process with DNNs. In Section 3, the performances of
deep learning with two different types of neural networks and with three types of accelera-
tion signal databases (x-, y- and z-axis) are compared. In order to improve classification
accuracy, a network with multiple input that can simultaneously learn 2-axis or 3-axis
acceleration signals is proposed. In addition, how the accuracy is affected by the number of
classes and length of input data is analyzed so that the network can be selected according
to the application. By applying the trained network to actual intelligent tire sensor experi-
Sensors 2021, 21, 3233 3 of 17
mental data, the classification performance in real time was also confirmed. In Section 4,
we summarize the findings of this study and provide suggestions for future work.
y
x
z
Hermetic
connecter
Telemetry module
with battery
(a) (b)
Figure 1. Intelligent tire system setup: (a) Attachment of accelerometer inside the tire. (b) Hermetic
connector and telemetry installation on the vehicle wheel.
To collect the vehicle driving status information, the steering angle, wheel speed, and
brake on/off signals were acquired from the vehicle internal controller area network (CAN)
at a sampling rate of 100 Hz in synchronization with the intelligent tire sensor signal. Since
the vehicle internal CAN signal was used, additional sensors other than the intelligent tire
sensor were not utilized. To receive signals from intelligent tire sensors and vehicle CAN
network simultaneously with time synchronization, DAQ system composed of cRIO-9063
+ NI-9853 of National Instrument was utilized.
driving command, but the driving test was conducted with an error of less than 10 kph
from the target vehicle speed. Due to operational limitations on the gravel and unpaved
roads, the experiments on these surfaces were conducted at up to only 50 and 20 kph,
respectively. For each road surface and vehicle speed, the experiment was repeated until
sufficient experimental data with a total effective driving distance greater than 800 m,
which corresponds to more than 400 instances of contact of the intelligent tire sensor with
the road surface.
Figure A1 shows the x-, y- and z-axis accelerometer measurement results of the 30 kph
experiment (20 kph for the unpaved roads), and the characteristics of the acceleration
signals from each road surface type can be compared with each other. To easily compare
the acceleration signals for each road surface, a time band with vehicle speed as close
as 30 kph was selected to reduce the effect of vehicle speed on the signal amplitude and
frequency. In the case of the unpaved road, since the experiment was performed up to only
20 kph, the signals of the 20 kph experiment were compared. The comparative analysis
will be described with the learning result in the next section.
(d) (e)
Figure 2. Data acquisition experiment environment of (a) dry asphalt, (b) wet asphalt with 1 mm
waterfilm, (c) wet asphalt with 4 mm waterfilm, (d) gravel, and (e) unpaved road.
q
µ x = ∑ik=+kn−1 A x (ti )/n, s x = ∑ik=+kn−1 ( A x (ti ) − µ x )2 )/n,
(1)
A0x (ti ) = ( A x (ti ) − µ x )/s x , i = k, k + 1, ...k + n − 1.
After normalizing the data, each sample data was labeled with the types of road
surfaces (1: dry asphalt, 2: wet asphalt 1, 3: wet asphalt 2, 4: gravel, 5: unpaved) and
stacked on the training or testing dataset. Figure 3 shows the process of obtaining the
training and testing dataset from the experimental results.
50
Steer/5 (degree)
Wheel speed (kph)
40
Brake on/off Wheel speed
Effective learning period
bound
30
steering angle/5 (deg)
Wheel speed (kph) &
Brake on/off
20
10 Steering angle On
bound
0 Off
⇒
Time window:
−10 [𝒕𝒌 𝒕𝒌+𝒏−𝟏 ]
−20
0 20 40 60 80 100 120 140
time(sec)
Data
Raw experiment 1: Dry, 2: Wet1, Training &
normalization
data 3: Wet2, 4: Gravel, Testing
5: Unpaved
(Ax, Ay, Az) Dataset
Sample Data Data Labeling
Through this process, a training dataset with 100,000 samples (20,000 samples per
road surface) and a testing dataset with 20,000 samples (4000 samples per road surface)
were acquired for each x-, y- and z-axis acceleration sensor signal. The training and testing
datasets were acquired from independent effective learning data periods to prevent the
network from learning the testing dataset in advance.
convolution and pooling and fully-connected layers [29]. Figure 4b shows the schematic
of a typical 1-dimensional CNN. The convolutional layer is a special form of network that
can extract the spatial features from images (or sequential features from 1-dimensional data)
by training the filters (kernels) inside the layer so that the CNN can identify the patterns or
objects. The pooling layer receives the output data from the convolution layer as the input
and is used to reduce the size of the output data or to highlight specific data. After sufficiently
recognizing the features of the image or sequential data with the convolutional layer, the
output of the last pooling layer is flattened and connected to the fully connected layer and the
softmax layer to finally output the classification result.
dry
wet 1
wet 2
⋯
⋯
gravel
⋯
⋯
unpaved
(a)
Input Conv. Layer Pooling Layer Conv. Layer Pooling Layer Flatten FC Layer FC Layer Softmax Layer
(1000) (1000 × 32) (200 × 32) (200 × 32) (40 × 32) Layer (50) (5)
(1280)
dry
wet 1
wet 2
⋯
⋯
gravel
unpaved
0 1
𝐶𝑜𝑛𝑣𝑜𝑙𝑢𝑡𝑖𝑜𝑛
𝐶𝑜𝑛𝑣𝑜𝑙𝑢𝑡𝑖𝑜𝑛 𝑀𝑎𝑥𝑝𝑜𝑜𝑙𝑖𝑛𝑔 𝑀𝑎𝑥𝑝𝑜𝑜𝑙𝑖𝑛𝑔
𝑓𝑖𝑙𝑡𝑒𝑟𝑠 # = 32
𝑓𝑖𝑙𝑡𝑒𝑟𝑠 # = 32 𝑓𝑖𝑙𝑡𝑒𝑟 𝑠𝑖𝑧𝑒 =5 𝑓𝑖𝑙𝑡𝑒𝑟 𝑠𝑖𝑧𝑒 = 5 𝑓𝑖𝑙𝑡𝑒𝑟 𝑠𝑖𝑧𝑒 = 5
𝑓𝑖𝑙𝑡𝑒𝑟 𝑠𝑖𝑧𝑒 = 20
𝑠𝑡𝑟𝑖𝑑𝑒 = 1
𝑠𝑡𝑟𝑖𝑑𝑒 = 1
𝑝𝑎𝑑𝑑𝑖𝑛𝑔 = 𝑠𝑎𝑚𝑒
𝑝𝑎𝑑𝑑𝑖𝑛𝑔 = 𝑠𝑎𝑚𝑒
𝑎𝑐𝑡. 𝑓𝑢𝑛𝑐. = 𝑅𝑒𝐿𝑈
𝑎𝑐𝑡. 𝑓𝑢𝑛𝑐. = 𝑅𝑒LU
(b)
Figure 4. Structure and parameters of each layer of the (a) FCNN and (b) 1-D CNN.
optimizer, a batch size of 64 and a learning rate of 0.01. To compare the performances of
the two networks, the time window length was fixed to 1000; that is, the input data for
training corresponded to a 1 s duration in the experiment. Of the training dataset, 20% was
separated as a validation set to check overfitting and to early-stop the training process. If,
during training, there is a mismatch between the values of the loss function in the training
set and the validation set, training stops before the network overfits.
To find the optimal network with the highest accuracy, the same dataset was repeatedly
trained by FCNNs and CNNs by changing the number of total layers, number of nodes in
fully-connected layer, number of filter and filter size of convolution layer, and filter size of
maxpooling layers. A network with a too-complex structure overfits the training dataset,
reducing the accuracy of the test dataset, and a network with a too simple structure cannot
acquire all the information in the training dataset, so it is important to find a network with
an appropriate complexity.
The final structure and parameters of each network are expressed in Figure 4, and the
performance of trained networks with different training datasets are expressed with the
confusion matrix results with the test dataset as shown in Figure 5.
The accuracy of the network was calculated as the proportion of cases where predicted
class by the network and actual class matched among the entire test dataset.
2: Wet 2: Wet
Actual Class
Actual Class
Asphalt 1 647 2377 680 235 61 59.4% Asphalt 1 949 1647 727 614 63 41.2%
3: Wet 3: Wet
Asphalt 2 418 315 3076 113 78 76.9% 724 216 2228 827 5 55.7%
Asphalt 2
4: Gravel 54 44 126 3266 510 81.7% 4: Gravel 355 174 968 2047 456 51.2%
2: Wet 2: Wet
Actual Class
Actual Class
477 2637 226 66 594 65.9% Asphalt 1 149 3478 64 283 26 87.0%
Asphalt 1
3: Wet 3: Wet
191 298 3237 79 195 80.9% Asphalt 2 1 32 3870 97 0 96.8%
Asphalt 2
2: Wet 2: Wet
Actual Class
Actual Class
3: Wet 3: Wet
Asphalt 2 31 70 3787 112 0 94.7% 0 6 3994 0 0 99.9%
Asphalt 2
Figure 5 shows that the training results from Ay has lower accuracy than the training
result from A x and Az because the tire of a straight driving vehicle has less interaction
with the road surface in the lateral direction than in the longitudinal or vertical directions.
Sensors 2021, 21, 3233 9 of 17
However, except for class 2, the trained network can classify with a meaningful level of
accuracy (over 80%) and especially for classes 3 and 4, an accuracy of over 94% is obtained.
This result shows that even for straight driving, the Ay signal has information that can
distinguish the road surface type, and some studies also confirm this result [31]. As seen in
Figure A1, although the waveform of Ay does not have distinct waveform characteristics
for each road surface compared to A x or Az , it can be distinguished to some extent with
respect to the size and frequency of the signal.
Although fully connected networks can be applied very widely due to their agnostic
nature, FCNN tend to perform worse than special-purpose networks, such as CNN that
extract patterns from sequential data in this case. Therefore, it is difficult to improve the
accuracy of FCNN beyond a certain level, while trained CNN achieves higher accuracy by
identifying important signal features that occur during the contact period for each type of
road surface.
Not only the type of network and dataset but also the normalization process of the
training dataset greatly affects the performance of the overall learning result. If the CNN is
trained using A x and Az training datasets without normalization process while maintaining
the network architectures and hyperparameters, the accuracies of the trained CNNs drop to
64.1% and 61.9%, respectively. These results indicate that the variations in amplitude and
offset value in the acceleration signals due to changes in vehicle speed degrade the quality
of the training data and the normalization process can eliminate such negative effects.
As a result of training the same CNN structure on four classes by combining dry
asphalt and wet asphalt with 1mm waterfilm, which are the classes with many confusions
at 5 class classification, trained network have accuracy of 95.7%, 86.1%, and 97.9% for A x ,
Ay , and Az dataset as shown in Figure 6. The overall accuracy of the trained networks is
clearly improved because the classes that have many confusions are combined into one
class and the total number of classes has decreased. In addition, confusion between gravel
and unpaved roads is also reduced by combining classes 1 and 2, apparently because
networks do not need to train the difference between classes 1 and 2, so the network can
allocate more capacity to distinguishing the rest of the classes.
As much as selecting a network with an appropriate structure, it is important to
correctly determine the number of classes and the type of dataset according to the purpose
of use. If the road classification result is used in the vehicle chassis stability control logic
used in extreme conditions, such as heavy rain and off-road conditions, it is desirable
to have high accuracy for roads under these conditions and it would be more proper to
have a 4-class classification strategy, as shown in Figure 6. In addition, to ensure that
per class accuracy for all classes is guaranteed more than 90%, the A x rather than the Ay
or Az dataset can be chosen as an input signal. Moreover, the gravel and unpaved road
can be combined into one off-road class and the network is trained with three classes for
higher accuracy.
Actual Class
3: Wet 3: Wet
0 3919 81 0 98.0% 24 3835 141 0 95.9%
Asphalt 2 Asphalt 2
5: Unpaved 5: Unpaved
Road 0 0 173 3827 95.7% Road 0 0 555 3445 86.1%
3: Wet
5 3979 2 14 99.5%
Asphalt 2
5: Unpaved
Road 0 0 490 3510 87.8%
Per Class
Accuracy
Predicted Class
Figure 6. Confusion matrix of 4-class classification results from CNN with x-, y- and z-axis acceleration test dataset.
with fast enough sampling time, the proposed algorithm has sufficient performance to be
utilized in vehicle controllers as a real-time road surface classifier.
Ax
Input
dry
Ay wet1
Input
wet2
⋯
⋯
gravel
unpaved
0 1
Az
Input
Figure 7. Structure and parameters of each layer of modified 1-D convolutional neural network with multiple input.
CNN with Ax, Az (Accuracy 95.3%) CNN with Ax, Ay, Az (Accuracy 95.9%)
1: Dry 1: Dry
3927 72 1 0 0 98.2% 3951 48 1 0 0 98.8%
Asphalt Asphalt
2: Wet 2: Wet
Actual Class
Actual Class
3: Wet 3: Wet
Asphalt 2 1 4 3995 0 0 99.9% 0 0 3991 5 4 99.8%
Asphalt 2
The Figure 9 shows the result of applying the 5-class classification CNN with multi-
axis acceleration input (A x , Ay and Az ) in Figure 8 to driving experimental data for each
road condition, and the result is the same as the result of applying the road classification
algorithm in real time. Several characteristics of the developed classification algorithm are
shown by the results. First of all, there is a 1 s delay at the beginning of all experiments
because sequential data corresponding to 1 s are required as an input to the network. In
Sensors 2021, 21, 3233 12 of 17
addition, as seen from the Figure 9a,d,e, since the learning was performed for 10–70 kph,
the estimation is not accurate when the vehicle speed is less than 10 kph. Last, as shown in
Figure 9d, the road classification algorithm allows us to clearly identify the transition from
dry asphalt to gravel in 10 s.
50 5
Wheel speed (kph) &
Steering angle (deg)
40 4
Road class
30 3
20 2
10 1
−10
0 10 20 30 40 50 60 0 2 4 6 8 10 12 14 16
time(sec) time(sec)
(a) (b)
50 5
Wheel speed (kph) &
Steering angle (deg)
40 4
Road class
30 3
20 2
10 1
−10
0 5 10 15 0 5 10 15 20 25
time(sec) time(sec)
(c) (d)
50 5
Wheel speed (kph) &
Steering angle (deg)
40 4
Road class
30 3
20 2
10 1
−10
0 10 20 30 40 50 60 70
time(sec)
(e)
Figure 9. Result of applying road classification network to driving experiment data at (a) dry asphalt road, (b) wet asphalt
with 1 mm waterfilm, (c) wet asphalt with 4 mm waterfilm, (d) gravel, and (e) unpaved road.
should be reduced as much as possible. Therefore, a CNN model with the same structure
and parameters is trained while steadily reducing the time window size. Figure 10 shows
the results of training after reducing the length of the CNN’s input data from 1000 to 100.
The results show that the model accuracy decreases as the time window size is reduced
and the CNN cannot guarantee the learning results with fewer than 400 input data points.
As a result, it will be the most optimal network to adopt the shortest time window
length that satisfies the minimum accuracy required for the network referring to Figure 10.
For example, if minimum requirement of overall accuracy for a road classification network
is 90%, the time window length of 400 is the most optimal choice and, if an accuracy of 95%
or more is required, the time window length of 1000 is optimal.
100
90
80
Accuracy (%)
70
60
𝑨𝒙
50 𝑨𝒚
40 𝑨𝒛
𝑨𝒙 + 𝑨𝒚 + 𝑨𝒛
30
20
0 100 200 300 400 500 600 700 800 900 1000
Length of Time Window
Figure 10. Accuracy analysis under different time window lengths.
4. Conclusions
In this study, a database containing tire–road interaction information in contact patches
was collected on various road surfaces by using an intelligent tire system with an acceler-
ation sensor attached to the inside of the tire. Based on this database, two types of deep
learning algorithms are trained and their performance is compared to classify the road
surface type. The results showed that a 1-dimensional CNN is more suitable than a FCNN
for extracting road surface identification information from acceleration sensor signals. The
results also indicate that x- and z-axis acceleration signals contain more information on the
type of road surface than the y-axis acceleration signal, and the learning result based on the
z-axis acceleration dataset has the highest overall accuracy. Additionally, by analyzing how
the accuracy of the trained network is affected by the number of classes and the length of
the time window, a criterion for selecting an appropriate network was suggested.
Compared with the existing research on road surface classification, the proposed
algorithm is more practical because it can provide road surface conditions continuously
and fast enough to be utilized in the chassis controller of a vehicle. Moreover, due to
the characteristics of the DNN, the developed road classification network can improve
its performance by using an additional experimental dataset. For example, for a rapid
acceleration period where the accuracy of road surface estimation is currently inferior, the
performance can be improved by additional learning with the experimental datasets for
the acceleration driving condition. In the same way, by securing additional datasets for
various vehicle conditions such as tire type, tire pressure, vehicle load, and various driving
Sensors 2021, 21, 3233 14 of 17
conditions, the road classification network can be improved so that it can operate robustly
in all situations.
The currently developed algorithm classifies the road surface condition with respect
to the acceleration sensor signal for only 1 axis but future research will be conducted to
improve the accuracy and usability of the road surface classification by using the 3-axis
acceleration sensor signal at the same time or by using the driving condition, vehicle body
acceleration and gyro signal.
Author Contributions: Conceptualization, methodology: D.L. and J.-C.K.; formal Analysis, software:
D.L. and M.K.; data curation, writing–original draft, writing–review and editing: D.L.; validation:
J.-C.K. and H.L.; supervision, project administration, funding acquisition: H.L. All authors have read
and agreed to the published version of the manuscript.
Funding: This research was supported by Unmanned Vehicles Core Technology Research and
Development Program through the National Research Foundation of Korea (NRF) and Unmanned
Vehicle Advanced Research Center (UVARC) funded by the Ministry of Science and ICT, the Republic
of Korea (grant number: 2020M3C1C1A02081912).
Conflicts of Interest: The authors declare that there is no conflict of interest regarding the publication
of the article.
Abbreviations
The following abbreviations are used in this manuscript:
Appendix A
10 5 40
30
0 0
20
−10 −5
10
−20 − 10 0
24.8 24.9 25 25.1 25.2 24.8 24.9 25 25.1 25.2 24.8 24.9 25 25.1 25.2
time[sec] time[sec] time[sec]
(a)
x-axis acceleration y-axis acceleration z-axis acceleration
20 10 50
Acceleration[g]
10 5 40
30
0 0
20
−10 −5
10
−20 −10 0
5 5.1 5.2 5.3 5.4 5 5.1 5.2 5.3 5.4 5 5.1 5.2 5.3 5.4
time[sec] time[sec] time[sec]
(b)
x-axis acceleration y-axis acceleration z-axis acceleration
20 10 50
Acceleration[g]
10 5 40
30
0 0
20
−10 −5
10
−20 −10 0
4.45 4.55 4.65 4.75 4.85 4.45 4.55 4.65 4.75 4.85 4.45 4.55 4.65 4.75 4.85
time[sec] time[sec] time[sec]
(c)
x-axis acceleration y-axis acceleration z-axis acceleration
20 20 50
Acceleration[g]
10 10 40
30
0 0
20
−10 −10
10
−20 −20 0
12.1 12.2 12.3 12.4 12.5 12.1 12.2 12.3 12.4 12.5 12.1 12.2 12.3 12.4 12.5
time[sec] time[sec] time[sec]
(d)
x-axis acceleration y-axis acceleration z-axis acceleration
10 10 20
Acceleration[g]
5 5 15
0 0 10
−5 -5 5
− 10 -10 0
24.7 24.8 24.9 25 25.1 25.2 24.7 24.8 24.9 25 25.1 25.2 24.7 24.8 24.9 25 25.1 25.2
time[sec] time[sec] time[sec]
(e)
Figure A1. x-, y-and z-axis accelerometer measurement results measured for 30 kph experiment (20 kph for unpaved
roads) at (a) dry asphalt, (b) wet asphalt with 1 mm waterfilm, (c) wet asphalt with 1 mm waterfilm, (d) gravel, and
(e) unpaved road.
Sensors 2021, 21, 3233 16 of 17
0.2
0
−0.2
0.2
0
−0.2
0.2
0
−0.2
0.2
0
−0.2
0 20 0 20 0 20 0 20 0 20 0 20 0 20 0 20
Figure A2. Filters of the First convolutional layer of the trained CNN based on x-axis acceleration
signal dataset.
0.2
0
−0.2
0.2
0
−0.2
0.2
0
−0.2
0.2
0
−0.2
0 20 0 20 0 20 0 20 0 20 0 20 0 20 0 20
Figure A3. Filters of the First convolutional layer of the trained CNN based on y-axis acceleration
signal dataset.
0.2
0
−0.2
0.2
0
−0.2
0.2
0
−0.2
0.2
0
−0.2
0 20 0 20 0 20 0 20 0 20 0 20 0 20 0 20
Figure A4. Filters of the First convolutional layer of the trained CNN based on z-axis acceleration
signal dataset.
Sensors 2021, 21, 3233 17 of 17
References
1. Sun, Z.; Bebis, G.; Miller, R. On-road vehicle detection: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 694–711.
[PubMed]
2. Dong, Y.; Hu, Z.; Uchimura, K.; Murayama, N. Driver inattention monitoring system for intelligent vehicles: A review. IEEE
Trans. Intell. Transp. Syst. 2010, 12, 596–614. [CrossRef]
3. Pevec, D.; Babic, J.; Podobnik, V. Electric Vehicles: A Data Science Perspective Review. Electronics 2019, 8, 1190. [CrossRef]
4. Khaleghian, S.; Emami, A.; Taheri, S. A technical survey on tire-road friction estimation. Friction 2017, 5, 123–146. [CrossRef]
5. Acosta, M.; Kanarachos, S.; Blundell, M. Road friction virtual sensing: A review of estimation techniques with emphasis on low
excitation approaches. Appl. Sci. 2017, 7, 1230. [CrossRef]
6. Zhang, X.; Gohich, D. A hierarchical estimator development for estimation of tire-road friction coefficient. PLoS ONE 2017, 12,
e0171085. [CrossRef]
7. Arat, M.A.; Saied, T. Identification of Road Surface Friction for Vehicle Safety Systems; SAE Technical Paper; SAE: Warrendale, PA,
USA, 2014. [CrossRef]
8. Zhao, J.; Zhang, J.; Zhu, B. Development and verification of the tire/road friction estimation algorithm for antilock braking
system. Math. Probl. Eng. 2014, 2014, 786492. [CrossRef]
9. Cho, W.; Yoon, J.; Yim, S.; Koo, B.; Yi, K. Estimation of tire forces for application to vehicle stability control. IEEE Trans. Veh.
Technol. 2009, 59, 638–649.
10. Li, B.; Du, H.; Li, W. Comparative study of vehicle tyre–road friction coefficient estimation with a novel cost-effective method.
Veh. Syst. Dyn. 2014, 52, 1066–1098. [CrossRef]
11. Bian, M.; Chen, L.; Luo, Y.; Li, K. A Dynamic Model for Tire/Road Friction Estimation under Combined Longitudinal/Lateral Slip Situation;
SAE Technical Paper; SAE: Warrendale, PA, USA, 2014. [CrossRef]
12. Singh, K.B.; Taheri, S. Estimation of tire–road friction coefficient and its application in chassis control systems. Syst. Sci. Control
Eng. 2015, 3, 39–61. [CrossRef]
13. Gao, L.; Xiong, L.; Lin, X.; Xia, X.; Liu, W.; Lu, Y.; Yu, Z. Multi-sensor fusion road friction coefficient estimation during steering
with lyapunov method. Sensors 2019, 19, 3816. [CrossRef]
14. Kuno, T.; Susira, H. Detection of road conditions with CCD cameras mounted on a vehicle. Syst. Comput. Jpn. 1999, 30, 88–99.
[CrossRef]
15. Howard, A.; Seraji, H. Vision-based terrain characterization and traversability assessment. J. Robot. Syst. 2001, 30, 577–587.
[CrossRef]
16. Holzmann, F.; Bellino, M.; Siegwart, R.; Bubb, H. Predictive estimation of the road-tire friction coefficient. In Proceedings of the
IEEE Conference on Computer Aided Control System Design, Munich, Germany, 4–6 October 2006; pp. 885–890.
17. Jokela, M.; Kutila, M.; Le, L. Road condition monitoring system based on a stereo camera. In Proceedings of the IEEE International
Conference on Intelligent Computer Communication and Processing, Napoca, Romania, 27–29 August 2009.
18. Pereira, V.; Tamura, S.; Hayamizu, S.; Fukai, H. Classification of paved and unpaved road image using convolutional neural
network for road condition inspection system. In Proceedings of the International Conference on Advanced Informatics: Concept
Theory and Applications, Krabi, Thailand, 14–17 August 2018; pp. 165–169.
19. Umeno, T. Estimation of tire-road friction by tire rotational vibration model. RD Rev. Toyota CRDL 2002, 37, 53–58.
20. Tuononen, A.J. Optical position detection to measure tyre carcass deflections. Veh. Syst. Dyn. 2008, 46, 471–481. [CrossRef]
21. Alonso, J.; López, J.M.; Pavón, I.; Recuero, M.; Asensio, C.; Arcas, G.; Bravo, A. On-board wet road surface identification using
tyre/road noise and support vector machines. Appl. Acoust. 2014, 76, 407–415. [CrossRef]
22. Khaleghian, S.; Taheri, S. Terrain classification using intelligent tire. J. Terramechanics 2017, 71, 15–24. [CrossRef]
23. Masino, J.; Foitzik, M.-J.; Frey, M.; Gauterin, F. Pavement type and wear condition classification from tire cavity acoustic
measurements with artificial neural networks. J. Acoust. Soc. Am. 2017, 141, 4220–4229. [CrossRef]
24. Lee, H.; Taheri, S. Intelligent tires: A review of tire characterization literature. IEEE Intell. Transp. Syst. Mag. 2017, 9, 114–135.
[CrossRef]
25. Matilainen, M.; Tuononen, A. Tyre contact length on dry and wet road surfaces measured by three-axial accelerometer. Mech.
Syst. Signal Process. 2015, 52, 548–558. [CrossRef]
26. Xu, N.; Huang, Y.; Askari, H.; Tang, Z. Tire slip angle estimation based on the intelligent tire technology. IEEE Trans. Veh. Technol.
2021, 70, 2239–2249. [CrossRef]
27. Niskanen, A.J.; Tuonomen, A.J. Three three-axis IEPE accelerometers on the inner liner of a tire for finding the tire-road friction
potential indicators. Sensors 2015, 15, 19251–19263. [CrossRef]
28. Braghin, F.; Brusarosco, M.; Cheli, F.; Cigada, A.; Manzoni, S.; Mancosu, F. Measurement of contact forces and patch features by
means of accelerometers fixed inside the tire to improve future car active control. Veh. Syst. Dyn. 2006, 44, 3–13. [CrossRef]
29. Aggarwal, C.C. Neural Networks and Deep Learning: A Textbook, 1st ed.; Springer: New York, NY, USA, 2018.
30. Sainath, T.; Kingbury, B.; Saon, G.; Soltau, H.; Mohamed, A.-R.; Dahl, G.; Ramabhadran, B. Deep convolutional neural networks
for large-scale speech tasks. Neural Netw. 2015, 64, 39–48. [CrossRef]
31. Niskanen, A.J.; Tuononen, A.J. Accelerometer tyre to estimate the aquaplaning state of the tyre-road contact. In Proceedings of
the IEEE Intelligent Vehicles Symposium, Seoul, Korea, 28 June–1 July 2015.