CNN Based Lane Detection With Instance Segmentation in Edge-Cloud Computing
CNN Based Lane Detection With Instance Segmentation in Edge-Cloud Computing
CNN Based Lane Detection With Instance Segmentation in Edge-Cloud Computing
Abstract
At present, the number of vehicle owners is increasing, and the cars with autonomous driving functions have
attracted more and more attention. The lane detection combined with cloud computing can effectively solve the
drawbacks of traditional lane detection relying on feature extraction and high definition, but it also faces the problem
of excessive calculation. At the same time, cloud data processing combined with edge computing can effectively
reduce the computing load of the central nodes. The traditional lane detection method is improved, and the current
popular convolutional neural network (CNN) is used to build a dual model based on instance segmentation. In the
image acquisition and processing processes, the distributed computing architecture provided by edge-cloud
computing is used to improve data processing efficiency. The lane fitting process generates a variable matrix to
achieve effective detection in the scenario of slope change, which improves the real-time performance of lane
detection. The method proposed in this paper has achieved good recognition results for lanes in different scenarios,
and the lane recognition efficiency is much better than other lane recognition models.
Keywords: Double layer network, Lane line detection, Edge computing
© The Author(s). 2020 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which
permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit
to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The
images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated
otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended
use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the
copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Wang et al. Journal of Cloud Computing: Advances, Systems and Applications (2020) 9:27 Page 2 of 10
analysis, and finally feed the calculation results to the with lane image recognition, the real-time performance of
central node [5, 6]. Cloud computing system has a new lane image detection is greatly improved, and the delay is
resource scheduling method, and the use of the cloud reduced.
computing system can effectively reduce the system com- 2) Based on convolutional neural network and semantic
puting load and improve the data processing efficiency segmentation, the lane line problem in multi-lane scenes
with current computer systems, which is helpful to lane is effectively identified, and when vehicles change lanes,
detection [7, 8]. At the same time, a dynamic resource allo- the current lane scene can be dynamically identified.
cation method for intensive data in the cloud, which has 3) During the lane fitting process, the fixed inverse per-
greatly improved the fault tolerance of cloud computing spective transformation matrix is changed. The neural
has been proposed [9]. These are the factors that cloud network was used to dynamically generate the inverse per-
computing has become the current mainstream frame- spective transformation matrix, which effectively solved
work [10]. Based on the advantages of cloud computing the problem of lane line deviation of the fixed inverse per-
technology in large-scale data processing, we decided to spective transformation matrix fitting under the gradient
use cloud computing technology to improve data process- road scene.
ing efficiency in lane detection. The remainder of this paper is organized as fol-
Edge computing refers to a computing platform that is lows: In “Related work” section, we review the peer
close to the object or data source and can integrate core research and work. “Prior knowledge” section describes
capabilities such as networking, computing, and applica- a priori knowledge of lane detection, including the two
tions [11]. It has a very prominent advantage in the field main frameworks of lane detection techniques in this
of intelligent transportation [12]. To achieve safe and reli- article. “Method” section describes the specific steps
able operation, smart cars need to process a large amount of lane detection. The experimental environment and
of data collected by multiple sensor in real times [13]. results analysis are included in “Experimental evaluation”
Through edge computing, data processing will be made section. Finally, we conclude our work in “Conclusion”
closer to the source [9]. Compared with cloud comput- section.
ing, centralizing data processing in the cloud can greatly
reduce latency time and effectively improving the data Related work
processing efficiency of smart cars. In terms of secu- The current methods for lane detection can be divided
rity and privacy protection, edge computing solves the into three main categories: road-based models, road-
risk of transferring private data to the cloud and leak- based features, and neural network-based models. The
ing user privacy under the cloud computing model, which detection method based on the road model mainly
can effectively protect data and privacy security using abstracts the lane lines into geometric shapes such as
its distributed and mobile features [14–16]. A data pri- straight lines, curves, parabolas, and splines, and uses dif-
vacy optimization scheme that supports edge computing ferent two-dimensional or three-dimensional models to
[17, 18], which further strengthens the data security of determine each model parameter.
edge computing has been proposed and we have com- Wang et al. [20] first located the initial area of the lane
bined this scheme in the lane detection process in order line, and then converted the lane line detection problem
to further improve the efficiency of data processing. into the problem of determining the spline model in the
However, it is difficult for cloud computing or edge initial area. Tan et al. [21] proposed a robust curve lane
computing to play a prominent role in lane recognition detection method, which uses the improved river method
of smart cars. Current lanes processed by convolutional to search for feature points in the far field of view, guided
neural networks face the problem of low real-time per- by a straight line detected in the near field or the curve of
formance, especially in complex scenarios [19]. Currently, the last frame, and can connect dotted lane marks or fuzzy
few researchers use cloud computing or edge comput- lane marks. Shin et al. [22] used parallel features with con-
ing to analyze lane lines, while only use cloud computing stant distance between the left and right lane lines in the
needs to upload a large amount of data to the cloud for top view, and use parallel lines to detect and track lane
processing, which increases the load on the network and is lines. The above method has high accuracy for detect-
difficult to obtain better results under the circumstances ing a specific lane line, but needs to select an appropriate
of the Internet. Therefore, we combine cloud computing model and parameters. One model is often difficult to
and edge computing to process lane data, which greatly cope with multiple road scenarios, and its generalization
improves the real-time performance of lane detection ability and real-time performance are poor. The detec-
during the driving process of smart cars. tion method based on road features uses lane line and
The main contributions of this paper are as follows: environment differences to extract lane line features for
1) Inspired by cloud computing and edge computing, recognition and segmentation. The main features include
adopt more efficient data processing methods to deal lane line edges, color, and geometric features. Yoo et al.
Wang et al. Journal of Cloud Computing: Advances, Systems and Applications (2020) 9:27 Page 3 of 10
[23] proposed a gradient-enhanced conversion method for has greatly improved the robustness and accuracy com-
robust lane detection, which converts RGB space images pared with traditional detection methods, but in order to
into grayscale images and combines Canny’s algorithm train the network, a huge data set is required as support,
to generate large gradients at the boundaries. Son et al. and the requirements for hardware facilities are corre-
[24] realized lane line detection by extracting white and spondingly increased. Xu et al. [35, 36] proposed a video
yellow lane lines respectively. Jung et al. [25] proposed surveillance resource method when studying the Internet
an effective method for reliably detecting lanes based of Vehicles (IoV), which supports edge computing and can
on spatiotemporal images, which improved the detec- effectively use edge computing to solve the problem of
tion accuracy. Niu et al. [26] proposed a lane detection lane image acquisition and training requirements and high
method with two-stage feature extraction in order to solve network bandwidth, we combine this method in the data
the problem of robustness caused by the inconsistency of processing process to distribute computing power from
lighting and background clutter, where each lane has two the central node to devices with image acquisition and
boundaries. The above-mentioned detection methods are processing capabilities.
easy to implement, have low complexity, and can obtain At the same time, we choose to convert the lane line
high real-time performance, but are extremely suscepti- detection problem into an instance segmentation prob-
ble to environmental influences. In rainy days or under lem, and combines edge computing to design a two-
insufficient light, misjudgment of lane lines is prone to instance segmentation branch network, where the lane
occur. segmentation branch outputs the background or lane line,
With the continuous development of science and tech- and the lane embedding branch separates the lane lines
nology and the continuous improvement of hardware obtained by the segmentation branch into different lane
equipment, neural network-based detection methods examples to improve the accuracy of lane detection. At the
have made breakthrough progress. Using deep learning to same time, a custom network is designed, and the training
detect lane lines can ensure good recognition accuracy in data is marginalized to enhance the real-time nature of the
most scenarios [27]. Insteading of relying on highly spe- data processing. A perspective transformation matrix is
cialized manual features and heuristics to identify lane proposed to solve the problem of lane detection for classic
breaks in traditional lane detection methods, target fea- neural networks when the road plane changes and helps
tures under deep learning can automatically learn and us improve the real-time and accurate performance of the
modify parameters during the training process. network used in this paper.
Liu et al. [28] used the mobile edge computing frame-
work to greatly reduce the time loss of learning. Qi et al. Prior knowledge
[29, 30] significantly improved the data processing rate Lane prediction
through the method of distributed location and multiple Lane detection usually requires the use of relevant algo-
data sources. John et al. [31] used the learned CNN model rithms to extract the pixel features of the lane line,
to predict the lane line position in the image, especially in and then the appropriate pixel fitting algorithm is used
the case of occlusion and missing, CNN can extract robust to complete the lane detection. Traditional lane detec-
features from the road image, and then train an additional tion uses Canny edge extraction algorithm or Sobel edge
tree-like regression model directly from the features esti- extraction algorithm to obtain lane line candidate points
mated by lane line position. Kim et al. [32] used serial and use Hough transform for lane feature detection, but
end-to-end transfer learning to directly estimate and sep- most operations are based on manual feature extraction
arate the left and right lanes, and collects a dataset that [37, 38]. The latest research is based on deep neural
includes multiple road conditions to train a CNN model. networks to make dense predictions instead of artificial
Pan et al. [33] proposed a spatial CNN (SCNN) and the feature extraction.
convolution of the extended layer structure in the fea- In order to enable the model to adapt to more road
ture map is a slice structure convolution. In this way, the scenes, by analyzing the structure of classic convolutional
information transfer between pixels is performed between neural networks and semantic segmentation, we use an
the rows and columns of each layer, which is suitable for improved two-branch network and custom function net-
strong spatial relationships, but long continuous-shaped work to convert the lane line detection problem into an
structures or large targets lack apparent clues in this instance segmentation problem detection lane line, where
method, such as lane lines, walls, and pillars. Zhang et each lane line forms its own instance. In this paper, we use
al. [34] applied deep learning-based convolutional neural a multi-branch network with two branches and the two-
networks to instance segmentation of monocular vision to branch network contains a lane segmented branch and a
achieve segmentation of different objects in actual scenes, lane embedded branch that can be trained end-to-end.
and achieved better segmentation results, but less robust. Lane segmentation branches output backgrounds or lane
The use of deep learning methods for lane line detection lines, and lane embedding branches further decompose
Wang et al. Journal of Cloud Computing: Advances, Systems and Applications (2020) 9:27 Page 4 of 10
the divided lane pixels into different lane proportions. By lane points near the horizon to be projected to infin-
dividing the lane detection problem into the above two ity, which affects lane line simulation total accuracy. In
tasks, the function of lane segmentation can be fully uti- order to solve this situation, we apply an inverse per-
lized without having to assign different classes to different spective transformation to the image before fitting the
lanes. Lane embedding branches trained with the clus- curve, and uses a loss function customized for the lane
tering loss function assign lanes to each pixel from the fitting problem to optimize it. A custom function net-
lane segmentation branch, ignoring background pixels. By work is used to generate a transformable matrix and
doing so, the problem of lane change is solved, a vari- transform the lane pixels, then use curve fitting polyno-
able number of lanes can be handled, and the practice of mials to perform pixel fitting on the converted pixels, and
detecting a fixed lane line is improved. finally convert the fitted pixels into the input image. The
At the same time, combined with edge computing, advantage of this network is that the detection algorithm
a blockchain-based edge offloading method is used to can fit the pixels of distant lanes with good robustness
ensure real-timeness and integrity in the data transmis- when the road surface changes, and better adapt to lane
sion process [39, 40]. For real-time processing and feed- changes.
back, edge computing, as a new computing paradigm, has The overall framework of the model is shown in Fig. 1.
a better ability to solve some problems such as high trans- The camera on the car collects the image and transmits it
mission delays, high bandwidth expenditures, and privacy to the cloud data processing center. After the image is sub-
leaks. jected to binary segmentation and embedded segmenta-
tion, the clustering operation is performed and combined
Lane fitting with the transformable matrix generated by the custom
To estimate the lane instance, determine which pixel network to generate the final lane detection image.
belongs to which lane, we need to convert each of them
into a parameter description. To this end, we use a Method
widely used fitting algorithm. At present, the widely used Binary segmentation
lane fitting models are mainly cubic polynomials, spline In lane detection, in order to save computing resources,
curves or arc curves. Inverse perspective transforma- we binarize the image. The segmented branch of the
tion is used to transform the image into a “bird’s-eye two-branch network outputs a binary segmentation map,
view” to improve the quality of the fitting while main- which divides the lane line and non-lane line parts. In this
taining the computational efficiency, and then adopt the paper, all ground-level car-to-car points are connected
method of curve fitting [41]. The fit lines in the “bird’s- together to form the connecting line of each lane to avoid
eye view” can be re-projected into the original image situations where lane lines are unpredictable due to lane
through an inverse transformation matrix. Generally, the line degradation or road obstacles blocking the lane lines.
inverse perspective transform calculates the transforma- At the same time, due to the high imbalance between lane
tion matrix on a single image and can remain fixed, but if and background, we use bounded inverse class weighting
the road plane changes, the fix is no longer valid, causing for lane pixel segmentation.
Lane instance embedding segmentation represents the threshold of the variance term, δd repre-
The lane loss embedding branch function of the dual sents the threshold of the distance term, and the total loss
branch network is used to output the pixel embeddings L = Lvar + Ldist . After the network converges, the lane
of each lane, so that the pixel embeddings belonging pixel embeddings will be clustered together so that the
to the same lane are clustered together, and the pixel distance between each cluster is greater than δd and the
embeddings belonging to different lanes have the largest radius of each cluster is less than δv .
distance. In the way, the pixel embeddings of the same In Fig. 2a and b are the lane images collected, and the
lane will come together to form a unique cluster for each images generated by the example embedding segmenta-
lane and this is achieved by introducing two terms, one tion are (c) and (d), respectively.
is the variance term (Lvar ), which applies a pulling force
to each embedding and tends to be closer to the aver- Clustering
age embedding of the lane. The other is the distance term For a large number of unlabeled data sets, the data set is
(Ldist ), which clusters the centers push away from each divided according to the inherent similarity of the data,
other. The pull force is active only when the distance from and the data set is divided into different categories with
the embedding point to the cluster center is greater than inherent differences. The similarity between the data in
δv . The thrust force is active only when the distance from the same category is large and the similarity between the
the embedding point to the cluster center is less than δd . data in different categories is small. In order to improve
⎧ the robustness of lane line detection in different envi-
⎪ 1 C 1 NC
⎨ Lvar = C c=1 NC i=1 [ μc − μi − δv ]+
2
ronments, it is necessary to perform cluster analysis on
C
Lsum = CA,CB=1 [ δd − μCA − μCB ]+ 2 (1) the lane line pixels, embed the pixels of the same lane
⎪
⎩L 1
dist = L
C(C−1) sum
line together, and distinguish the pixels of different lane
lines. In the future, clustering can ensure the recogni-
In (1), C is the number of lane lines, NC is the num- tion accuracy under the conditions of insufficient lighting
ber of elements in C, μi is the lane pixel embedding, μc conditions and poor traffic conditions.
is the average embedding of C, “ μc − μi ” represents In this paper, we complete clustering by iteration, com-
the average embedding and pixel embedding distance, and bining the clustering algorithm with the loss function. In
CA and CB represent two lane line, μCA − μCB repre- (1), by setting 6δv < δd , we take a random lane with a
sents the average embedding distance between CA lane radius of 2δv and its surrounding threshold to select the
line and CB lane line, [ δd − μCA − μCB −δv ]2+ means ones that belong to the same lane all embedded. Repeat
0 and δd − μCA − μCB has the largest positive num- the above operation until all lanes are embedded and
ber value, [ μc − μi − δv ]2+ represents the maximum value assigned to one lane. In order to avoid choosing the dis-
of a positive number between 0 and μc − μi − δv , δv tance from the outlier to the threshold, we first use the
a b
c d
mean shift to approach the center of the cluster, and then transformation, P(Pi =[ Xi , Yi , 1]T ∈ P, i = 1, 2, . . .) rep-
sets the threshold. resents the pixels of real lane pixels, Xi , Yi represents the
horizontal and vertical coordinates of pixels of real lanes,
Pixel fitting Xi , Yi represents the horizontal and vertical coordinates
The output of the two-branch network is a set of pixels for of lane pixel points after inverse perspective transforma-
each lane line. It is not ideal to fit a polynomial with these tion. In order to obtain the position of the lane x at a
pixels in the input image space, because a higher-order given y position, the real lane pixel extraction point Pi =
polynomial is needed to process the curve lane. Several [ −, yi , 1]T is transformed into a transformed lane pixel
commonly used fitting methods are RANSAC algorithm, point Pi =[ −, yi , 1]T after perspective transformation ,
Bezier curve fitting, curve fitting based on spline interpo- predict lane pixel points and coordinates Pi∗ = (x∗
i , yi )
lation, and polynomial fitting based on least squares. In
at each yi position. A third-order polynomial is used to
this paper, the image is first transformed by inverse per- fit the predicted lane pixel point Pi∗ by the least square
spective and projected into the “bird’s eye view", where method. Equations (4) and (5) are third-order polynomials
the lanes are parallel to each other, so the curve lane can and least squares formulas, respectively.
be fitted by a second to third order polynomial. How-
ever, in these cases, the fixed transformation matrix H f y = αy2 + βy + γ (4)
is calculated only once and applied to all images. This
results in errors in the case of ground-level changes, where −1
the vanishing point projected to infinity moves up or w = YTY Y T x (5)
down as shown. To solve this problem, this paper trains
In (4) and (5), α, β, and γ are the given optimal param-
a neural network with a custom loss function. The net-
eters, T is the transpose of the matrix, w =[ α, β, γ ]T is
work is optimized end-to-end to predict the parameters
the least square method to represent the letter, x =
of the perspective transformation matrix H. In the per-
[ x1 , x2 , x3 , · · · xn ]T represents the lane coordinate point P
spective transformation matrix H, the transformed lane
abscissa combination, y =[ y1 , y2 , y3 , · · · yn ]T represents
points can be optimally fitted with second or third order
the transformable parameter matrix ⎡ transform lane
⎤ pixel
polynomials. The prediction is based on the input image,
y2
1 y 1
1
allowing the network to adjust the projection parameters
point P ordinate combination, Y = ⎣ · · · · · · · · · ⎦ is the
when the ground plane changes, so that the lane line fit
y2
n y1 1
is still correct. In (2), the transformation matrix H has 6
required matrix in the least squares formula. Fitting the
degrees of freedom, of which six variables a-f represent 6
predicted lane pixels Pi∗ =[ x∗ T
i , yi , 1] perspective trans-
degrees of freedom parameters. The purpose of placing 0
formation to the road lane image to obtain lane pixels Pi∗ at
is to enforce the constraint and make the horizontal line
different lane positions. The cluster loss function is used
remains horizontal under the transformation.
to segment the lane line instance into a pixel map, and
⎡ ⎤
a b c the predicted pixels of the fitted lane are projected into
H = ⎣0 d e ⎦ (2) the road lane image through perspective transformation
0 f 1 to obtain the final detection result image, and the perspec-
tive transformation formula of the cluster loss function is
The network architecture of the custom function net- shown in (6).
work remains small in size and consists of a continu-
ous block of 3 × 3 convolutional layers, Batchnorm, and Pi∗ = H −1 Pi∗ (6)
ReLU activation functions, and the largest pooling layer to In (6), Pi∗ =[ x∗i , yi , 1]T is a lane pixel projected into the
reduce the network size, and finally add 2 fully connected input road lane image, Pi∗ =[ x∗ T
i , yi , 1] is the predicted
layers. Before a pixel fits a lane line, a custom function −1
lane pixel point, and H represents the perspective trans-
network output transformable matrix is used to transform formation matrix. In order to train the output of the
the lane pixels. The inverse perspective transformation is custom function network H, which is the most suitable
performed on the road image to a new top view to gener- transformation matrix H for pixel fitting polynomials, the
ate pixels in the top view lane after the inverse perspective following loss function is constructed. Since lane fitting
transformation and the inverse perspective transforma- is done by using the least squares method, the loss func-
tion formula is shown in (3). tion is differentiable and the loss function of the custom
function network is shown in (7).
P = HP (3)
1 ∗
N
In (3), P (Pi =[ Xi , Yi , 1]T ∈ P , i = 1, 2, . . .) represents Loss = (xi − xi )2 (7)
the transformed lane line pixels after inverse perspective N
i=1
Wang et al. Journal of Cloud Computing: Advances, Systems and Applications (2020) 9:27 Page 7 of 10
a b
c d
detection results. It can be seen that even in some cases lines well in the case of obstacles and shadow, and has
the lane line has been degraded or blurred, the lane line strong robustness.
can still be accurately detected under the deep training
lane network model, which proves that the network model Conclusion
used in this paper has better robustness. In this paper, the lane detection algorithm is carefully
In Fig. 5a and b are lane images collected in a scene studied. The lane detection method based on traditional
with shadows and obstacles, and (c) and (d) are corre- image processing and the lane detection method based
sponding detection results. It can be found that the lane on deep learning are analyzed and compared to solve
detection model used in this paper can still detect lane the problem of lane detection under complex conditions
a b
c d
a b
c d
tion of vehicle edge calculation and actual engineering can Science and Technology, China No.219 Ningliu Road, 210044 Nanjing, China.
3 Key Laboratory for Virtual Geographic Environment, Ministry of Education,
improve the effect of lane line detection network. Nanjing Normal University, Nanjing, China. 4 Jiangsu Center for Collaborative
In the inverse perspective transformation, the use of Innovation in Geographical Information Resource Development and
a fixed transformation matrix will cause errors when Application, Nanjing, China.
ground changes, which will cause the vanishing point pro- Received: 4 March 2020 Accepted: 22 April 2020
jected to infinity to move up or down. We trained a neural
network with a custom loss function that can dynamically
predict the parameter values of the transformable matrix. References
1. Smuda P, Schweiger R, Neumann H, Ritter W (2006) Multiple cue data
The transformed lane points are fitted with second or
fusion with particle filters for road course detection in vision systems. In:
third-order polynomials. The prediction is based on the 2006 IEEE Intelligent Vehicles Symposium. IEEE. pp 400–405. https://doi.
input image, allowing the network to adjust the projection org/10.1109/ivs.2006.1689661
2. Aly M (2008) Real time detection of lane markers in urban streets. In: 2008
parameters when the ground plane changes, making the
IEEE Intelligent Vehicles Symposium. IEEE. pp 7–12. https://doi.org/10.
model excellent stickiness. The final tests show that the 1109/ivs.2008.4621152
model in our paper has better performance in scenarios 3. Teng W, Wang Y, Yu B, Liu J (2020) Icciu: A new real-time lane-level
positioning algorithm. IEEE Access. https://doi.org/10.1109/access.2020.
such as insufficient lighting and lane line degradation.
2970503
4. Bertozzi M, Broggi A (1996) Real-time lane and obstacle detection on the
Authors’ contributions
gold system. In: Proceedings of Conference on Intelligent Vehicles. IEEE.
Wei Wang, Hui Lin, Junshu Wang conceived and designed the study. Wei
pp 213–218. https://doi.org/10.1109/ivs.1996.566380
Wang, Hui Lin and Junshu Wang performed the simulations. Wei Wang, Hui
5. Xue C, Lin C, Hu J (2019) Scalability analysis of request scheduling in cloud
Lin and Junshu Wang wrote the paper. All authors reviewed and edited the
computing. Tsinghua Sci Technol 24(3):249–261
manuscript. All authors read and approved the final manuscript.
6. Shen D, Luo J, Dong F, Zhang J (2019) Virtco: joint coflow scheduling and
virtual machine placement in cloud data centers. Tsinghua Sci Technol
Funding 24(5):630–644
This research is supported by National Natural Science Foundation of China 7. Xu X, He C, Xu Z, Qi L, Wan S, Bhuiyan M (2019) Joint optimization of
under Grant No. 41671055, Jiangsu Natural Science Foundation of China offloading utility and privacy for edge computing enabled IoT. IEEE
under Grant No. BK20171037, the Program of Natural Science Research of Internet of Things J. https://doi.org/10.1109/jiot.2019.2944007
Jiangsu colleges and universities under Grant No.17KJB170010. 8. Xu X, Zhang X, Khan M, Dou W, Xue S, Yu S (2017) A balanced virtual
machine scheduling method for energy-performance trade-offs in
Availability of data and materials cyber-physical cloud systems. Futur Gener Comput Syst
The raw data required to reproduce these findings cannot be shared at this 9. Xu X, Mo R, Dai F, Lin W, Wan S, Dou W (2019) Dynamic resource
time as the data also forms part of an ongoing study. provisioning with fault tolerance for data-intensive meteorological
Wang et al. Journal of Cloud Computing: Advances, Systems and Applications (2020) 9:27 Page 10 of 10
workflows in cloud. IEEE Trans Ind Inform. https://doi.org/10.1109/tii.2019. 33. Pan X, Shi J, Luo P, Wang X, Tang X (2018) Spatial as deep: Spatial cnn for
2959258 traffic scene understanding. In: Thirty-Second AAAI Conference on
10. Yang S, Wu J, Shan Y, Yu Y, Zhang S (2019) A novel vision-based Artificial Intelligence
framework for real-time lane detection and tracking. Technical report. 34. Zhang Z, Schwing AG, Fidler S, Urtasun R (2015) Monocular object
SAE Technical Paper. https://doi.org/10.4271/2019-01-0690 instance segmentation and depth ordering with cnns. In: Proceedings of
11. Xu X, Zhang X, Gao H, Xue Y, Qi L, Dou W (2020) Become: the IEEE International Conference on Computer Vision. pp 2614–2622.
Blockchain-enabled computation offloading for iot in mobile edge https://doi.org/10.1109/iccv.2015.300
computing. IEEE Trans Ind Inform 16(6):4187–4195 35. Xu X, Fu S, Yuan Y, Luo Y, Qi L, Lin W, Dou W (2019) Multiobjective
12. Ganin AA, Mersky AC, Jin AS, Kitsak M, Keisler JM, Linkov I (2019) Resilience computation offloading for workflow management in cloudlet-based
in intelligent transportation systems (its). Transp Res Part C Emerg mobile cloud using nsga-ii. Comput Intell 35(3):476–495
Technol 100:318–329 36. Xu X, Liu X, Xu Z, Dai F, Zhang X, Qi L (2019) Trust-oriented iot service
13. Kuo C, Lu Y, Yang S (2019) On the image sensor processing for lane placement for smart cities in edge computing. IEEE Internet Things J.
detection and control in vehicle lane keeping systems. Sensors 19(7):1665 https://doi.org/10.1109/jiot.2019.2959124
14. Xu Y, Qi L, Dou W, Yu J (2017) Privacy-preserving and scalable service 37. Aminuddin NS (2020) A new approach to highway lane detection by
recommendation based on simhash in a distributed cloud environment. using hough transform technique. J Inf Commun Technol 16(2):244–260
Complexity 2017. https://doi.org/10.1155/2017/3437854 38. Wang J, Kong B, Mei T, Wei H (2019) Lane detection algorithm based on
15. Xu X, Liu Q, Luo Y, Peng K, Zhang X, Meng S, Qi L (2019) A computation temporal–spatial information matching and fusion. CAAI Trans Intell
offloading method over big data for iot-enabled cloud-edge computing. Technol 2(4):154–165
Futur Gener Comput Syst 95:522–533 39. Xu X, Liu Q, Zhang X, Zhang J, Qi L, Dou W (2019) A blockchain-powered
16. Chi X, Yan C, Wang H, Rafique W, Qi L Amplified locality-sensitive crowdsourcing method with privacy preservation in mobile environment.
hashing-based recommender systems with privacy protection. Concurr IEEE Trans Comput Soc Syst 6(6):1407–1419
Comput Pract Experience:5681. https://doi.org/10.1002/cpe.5681 40. Nguyen TNA, Phung SL, Bouzerdoum A (2020) Hybrid deep
17. Xu X, Chen Y, Zhang X, Liu Q, Liu X, Qi L (2019) A blockchain-based learning-gaussian process network for pedestrian lane detection in
computation offloading method for edge computing in 5g networks. unstructured scenes. IEEE Trans Neural Netw Learn Syst. https://doi.org/
Softw Pract Experience. https://doi.org/10.1002/spe.2749 10.1109/tnnls.2020.2966246
18. Xu X, Cao H, Geng Q, Liu X, Dai F, Wang C (2020) Dynamic resource 41. Barrera A, Guindel C, Beltrán J, García F (2020) Birdnet+: End-to-end 3d
provisioning for workflow scheduling under uncertainty in edge object detection in lidar bird’s eye view. arXiv preprint. arXiv:2003.04188
computing environment. Concurr Comput Pract Experience:5674.
https://doi.org/10.1002/cpe.5674 Publisher’s Note
19. Son Y, Lee ES, Kum D (2019) Robust multi-lane detection and tracking Springer Nature remains neutral with regard to jurisdictional claims in
using adaptive threshold and lane classification. Mach Vis Appl published maps and institutional affiliations.
30(1):111–124
20. Wang Y, Teoh EK, Shen D (2004) Lane detection and tracking using
b-snake. Image Vis Comput 22(4):269–280
21. Tan H, Zhou Y, Zhu Y, Yao D, Li K (2014) A novel curve lane detection
based on improved river flow and ransa. In: 17th International IEEE
Conference on Intelligent Transportation Systems (ITSC). IEEE.
pp 133–138. https://doi.org/10.1109/itsc.2014.6957679
22. Shin B-S, Tao J, Klette R (2015) A superparticle filter for lane detection.
Pattern Recogn 48(11):3333–3345
23. Yoo H, Yang U, Sohn K (2013) Gradient-enhancing conversion for
illumination-robust lane detection. IEEE Trans Intell Transp Syst
14(3):1083–1094
24. Son J, Yoo H, Kim S, Sohn K (2015) Real-time illumination invariant lane
detection for lane departure warning system. Expert Syst Appl
42(4):1816–1824
25. Jung S, Youn J, Sull S (2015) Efficient lane detection based on
spatiotemporal images. IEEE Trans Intell Transp Syst 17(1):289–295
26. Niu J, Lu J, Xu M, Lv P, Zhao X (2016) Robust lane detection using
two-stage feature extraction with curve fitting. Pattern Recogn
59:225–233
27. De Brabandere B, Van Gansbeke W, Neven D, Proesmans M, Van Gool L
(2019) End-to-end lane detection through differentiable least-squares
fitting. arXiv preprint. arXiv:1902.00293
28. Liu L, Chen X, Lu Z, Wang L, Wen X (2019) Mobile-edge computing
framework with data compression for wireless network in energy
internet. Tsinghua Sci Technol 24(3):271–280
29. Qi L, Zhang X, Dou W, Ni Q (2017) A distributed locality-sensitive
hashing-based approach for cloud service recommendation from
multi-source data. IEEE J Sel Areas Commun 35(11):2616–2624
30. Qi L, Zhang X, Dou W, Hu C, Yang C, Chen J (2018) A two-stage
locality-sensitive hashing based approach for privacy-preserving mobile
service recommendation in cross-platform edge environment. Futur
Gener Comput Syst 88:636–643
31. John V, Liu Z, Guo C, Mita S, Kidono K (2015) Real-time lane estimation
using deep features and extra trees regression. In: Image Video Technol.
Springer. pp 721–733. https://doi.org/10.1007/978-3-319-29451-3_57
32. Kim J, Park C (2017) End-to-end ego lane estimation based on sequential
transfer learning for self-driving cars. In: Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition Workshops.
pp 30–38. https://doi.org/10.1109/cvprw.2017.158