Enhancing GPS Positioning Accuracy Using Machine Learning Regression
Enhancing GPS Positioning Accuracy Using Machine Learning Regression
1
algorithms, addressing the critical need for precise ground truth records, satellite positions, and corrected
velocity determination in applications such as pseudoranges, followed by meticulous preprocessing
autonomous navigation and geodesy [4]. to ensure data integrity. Feature engineering extracts
Complementing these efforts, Kong (2021) introduces crucial features like corrected pseudoranges and
an innovative satellite positioning method employing satellite positions.
the Total Least Squares (TLS) algorithm, enhancing In the realm of traditional and improved position
accuracy and efficiency in satellite positioning [5]. estimation, both Traditional Least Square Estimation
Building upon these foundational studies, this research and Weighted Least Square Estimation are employed.
explores the application of advanced machine learning Traditional Least Square Estimation establishes a
regression techniques. The methods include ‘Support baseline, while Weighted Least Square Estimation
Vector Regression (SVR)’, XGBoost, ‘Decision Tree enhances accuracy by incorporating weightings to
prioritize more reliable measurements and mitigate the
Regression (DTR)’, and ‘Random Forest Regression
impact of outliers.
(RFR)’. Despite facing challenges in applying
The innovation lies in the integration of machine
Gaussian Process Regression (GPR) due to
learning regression techniques. ‘Support Vector
computational limitations, the study aims to contribute
Regression (SVR)’, XGBoost, ‘Decision Tree
to the ongoing development and improvement of
Regression (DTR)’, and ‘Random Forest Regression
GNSS-based navigation and positioning systems. (RFR)’ are employed for improved position
This research integrates insights from the literature estimation. SVR optimizes a hyperplane to minimize
survey to address the crucial aspects of precision and errors, XGBoost employs an ensemble of decision
reliability in GNSS systems. The application of trees, DTR leverages decision trees, and RFR
machine learning regression techniques opens new aggregates predictions from multiple decision trees.
avenues for achieving higher accuracy in diverse fields, Given below are the equations used for the respective
including autonomous vehicles, surveying, and algorithms that are used for the project.
geodetic applications.
4. SYSTEM ARCHITECTURE
2. PROPOSED SYSTEM This architecture represents an approach to GNSS
The proposed system architecture for enhancing GPS positioning that leverages machine learning to improve
positioning accuracy integrates traditional methods the accuracy and robustness of error mitigation
with advanced machine learning techniques. In the data techniques.
collection module, diverse GNSS data, including
ground truth records, satellite positions, and corrected
pseudo ranges, are acquired to represent various
environmental conditions. The data pre-processing
module employs robust techniques such as outlier
removal and noise reduction, ensuring the integrity of
the dataset. Feature engineering identifies critical
features like corrected pseudo ranges, forming the
foundation for both traditional Least Square Estimation
and advanced machine learning models. The system
incorporates ‘Support Vector Regression (SVR)’,
XGBoost, ‘Decision Tree Regression (DTR)’, and
‘Random Forest Regression (RFR)’ in the machine
learning integration module. Additionally, a ‘Gaussian
Process Regression (GPR)’ module attempts to model
non-linear relationships in the data. The quantitative
evaluation module employs metrics such as ‘mean
absolute error (MAE)’ and ‘mean squared error
(MSE)’ for a comprehensive analysis of model
accuracy. Results are then analyzed to understand the
strengths and limitations of each algorithm. The
proposed architecture aims to provide a robust and
adaptable solution for improving GPS accuracy in
diverse environments.
3. METHDOLOGY
This methodology integrates traditional techniques Figure 4.1: System Architecture
and advanced machine learning methodologies to
enhance GPS positioning accuracy. The initial phase
involves ingesting diverse raw GNSS data from
2
Where, GNSS OBS: This block likely refers to GNSS where X is the estimated position vector, ‘A’ is the
observations, which are the raw data received from design matrix, ‘W’ is the weight matrix, and ‘b’ is the
GNSS satellites. pseudo range residual vector.
Data Collection: This block collects GNSS
observations. 5.2 Weighted Least Square Estimation:
GPS/GNSS EPH: This block likely refers to ephemeris
data, which is broadcast by GNSS satellites and Weighted least squares estimation adjusts the
contains information about their orbits and positions. traditional method to account for measurement
Raw Pseudorange: This block indicates raw uncertainties. The estimation is given by:
pseudorange measurements, which are estimates of the
X = (ATWA)-1 AT Wb (5.2)
distance between a GNSS receiver and a satellite,
Where W is a diagonal matrix of weights, allowing for
affected by factors like satellite clock errors and
the down-weighting of less reliable measurements.
propagation delays.
GNSS Position Here, the GNSS position is calculated
6. MACHINE LEARNING REGRESSION
based on the raw pseudorange measurements.
TECHNIQUES
Basic Estimations: This block performs basic
estimations of the user's position using the 6.1. Support Vector Regression (SVR):
pseudorange measurements. ‘Support Vector Regression’ aims to find a hyperplane
Position Calculation: This block refines the position that best represents the underlying mapping function
calculation by incorporating error mitigation between the input features and the output. The SVR
techniques. objective function for regression can be defined as
Errors and Effects Mitigation: This block addresses follows:
various errors that affect the accuracy of GNSS
positioning, including: Objective Function:
Clock Synchronization Errors: These errors arise from Min V, b, l, l* ½ ||V||2 + C ∑ ( li + li* ) (6.1)
differences between the GNSS receiver clock and the
Subject to:
satellite clocks.
Yi – vT ϕ (Xi) – b ≤ € + li (6.2)
Ionospheric Errors: The ionosphere delays GNSS
vT ϕ (Xi) + b - Yi ≤ € + li* (6.3)
signals, affecting positioning accuracy.
li , li* ≥ 0 (6.4)
Tropospheric Errors: The troposphere, the Earth's
Here, ‘ϕ (xi) represents the feature mapping of the
lower atmosphere, also delays GNSS signals.
input data Xi’, ‘w is the weight vector’, ‘b is the bias
Relativistic Effects: Relativistic effects, based on
term’, ‘C is the regularization parameter’, and ϵ
Einstein's theory of relativity, can introduce minor
controls the width of the ϵ -insensitive tube.
errors in GNSS positioning.
Weighted Least Squares Estimator: This statistical
6.2. XGBoost:
method is likely used to refine the position calculation XGBoost is an ‘ensemble learning’ method that
by accounting for the precision of different combines the predictions from multiple decision trees.
measurements. The objective function for XGBoost regression can be
Machine Learning Layer: This block incorporates expressed as a sum of a loss term and regularization
machine learning algorithms to potentially improve the terms:
accuracy and robustness of error mitigation, especially
Objective Function:
regarding complex error sources that are difficult to
Obj = ∑ ( Yi + Yi* ) + ∑ ( fk ) (6.5)
model mathematically. The specific machine learning
Where ‘L (Yi + Yi* )’ is the loss term measuring the
algorithms mentioned include Support Vector
difference between the ‘Yi’ predicted and ‘Yi*’ actual
Regression (SVR), Extreme Gradient Boosting
values. ‘ (fk)’ is the regularization term for each tree.
(XGBoost), Decision Tree Regression (DTR),
Random Forest Regression (RFR). These algorithms
are likely trained on datasets containing GNSS 6.3. Decision Tree Regression (DTR):
‘Decision Tree Regression’ builds a tree structure to
observations, position information, and various error
predict the output value for a given set of input
sources.
features. The prediction for a new input \(x\) is
Output: The output of the system is a corrected GNSS
obtained by traversing the tree from the root to a leaf
position that is more accurate than the basic estimates
and outputting the average of the training target values
obtained using only pseudorange measurements.
in that leaf.
Prediction for Decision Tree Regression:
5. TRADITIONAL AND IMPROVED
Y*(X) – 1/Nk ∑i=leaf(X)Yi (6.6)
POSITION ESTIMATION Where ‘Y*(X)’ is predicted output for input X,
5.1 Traditional Least Square Estimation:
‘leaf(X)’ denotes the leaf node that input X falls into
X = (AT WA)-1 AT Wb (5.1) and ‘Nk’ is the number of training samples in leaf k.
Table 7.1: Position Comparison Figure 7.2: Plot for XGBoost and Actual Positions on
Map
ML Model Latitude Longitude
When comparing the predicted position of the
Actual values 37.46758689 -122.1523673 XGBoost model to the actual coordinates, it is evident
Random Forest that XGBoost exhibits significant deviation, with a
37.46745277 -122.1527937 distance of 947.25 meters from the actual values. This
Regression (RFR)
large discrepancy highlights the model's lower
Support Vector
37.468082 -122.15336 accuracy in predicting geographical positions
Regression (SVR)
compared to other models such as Random Forest
Decision Tree
37.46716829 -122.1523673 Regression (RFR), Support Vector Regression (SVR),
Regression (DTR)
and Decision Tree Regression (DTR). The substantial
XGBoost 37.46513822 -122.1626473
error suggests that XGBoost may not be well-suited
for tasks requiring precise location predictions in this
Based on the distances calculated between the actual
specific context. Thus, while XGBoost is a powerful
coordinates and the predicted coordinates by each
and versatile model in many scenarios, its performance
model, here are the results:
4
in this case indicates room for improvement or the
need for further fine-tuning to achieve better accuracy
in predicting geographical coordinates.
5
The integration of machine learning within the
framework of traditional techniques showcases the 9. REFERENCES
potential for achieving higher accuracy and reliability
in GPS positioning. Future work will explore further [1] Boguspayev, A., Zhilong, L., & Zhang, X. (2023).
Integration of GNSS/INS techniques: A survey. Journal of
optimization and real-world implementation, Navigation and Control, 10(2), 101-11
contributing to the ongoing development of GNSS-
based navigation and positioning systems.
[2] Zawislak, P., Borkowski, M., & Wrobel, J. (2022). [6] Zhipeng Wang, Xiaopeng Hou, Zhiqiang Dan & Kun
Enhancing GNSS Multipath Detection with Unsupervised Fang, "Adaptive Kalman filter based on integer ambiguity
Domain Adaptation. In Proceedings of the ION-GNSS+ validation in moving base RTK".
Conference 2022 (pp. 1234-1245).
[7] Min Li, Tianhe Xu, Yali Shi, Kai Wei, Xianming Fei,
[3] Maghdid, H. J., Lami, K. S., & Ghafoor, K. Z. (2021). Dixing Wang, 1"Adaptive Kalman Filter for Real-Time
Optimization approaches in GNSS-based positioning: A Precise Orbit Determination of Low Earth Orbit Satellites
review. GPS Solutions, 25(3), 89-102. Based on Pseudorange and Epoch-Differenced Carrier-Phase
[4] Ji, X., Zhiqiang, W., & Lu, Y. (2022). A review of Measurements".
GNSS-based velocity estimation algorithms. Journal of
Navigation, 75(1), 1-19. [8] Haibo Ge, Bofeng, Icon,Song, Jia,Liangwei Nie, Tianhao
Wu, Zhe Yang, Jingzhe Shang, Yanning Zheng & Maorong
[5] Kong, L. (2021). An improved satellite positioning Ge, "LEO Enhanced Global Navigation Satellite System
method based on the total least squares algorithm. GPS (LeGNSS): progress, opportunities, and challenges".
World, 32(4), 54-63.
[9] Eugenio Realini, "goGPS free and constrained relative
kinematic positioning with low cost receivers"