Per Capita Expenditure Prediction Using Model Stacking Based On Satellite Imagery
Per Capita Expenditure Prediction Using Model Stacking Based On Satellite Imagery
Heri Kuswanto, Asva Abadila Rouhan, Marita Qori’atunnadyah, Supriadi Hia, Kartika Fithriasari,
Tintrim Dwi Ary Widhianingsih
Department of Statistics, Institut Teknologi Sepuluh Nopember, Surabaya, Indonesia
Corresponding Author:
Heri Kuswanto
Department of Statistics, Institut Teknologi Sepuluh Nopember
St. Raya ITS Sukolilo, Surabaya 60111, Indonesia
Email: [email protected]
1. INTRODUCTION
Poverty is a crucial issue on a global scale which becomes a big concern of all countries, including
Indonesia. Poverty is characterized by an economic incapacity to fulfill essential food and non-food needs as
measured through the per capita expenditure. A population is classified as poor if its average monthly per
capita expenditure falls below the poverty threshold. Policymakers rely on measures of poverty metrics,
assessed at the household or district level, to allocate government resources and assistance effectively.
In most countries, household income or household consumption serves as the primary basis for determining
the welfare of the poor [1]. Unfortunately, collecting reliable data to measure poverty based on per capita
expenditure is both time-consuming and costly, as it requires detailed household surveys [2].
Statistics Indonesia (BPS) is a government agency in Indonesia which is responsible to provide
required data for policymaking, as well as to assist statistical activities in other government institutions.
One of BPS main activities is to conduct the National Socio-Economic Survey (Survei Sosial Ekonomi
Nasional or SUSENAS) twice a year, where the consumption expenditure data is collected. Given the long
time of data updating process and the high cost of survey for data collection, a new strategy of collecting the
required data is indispensible. One of the proposed strategies to estimate poverty is to leverage the sattelite
image or mobile data. The satellite image data can be obtained from Google Earth. For instance, it has been
showed that light levels observed from sattelite image data is correlated with asset-based wealth measures in
37 African countries [3]. Another research showed that luminuousity from nighttime light can improve the
estimates at regional level of countries with the poor statistical system [4]. Although the processing of
satellite images can enhance the estimate of economic features, the methods used to capture surface
conditions have yet to achieve convincing results, as the level of light captured by the satellite is
indistinguishable from the noise especially when measuring low-income areas [4]. Mobile data is introduced
as an alternative to infer socioeconomic status and utilized to estimate poverty [5]. Although the method
shows convincing results, it can not be generalized to the case of other countries due to different features
available in the mobile phone. Several other work examine the implementation of satellite imagery are the
following [6]–[8]. Considering the benefits of using satellite imagery, this study aims to evaluate the
effectiveness of this data in estimating poverty levels.
Daytime digital images have a higher image resolution than nighttime digital images. This image
contains visible objects such as buildings, roads, cars, farm fields, and roofs that allow identifying the
well-being of an area. Image color is usually used as the feature for object identification, such as plants have
shades of green, urban areas are usually silver or gray which comes from concentration of concrete or other
building materials. In addition to color, objects that can be identified are patterns, shapes, and textures. Farms
usually have prominent geometric shapes rather than naturally occurring patterns. Deforestation is often
square. The straight lines that appear are almost certainly roads, canals, to regional boundaries. It is found
that visible visual objects have correlation with expenditure level, so that it can capture the heterogeneity of
poverty level in those specific areas [9], e.g., roads are correlated with economic development, which thus
provide useful information about the level of development of an area [10]. Related research on the use of
digital images can be found in references [9], [11]–[15].
Leveraging sattelite image for poverty estimation is a complex work which requires a fast and
realiable method. Machine learning approach is one of the solutions to solve this issue. VGG16 architecture
implemented to predict socioeconomic indicators in the context of poverty mapping in Philippine archipelago
[12]. More machine and deep learning approaches, e.g., decision tree regression (DTR), support vector
machine (SVM), random forest regression (RFR), multi-layer perceptron (MLP), and one-dimensional
convolutional neural network (CNN-1D) are utilized to process and analyze multi-source satellite imagery
for poverty mapping in East Java, Indonesia [16]. However, performance of those methods needs to be
improved. One of the techniques to optimize prediction is by stacking the methods. Prior research implement
this technique integrate multiple classifiers to the heart failure study [17]. The advantages of stacking
methods in these contexts include improved prediction accuracy and the ability to harness diverse classifier
strengths, leading to more robust and reliable models [18]. With these advantages, this research proposes
stacking of several machine learning methods to improve the prediction performance. We propose to use
least absolute shrinkage and selection operator (LASSO) regression [19] as a meta-learner. LASSO
regression has been used in economic [19] and biology fields [20]. Due to frequently update of Google digital
imagery, using available satellite imagery data combined with machine learning tools to identify image
features is expected to effectively improve the predictability of per capita expenditures. We employ root
mean square error (RMSE) and R2 to evaluate the comparison of the proposed method with other machine
learning techniques.
The structure of this paper is organized as follows: section 2 outlines the materials and methods
applied in this study, detailing the data, analytical tools, and techniques employed. In section 3, we discuss
the results of our analysis, accompanied by a thorough discussion of their implications. Lastly, section 4
presents the conclusion, summarizing the main findings and their potential implications.
Per capita expenditure prediction using model stacking based on satellite imagery (Heri Kuswanto)
1222 ISSN: 2252-8938
Where 𝑎 and 𝑏 indicates latitude and longitude of the specific location respectively.
2.5. VGG16
VGGNet was developed by the Visual Geometric Group of Oxford University. This architecture
secured the second place in the ImageNet large scale visual recognition challenge 2014 (ILSVRC-2014)
competition with a test error of 7.3%. VGG16, one of the best-performing networks, comprises 16
convolutional and pooling layers followed by three fully connected layers and a softmax function [22].
The input to this architecture is a fixed 224×224 RGB image size. It uses a 3×3 a convolution layer with a
stride of 1 pixel and a padding of 1 pixel. Spatial pooling is performed by 5 layers of max-pooling layers,
each using a 2×2-pixel window with a stride 2. All hidden layers utilize rectified linear unit (ReLu)
activation. In total, the VGG16 has 138 million parameters.
𝑤 𝑇 𝑥
𝑓(𝑥) = [ ] [ ] = 𝑤 𝑇 𝑥 + 𝑏, 𝑥, 𝑤 ∈ ℝ𝑀+1 (2)
𝑏 1
The SVR tries to identify the narrowest tube surrounding the surface while minimizing prediction errors.
This condition leads to the formulation of the objective function presented in (3), where the magnitude of the
vector is normal to the approximated surface.
1
min ‖𝑤‖2 (3)
𝑤 2
The SVR utilizes an ϵ-insensitive loss function that penalizes predictions deviating from desired
output. The value of ϵ determines the width of the tube; decreasing ϵ leads to an increase in the number of
support vectors, whereas increasing ϵ results in fewer support vectors. Reducing ϵ will result in more support
vectors, while increasing ϵ will result in fewer supporting vectors. The standard formulation of SVR for
addressing the approximation problem is represented in (4) [27].
𝑓(𝑥) = ∑𝑁 ∗
𝑖=1(𝛼𝑖 − 𝛼𝑖 ) 𝑘(𝑥𝑖 , 𝑥𝑗 ) + 𝑏 (4)
2
‒ Radial basis function (RBF) where 𝑘(𝒙𝑖 , 𝒙𝑗 ) = 𝑒𝑥𝑝 (−𝛾‖𝒙𝑖 − 𝒙𝑗 ‖ ). If 𝒙𝑖 = 𝒙 and −∞ < 𝑧 < +∞, then
the RBF kernel 𝑘(𝒙, 𝒛) can be illustrated in Figure 3. Generally, this kernel has fine performance [28].
Per capita expenditure prediction using model stacking based on satellite imagery (Heri Kuswanto)
1224 ISSN: 2252-8938
During the training process, SVR finds the margin hyperplane by estimating sets of parameter
𝛼𝑖 and 𝑏 [29]. SVR performance is also determined by another set of parameters, the so-called
hyperparameters. These are the soft margin constant 𝐶 and the parameter of the kernel, 𝛾, 𝜎, or 𝑛.
2.8. XGBoost
Extreme gradient boosting (XGBoost) is a machine learning technique employed for regression
analysis and classification, derived from gradient boosting decision tree (GBDT). This method was first
introduced by [31], who linked boosting and optimization in the development of gradient boosting machine
(GBM). The boosting approach involves creating a new model to predict the errors made by the previous
model, and this process is used in the boosting method. The addition of new models is carried out until no
more error corrections can be made. The gradient boosting method uses gradient descent to minimize errors
when creating a new model. The computational process of the XGBoost algorithm is shown in Figure 4 [32].
A new tree is created in the direction of the negative gradient of the loss function. As the number of tree
models increases, the loss diminishes progressively.
For the XGBoost algorithm, determining the number of trees and depth is important. The problem in finding
the optimum algorithm can be changed by searching for a new classification that can reduce the loss function,
with the target loss function shown in (6).
(𝑡)
𝑂𝑏𝑗 (𝑡) = ∑𝑡𝑖=1 𝑙(𝑦𝑖 , 𝑦̂𝑖 ) + ∑𝑡𝑖=1 Ω(𝑓𝑖 ) (6)
Because the ensemble tree model in (6) is a function as a parameter and cannot be optimized using traditional
optimization methods on Euclidean space. So it was replaced with a model trained in an additive way, with 𝑦
using the 𝑖-th prediction and t-th iteration [33]. In minimizing the loss function, 𝑓𝑡 is added so that the (7) is
obtained as (7).
(𝑡)
𝑂𝑏𝑗 (𝑡) = ∑𝑡𝑖=1 𝑙(𝑦𝑖 , 𝑦̂𝑖 ) + (𝑓𝑡 (𝑥𝑖 )) + Ω(𝑓𝑡 ) + 𝑐𝑜𝑛𝑠𝑡𝑎𝑛𝑡 (7)
The regularization term Ω(𝑓𝑖 ) can be calculated using (8) which is used to reduce model complexity and can
improve usability on other datasets.
y = Xβ + ε (9)
Where y is the observation vector of the response variables with size n×1, 𝑋 = (𝑥1 , 𝑥2 , . . . , 𝑥𝑛 )𝑇 ,
𝑥𝑖 = (𝑥𝑖1 , 𝑥𝑖2 , . . . , 𝑥𝑖𝑝 ), i=1,2,…, n is the n×p matrix of predictor variables and 𝜀 = (𝜀1 , 𝜀2 , . . . , 𝜀𝑛 )𝑇 is the
𝑇
error vector. E(εi)=0 and Var(εi)=σ2. Then it will be searched 𝛽̂ = (𝛽̂0 , 𝛽̂1 , . . . , 𝛽̂𝑝 ) as an estimator for β
which will find the sum of the squares of the errors, namely ε Tε.
One of the assumptions that must be met when using linear regression is that the predictor variables
used should not be multicollinear. LASSO regression is one of the methods used to select predictor variables
when multicollinearity occurs in the regression model. By standardizing column X so that 𝑛−1 ∑𝑛𝑖=1 𝑥𝑖𝑗 = 0
and 𝑛−1 ∑𝑛𝑖=1 𝑥𝑖𝑗2 = 1, the LASSO 𝛽̂ estimator is defined as (10).
𝑝
With the constraint function ∑𝑗=1|𝛽𝑗 | ≤ 𝑠, where s is a quantity that controls the amount of shrinkage in the
LASSO regression coefficient. The (10) can also be written as (11).
𝑝
β̂ = arg min{∑𝑛𝑖=1(𝑦𝑖 − 𝑥𝑖𝑇 𝛽)2 + 𝜆 ∑𝑗=1|𝛽𝑗 |} (11)
𝛽
With λ is the tuning paramater of LASSO regression. λ has a maximum value of 0. If the value of alpha close
to 0, the LASSO estimate is equal to ordinary least squares (OLS) estimate. The LASSO estimation does not
have an explicit solution because the LASSO regression constraint function is an absolute function that
cannot be derived at the inflection point. One way to get the LASSO estimate is to use the least-angle
regression (LARS) algorithm [34].
𝑌 = 𝑓(𝑋) + 𝜀 (12)
1
Where 𝑓̂(𝑥) = ∑𝑥 𝑖𝑛 𝑁 𝑦𝑗 and K are the numbers of neighbors.
𝐾
The KNN algorithm is: i) determine the metric distance; ii) determine the number of nearest
neighbors (k<n); iii) calculate the distance from another data point to the desired point; iv) sort the distance
from smallest to largest; and v) calculate the average of the KNN. Generally, KNN uses Euclidean distance
in distance calculation. For p dimensions and data samples, the Euclidean distance between 2 points
𝑎1 , 𝑎2 , 𝑎3 , . . . , 𝑎𝑝 and 𝑏1 , 𝑏2 , 𝑏3 , . . . , 𝑏𝑝 is (13) [35].
2
𝐸𝑢𝑐𝑙𝑖𝑑𝑒𝑎𝑛 = √(𝑎1 − 𝑏1 )2 + (𝑎2 − 𝑏2 )2 + ⋯ + (𝑎𝑝 − 𝑏𝑝 ) (13)
While the response variable value for new data (xnew) can be calculated by (14).
1
𝑦𝑛𝑒𝑤 (𝑥𝑛𝑒𝑤 ) = ∑𝑖∈𝑘(𝑥𝑛𝑒𝑤 ) 𝑦𝑖 (14)
𝑘
The value of k determines the accuracy of the model so choosing the optimal value of k is very important.
Per capita expenditure prediction using model stacking based on satellite imagery (Heri Kuswanto)
1226 ISSN: 2252-8938
∑(𝑦𝑖 −𝑦̂𝑖 )2
𝑅𝑀𝑆𝐸 = √ (16)
𝑛
Each image obtained from Google Maps is processed using the VGG16 model, extracting a 4096-
dimensional feature vector that serves as the predictor variable. The data set, consisting of 117 observations
representing consumption expenditure per capita at the city/regency level, is divided into a training set
(80% of the data) and a validation set (20% of the data). Given the small size of the data, we employ
bootstrap sampling with 25 repetitions as a cross-validation technique.
Our base models include XGBoost, KNN, RF, and SVM. Each model undergoes hyperparameter
tuning, with RF, XGBoost, and KNN having 6, 16, and 10 hyperparameter combinations, respectively.
The SVM model is tuned across 20 hyperparameter combinations, distributed equally among four kernels:
linear, 2nd-degree polynomial, 3rd-degree polynomial, and RBF. The predictions from these seven base
models are used as new predictor variables in a stacked model, where LASSO regression is employed to
generate the final prediction. LASSO regression is chosen due to the high correlation among the base model
predictions, ensuring a more robust final model. Figure 7 illustrated the performance result of all models’;
training set is illustrated in Figure 7(a) and validation set is illustrated in Figure 7(b).
(a)
(b)
Figure 7. Performance results of (a) training set and (b) validation set
The stacked model achieved a lower RMSE than RF and the other models in both the training and
validation sets. However, it exhibited the highest standard deviation in both RMSE and R², suggesting that
the stacked model (LASSO regression) struggled to generalize well to the training data. In terms of R² values,
the stacked model was outperformed by RF and performed similarly to XGBoost and KNN in the training
set, indicating it did not achieve the best R². In the validation set, the stacked model's performance was
Per capita expenditure prediction using model stacking based on satellite imagery (Heri Kuswanto)
1228 ISSN: 2252-8938
comparable to that of RF. On the other hand, the SVM with an RBF kernel was the poorest-performing
model, likely due to suboptimal hyperparameter combinations. Optimizing the RBF SVM would require
more time and resources.
The model stacking approach yields excellent results when applied to the entire dataset, as shown in
Figure 8, with an RMSE of 936.7027 and an R² of 0.9385487. The stacked model appears to handle outliers
reasonably well, as 4 out of the 7 outliers have residuals less than twice the RMSE. However, the model
tends to underpredict when the consumption expenditure per capita exceeds 15,000. While the results are
promising, the validation phase suggests that the model may not be sufficiently robust for predicting new
data. Figure 9 presents a map of the stacked model's predictions, which clearly shows that high degree of
error observed only for few districts.
4. CONCLUSION
The stacked model is the most accurate but also the least precise among the models tested.
The stacking method provides only marginal improvements in RMSE and R² compared to individual models,
such as RF, suggesting that model selection may yield performance comparable to model stacking.
This implies that the benefits of stacking are limited in this context. Our model's performance aligns with
findings from various previous studies. However, predicting economic indicators using feature extraction
from satellite images does not outperform traditional econometric methods. To enhance the predictive power
of satellite images, fine-tuning a CNN through transfer learning could be explored in future work. Although
their results did not surpass traditional econometric methods, their approach highlights that selecting the right
CNN architecture and optimizing its parameters could potentially close this gap. Fine-tuning the CNN is
expected to produce results that are more competitive with traditional methods.
ACKNOWLEDGEMENTS
The authors gratefully acknowledge financial support from the Ministry of Education, Culture,
Research and Technology-Indonesia for this work, under project scheme of Fundamental research 2023.
REFERENCES
[1] World Bank, “Introduction to poverty analysis,” World Bank Group, 2014. Accessed: Jan. 07, 2024. [Online]. Available:
http://documents.worldbank.org/curated/en/775871468331250546/Introduction-to-poverty-analysis
[2] M. Jerven, “Poor numbers-how we are misled by African development statistics and what to do about it (Uzuazo Etemire),” in
Verfassung in Recht und Übersee, vol. 46, no. 3, 2013, pp. 336–340.
[3] A. M. Noor, V. A. Alegana, P. W. Gething, A. J. Tatem, and R. W. Snow, “Using remotely sensed night-time light as a proxy for
poverty in Africa,” Population Health Metrics, vol. 6, 2008, doi: 10.1186/1478-7954-6-5.
[4] X. Chen and W. D. Nordhaus, “Using luminosity data as a proxy for economic statistics,” Proceedings of the National Academy
of Sciences of the United States of America, vol. 108, no. 21, pp. 8589–8594, 2011, doi: 10.1073/pnas.1017031108.
[5] J. Blumenstock, G. Cadamuro, and R. On, “Predicting poverty and wealth from mobile phone metadata,” Science, vol. 350,
no. 6264, pp. 1073–1076, 2015, doi: 10.1126/science.aac4420.
[6] R. Engstrom, J. Hersh, and D. Newhouse, “Poverty from space: using high resolution satellite imagery for estimating economic
well-being,” World Bank Economic Review, vol. 36, no. 2, pp. 382–412, 2022, doi: 10.1093/wber/lhab015.
[7] T. Stark, M. Wurm, X. X. Zhu, and H. Taubenbock, “Satellite-based mapping of urban poverty with transfer-learned slum
morphologies,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 13, pp. 5251–5263,
2020, doi: 10.1109/JSTARS.2020.3018862.
[8] C. Yeh et al., “Using publicly available satellite imagery and deep learning to understand economic well-being in Africa,” Nature
Communications, vol. 11, no. 1, 2020, doi: 10.1038/s41467-020-16185-w.
[9] N. Jean, M. Burke, M. Xie, W. M. Davis, D. B. Lobell, and S. Ermon, “Combining satellite imagery and machine learning to
predict poverty,” Science, vol. 353, no. 6301, pp. 790–794, 2016, doi: 10.1126/science.aaf7894.
[10] S. M. Pandey, T. Agarwal, and N. C. Krishnan, “Multi-task deep learning for predicting poverty from satellite images,”
Proceedings of the 30th Innovative Applications of Artificial Intelligence Conference, IAAI 2018, pp. 7793–7798, 2018, doi:
10.1609/aaai.v32i1.11416.
[11] A. Head, M. Manguin, N. Tran, and J. E. Blumenstock, “Can human development be measured with satellite imagery?,”
Proceedings of the Ninth International Conference on Information and Communication Technologies and Development, pp. 1–11,
2017, doi: 10.1145/3136560.3136576.
[12] I. Tingzon et al., “Mapping poverty in the Philippines using machine learning, satellite imagery, and crowd-sourced geospatial
information,” International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives,
vol. 42, no. 4/W19, pp. 425–431, 2019, doi: 10.5194/isprs-archives-XLII-4-W19-425-2019.
[13] J. Yin, Y. Qiu, and B. Zhang, “Identification of poverty areas by remote sensing and machine learning: A case study in guizhou,
southwest China,” ISPRS International Journal of Geo-Information, vol. 10, no. 1, pp. 1-19, Dec. 2021, doi:
10.3390/ijgi10010011.
[14] J. Hersh, R. Engstrom, and M. Mann, “Open data for algorithms: mapping poverty in Belize using open satellite derived features
and machine learning,” Information Technology for Development, vol. 27, no. 2, pp. 263–292, 2021, doi:
10.1080/02681102.2020.1811945.
[15] B. Babenko, J. Hersh, D. Newhouse, A. Ramakrishnan, and T. Swartz, “Poverty mapping using convolutional neural networks
trained on high and medium resolution satellite images, with an application in Mexico,” arXiv-Statistics, pp. 1-4, 2017.
[16] S. R. Putri, A. W. Wijayanto, and S. Pramana, “Multi-source satellite imagery and point of interest data for poverty mapping in
East Java, Indonesia: Machine learning and deep learning approaches,” Remote Sensing Applications: Society and Environment,
vol. 29, 2023, doi: 10.1016/j.rsase.2022.100889.
[17] C. C. Chiu, C. M. Wu, T. N. Chien, L. J. Kao, C. Li, and H. L. Jiang, “Applying an improved stacking ensemble model to predict
the mortality of ICU patients with heart failure,” Journal of Clinical Medicine, vol. 11, no. 21, pp. 1-20, Oct. 2022, doi:
10.3390/jcm11216460.
[18] N. Kardani, A. Zhou, M. Nazem, and S. L. Shen, “Improved prediction of slope stability using a hybrid stacking ensemble method
based on finite element analysis and field data,” Journal of Rock Mechanics and Geotechnical Engineering, vol. 13, no. 1,
pp. 188–201, 2021, doi: 10.1016/j.jrmge.2020.05.011.
[19] R. Tibshirani, “Regression Shrinkage and selection via the LASSO,” Journal of the Royal Statistical Society. Series B:
Methodological, vol. 58, no. 1, pp. 267–288, 1996, doi: 10.1111/j.2517-6161.1996.tb02080.x.
[20] B. W. Sloboda, D. Pearson, and M. Etherton, “An application of the LASSO and elastic net regression to assess poverty and
economic freedom on ECOWAS countries,” Mathematical Biosciences and Engineering, vol. 20, no. 7, pp. 12154–12168, 2023,
doi: 10.3934/mbe.2023541.
[21] R. C. Gonzalez and R. E. Woods, Digital image processing. New York: Pearson, 2018.
Per capita expenditure prediction using model stacking based on satellite imagery (Heri Kuswanto)
1230 ISSN: 2252-8938
[22] M. R. Karim, M. Sewak, and P. Pujari, Practical convolutional neural networks: implement advanced deep learning models using
Python. Packt Publishing, 2018.
[23] I. Goodfellow, Y. Bengio, and A. Courville, Deep learning in adaptive computation and machine learning. The MIT Press, 2016.
[24] J. Patterson and A. Gibson, Deep learning: a practitioner’s approach, First edition. O’Reilly, 2017.
[25] V. N. Vapnik, The nature of statistical learning theory. New York: Springer, 1995.
[26] M. Awad and R. Khanna, “Efficient learning machines: Theories, concepts, and applications for engineers and system designers,”
Efficient Learning Machines: Theories, Concepts, and Applications for Engineers and System Designers, pp. 1–248, 2015, doi:
10.1007/978-1-4302-5990-9.
[27] D. Basak, S. Pal, and D. C. Patranabis, “Support vector regression,” Neural Information Processing-Letters and Reviews,
pp. 203–224, 2008.
[28] S. Makridakis, E. Spiliotis, and V. Assimakopoulos, “Statistical and machine learning forecasting methods: concerns and ways
forward,” PLoS ONE, vol. 13, no. 3, 2018, doi: 10.1371/journal.pone.0194889.
[29] J. S. Racine, L. Su, A. Ullah, W. K. Härdle, D. D. Prastyo, and C. M. Hafner, “Support vector machines with evolutionary model
selection for default prediction,” The Oxford Handbook of Applied Nonparametric and Semiparametric Econometrics and
Statistics, 2014, doi: 10.1093/oxfordhb/9780199857944.013.011.
[30] Y.-W. Chiu, Machine learning with R cookbook: explore over 110 recipes to analyze data and build predictive models with the
simple and easy-to-use R code. 2015.
[31] J. H. Friedman, “Greedy function approximation: A gradient boosting machine,” Annals of Statistics, vol. 29, no. 5, pp. 1189–
1232, 2001, doi: 10.1214/aos/1013203451.
[32] H. Mo, H. Sun, J. Liu, and S. Wei, “Developing window behavior models for residential buildings using XGBoost algorithm,”
Energy and Buildings, vol. 205, 2019, doi: 10.1016/j.enbuild.2019.109564.
[33] T. Chen and C. Guestrin, “XGBoost: A scalable tree boosting system,” Proceedings of the ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining, pp. 785–794, 2016, doi: 10.1145/2939672.2939785.
[34] W. K. Härdle and L. Simar, Applied multivariate statistical analysis. Berlin, Heidelberg: Springer, 2019.
[35] V. Kumar and M. Sahu, “Evaluation of nine machine learning regression algorithms for calibtion of low-cost PM2.5 sensor,”
Journal of Aerosol Science, vol. 157, 2021, doi: 10.1016/j.jaerosci.2021.105809.
BIOGRAPHIES OF AUTHORS
Asva Abadila Rouhan earned his bachelor's and master's degrees in statistics
from Institut Teknologi Sepuluh Nopember (ITS), Indonesia. He is interested in statistical
computation, machine learning, and large language models. He has gained notable experience
as a research assistant in the Covid-19 behavior change survey. He can be contacted at email:
[email protected].
Supriadi Hia earned his Bachelor of Applied Science degree in Statistics (2015)
from Sekolah Tinggi Ilmu Statistik (STIS), Jakarta, Indonesia, and Master of Statistics (2022)
from Institut Teknologi Sepuluh Nopember (ITS), Surabaya, Indonesia. He started working at
BPS-Statistics since 2016 until now. During his work at BPS-Statistics, he was part of the
'Regional balance and statistical analysis' team and the 'Integration of statistical processing and
dissemination' team. Currently working at BPS-Statistics of Nias Utara Regency as a 'young
statistician', and team leader of 'Integration of statistical processing and dissemination'. He can
be contacted at email: [email protected].
Tintrim Dwi Ary Widhianingsih received her bachelor's degree in statistics from
Institut Teknologi Sepuluh Nopember (ITS), Surabaya, Indonesia, in 2016, followed by a
master's degree in statistics from the same institution in 2018. She completed her Ph.D. in
computer engineering at Dongseo University, South Korea, in 2021. Her doctoral research
focused on the application of machine learning and deep learning techniques in text mining,
high-dimensional data analysis, and image processing. From 2016 to 2018, she served as a
research assistant in the Computational Statistics and Data Science Laboratory. Since 2022, she
has been an Assistant Professor in the Department of Statistics, ITS. Her research interests
encompass machine learning, deep learning, and computational statistics. She can be contacted
at email: [email protected].
Per capita expenditure prediction using model stacking based on satellite imagery (Heri Kuswanto)