Thesis - 1
Thesis - 1
RESERVOIR ENGINEERING
A Thesis
Doctor of Philosophy
in
University of Regina
By
Regina, Saskatchewan
February, 2012
Saber Khaled Elmabrouk, candidate for the degree of Doctor of Philosophy in Petroleum
Systems Engineering, has presented a thesis titled, Application of Function Approximations
to Reservoir Engineering, in an oral examination held on January 18, 2012. The following
committee members have found the thesis acceptable in form and content, and that the
candidate demonstrated satisfactory knowledge of the subject material.
Chair of Defense: Dr. Dongyan Blachford, Faculty of Graduate Studies & Research
The need for function approximations arises in many branches of engineering, and
function, which then finds an underlying relationship within a given set of input-output
data. In this study, a Multiple Regression Analysis (MRA), Artificial Neural Network
(ANN), and Least Squares Support Vector Machine (LS-SVM) are applied to address
some of the most three important ongoing challenges in reservoir engineering. The three
ongoing challenges are reservoir fluid properties in the absence of PVT analysis, average
1. The PVT properties of crude-oil such as the bubble point pressure (Pb), oil
formation volume factor (Bob), dissolved GOR (Rsob) and stock tank vent GOR (RST) play
The properties are traditionally determined from laboratory analyses of reservoir oil
used to estimate reservoir fluid properties. In this study, MRA, ANN and LS-SVM are
applied to develop PVT models with which to estimate the Pb, Bob, Rsob and RST. Unlike
the present PVT models, the proposed models can be applied in a straightforward manner
unnecessary.
2. Insight into average reservoir pressure (Pavg), and its change over time, plays a
critical role in reservoir development. However, in order to determine the Pavg, the well is
shut-in for a build up test, resulting in loss of production. In high permeability reservoirs,
this may not be a significant issue, but in medium to low permeability reservoirs, the
i
shut-in period during the entire test may last several weeks before a reliable reservoir
pressure can be estimated. This loss of production, and cost of monitoring the shut-in
pressure, is often unacceptable. It is of great practical value if the Pavg can be obtained
from the historical production and reservoir pressure data without having to shut-in the
well. Three different models (BP-ANN, ANN-GA and LS-SVM) are obtained to predict
and interpolate Pavg without closing the producing wells. The results indicate the
proposed approach has the ability to accurately interpolate and predict current average
Traditionally, numerical simulations and decline curve analysis have been used to predict
a reservoir’s future performance based on its current and past performance. Reservoir
simulation is very time consuming and offers a non-unique solution with a high degree of
uncertainty. On the other hand, decline curve analysis fits the observed production rates
predict future production by extrapolating the declining function. In this study, ANN and
LS-SVM are presented to predict the performance of oil production within water
injection reservoirs. The historical production and injection data are used as inputs. The
approach can be categorized as a new and rapid method with reasonable results. Another
application of these models is that it can be utilized to find the most economical scenario
of water injection to maximize ultimate oil recovery. This method could be a new
window for fast simulators. It has reasonable accuracy, requires little data and can
forecast quickly.
ii
ACKNOWLEDGEMENTS
A very special thanks goes out to Dr. Amr Henni, without whose motivation and
encouragement I would not have made it this far. He provided me with direction,
technical support and became more of a mentor and friend, than a professor. I doubt that I
will ever be able to convey my appreciation fully, but I owe him my eternal gratitude.
I would like to thank the other members of my committee, Dr. Koorosh Ashagari for the
assistance he provided and taking time out from his busy schedule and Dr. Malek
Mouhoub from the Department of Computer Science for his valuable suggestions and
discussions regarding my thesis study.
I gratefully acknowledge the Waha Oil Company for giving permission to use some
of production data in chapters VI, VII, VIII, IX, and X. Moreover, I would also like to
thank my family for the support they provided me through my entire life and in
particular, I must acknowledge my wife and kids, without whose love, and
encouragement I would not have finished my PhD.
In conclusion, I recognize that this research would not have been possible without the
financial assistance of Faculty of Graduate Studies and Research, and the Faculty of
Engineering and Applied Science, University of Regina (Teaching Assistantships,
Graduate Research Scholarships). I express my gratitude to them.
iii
This thesis is dedicated
to
and
with a smile.
iv
TABLE OF CONTENTS
CHAPTER I
Introduction
1.1 Background .............................................................................................................. 1
1.2 Function Approximation .......................................................................................... 3
1.2.1 Regression Analysis ......................................................................................... 3
1.2.2 Artificial Neural Network ................................................................................. 4
1.2.3 Least squares support vector machines ............................................................ 4
1.3 Key Ongoing Challenges ....................................................................................... 5
1.3.1 Why Reservoir Fluid Properties? ..................................................................... 5
1.3.2 Why Average Reservoir Pressure? ................................................................... 7
1.3.3 Why Reservoir Oil Production Prediction? ...................................................... 8
1.4 Goals of the Thesis............................................................................................... 10
1.5 Chapter by Chapter Overview .............................................................................. 11
1.6 References ............................................................................................................. 14
CHAPTER II
Prediction of Bubble Point Pressure, Bubble Point Oil Formation Volume Factor in the
Absence of PVT Analysis by Utilizing Multiple Regression Analysis
2.1 Introduction ............................................................................................................ 16
2.2 Literature Review................................................................................................ 19
2.3 Newly Developed Pb and Bob Correlations ......................................................... 25
2.3.1 Bubble point pressure correlation ................................................................... 26
2.3.2 Bubble Point Oil FVF Correlation.................................................................. 28
2.4 Correlation Verification and Validation ............................................................. 28
2.5 Comparison with Other Correlations .................................................................. 35
2.6 Summary and Conclusions ................................................................................. 39
2.7 References ........................................................................................................... 41
CHAPTER III
Prediction of Bubble Point Solution Gas/Oil Ratio in the Absence of PVT Analysis
3.1 Introduction .......................................................................................................... 44
3.2 Literature Review................................................................................................... 46
3.3 Newly Developed Correlations ............................................................................. 48
3.3.1 PVT Data ........................................................................................................ 49
3.3.2 Bubble Point Solution Gas/Oil ratio, Rsob Correlation ................................... 50
3.3.3 Stock-Tank Vent Gas/Oil Ratio, RST Correlation ........................................... 53
v
Figure 3.2 - Estimated Rsob from Eq. (3.7) vs. Actual Rsob (431 Data Points)3.4
Comparison with Other Correlations 56
3.5 Conclusions ............................................................................................................ 66
3.6 References .............................................................................................................. 68
CHAPTER IV
Development of PVT Models Utilizing ANNs
4.1 Introduction ............................................................................................................ 71
4.2 Review of Recent PVT Neural Network Models ............................................... 72
4.3 PVT Datasets ...................................................................................................... 79
4.4 Bubble point Pressure ANN Model ....................................................................... 80
4.4.1 Bubble Point Pressure ANN Model Evaluation ............................................. 83
4.4.2 Contribution of Input Parameters ................................................................... 91
4.4.3 Deployment Test............................................................................................. 91
4.5 Bubble Point Oil FVF ANN Model .................................................................... 94
4.5.1 Graphical Error Analysis .............................................................................. 102
4.5.2 Deployment Test........................................................................................... 117
4.6 Bubble Point Solution Gas/Oil ratio (Rsob) ANN Model .................................. 117
4.6.1 Model Design ................................................................................................... 122
4.6.2 Model Evaluation ......................................................................................... 126
4.6.3 Model Deployment ....................................................................................... 138
4.7 Stock-Tank Vent Gas Oil Ratio ANN Model ..................................................... 138
4.7.1 Stock-Tank Vent Gas/Oil Ratio ANN Model Evaluation ............................ 142
4.8 Conclusions ....................................................................................................... 154
4.9 References .......................................................................................................... 156
CHAPTER V
Development of PVT Models Utilizing LS-SVM Approach
5.1 Introduction ....................................................................................................... 158
5.2 LS-SVM for Function Approximation................................................................. 159
5.3 Model Parameter Selecting ............................................................................... 161
5.4 Newly Developed PVT LS-SVM Models ........................................................ 162
5.5 PVT Data Preparation ....................................................................................... 163
5.5.1 Accuracy of the Bubble Point Pressure LS-SVM Model ............................. 171
5.6 Bubble Point Oil Formation volume Factor, Bob LS-SVM Model ..................... 175
5.6.1 Accuracy of the Bubble Point Oil FVF LS-SVM Model ............................. 175
5.7 Bubble Point Solution Gas/Oil Ratio LS_SVM Model ....................................... 176
vi
5.7.1 Accuracy of the Bubble Point Solution Gas/Oil Ratio LS-SVM Model ...... 184
5.8 Stock-Tank Vent Gas/Oil Ratio LS-SVM Model ................................................ 185
5.8.1 Accuracy of the RST LS-SVM Model ........................................................... 198
5.9 Conclusions .......................................................................................................... 198
5.10 References ............................................................................................................ 202
CHAPTER VI
Estimating Average Reservoir Pressure: A Neural Network Approach
6.1 Introduction .......................................................................................................... 204
6.2 Artificial Neural Networks (ANNs)..................................................................... 205
6.3 The Proposed Model ......................................................................................... 206
6.3.1 Reservoir Production History and Pressure Data ......................................... 208
6.3.2 Data Preparation ........................................................................................... 213
6.3.3 The Architecture of a Neural Network ......................................................... 216
6.4 Average Reservoir Pressure ANN Model ........................................................... 219
6.5 Estimation of Average Reservoir Pressure ......................................................... 232
6.6 Conclusions ......................................................................................................... 235
6.7 References ........................................................................................................... 236
CHAPTER VII
A Hybrid Genetic Algorithm to Estimate Average Reservoir Pressure
7.1 Introduction ......................................................................................................... 238
7.2 Genetic Algorithm .............................................................................................. 239
7.3 A hybrid Genetic Algorithm ............................................................................... 239
7.4 BP-ANN-GA Model Building ............................................................................ 241
7.4.1 Evolving PB-ANN weights using GA .......................................................... 242
7.5 Prediction of Average Reservoir Pressure .......................................................... 244
7.6 Conclusion .......................................................................................................... 244
7.7 References ........................................................................................................ 250
CHAPTER VIII
Estimating Average Reservoir Pressure: A LS-SVM Approach
8.1 Introduction ......................................................................................................... 251
8.2 Historical Production Data .................................................................................. 252
8.3 LS-SVM Model .................................................................................................. 255
8.3.1 Modeling Process ......................................................................................... 256
vii
8.3.2 Tuning Model Parameters ............................................................................ 258
8.4 Interpolating Average Reservoir Pressure .......................................................... 261
8.5 Forecasting Process ............................................................................................. 261
8.6 Conclusions ......................................................................................................... 268
8.7 References ........................................................................................................... 269
CHAPTER IX
Artificial Neural Network Modeling for Oil Production Prediction
9.1 Introduction ......................................................................................................... 270
9.2 Oil Production Prediction Neural Networks ....................................................... 273
9.2.1 Oil Reservoir Production Data ..................................................................... 274
9.2.2 Data Pre-Processing ...................................................................................... 276
9.2.3 The Network Model Architecture................................................................. 277
9.3 Results and Discussions ...................................................................................... 277
9.4 Conclusions ......................................................................................................... 285
9.5 References ........................................................................................................... 286
CHAPTER X
Forecasting of Oil Production under Water Injection: LS-SVM Approach
10.1 Introduction .......................................................................................................... 287
10.2 Description of the Production and Injection Data................................................ 289
10.3 LS-SVM Models .................................................................................................. 289
10.4 LS-SVM Model 1, (M1) ...................................................................................... 291
10.4.1 Prediction Performance of M1 ..................................................................... 296
10.5 LS-SVM Model 2, (M2) ...................................................................................... 296
10.5.1 Prediction Performance of M2 ..................................................................... 301
10.6 Results and Discussions ....................................................................................... 307
10.7 Conclusions .......................................................................................................... 307
CHAPTER XI
Conclusions and Recommendations
11.1 Conclusions .......................................................................................................... 309
11.2 Recommendations ................................................................................................ 313
viii
LIST OF FIGURES
Figure 2.1− Fluid flow of Oil and Gas from the Reservoir to the Separator and Stock-
Tank (Two-Stage Separation) ................................................................................... 18
Figure 2.2 −Error Distribution for Bubble point Pressure Correlation ............................. 31
Figure 2.3 −Error Distribution for Bob Correlation ........................................................... 32
Figure 2.4 −A 45o Straight Line Cross Plot for Bubble point Pressure Correlation ......... 33
Figure 2.5 − A 45o Straight Line Cross Plot for Bubble point Oil FVF Correlation ........ 34
Figure 2.6 − Evaluation and Validation of Bubble Point Oil FVF Correlation ................ 38
Figure 3.1 − Residual Distributions of Rsob Obtained from Eq. (3.7) (431 Data Points) .. 55
Figure 3.3 − Evaluation Test; Estimated vs. Actual Rsob (355 Data Points) ..................... 59
Figure 3.4 − Evaluation Test; Relative Error Distributions (355 Data Points) ................. 60
Figure 3.5 − Validation Test; Actual vs. Estimated Rsob (49 Data Points) ....................... 64
Figure 3.6− Validation Test; Relative Error Distributions (49 Data Points) .................... 65
Figure 4.1 – Cross Plot of Modeled Bubble point Pressure on Training Dataset versus
Laboratory PVT Values ............................................................................................ 87
Figure 4.2 – Cross-plot of Modeled Bubble point Pressure on Validation Dataset versus
Laboratory PVT Analysis ......................................................................................... 88
Figure 4.3 - Cross-plot of Modeled Bubble point Pressure on Testing Dataset versus
Laboratory PVT Analysis ......................................................................................... 89
Figure 4.4 - Cross-plot of Modeled Bubble point pressure on Overall Points versus
Laboratory PVT Analysis ......................................................................................... 90
Figure 4.5 – Deployment Test: Cross-plot of Modeled Bubble point versus Laboratory
PVT Analysis ............................................................................................................ 92
Figure 4.6 – Bob BP-Model Architecture, [4-5-4-1] .......................................................... 97
Figure 4.7 (a, b) – Graphical Error Analysis on Training Dataset .................................. 105
Figure 4.7 (c, d) – Graphical Error Analysis on Training Dataset .................................. 106
Figure 4.7 (e, f) – Graphical Error Analysis on Training Dataset .................................. 107
Figure 4.8 (a, b) – Graphical Error Analysis on Validation Dataset............................... 108
Figure 4.8 (c, d) – Graphical Error Analysis on Validation Dataset............................... 109
Figure 4.8 (e, f) – Graphical Error Analysis on Validation Dataset ............................... 110
Figure 4.9 (a, b) – Graphical Error Analysis on Testing Dataset ................................... 111
Figure 4.9 (c, d) – Graphical Error Analysis on Testing Dataset ................................... 112
Figure 4.10 (a, b) – Graphical Error Analysis on Overall Dataset ................................. 114
Figure 4.10 (c, d) – Graphical Error Analysis on Overall Dataset ................................. 115
Figure 4.10 (e, f) – Graphical Error Analysis on Overall Dataset .................................. 116
Figure 4.11 – Deployment Test of Bob BP-ANN Model ................................................. 119
Figure 4.12 – Rsob BP-ANN Model Architecture [3-7-4-1] ............................................ 125
Figure 4.13 – Cross-plot of Modeled Rsob on Training Dataset vs. Laboratory PVT Values
................................................................................................................................. 130
Figure 4.14 – Cross-plot of Modeled Rsob on Validation Dataset vs. Laboratory PVT
Values ..................................................................................................................... 131
Figure 4.15 – Cross-plot of Modeled Rsob on Testing Dataset vs. Laboratory PVT Values
................................................................................................................................. 132
Figure 4.16 – Cross-plot of Modeled Rsob on Overall Data Points vs. Laboratory PVT
Values ..................................................................................................................... 133
ix
Figure 4.17 – Histogram of the Residuals of Training Dataset ...................................... 134
Figure 4.18 – Histogram of the Residuals of Validation Dataset ................................... 135
Figure 4.19 – Histogram of the Residuals of Testing Dataset ........................................ 136
Figure 4.20 – Histogram of the Residuals of Overall Data Points.................................. 137
Figure 4.21 – Rsob BP-ANN Model Deployment Test .................................................... 139
Figure 4.22 – Cross-plot of Modeled RST on Training Dataset vs. PVT Values ............ 146
Figure 4.23 – Cross-plot of Modeled RST on Validation Dataset vs. PVT Values ......... 147
Figure 4.24 – Cross-plot of Modeled RST on Testing Dataset vs. PVT Values .............. 148
Figure 4.25 – Cross-plot of Modeled RST on Overall Data Points vs. PVT Values........ 149
Figure 4.26 – Histogram of the Residuals of Training Dataset ...................................... 150
Figure 4.27– Histogram of the Residuals of Validation Dataset .................................... 151
Figure 4.28 – Histogram of the Residuals of Testing Dataset ........................................ 152
Figure 4.29 – Histogram of the Residuals of Overall Data Points.................................. 153
Figure 5.1− Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Pb when Optimizing the Parameters γ [exp(5.627)] and σ2 [exp(1.756)].
The Red Square Indicates the Selected Optimal Settings. ...................................... 167
Figure 5.2− Performance of Training Dataset for Pb LS-SVM Model. Upper: Calculated
Pb vs. Data Observation Order. Lower: Calculated Pb vs. Experimental Values. .. 168
Figure 5.3− Error Distribution Resulting from the Training Dataset for Pb LS-SVM
Model ...................................................................................................................... 170
Figure 5.4− Performance of Testing Dataset for Pb LS-SVM Model. Upper: Calculated Pb
vs. Data Observation Order. Lower: Calculated Pb vs. Experimental Values. ....... 172
Figure 5.5− Error Distribution Resulting From Testing Dataset for Pb LS-SVM Model173
Figure 5.6− Performance of Validation/Testing Dataset for Pb LS-SVM Model........... 174
Figure 5.7 - Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Bob when Optimizing the Parameters γ [exp(8.874)] and σ2 [exp(6.893)].
The Black Square Indicates the Selected Optimal Settings. ................................... 177
Figure 5.8− Performance of the Optimal Bob LS-SVM Model in Training Dataset ....... 179
Figure 5.9− Error Distribution Resulting from Training Dataset for Bob LS-SVM Model
................................................................................................................................. 180
Figure 5.10− Performance of the Optimal Bob LS-SVM Model in Testing Dataset ....... 181
Figure 5.11 –Error Distribution Resulting from Testing dataset for Bob LS-SVM Model
................................................................................................................................. 182
Figure 5.12− Performance of Validation/Testing Dataset for Bob LS-SVM Model ....... 183
Figure 5.13 - Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Rsob when Optimizing the Parameters γ [exp(14.89)] and σ2 [exp(8.97)].
The Black Square Indicates the Selected Optimal Settings. ................................... 186
Figure 5.14 − Performance of the Optimal Rsob LS-SVM Model in the Training Dataset
................................................................................................................................. 187
Figure 5.15 –Error Distribution Resulting from Training Dataset for Rsob LS-SVM Model
................................................................................................................................. 189
Figure 5.16 − Performance of the Optimal Rsob LS-SVM Model in the Testing Dataset 190
Figure 5.17 – Error Distribution Resulting from Testing Dataset for Rsob LS-SVM Model
................................................................................................................................. 191
Figure 5.18− Performance of Validation/Testing Dataset for Rsob LS-SVM Model ...... 192
x
Figure 5.19 - Contour Plot of the Optimization Error of Cross Validation of RST when
Optimizing the Parameters γ [exp(3.864)] and σ2 [exp(0.298)]. The Black Square
Indicates the Selected Optimal Settings. ................................................................. 195
Figure 5.20 − Performance of the Optimal RST LS-SVM Model in the Training Dataset
................................................................................................................................. 196
Figure 5.21 –Error Distribution Resulting from RST LS-SVM Model ........................... 197
Figure 5.22− Performance of Rsob Calculated from RST LS-SVM Model ...................... 200
Figure 5.23 − Performance of Rsob Calculated from RST LS-SVM Model using Testing
Dataset..................................................................................................................... 201
Figure 6.1 – Impact on the Decline of an Average Reservoir Pressure According to the
Oil Production Rate................................................................................................. 209
Figure 6.2 – Impact on the Decline of an Average Reservoir Pressure According to the
Gas Production Rate ............................................................................................... 210
Figure 6.3 – Impact on the Decline of an Average Reservoir Pressure According to the
Water Production Rate ............................................................................................ 211
Figure 6.4 – Impact on the Decline of an Average Reservoir Pressure According to the
Number of Producers .............................................................................................. 212
Figure 6.5 – Actual Average Reservoir Pressure versus Modeled ANN Values for
Training Dataset ...................................................................................................... 225
Figure 6.6 – Actual Average Reservoir Pressure versus Modeled ANN Values for
Validation Dataset ................................................................................................... 226
Figure 6.7 – Actual Average Reservoir Pressure versus Model ANN Values for Testing
Dataset..................................................................................................................... 227
Figure 6.8 – Actual Average Reservoir Pressure versus Modeled ANN Values for Overall
Dataset..................................................................................................................... 228
Figure 6.9 − Error Distribution for Training Dataset ...................................................... 230
Figure 6.10 − Error Distribution for Overall Dataset ..................................................... 231
Figure 6.11 – ANN model performance ......................................................................... 233
Figure 7.1 – Modeled Versus Field Average Reservoir Pressure ................................... 247
Figure 8.1 - Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Calculating Average Reservoir Pressure. The Black Square Indicates the
Selected Optimal Settings of γ [exp(2.837)] and σ2 [exp(1.418)]. .......................... 260
Figure 8.2 − Histogram of the Errors of the LS-SVM Average Reservoir Pressure Model
................................................................................................................................. 262
Figure 8.3 − Performance of the Optimal LS-SVM Model in the Training Dataset ...... 263
Figure 8.4 − Performance of the LS-SVM Model. Both training and Testing Dataset are
Scattered around the y = x Line indicating No Overfit Existing in the Model ....... 264
Figure 8.5 − LS-SVM Model Performance. The Plot can be used to interpolate the
missing average reservoir pressure values. ............................................................. 265
Figure 9.1− Reservoir Oil Production Performance from Sept. 1978 to July 2010........ 275
Figure 9.2− ANN Model Output versus the Actual Reservoir Production (1st Dataset) 279
Figure 9.3− Performance of the ANN Model on Training Dataset ................................ 280
Figure 9.4− Performance of the ANN Model on Validation Dataset ............................. 281
Figure 9.5− Performance of the ANN Model on Testing Dataset .................................. 282
xi
Figure 9.6− Prediction of average reservoir oil production (one-year-ahead from Aug.
2009 to Jul. 2010) ................................................................................................... 284
Figure 10.1 - Contour Plot of the Optimization Error for LS-SVM Regression M1 when
Optimizing the Parameters. The red square indicates the selected optimal Settings of
γ = exp(3.53) & σ2 = exp(4.21)................................................................................ 293
Figure 10.2 – History Matching of LS-SVM M1 ........................................................... 294
Figure 10.3 - Cross Plot of Modeled Oil Production Rates vs. Field Values for M1 ..... 295
Figure 10.4 - Flow Chart of Load Forecasting ............................................................... 297
Figure 10.5 – Cross Plot of Predicted Oil Production Rates vs. Field Values for M1 ... 298
Figure 10.6 - Contour Plot of the Optimization Error for M2 Model when Optimizing the
Parameters. The red square indicates the selected optimal settings of γ = exp(7.201)
& σ2 = exp(5.532) .................................................................................................... 300
Figure 10.7 – History Matching of M2 ........................................................................... 302
Figure 10.8 – Correlation between Modeled and Field Oil Flow Rates for M2 ............. 303
Figure 10.9 – Forecasting Oil Production against Field Values for M2 Model ............. 304
Figure 10.10 − Illustrates the Comparison of ARE between the Two Models in
Forecasting Data Domain ....................................................................................... 306
Figure A.1 - Different Features in Residual Plots, Where the Residuals Plotted versus
fitted values (x-axis). After Draper and Smith (1998) ............................................ 320
Figure A.2 - Further Different Features in Residual Plots .............................................. 322
Figure A.3 - Histogram of the Residuals. The Residuals Close Normal Distribution .... 324
Figure A.4 - Tests of the Normality. The Histogram and the Normal Probability Plot in (c
and e) Show that the Error Terms are Normally Distributed. ................................. 325
Figure A.5 - Box Plot ...................................................................................................... 327
Figure A.6 - ANOVA Table ........................................................................................... 331
Figure A.7 - Different Features in Correlation Coefficient ............................................ 333
Figure A.8 - Probability Density Function Graph of a T-distribution. The P-value is the
sum of the two shaded areas under the curve. ........................................................ 336
Figure B.1- Schematic of Typical Multilayer Neural Network Model. It consists of an
input layer with four neurons (xp1, xp2, xp3, and xp4), one hidden layer with five
neurons (hn1, hn2, hn3, hn4, and hn5) and an output layer with single neuron, y. ... 346
Figure B.2- A Neural Net Perceptron ............................................................................. 350
Figure B.3- Illustration of the Early-stopping Rule. The trained network is saved at the
point defined by the early-stopping criterion, and training continues exclusively to
check whether the error will fall again. This will ensure that the increase in the
validation error is not a temporary event. ............................................................... 352
Figure B.4- (a) Hypothetical Surface in Weight Space with Many Local Minimums. (b)
Cross-section of a Hypothetical Error Surface in Weight Space. ........................... 356
Figure C.1 - Illustration of the Key Ingredient of LS-SVM: Transformation Φ of the input
data to the feature space. ......................................................................................... 360
Figure C.2 − LS-SVM Modeling Procedure ................................................................... 364
Figure C.2 – One-iteration of a 3-fold Cross Validation Algorithmic Pattern ............... 366
xii
xiii
LIST OF TABLES
Table 2.1 - Range of Data in Group A .............................................................................. 27
Table 2.2 - Quantitative Statistical Error Analysis ........................................................... 30
Table 2.3 - Data Range of Group B .................................................................................. 36
Table 2.4 - Comparison of Proposed Bubble point Oil FVF Correlation ......................... 37
Table 3.1− Range of Data Points for Rsob Correlation (431 Data Points) ......................... 51
Table 3.2− Range of Data points for RST Correlation (355 Data Points) .......................... 52
Table 3.3 − Quantitative Statistical Analysis of the Present Correlations ........................ 54
Table 3.4− Statistical Analysis of Evaluation Test (355 Data Points) .............................. 58
Table 3.5 − Range of Data used for Validation Test (49 Data Points) ............................. 62
Table 3.6−Statistical Accuracy of Validation Test (49 Data Points) ................................ 63
Table 4.1 − Ranges of Datasets used in Training, Validation, Testing and Deployment Pb
ANN Model. ............................................................................................................. 81
Table 4.2 − Comparison of Pb Network Architectures - The 5 Best Architectures .......... 84
Table 4.3 − Average Error during Training, Validation and Testing Pb Model ............... 85
with Various Training Algorithms (Quasi-Newton (Q-N), Conjugate Gradient Descent
(CGD), Levenberg-Marquardt (L-M) and Limited Memory Quasi-Newton (LM Q-
N) .............................................................................................................................. 85
Table 4.4 − Ranges of Pb ANN Model Output with Statistical Error Analysis ................ 86
Table 4.5- Statistical Error Analysis for Deployment Test............................................... 93
Table 4.6 – Ranges of Datasets used in Training, Validation, Testing and Deployment of
the Bob ANN Model .................................................................................................. 95
Table 4.7 – Comparison of Bob Network Architectures - The 5 Best Architectures ......... 98
Table 4.8 - Average Error during Training, Validation and Testing of Architecture [4-5-4-
1] with Various Training Algorithms and Input/Output Activation Functions.
(Illustrates the Effect of Choice of Training Algorithm and Activation Function) 103
Table 4.9 – Comparison between the Four Best Network Models ................................. 104
Table 4.10- Statistical Error Analysis for Deployment Test........................................... 120
Table 4.11 − Ranges of Datasets used in Training, Validation, Testing and Deployment
Rsob ANN Model. .................................................................................................... 121
Table 4.12 − Comparison of Rsob Network Architectures - The 5 Best Architectures ... 124
Table 4.13 - Average Error during Training, Validation and Testing Rsob Model with
Various Training Algorithms (Quick Propagation (QP), Quasi-Newton (Q-N),
Conjugate Gradient Descent (CGD), and Limited Memory Quasi-Newton (LM Q-N)
................................................................................................................................. 127
Table 4.14 – Statistical Error Analysis of Rsob ANN Model Output............................... 128
Table 4.15- Statistical Error Analysis for Deployment Test........................................... 140
Table 4.16 - Comparison of RST Network Architectures - The Best Architectures ........ 143
Table 4.17 - Average Error during Training, Validation and Testing RST Model with
Various Training Algorithms (Quasi-Newton (Q-N), Conjugate Gradient Descent
(CGD), and Limited Memory Quasi-Newton (LM Q-N) ....................................... 144
Table 4.18 - Ranges of RST ANN Model Output with Statistical Error Analysis ........... 145
Table 5.1 – Ranges of PVT Data .................................................................................... 166
Table 5.2 − Statistical Error analysis of Pb LS-SVM Model .......................................... 169
Table 5.3 − Statistical Error Analysis of Bob LS-SVM Model ....................................... 178
xiv
Table 5.4 − Statistical Error Analysis of Rsob LS-SVM Model ...................................... 188
Table 5.5 – Ranges of PVT Data used in RST LS-SVM Model ...................................... 194
Table 5.6 − Error of Analysis of Rsob Obtained by Adding LS-SVM RST and Field RSP 199
Table 6.1- Range of Overall Data ................................................................................... 214
Table 6.2 – Range of Data Used in Training Process ..................................................... 215
Table 6.3 - Network Architecture Parameters ................................................................ 217
Table 6.4 - Top Five Architectures - The best results are emphasized in bold............... 218
Table 6.5 - Error Analysis of Architecture [4-10-1] under Various Training Algorithms
................................................................................................................................. 220
Table 6.6 - Error analysis of Architecture [4-8-3-1] under Various Training Algorithms
................................................................................................................................. 221
Table 6.7 - Error Analysis of Architecture [4-8-3-1] under Various Training Algorithms
................................................................................................................................. 222
Table 6.8 - Error Analysis of Architecture [4-8-1-4-1] under Various Training Algorithms
................................................................................................................................. 223
Table 6.9 - Error Analysis of Architecture [4-7-6-4-1] under Various Training Algorithms
................................................................................................................................. 224
Table 6.10 – Statistical Error Analysis of the Proposed Model ...................................... 229
Table 6.11 – Model Prediction Performance .................................................................. 234
Table 7.1 – BP-ANN-GA Model Parameters ................................................................. 246
Table 7.2 – Statistical Error Analysis of the Proposed Model ........................................ 248
Table 7.3 – Model Prediction Performance .................................................................... 249
Table 8.1 - Average Reservoir Pressure Data; ................................................................ 254
Testing Dataset is emphasized in Bold ........................................................................... 254
Table 8.2 − Statistical Error Analysis of the Proposed Model ....................................... 266
Table 8.3 – LS-SVM Model Forecasting Performance .................................................. 267
Table 9.1 – Error Analysis of the Proposed Model ........................................................ 283
Table 10.1 − Statistical Error Analysis of M1 and M2 LS-SVM Models ...................... 305
xv
LIST OF APPENDICES
A.1 Introduction ............................................................................................................. 315
A.2 Simple Linear Regression Model............................................................................ 316
A.3 Correlation and Regression Modeling in the Oil and Gas Industry ........................ 316
A.4 Correlation and Regression Modeling Building Procedure .................................... 317
A.5 Multiple Linear Regression Model Assumptions ................................................... 318
A. 5.1 Linearity - Assumption (1)................................................................................ 318
A5.2 Independence of Error terms - Assumption (2)....................................................... 319
A.5.2.1 Standard Error of Estimate .......................................................................... 321
A.5.3 Normality - Assumption (3) .............................................................................. 323
A.5.3.1 Outliers, Leverage and Influential Points in Regression ............................. 326
A.5.4 Multicollinearity - Assumption (4) ................................................................... 328
A.5.4.1 Coefficient of Determination R2 .................................................................. 329
A.5.4.2 Adjusted R2 .................................................................................................. 330
A.5.4.3 Correlation Coefficient among Variables, Cor(X1,X2)................................. 332
A.6 Stepwise Regression (Variables Selection): ........................................................... 332
A.7 T-statistics and P-value ........................................................................................... 334
A.8 Analysis of Variance (ANOVA Table)................................................................... 335
A.9 Comparing Models.................................................................................................. 337
A.10 References ............................................................................................................... 339
A.10.1 Regression Analysis......................................................................................... 339
A.10.2 Regression Analysis in Seismic ....................................................................... 339
A.10.3 Regression Analysis in Petrophysical ............................................................. 339
A.10.4 Regression analysis in Natural Gas Engineering ........................................... 340
A.10.5 Regression analysis in Relative Permeability ................................................. 340
A.10.6 Regression Analysis in Reservoir Recovery Factor ........................................ 342
A.10.7 Regression Analysis in Drilling Engineering .................................................. 342
A.10.8 Regression analysis in Flow in Porous Media ................................................ 343
B.1 Introduction ............................................................................................................. 344
B.2 Backpropagation Neural Network .......................................................................... 344
B.3 Mathematic Description of Backpropagation Algorithm ....................................... 345
B.4 Design of Neural Networks Architecture ............................................................... 348
B.4.1 Number of Input Variables ............................................................................... 349
B.4.2 Data Pre-processing ........................................................................................ 349
B.4.3 Type of Neural Network Architecture .............................................................. 351
B.4.4 Number of Hidden Layers and Hidden Nodes per Hidden Layer .................... 353
B.4.5 Activation Function .......................................................................................... 353
B.4.6 Type of Optimization Technique ...................................................................... 354
B.5 Selections of the Best Architecture ......................................................................... 357
B.6 Model Verification .................................................................................................. 357
C.1 Introduction ............................................................................................................. 359
C.2 LS-SVM for Function Approximation.................................................................... 361
C.3 LS-SVM Regression Modeling Procedure ............................................................. 362
C.4 Leave-one-out Cross-validation for LS-SVM ........................................................ 363
C.5 Grid-search Technique ............................................................................................ 365
C.6 Statistical Measures for the Evaluation of Prediction Performance ....................... 365
xvi
C.7 References ............................................................................................................... 367
D.1 Introduction .............................................................................................................. 368
D.2 Mean Squared Error (MSE) .................................................................................... 368
D.2 Root Mean Square Error (RMSE)........................................................................... 369
D.3 Relative Error (RE) ................................................................................................. 369
D.4 Absolute Relative Error (ARE)............................................................................... 369
D.5 Standard Deviation (SD) ......................................................................................... 369
D.6 Correlation Coefficient (R2).................................................................................... 369
xvii
NOMENCLATURE
xviii
psi Pounds per Square Inch
psia Pounds per Square Inch Absolute
ln Natural Logarithm
API API Stock Tank Oil Gravity, degree
Bob Bubble point oil formation volume factor, bbl/STB
Pavg Average reservoir pressure, psi
pb Bubble point pressure, psia
PR Reservoir pressure, psi
PSP Separator pressure, psia
2
R Correlation coefficient, 100%
Rsob Bubble point solution gas oil ratio, SCF/STB
Rso Solution gas oil ratio, SCF/STB
RsFb Flash bubble point solution GOR, SCF/STB
RsDb Differential bubble point solution GOR, SCF/STB
RSP Separator gas oil ratio, SCF/STB
RST Stock-tank vent gas oil ratio, SCF/STB
TR Reservoir temperature, °F
TSP Separator temperature, °F
TST Stock-tank temperature, °F
γ LS-SVM regularization parameter
γg Gas specific gravity, air = 1
γgSP Separator gas specific gravity, air =1
γgST Stock-Tank gas specific gravity, air = 1
γgTotal Total gas specific gravity, air = 1
γoST Stock-Tank oil specific gravity, water = 1
σ2 kernel parameter for LS-SVM
xix
CHAPTER I
INTRODUCTION
1.1 Background
as regression analysis, Artificial Neural Networks (ANNs) and Support Vector Machine
(SVM). The need for function approximations arises in many branches of engineering,
and in particular, petroleum engineering. The field of petroleum engineering is filled with
Many problems in the oil and gas industry are extremely difficult, if not impossible,
to solve using the existing analytical tools, and the same problems prove to be difficult to
solve numerically. As an example, the partial differential equation (Equation 1.1) that
describes the constant flow of a fluid in reservoir rocks, and is independent of fluid type
is:
= ∅
(1.1)
1
3. Gravitational effects are negligible.
that the diffusivity equation would never apply to actual reservoirs. If it is assumed this is
true, and if the limitations are kept in mind, the results obtained using Equation (1.1) are
within acceptable engineering accuracies (Chen at all, 2006). Thus, even when the exact
need for function approximations arises in almost all branches of the oil and gas industry.
defined by y = f (x, t) in which x and y are vectors and t denotes time. The oil and gas
production rates change as a function of the passing of time and, therefore, can be viewed
as a function of time and spatial location for certain wells and production operations. If
the function y = f (x, t) is approximated on the basis of known data, the production
problem asks that a function among a well-defined class that closely matches a target
function (value of the function or output) in a task-specific way be selected. However, the
task of function approximation is to find an estimate of the unknown function which then
2
1.2 Function Approximation
Consider a given set of training samples {xk, yk}k=1….N, in which xk is the input vector
(predictors) and yk is the corresponding target value (output or response) for sample k.
The goal of function approximation is to learn the underlying relation between the input
and the target value. Following the retrieval of this information, a function approximation
defined as:
ek = − (1.2)
MRA, ANN and LS-SVM and the appropriate way of applying each method. In all
cases, the approximation target is a function of the input variables called “a model”.
Thus, function approximations include any techniques with which to model and analyze
several variables searching for a relationship between a dependent variable and one or
more independent variables. Regression analysis, artificial neural networks, and least
squares support vector machine modeling techniques, for example, can be applied.
estimate the value of a continuous random variable (dependent variable or output) using
simple regression is one in which a single independent variable (just one input) is used to
3
determine a dependent variable. The MRA method learns a relation between a set of
In this equation, the α's are the n parameters which receive a value during the training
by minimizing the summed squared approximation error over all examples. The
the ANN method is growing rapidly in the oil and gas industry. The multilayer feed
forward neural network is one of the most widely used forms of ANN and is capable of
Least Squares Support Vector Machines (LS-SVM) are least squares versions of
Support Vector Machines (SVM), a class of kernel based learning methods used for
regression analysis and classification. In LS-SVM the solution is found by solving a set
4
1.3 Key Ongoing Challenges
Key Ongoing Challenges highlight the area of reservoir engineering that describes the
key individual challenges that affect reservoir management and their ability to implement
new strategies. The narratives in this brief focus on three key sets of issues, i.e. reservoir
This research regarding the application of MRA, ANN and LS-SVM, in order to
innovative because it presents a very valuable tool to petroleum engineering. This work is
poised to support the oil and gas industry, especially in the areas of reservoir
The physical and chemical properties of reservoir oils vary considerably and are
dependent upon the concentration of the various types of hydrocarbons and minor
constituents present in the well. Reservoir fluid studies are essentially based on Pressure-
procedures designed to provide the values of the reservoir fluid properties as required in
material balance calculations, well test analysis, reserve estimates, inflow performance
5
to estimate reservoir fluid properties. The reasons for using empirical PVT correlations
• economic issues
A literature survey has shown that very little attention is paid to estimating reservoir
analysis. Until recently, not one specific correlation has been published to directly
estimate the bubble point pressure (Pb) in the absence of PVT analysis and, at the
moment, only one published correlation is available to estimate the bubble point of oil
FVF (Bob) directly in the absence of PVT analysis (Labedi 1990). The majority of the Pb
pressure and Bob correlations cannot be applied directly as the correlations of bubble
point solution gas/oil ratio (Rsob) must be known along with gas specific gravity (γg) as
part of the input variables, both of which are rarely measured field parameters. However,
Pb and Bob correlations are essentially based on the relationship illustrated in Equations
1.4 and 1.5, respectively. Rsob and γg can be obtained either experimentally or estimated
from correlations. However, the oil and gas industry offers numerous correlations with
which to estimate Rsob. The basic assumption of Rsob correlations is expressed in Equation
1.6:
6
Pb = f1 (Rsob, γg, API gravity, TR) (1.4)
Accordingly, in order to apply those correlations in the absence of PVT analysis, the
Pb/Rsob and γg must be known in advance (they are rarely measured as field data).
Insight into average reservoir pressure and its change in production (time), plays a
hydrocarbon in place, computing rock and fluid characteristics, reservoir material balance
calculations, pressure maintenance projects, surface and subsurface facility designs and
which measures long-term built-up pressure when the producing wells are shut in. The
test is initiated after the well is opened to a constant flow rate over an extended period of
time. This will cause the reservoir pressure to drawdown due to the withdrawal of fluids
from the well. Afterward, the well is shut-in to allow the pressure to build up. During the
period of time, the build-up pressure reaches average reservoir pressure. This time period
depends mainly upon the reservoir permeability and thus, may result in extensive time
7
consumption in low permeability reservoirs. The resulting pressure build-up is then
pressure. Currently, the only available method with which to obtain average reservoir
because reservoir pressure changes as fluids (oil, gas and water) are released from the
Traditionally, numerical simulations and decline curve analysis have been used to predict
geological model. Once the oil field enters the production stage, many changes take place
in the reservoir and thus, the changes must be reflected in the reservoir model. Usually,
simulator programs are used to generate a model in order to simulate the reservoir. In
reservoir simulation, the simulator takes a set of reservoir parameter values as inputs and
returns a set of fluid flow information as outputs. The outputs are usually a time series
over specific period of time. This time series is then compared with the historical
production data to evaluate their match. Experts modify the input model parameters and
re-run the simulation model. This process (history matching) continues until the model
output rates match the historical production rates. History matching is the process of
8
updating reservoir descriptor parameters to reflect such changes, based on production
data. The process of history matching is an inverse problem. In this problem, a reservoir
is a black box with unknown parameter values. Given the rock, fluid properties and other
production information collected from the field, the task is to identify these unknown
parameter values such that the reservoir model gives flow outputs matching production
data. Since the inverse problem does not have a unique solution, a large number of
models with good history matching can be achieved. In other words, in the reservoir
simulation process, more than one combination of reservoir parameter input values can
give the same output. As a result, history matching is a challenging task for the following
reasons:
• Highly non-linear flow equations with a large number of assumptions and input
parameters
• Inverse problem (non-unique solution), where more than one model can produce
acceptable outputs
• Time consuming
On the other hand, decline curve analysis fits the observed production rates of
predict future production by extrapolating the declining function. The basis of decline
production continues to follow the historical trend. In some cases, production decline data
do not follow a particular model but crossover the entire set of curves. In addition,
9
decline curve analysis does not take into account opening or closing intervals and varying
• MRA, ANN and LS-SVM methods are applied in order to develop four novel
models with which to estimate the bubble point pressure (Pb), the bubble point oil
formation volume factor (Bob), bubble point solution gas/oil ratio (Rsob), and
stock-tank vent gas/oil ratio (RST). These new models can be applied in a
models are performed to interpolate the missing values, forecast and to predict
• Design a neural network and least square support machine models to predict oil
injection wells in service and varying water injection rates. The proposed models
10
1.5 Chapter by Chapter Overview
The core of this thesis is divided into nine chapters along with the Introduction
(Chapter 1) and Conclusions (Chapter 11). Each chapter deals with the application of one
type of the aforementioned function approximation techniques. Figure 1.1 illustrates the
general structure of this thesis. The content of each chapter is organized as follows:
Chapter 2 provides four multiple regression analysis models with which to estimate
bubble point pressure and the bubble point formation volume factor. The correlations can
be applied in the absence of PVT analysis to overcome the limitations faced by the
published PVT correlations. Until recently, however, there has not been one specific
correlation published to directly estimate bubble point pressure in the absence of PVT
analysis. However, one published correlation (labedi 1990) to estimate oil formation
volume factor at bubble point pressure directly in the absence of such laboratory analysis
does exist.
Chapter 3 extended from Chapter 2 and proposes another two novel empirical
correlation to estimate the bubble point solution gas/oil ratio, and the stock tank vent
gas/oil ratio. However, the majority of the available PVT empirical equations are used to
estimate the solution gas/oil ratio at various temperatures rather than the reservoir
performance. Since these correlations required prior knowledge of bubble point pressure
and gas gravity, it is difficult to apply them in the absence of PVT analysis. Bubble point
pressure and gas gravity can be obtained from either PVT analysis or can be estimated
11
Chapter 4 covers the use of ANN as an alternative technique to estimate the same
three reservoir fluid properties and the stock-tank vent gas/oil ratio that are addressed in
Chapter 2 and 3. The four ANN models are proposed as a function of readily available
analysis.
Chapter 5 describes a LS-SVM approach to design four models in order to predict the
same fluid properties mentioned in Chapters 2 and 3 which would increase accuracy and
to map the relationship which controls reservoir oil, gas and water production
performance in order to interpolate and estimate the current average reservoir pressure
without closing the producing wells. This method is suitable for constant and variable
flow rates. After interpolating, predicting and/or obtaining the current average reservoir
pressure, this method can be utilized in almost all reservoir and production engineering
studies.
Since the backpropagation network is a gradient descent method, it tends to become stuck
in the local minima of the error surface and thus, does not find the global minimum. A
backpropagation with a genetic algorithm is proposed in this chapter for tuning the link
switches and weight parameters between the input layer and hidden layer. The main
12
search space avoiding being trapped in local minima which usually happens with local
search procedures.
reservoir pressure performance. Using this approach, the current average reservoir
pressure can be interpolated and the current average reservoir pressure can be estimated
Chapter 9 clarifies how ANN, as a function approximation, has the ability to predict
reservoir oil production with impressive accuracy. In this way, the history matching
addition, unlike traditional decline curve analysis, the designed ANN model can take into
account the number of producers and injectors in service and the varying water injection
rates. The training ANN model can serve as a practical and robust reservoir production
management tool.
production. The novelty in this chapter, however, is the introduction of LS-SVM as new
alternative tool to predict and interpolate average reservoir pressure without closing the
producing wells.
regression model.
13
• Appendix C addresses the basic steps of applying a LS-SVM regression
model.
model’s accuracy.
1.6 References
Chen, Z., Huan, G. and Ma, Yu., 2006. Computational Methods for Multiphase Flows in
Porous Media. Society for Industrial and Applied Mathematics, Dallas, Texas
Labedi, R.M., 1990. Use of Production Data to Estimate the Saturation Pressure, Solution
GOR, and Chemical Composition of Reservoir Fluids. SPE 21164 presented at the
SPE Latin American Petroleum Conference held in Rio de Janeiro, October 14-19.
14
Figure 1.1: Structure of the thesis. Arrows suggest the reading order of the chapters. The
main contributions are indicated in dashed box.
15
CHAPTER II
ANALYSIS
2.1 Introduction
The physical and chemical properties of reservoir oils vary considerably and are
dependent upon the concentration of the various types of hydrocarbons and minor
constituents present in the reservoir. Reservoir fluid studies are essentially based upon
laboratory procedures designed to provide the values of the reservoir fluid properties as
required in material balance calculations, well test analysis, reserve estimates, inflow
Ideally, reservoir fluid properties should be determined from laboratory studies with
respect to live oil samples collected from the bottom of the wellbore or from the surface.
Standard reservoir PVT fluid studies are designed to simulate the simultaneous fluid flow
of oil and gas from the reservoir to the surface. The production path of reservoir fluids
from the reservoir to the surface is simulated in the laboratory at reservoir temperature.
During this process, the bubble point pressure Pb is measured. Likewise, the oil volumes
16
and the amount of gas released are measured and used to determine the oil formation
volume factor (oil FVF) Bo, the gas formation volume factor (gas FVF) Bg, solution
gas/oil ratio Rs, oil compressibility Co, oil density ρo, gas specific gravity γg, and gas
reservoir fluid properties at temperatures other than the reservoir temperature. However,
the PVT analysis is usually carried out at the reservoir temperature only. Reservoir fluid
properties, at various temperatures other than the reservoir temperature, are required to
operations. In such cases, even though PVT analysis is available, the required reservoir
In the absence of such experimental analysis, empirical PVT correlations can be used
to estimate reservoir fluid properties. Reasons for using empirical PVT correlations could
be:
• economic issues
laboratory analysis
17
Figure 2.1− Fluid flow of Oil and Gas from the Reservoir to the Separator and Stock-
Tank (Two-Stage Separation)
18
Until recently, not one specific correlation has been published to directly estimate
bubble point pressure in the absence of PVT analysis and, at the moment, there is just one
published correlation available with which to estimate the bubble point oil FVF directly
in the absence of PVT analysis (Labedi, 1990). The majority of the published bubble
point pressure and bubble point oil FVF correlations cannot be applied directly because
the correlations require prior knowledge of the bubble point solution GOR and gas
specific gravity as part of the input variables. Both of which are rarely measured field
parameters. Solution GOR and gas specific gravity can be obtained either experimentally
applied in order to develop two novel correlations with which to estimate the bubble
point pressure and the bubble point oil FVF. These new correlations can be applied in a
straightforward manner by using direct field data. Separator GOR, separator pressure,
stock-tank oil gravity and reservoir temperature are the only key parameters required to
predict bubble point pressure and bubble point oil FVF. Additional correlations or
PVT correlations are ultimately required in order to estimate reservoir fluid properties
at temperatures other than the reservoir temperature. This is due to reservoir fluid
properties being found in the PVT analysis report at the reservoir temperature only.
Reservoir fluid properties, at various temperatures other than the reservoir temperature,
are required to design surface operation equipment and to study reservoir inflow
19
performance operations. In such cases, even though PVT analysis is available, the
Several correlations already exist within the oil and gas industry to estimate bubble
point pressures (Pb) and bubble point oil FVF (Bob) of reservoir oils. The correlations are
essentially based on the assumption that Pb and Bob are strong functions of bubble point
solution GOR (Rsob), reservoir temperature (TR), gas specific gravity (γg), and stock-tank
oil specific gravity (γoST). Bubble point solution GOR can be obtained as the sum of the
stock-tank vent GOR (RST) (seldom conducted field measurement) and the measured
separator GOR (RSP). This is valid only if the RSP and RST are measured while the
reservoir pressure is above the bubble point pressure. Sometimes, the sum of the two
producing GORs is called “flash bubble point solution GOR” (RsFb) or “total GOR”,
Equation 2.1. Some correlations use differential bubble point solution GOR (RsDb) rather
than RsFb. Examples include the correlations developed by Borden and Rzasa (1950),
Knopp and Ramsey (1960), Vasques and Beggs (1980), Al-Marhoun (1988), Dokla and
Osman (1992), Elsharkawy and Alikhan (1997), Almehaideb (1997), Hanafy et al.
(1997), McCain et al. (1998), Velarde et al. (1999), Boukadi et al. (2002), Gharbi and
Elsharkawy (2003) and Mazandarani and Asghari (2007). Others preferred to use flash
bubble point solution GOR (RsFb) e.g. Standing (1947), Lasater (1958), Tahrani (1967),
Glaso (1980), Macary and Batanony (1992), Al-Marhoun (1992), Frashad et al. (1996),
20
Moreover, several correlations developed by Labedi (1990), Rollins et al. (1990),
Dokla and Osman (1992), Macary and Batanony (1992), Velarde et al. (1999), Petrosky
and Farshad (1998) and McCain et al.(1998) use flash separator gas specific gravity
(γgSP), while still others such as Standing (1947), Borden and Rzasa (1950), Lasater
(1958), Elsharkawy and Alikhan (1997), Glaso (1980), Mazandarani and Asghari (2007)
use total gas specific gravity. Other correlations use weight average specific gas gravity
of the separator and stock-tank vent gas. Among the researchers are Al-Marhoun (1988),
Frashad et al. (1996), Al-Marhoun (1997), Al-shammasi (1999), Hemmati and Kharrat
(2007) and Ikiensikimama and Oboja (2009). Initially, some methods require adjusting
gas gravity to separator pressure at 100 psig before they can be used in the correlations
e.g. Vasquez and Beggs (1980). Others require compositional analysis as input variables
Accordingly, in order to apply bubble point pressure and bubble point oil FVF
correlations in the absence of PVT analysis, there must be prior knowledge of the bubble
point solution GOR (Rsob), differential bubble point GOR (RsDb) or stock-tank vent GOR
(RST) in addition to separator gas specific gravity (γgSP), total gas specific gravity (γgTotal),
or weight average specific gas gravity. The aforementioned parameters are rarely
measured using field data. A literature survey has shown little attention is paid to
predicting reservoir fluid properties in the absence of PVT analysis. Currently, not one
specific correlation has been published to directly estimate the bubble point pressure in
the absence of PVT analysis and, at the moment, there is just one published correlation
21
available to estimate the bubble point oil FVF (Labedi, 1990) directly in the absence of
PVT analysis.
Labedi (1990) proposed a bubble point pressure correlation (Equation 2.2) based
upon separator GOR (RSP), separator temperature (TSP), separator gas specific gravity
(γgSP), stock-tank API gravity and reservoir temperature (TR). Obviously, Labedi’s bubble
point pressure correlation cannot be used directly because of the separator gas gravity. To
work around the problem, Labedi proposed another correlation to estimate separator gas
gravity (Equation 2.3). In the same work, Labedi proposed a new correlation to estimate
bubble point oil FVF (Equation 2.4). Presently, this is the only existing correlation in the
literature which can be applied directly to obtain Bob in the absence of PVT analysis.
6 .0 R SP
0.6714
(T R / API ) 0.7097 TSP0.0893
pb = (2.2)
γ gSP 10 0.00007995 RSP
0.0659
T (2.3)
γ gSP = 0.0755 * 101.1938 / PSP
API 0.0621 SP
R SP
estimate bubble point pressure and bubble point oil FVF, he suggested Sanding (1977)
22
correlations (Equations 2.5 and 2.6) in conjunction with Rollins et al. (1990) stock-tank
0.83
R (10)0.00091T R
pb = 18.2 sb − 1 .4 (2.5)
γ (10 )0.0125 API
gTotal
1.2
γ gTotal
0.5
Bob=0.9759+12*10-5 Rs + 1.25T (2.6)
γ oST
LogRST = 0.3818–5.506 logγoST + 2.902 log γgSP +1.327 logPSP – 0.7355 logTSP (2.7)
In short, McCain recommended the following steps to estimate bubble point pressure
and bubble point oil FVF with the limitation of the separator temperature not exceeding
2. add the estimated RST to field separator GOR (Equation 2.1) to obtain a bubble
3. estimate the Pb and Bob from Standing correlations (Equations 2.5 and 2.6,
respectively). That is, use the Rsob obtained from step 2, and the γgSP instead of
23
Nevertheless, this approach is impractical because the γgSP is a rarely measured field
data and is usually based upon a gas composition analysis. In 2003, Valko and McCain
revisited reservoir oil bubble point pressure. They derived new correlations to estimate
bubble point pressure (Equation 2.8) and stock-tank vent GOR (Equation 2.9). The
bubble point correlation is a function of Rsob, API, γgSP and TR. The stock-tank vent GOR
4
where z = ∑ z n and
n =1
n VAR C0 C1 C2 C3
1 lnRsob -5.48 -0.0378 0.281 -0.0206
2 API 1.27 -0.0449 4.36*10-4 -4.76*10-6
3 γgSP 4.51 -10.84 8.39 -2.34
4 TR -0.7835 6.23*10-3 -1.22*10-5 1.03*10-8
3
where z = ∑ z n and
n =1
n VAR C0 C1 C2
1 lnPSP -8.005 2.7 -0.161
2 lnTSP 1.224 -0.5 0
3 API -1.587 0.0441 -2.29*10-5
24
Just as with the McCain (1991) approach, the work of Valko and McCain (2003) has
The main objective of this chapter is to overcome the limitations faced by previous
input variables in order to estimate Pb and Bob. Two correlations are proposed as
functions of four readily available field parameters (RSP, PSP, γoST and TR). By employing
the four parameters, engineers can estimate the bubble point pressure and the bubble
point oil FVF for crude oil straightforwardly in the absence of PVT analysis.
The PVT data used in this work were obtained from PVT analysis at two-stage and
single-stage flash separation tests. A total of 118 reservoir fluid studies (476 data points)
were collected from various Libyan oil fields in the Sirte Basin (AGOCO, Sirte Oil,
Waha Oil, Millita Oil and Gas, Repsol Oil, and Harouge Oil, 1971-1995). The majority
of the data points are taken from two-stage flash separation tests (355 data points). In the
single-stage separation test, the separator pressure is atmospheric pressure and the stock-
In order to study the validity of the proposed correlations, the 476 data points were
randomly divided into two groups. Group A includes a total of 413 data points. Group B
data (62 data points) was used to test the validity of the newly developed correlations.
25
2.3.1 Bubble point pressure correlation
Numerous models were tried as regression equations. Equation 2.10 was found to be
very accurate. The natural logarithm of bubble point pressure was regressed against the
natural logarithms of separator GOR, separator pressure, stock-tank oil gravity and
reservoir temperature.
conditions, the first and most important operation, before running a complete reservoir
fluid study, is to check the sample validity and select the most representative sample. The
bottom-hole sample, used for PVT study, is selected according to the results obtained
during the verification of the samples. It is worth mentioning that the proposed bubble
point pressure correlation can be used to select the most representative sample with
26
Table 2.1 - Range of Data in Group A
27
2.3.2 Bubble Point Oil FVF Correlation
Usually, the oil FVF obtained from a differential vaporization test should be adjusted
using flash separation oil FVF to properly approximate a combination liberation system.
However, at bubble point pressure, oil FVF (Bob) is equal to flash separation oil FVF
(BoFb). Accordingly, by using a multiple regression analysis technique, the Bob was
correlated as a function of PSP, RSP, γoST and TR. After trying many models, the following
model was found to be a very good prediction equation of bubble point oil FVF.
Verification and validation of the correlation are the most important steps in the
development process. Both quantitative and graphical analyses of the residuals are used
minimum and maximum of the Error (E), Relative Error (RE) and Absolute Relative
Error (ARE). Table 2.2 summarizes the quantitative statistical error analysis for the
proposed correlations. The analysis shows a small error and a small standard deviation
Following the estimation of a regression model, the graphical error analysis was
carried out by analyzing the residuals. The residual distributions for the bubble point
pressure and the bubble point oil FVF correlations are presented in Figures 2.2 and 2.3,
28
respectively. Both figures show that the error is normally distributed, and it can be
concluded that both correlations satisfy the normality assumption. Figures 2.4, and 2.5
display the computed values from the regression models versus the experimental values.
It is evident from both figures that the points are scattered around the y = x line.
29
Table 2.2 - Quantitative Statistical Error Analysis
30
Histogram of the Residual
Bubble Point Pressure Correlation
200
Mean 17.72
SD 435.6
150
Frequency
100
50
0
-1500 -1000 -500 0 500 1000 1500
Residual
31
Histogram Plot of the Residuals
Bubble Point oil FVF correlation
80
Mean -3.13249E-15
SD 0.02870
70
60
50
Frequency
40
30
20
10
0
-0.09 -0.06 -0.03 0.00 0.03 0.06 0.09
Residual
32
Scatter Diagram
Bubble Point Pressure Correlation
5000
Estimated Bubble Point, psia
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Actual Bubble Point Pressure, psia
Figure 2.4 −A 45o Straight Line Cross Plot for Bubble Point Pressure Correlation
33
Scatter Diagram
Bubblepoint oil FVF Correlation
1.8
1.7
1.6
1.5
Fitted Value
1.4
1.3
1.2
1.1
1.0
1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Experimental Bo Values, bbl/STB
Figure 2.5 − A 45o Straight Line Cross Plot for Bubble point Oil FVF Correlation
34
2.5 Comparison with Other Correlations
As mentioned previously, not one specific correlation has been published so far to
directly estimate bubble point pressure in the absence of PVT analysis. Earlier
correlations first require estimates of Rsob and γgSP (obtained from secondary correlations
or from experiments) before they can be implemented. As a result, none of the published
bubble point correlations could be compared with the proposed bubble point correlation
in this chapter. However, the proposed Bob correlation was subjected to evaluation and
validation. Its accuracy was tested solely against Labedi’s correlation because Labedi’s
bubble point oil FVF is presently the only published correlation available in the literature
to estimate bubble point oil FVF directly in the absence of PVT analysis. The Group B
data set (62 data points), as described in Table 2.3, was used in this test. However, the
data points were not switched in the model derivation process. SD, average, minimum
and maximum values of E, RE and ARE with a SD and 45 degree line cross plot were
used as comparative criteria. Figure 2.6 compares the behaviour of the proposed Bob
regression model to Labedi (1990). The Figure shows that the produced model provides
further reliable results. The majority of the points, estimated using the proposed
correlation, fall very close to the 45 degree line with less RE, less average RE and less
average ARE. Table 2.4 demonstrates the statistical analysis of this comparison.
35
Table 2.3 - Data Range of Group B
36
Table 2.4 - Comparison of Proposed Bubble point Oil FVF Correlation
37
Evaluation and Validation of Bubblepoint Oil FVF Correlation
1.0 1.2 1.4 1.6 1.8
This Study Labedi Bob
Estimated Bubblepoint Oil FVF, bbl/STB
1.8 1.8
1.7 1.7
1.6 1.6
1.5 1.5
1.4 1.4
1.3 1.3
1.2 1.2
1.1 1.1
1.0 1.0
1.0 1.2 1.4 1.6 1.8
Actual bubblepoint oil FVF, bbl/STB
Figure 2.6 − Evaluation and Validation of Bubble Point Oil FVF Correlation
38
2.6 Summary and Conclusions
• A review of the literature shows little attention has been paid to the prediction of
reservoir fluid properties in the absence of PVT analysis. Currently, not one specific
correlation has been published to directly estimate the bubble point pressure in the
absence of PVT analysis and, presently, there is just one published correlation (Labedi
1990) to estimate bubble point oil FVF directly in the absence of PVT analysis.
• The majority of the available PVT correlations are used to estimate the reservoir
oil properties at various temperatures rather than reservoir temperature in order to design
surface operation equipment and to study reservoir inflow performance. However, due to
the fact the correlations require prior knowledge of parameters such as solution GOR and
specific gas gravity; it is difficult to apply them in the absence of PVT analysis. Since
these parameters are rarely measured in the field, this study proposed correlations which
can be applied straightforwardly in the absence of PVT analysis. Thus, there is no need
for additional correlations. The only required input parameters are separator GOR,
The proposed correlations were developed based on a multiple regression analysis and
118 reservoir fluid studies (476 two-stage and single-stage flash separator tests). In the
single-stage separation test, the separator pressure is atmospheric pressure and the stock-
tank vent GOR value is equal to zero as per samples collected from various Libyan oil
order to verify the accuracy of the proposed correlations. The newly proposed bubble
39
point pressure correlation provides prediction values with an average percent relative
error of 2.83%, an absolute average absolute percent relative error of 16.757% and a
correlation coefficient of 95.67%. The proposed bubble point oil FVF correlation
provides prediction values, with an average percent relative error of 0.038%, an absolute
average percent relative error of 1.6874% and a correlation coefficient of 96.3%. The
residuals of both correlations are normally distributed, indicating that both models
• None of the previously published bubble point correlations were compared to the
proposed bubble point correlation in this study. This is because prior knowledge of some
of the data which are rarely measured in the field (bubble point solution GOR and gas
• Of the 476 data points used in the development of the new correlations, 62 were
used to validate and evaluate the accuracy of the proposed bubble point oil FVF
correlation. Its accuracy was tested against Labedi’s correlation. The newly proposed oil
FVF correlation provides better predictions and higher accuracy than Labedi’s
correlation.
conditions, the first and most important operation, before running a complete reservoir
fluid study, is to check the validity of the samples. The bottom-hole sample, used for the
PVT study, is selected according to the results obtained during the verification of sample
validity. It is worth mentioning, the proposed bubble point pressure correlation can be
used to select the most representative sample with which to run a PVT analysis.
40
2.7 References
AGOCO, Sirte Oil, Waha Oil, Millita Oil and Gas, Repsol Oil, and Harouge Oil
Companies (1971-1995). Reservoir Fluid Studies (PVT Analysis).
Al-Marhoun, M., 1988. PVT correlations for Middle East Crude Oils. JPT, 650–666.
Al-Marhoun, M.A., 1992. New correlations for formation volume factors of oil and gas
mixtures. J. Can. Pet. Technol. 31 3, pp. 22–26.
Al-Shammasi, A.A., 1999. Bubble point pressure and oil formation volume factor
correlations. Paper SPE 53185 Presented at the SPE Middle East Oil Show, Bahrain,
Feb. 20–23.
Borden, G. and Rzasa, M.J., 1950. Correlation of Bottom Hole Sample Data. Trans.
AIME (Am. Inst. Min. Metall. Eng.), 189; 345-348.
Boukadi, F. H., Bemani, A. S., Hashmi, A., 2002. PVT Emperical Models for Saturated
Omani Crude Oils. Journal of Petroleum Science and Technology, 20(1&2), 89–100.
Dokla, M.E. and Osman, m.E., 1992. Correlation of PVT Properties for UAE Crudes.
SPE Formation Evaluation March 7 pp 41-46.
Elsharkawy, Adel M. 2003. An empirical model for estimating the saturation pressures of
crude oils. Journal of Petroleum Engineering and Science, 38 PP-57-77.
Elsherkawy, A.M. and Alikhan, A.A., 1997. Correlation for Predicting Solution Gas/Oil
Ratio, Formation Volume Factore, and Undersaturated Oil Compressibility.
J.Petroleum Science and Engineering; 17 pp. 291-302.
Frashad, F., LeBlanc, J.L., Gruber, J.D., Osorio, J.G., 1996. Empirical PVT correlations
for Colombian crude oils. Paper SPE 36105 Presented at the Fourth Latin American
and Caribbean Petroleum Engineering Conference, Port-of-Spain, 23– 26Apr.
Gharbi, R. and Elsharkawy, A.A., 2003. Predicting the Bubble-Point Pressure and
Formation-Volume-Factor of Worldwide Crude Oil Systems. Petroleum Science and
Technology. Vol. 21, Nos. 1 & 2, pp. 53–79.
Glaso, O., 1980. Generalized pressure–volume temperature correlations. JPT May, 785.
Hanafy, H.H., Macary, S.M., ElNady, Y.M., Bayomi, A.A. and El Batanony, M.H., 1997.
Empirical PVT correlations applied to Egyptian crude oils exemplify significance of
41
using regional correlations. SPE 37295 presented at the 1997 SPE international
symposium on oilfield chemistry held in Houston, Texas, 18-20 February.
Hemmati, M.N. and Kharrat, R., 2007. A correlation approach for prediction of crude-oil
PVT properties. SPE 104543 presented at the 15th SPE Middle East show and
conference hold in Bahrain 11-14 March.
Ikiensikimama,S.S. and Ogboja, O., 2009. New Bubble Point Pressure Empirical PVT
Correlation. SPE 128893-MS presented at Nigeria Annual International Conference
and Exhibition, 3-5 August, Abuja, Nigeria.
Knopp, C.R. and Ramsey, L.A., 1960. Correlation for Oil Formation Volume Factor and
Solution Gas-Oil Ratio. JPT (Aug.), pp. 27-29.
Labedi, R.M., 1990. Use of Production Data to Estimate the Saturation Pressure, Solution
GOR, and Chemical Composition of Reservoir Fluids. SPE 21164 presented at the
SPE Latin American Petroleum Conference held in Rio de Janeiro, October 14-19.
Lasater, J.A., 1958. Bubble Point Pressure Correlation. Trans. AIME 231, pp 379-381.
Macary, S.M., El-Batanoney, M.H., 1992. Derivation of PVT correlations for the Gulf of
Suez crude oils. EGPC 11th Pet. Exp. and Prod. Conference.
Mazandarani, M.T. and Asghari, S.M., 2007. Correlations for predicting solution gas-oil
ratio, bubble point pressure and oil formation volume factor at bubble point of Iran
crude oils. European Congress of Chemical Engineering (ECCE-6) Copenhagen,
(Sep.) 16-20
McCain, W.D., Jr. 1991. Reservoir Fluid Properties Correlations-State of the Art. SPE
Reservoir Engineering, 266-270
Petrosky, G.E., Farshad, F.F., 1998. Pressure–volume–temperature correlations for Gulf
of Mexico crude oils. SPEREE (Oct.), 416– 420.
Rollins, J.B., McCain Jr., W.D., Creager, J.T., 1990. Estimation of the solution GOR of
black oils. JPT (Jan.), 92– 94.
Standing, M.B., 1 977. Volumetric and phase behaviour of oil field hydrocarbon systems.
SPE, Richardson, TX 124.
42
Standing, M.B., 1947. A pressure-volume-temperature correlation for mixtures of
California oils and gases. Drilling and Production Practice, API, pp. 275-287
Tehrani, H.D., 1968. Bubble point pressure correlation for reservoir fluids of Southwest
Iran. Second AIME reginal technical symposium, Saudi Arabia, 27-29 March.
Valko, P.P, McCain, W.D., 2003 Reservoir Oil Bubble point Pressures Revisited;
Solution Gas-Oil Ratios and Surface Gas Specific Gravities, Journal of Petroleum
Engineering and Science, 37 PP-153-169.
Vasquez, M. and Biggs, H.D., 1980. Correlation for Fluid Physical Property Prediction.
JPT June 32, pp 968-970.
Velarde, J., Blasingame, T.A., McCain Jr., W.D., 1999. Correlation of black oil
properties at pressures below bubble point pressure—a new approach. J. Can. Pet.
Technol., Spec. Ed. 38(13), 62–68.
43
CHAPTER III
3.1 Introduction
simulate the simultaneous fluid flow of oil and gas from the reservoir to the surface. This
process is illustrated in Figure 2.1 with notation of the used variables to enrich the
published correlations with which to estimate the bubble point pressure and bubble point
oil formation volume factor of reservoir oils require the value of the bubble point solution
additional correlation in order to estimate this value. As well, the majority of published
bubble point solution gas/oil ratio correlations are functions of bubble point pressure and
PVT analysis) or estimated from the correlations. Thus, it is difficult to apply the
A bubble point solution gas/oil ratio (Rsob) is a very important parameter in reservoir
and production engineering calculations. The solution gas/oil ratio is the amount of gas
44
dissolved in the oil at any pressure per unit volume of oil both measured at standard
conditions. The solution gas/oil ratio increases with pressure until the bubble point
pressure is reached, after which it is a constant and is said to be a bubble point solution
gas/oil ratio. A bubble point pressure (Pb) for reservoir oil is defined as the pressure at
which gas first begins to exit a solution at a constant temperature. The important
a basic parameter in many PVT correlations in order to estimate other fluid properties
such as bubble point pressure, Pb and the bubble point oil formation volume factor (Bob).
Numerous correlations, offered in the oil and gas industry, to estimate reservoir oil Pb and
Bob require a value of Rsob as an input variable. On the other hand, several correlations
presently exist to obtain the Rsob. Examples include Standing (1947), Vasques and Beggs
(1980), Petrosky and Farshad (1993), Frashad et al. (1996), Elsharkawy and Alikhan
(1997), Boukadi et al. (2002), and most recently, Hemmati and Kharrat (2007),
Mazandarani and Asghari (2007) and Khamehchi et al. (2009). The basic assumption of
The correlations are essentially based on the assumption that Rsob is a strong function
of bubble point pressure (Pb), gas specific gravity (γg), stock-tank oil API gravity and
correlations in the absence of PVT analysis, there must be previous knowledge of the Pb
45
and γg. Parameters Pb and γg are rarely measured as field data and therefore, engineers
correlations are essentially based on the relationship illustrated in Equation (3.2) which
includes Standing (1947), Lasater (1958), Vazquez and Beggs (1980), Glaso (1980), Al-
Marhun (1988), McCain (1990), Dokla and Osman (1992), Macary and El- Batanoney
(1992), Petrosky and Farshad (1993), Omar and Todd (1993), De Ghetto and Villa
(1994), Kartoatmodjo and Schmidt (1994), Frashad et al. (1996), Almehaideb (1997),
Hanafy et al. (1997), Boukadi et al. (1999), Velarde et al. (1999), Al-Shammasi (1999),
Dindoruk and Christman (2001), Mehran et al. (2006), and most recently, Hemmati and
Kharrat (2007), Ikiensikimama and Ogboja (2009) and Moradi et al. (2010). However,
the foresaid correlations are ultimately required in order to estimate Pb and Rsob at various
temperatures other than TR in order to design surface and subsurface facilities. This is due
to reservoir fluid properties being found in the PVT analysis reports at TR only. In such
cases, even though PVT analysis is available, the required reservoir fluid properties must
A review of the literature indicates little attention has been given to the prediction of
reservoir fluid properties in the absence of PVT analysis. However, in most cases, Rsob
46
correlations are derived by mathematically solving the Pb correlation for a Rsob. An
exception to this precedent are the correlations presented by Vazquez and Beggs (1980),
Elsharkawy and Alikhan (1997), Boukadi et al. (2002), Hemmati and Kharrat (2007),
Mazandarani and Asghari (2007) and Khamehchi et al. (2009), where the Rsob was treated
available Rsob correlations are impractical in the absence of PVT analysis, simply because
Rsob correlations require knowledge of the Pb and γg. As a result, it is difficult to apply the
Furthermore, the Rsob can be obtained indirectly by adding the estimated stock-tank
vent gas/oil ratio, RST, to the field measured separator gas/oil ratio, RSP, as shown in
Equation (3.3). This is only valid if the reservoir pressure is above the Pb. Accordingly,
Rollins et al. (1990) developed a correlation (Equation (3.4)) to estimate the RST as a
function of γoST, γgSP, PSP and TSP. However, this work has a practical limitation in the
absence of PVT analysis due to the γgSP which is rarely measured in the field and it is
ln RST = 0.2126 − 4.916 lnγ oST + 3.469 ln γ gSP + 1.501 ln pSP − 0.9213 ln TSP (3.4)
47
In 2003, Valko and McCain derived a correlation to estimate the RST as a function of
PSP, TSP and API. In applying a non-parametric regression analysis method, with 881 data
3
where, z = ∑ z n ,
n =1
Valko and McCain estimated the Rsob indirectly using Equation (3.3) and reported an
correlation (Equation (3.6)) to estimate the Rsob directly when the separator conditions are
unknown. In estimating the Rsob, they reported an average RE of 0.0% and an average
ARE of 9.9%.
This study presents two multiple regression analysis models to predict the Rsob in the
absence of PVT analysis. Both correlations are proposed as a function of readily available
48
analysis. An additional and important application of the proposed stock-tank vent gas/oil
A total of 480 separator tests, measured at the Pb, were obtained from 118 PVT
reports collected from various Libyan oil reservoirs located in the Sirte Basin. The
majority of the data points were taken from a two-stage separation test. Normally, the
separation flash test is carried out for two stages of separation (Figure 2.1). The volume
of gas liberated at each stage and the volume of remaining liquid are measured. The first-
stage separator pressure is generally varied to include at least four possible separator
pressures at ambient temperature, and the second stage is generally carried out under
stock tank conditions of 14.7 psia (1 atmosphere) at ambient temperature. The stock tank
the test is carried out under separator conditions of 14.7 psia at ambient temperature. In
Prior to the regression analysis, the collected datasets were divided randomly into two
sets. The first set of 431 data points was used to construct the regression models, while
the second set of 49 data points was used to test the validity and compare the regression
models against similar available correlations (Valko and McCain 2003). The first dataset
was used to generate the Rsob correlation. The range of this data is presented in Table 3.1.
Whereas, only 355 two-stage flash separation tests out of the 431 separator tests were
used to generate the RST empirical correlation, since data from the single-stage flash
49
3.3.2 Bubble Point Solution Gas/Oil ratio, Rsob Correlation
A novel correlation to estimate the Rsob was developed during this study. It was
correlated in terms of the RSP, PSP and γoST after the forward-backward search algorithm
is conducted to select the best subset input variables. Numerous models were attempted
as regression equations but the following Equation (3.7) was found to be the most
accurate:
Rsob = −152 + 0.99 RSP + 0.256 PSP + 19.543 lnPSP – 557.2 lnγoST (3.7)
The model shows a small average RE of 1.2% and an average ARE of 6.8% with a
high R2 of 99.3%. Figure 3.1 shows the histogram plot of the residuals indicating the
error terms follow a normal distribution with a mean almost equal to zero, and they
satisfy the normality assumption. The estimated Rsob was plotted as a function of the
actual values in Figure 3.2. The cross-plot shows the precision of the empirical
correlation without bias. Table 3.3 summarizes the results of the quantitative analysis of
the residuals.
50
Table 3.1− Range of Data Points for Rsob Correlation (431 Data Points)
51
Table 3.2− Range of Data points for RST Correlation (355 Data Points)
52
3.3.3 Stock-Tank Vent Gas/Oil Ratio, RST Correlation
Usually, the RSP is a field measurement, while the RST is rarely measured in the field.
Thus, in order to obtain the Rsob indirectly, we must first estimate the RST, and add it to
the field measured RSP. In this work, a novel correlation to estimate the RST was
developed. PSP, TSP and γoST are the only independent variables. A total of 355 two-stage
separator tests were employed to develop the RST correlation. The range of the dataset is
illustrated in Table 2.2. The natural logarithm of the RST was regressed as a function of
the natural logarithms of PSP, TSP and γoST and the following empirical Equation (3.8) was
obtained:
The model shows small values of average RE and average ARE of 1.1% and 7.6%
respectively, with a high R2 of 99.2%, indicating the model is sufficient to describe the
data.
53
Table 3.3 − Quantitative Statistical Analysis of the Present Correlations
54
Histogram Plot of the Residuals
Solution GOR Correlation
140
Mean -1.98419E-13
120 SD 21.57
100
Frequency
80
60
40
20
0
-60 -30 0 30 60
Residual
Figure 3.1- Residual Distributions of Rsob Obtained from Eq. (3.7) (431 Data Points)
55
Scatter Diagram
Solution GOR correlation
1400
1200
Estimated Rsb, SCF/STB
1000
800
600
400
200
0
0 200 400 600 800 1000 1200 1400
Actual Rsb, SCF/STB
Figure 3.2 - Estimated Rsob from Eq. (3.7) vs. Actual Rsob (431 Data Points)
56
3.4 Comparison with Other Correlations
In order to study the accuracy of the proposed correlations, the estimated Rsob values
from Equation (3.7) and those obtained from Equation (3.8), using 355 data points
described in Table 2.2, were compared with Rsob values calculated from Valko and
McCain correlations, Equations (3.5) and (2.6). This comparison is illustrated in Figure
3.3. The cross-plot shows the majority of the plotted data points, estimated using our
proposed correlations, fall on the y=x line, indicating the two proposed models have a
higher accuracy than the Valko and McCain correlations and provide far more reliable
results. Moreover, the histogram of the relative error of the proposed correlations, along
with the relative error from the Valko and McCain correlations, is plotted in Figure 3.4.
The error terms from Equations (3.5), (3.7) and (3.8) satisfy the normality assumption
with a mean equal to −1.958, 1.261 and 0.1568, respectively. However, the relative errors
of the Rsob values, obtained from Equation (3.6), deviate and do not meet the normal
distribution assumption. Table 3.4 summarizes the statistical results of this comparison. It
is evident the two proposed correlations give less average RE and average ARE than the
Valko and McCain correlations. However, the two proposed correlations have
57
Table 3.4− Statistical Analysis of Evaluation Test (355 Data Points)
58
Evaluation Test; Scatter Diagram
0 300 600 900 1200
This Study Eq (3.7) This Study Eq (3.8)
1200
900
Experimental solution GOR
600
300
0
Valko & McCain, Eq. (3.6) Valko & McCain, Eq (3.5)
1200
900
600
300
0
0 300 600 900 1200
Estimated solution GOR
Figure 3.3 − Evaluation Test; Estimated vs. Actual Rsob (355 Data Points)
59
Histogram of the Percent Relative Error ; Evaluation Test
This Study Eq 3.7 This Study Eq 3.8
160
100
120
75 This Study Eq 3.7
80 Mean 1.261
50
SD 11.09
0 0 SD 10.51
-30 -15 0 15 30 45 -30 -20 -10 0 10 20 30 Eq 3.6
Eq 3.6 Eq 3.5 Mean 13.31
150 160 StDev 29.59
Eq 3.5
120 Mean -1.958
100
SD 11.27
80
50
40
0 0
-40 0 40 80 120 160 200 -30 -15 0 15 30
Figure 3.4 − Evaluation Test; Relative Error Distributions (355 Data Points)
60
Moreover, a validation test was carried out using the 49 two-stage flash separation
tests described in Table 3.5, against the Valko and McCain correlations. The data points
were not switched in the process of correlation derivation. Figure 3.5 illustrates the
behaviour of Rsob, estimated from the two proposed correlations against Rsob, as estimated
from the Valko and McCain correlations. The cross-plot of the estimated values versus
experimental values shows the majority of the plotted data points, estimated using our
proposed models, fall on the 45 degree line. Our proposed models, as explained by
Equations (3.7) and (3.8), show a high R2 of 99.61% and 99.97%, respectively. However,
the Valko and McCain modes, expressed by Equations (3.5) and (3.6) show a R2 of
99.27% and 83.49%, respectively. Furthermore, the histogram in Figure 3.6 shows the
distribution of the relative errors obtained from Equations (3.5), (3.6), (3.7) and (3.8).
The error terms from Equation (3.5) deviate and skew to the right with a mean equal to
−5.474, while the error terms from Equation (3.6) are not normally distributed. However,
the error terms from Equations (3.7) and (3.8) are satisfying the normality assumption
with a mean equal to 2.675 and 0.581, respectively. The statistical results of this
validation test are illustrated in Table 3.6. Obviously, the calculated Rsob, obtained from
the two proposed empirical correlations, gives less average RE and average ARE,
indicating our proposed correlations have a higher accuracy than the Valko and McCain
61
Table 3.5 − Range of Data used for Validation Test (49 Data Points)
62
Table 3.6−Statistical Accuracy of Validation Test (49 Data Points)
63
Validation Test; Scatte Diagram
0 300 600 900 1200
This Study Eq 3.7 This Study Eq 3.8
1200 1200
900 900
Experimental solution GOR
600 600
300 300
0 0
Valko & MaCain Eq 3.6 Valko @ McCain Eq 3.5
1200 1200
900 900
600 600
300 300
0 0
0 300 600 900 1200 Estimated solution GOR
Figure 3.5 − Validation Test; Actual vs. Estimated Rsob (49 Data Points)
64
Validation Test; Histogram of the Percent Relative Error
This Study Eq (3.7) This Study Eq (3.8)
40
48
30
36
This Study Eq (3.7)
20 24 Mean -1.273
StDev 5.378
10 12 This Study Eq (3.8)
Frequency
Mean -0.7015
0 0 SD 6.313
-20 -15 -10 -5 0 5 10 -24 -12 0 12 24
Valko & McC ain Eq (3.6)
Valko & McCain Eq (3.6) Valko & McCain Eq (3.5) Mean 1.191
20 SD 22.26
48
Valko & McC ain Eq (3.5)
15 36 Mean 1.601
SD 6.670
10 24
5 12
0 0
-40 -20 0 20 40 60 -10 0 10 20 30
Figure 3.6− Validation Test; Relative Error Distributions (49 Data Points)
65
3.5 Conclusions
analysis to estimate Rsob and RST in the absence of PVT analysis, were proposed
• The proposed bubble point solution gas/oil ratio correlation was developed based
on 480 two-stage and single stage flash separation tests. The majority of the tests
are taken from a two-stage flash separation test. In the single-stage separation test,
the separator pressure is atmospheric pressure and the stock-tank vent gas/oil ratio
is equal to zero. However, the proposed stock-tank vent gas/oil ratio correlation
• Both correlations were developed specifically for Libyan oil reservoirs located in
Sirte Basin. However, they may be considered for use in other reservoirs around
the world.
the accuracy of the proposed correlations. The present Rsob correlation provides
calculated values with an average RE of 1.2% and an average ARE of 6.8% with
of 1.1% and an average ARE of 7.6% with a R2 of 99.2%. The residuals of both
66
presented correlations satisfy the normal distribution assumption and describe the
data well.
correlations from the literature indicate both presented correlations render a better
approximation.
• The two proposed correlations achieve approximately the same accuracy when
• The estimated Rsob values from the proposed correlations can be used as a basic
other fluid properties such as the Pb and Bob at various temperature rather than
reservoir temperature.
67
3.6 References
Al-Marhun, M.A., 1988. PVT correlations for Middle East crude oils. JPT (May) pp.
650-666.
Almehaideb, R.A., 1997. Improved PVT correlations for UAE crude oils. Paper SPE
37691 Presented at the Middle East Oil Conference and Exhibition, Bahrein, Mar. 17-
20.
Al-Shammasi, A.A., 1999. Bubble point pressure and oil formation volume factor
correlations. Paper SPE 53185 Presented at the SPE Middle East Oil Show, Bahrein,
Feb. 20-23.
Boukadi, F. H., Bemani, A. S. and Hashmi, A., 2002. PVT empirical models for saturated
Omani crude oils. J. of Petroleum Science and Technology, 20(1&2), 89–100.
De Ghetto, G. and Villa, M., 1994. Reliability analysis on PVT correlations. Paper SPE
28904 Presented at the European Petroleum Conference, London, Oct. 25-27.
Dindoruk, B. and Christman, P.G., 2001. PVT properties and viscosity correlations for
Gulf of Mexico oils. Paper SPE 71633 Presented at the SPE ATCE, New Orleans,
Sep. 30-Oct. 2.
Dokla, M. and Osman, M., 1992. Correlation of PVT Properties for UAE Crudes. SPEFE
(March), 41-46
Elsherkawy, A.M. and Alikhan, A.A., 1997. Correlation for predicting solution Gas/Oil
ratio, formation volume factor, and undersaturated oil compressibility. J. Petroleum
Science and Engineering; 17, pp. 291-302.
Frashad, F., LeBlanc, J.L., Gruber, J.D., Osorio, J.G., 1996. Empirical PVT correlations
for Colombian crude oils. Paper SPE 36105 Presented at the Fourth Latin American
and Caribbean Petroleum Engineering Conference, Port-of-Spain, Apr 23–26.
Glaso, O. 1980, Generalized pressure-volume-temperature correlations. JPT (May) pp.
785-795.
68
Hanafy, H.H., Macary, S.M., El-Nady, Y.M., Bayomi, A.A. and Batanony, M.H., 1997.
A new approach for predicting the crude oil properties. Paper SPE 37439 Presented at
the SPE Production Operations Symposium, Oklahoma City, Mar. 9-11.
Hemmati,M.N. and Kharrat, R., 2007. A correlation approach for prediction of crude-oil
PVT properties. Paper SPE 104543 presented at the SPE Middle East Oil Show and
Conference held in Bahrain Mar 11-14.
Ikiensikimama, S.S. and Ogboja, O., 2009. New bubble point Pressure empirical PVT
correlation. SPE 128893 presented at the Nigeria Annual International Conference
and Exhibition, Abuja, Nigeria Aug. 3-5.
Kartoatmodjo T. and Schmidt, Z., 1994. Large data bank improves crude physical
property correlations. Oil & Gas J. (4 July) pp. 51-55.
Khamehchi, H., Rashidi, F., Rasouli, H. and Ebrahimian, A., 2009. Novel empirical
correlation for estimation of bubble point pressure, saturation viscosity and gas
solubility of crude oils. Petroleum Science, Vol. 6.1 pp. 86-90
Mazandarani, M.T. and Asghari, S.M., 2007. Correlations for predicting solution gas-oil
ratio, bubble point pressure and oil formation volume factor at bubble point of Iran
crude oils. European Congress of Chemical Engineering (ECCE-6) Copenhagen,
(Sep.) pp.16-20
Macary, S.M. and El-Batanoney, M.H., 1992. Derivation of PVT correlations for the Gulf
of Sues crude oils. EGPC 11th Pet. Exp. and Prod. Conference.
McCain, W.D. Jr., 1990. The properties of petroleum fluids. 2nd Ed. PennWell Publishing
Company. Tulsa, Oklahoma.
Mehran, F., Movagharnejad, K. and Didanloo, A., 2006. New correlation for estimation
of formation vilume factor and bubble point pressure for Iranian oil Fields. 1st Iranian
Pet. Eng. Conference, Tehran.
Moradi, B., Malekzadeh, E., Amani, M., Boukadi, F.H. and Kharrat, R., 2010. Bubble
point pressure empirical correlation. Paper SPE 132756 presented at the Trinidad and
Tobago Energy Resources Conference held in Port of Spain, Trinidad, Jun. 27-30.
69
Omar, M.I. and Todd, A.C., 1993. Development of a modified black oil correlation for
Malaysian crudes. Paper SPE 25338 Presented at the SPE Asia Pacific Oil and Gas
Conference, Singapore, Feb. 08-10
Petrosky, G.E. and Farshad, F.F., 1993. Pressure–volume–temperature correlations for
Gulf of Mexico crude oils. SPEREE (Oct.), pp. 416-420.
Rollins, J.B., McCain Jr., W.D., Creager, J.T., 1990. Estimation of the solution GOR of
black oils. JPT (Jan.), pp. 92-94.
Standing, M.B., 1947. A pressure-volume-temperature correlation for mixtures of
California oils and gases. Drilling and Production Practice, API, pp. 275-287
Valko, P.P, and McCain, W.D.Jr, 2003 Reservoir oil bubble point pressures revisited;
solution gas-oil ratios and surface gas specific gravities. J. of Petroleum Engineering
and Science, 37 pp.153-169.
Vasquez, M. and Beggs, H.D., 1980. Correlation for fluid physical property prediction.
JPT June 32, pp. 968-970.
Velarde, J., Blasingame, T.A. and McCain Jr., W.D., 1999. Correlation of black oil
properties at pressures below bubble point pressure- a new approach. J. Can. Pet.
Technol., Spec. Ed. 38 (13), 62-68.
70
CHAPTER IV
4.1 Introduction
The most common and widely used artificial neural networks, ANNs, are found in the
literature and are known as the “feedforward neural networks” with a backpropagation
intelligence modeling scheme in both prediction and classification tasks. In recent years,
ANNs has gained popularity in petroleum applications. Many authors in the field of
petroleum engineering such as Ali (1994), Kumoluyi and Daltaban (1994), Mohaghegh
(1998), Varotsis et al. (1999) and Mohaghegh (2000) discussed the applications of
ANNs. Presently, ANNs is serving the petroleum industry in predicting PVT correlations
(Goda et al. (2003) and Shokir (2004)). Tips and guidelines regarding ANNs modeling
The main objective of Chapter IV is to realize whether artificial neural networks can
be used as
71
pressure, Pb, bubble point oil FVF, Bob, bubble point solution GOR, Rsob and
2. a tool to interpolate and model the average reservoir pressure in oil reservoirs
Presently, the feedforward neural network serves the petroleum industry in predicting
PVT Models. Numerous PVT ANN models have been published in the petroleum
engineering literature. In this section, some of the efforts related to the PVT ANN models
Gharbi and Elsharkawy (1997a) published ANN models to estimate Pb and Bob
Middle East crude oils. Separate models were used for each property based on Equations
1.4 and 1.5. The models’ architectures were two hidden layers and four nodes in the input
layer. The Pb model architecture has eight nodes in the first layer and four nodes in the
second, [4-8-4-4]. The Bo model architecture has six nodes in both layers, [4-6-6-1]. Both
models were trained using 498 data points collected from the literature and unpublished
sources. The models were tested using another dataset of 22 points from the Middle East.
The results showed improvement over the conventional correlation methods with at least
a 50% reduction in the average error for the Pb and a 30% reduction for Bob.
Gharbi and Elsharkawy (1997b) published another ANN model to estimate Pb and Bob
for universal use. The two properties were predicted by one model consisting of one
72
hidden layer of five neurons and the same inputs, [4-5-2]. The study used 5200 datasets
collected from all over the world and representing 350 different crude oils. Another set of
data, consisting of 234 datasets, was used to verify the model’s results. The results for the
universal NN model showed the Pb average error was lower than that of the conventional
correlations for both training and test data. The Bob, on the other hand, was better than the
conventional correlations in terms of correlation coefficient. The average error for the
NN model is similar to the conventional correlations for training data and higher for test
Elsharkawy (1998) presented a new technique to model the behaviour of crude oil
and natural gas systems using a radial-basis-function, RBF neural network model. Two
RBF models can predict Bo, Rso, µo, saturated ρo, undersaturated Co, and liberated γg. The
study used the differential PVT data of 90 samples for training and another 10 samples
for testing the model. The architecture of the RBF network model is arranged in four
layers: an input layer, two hidden layers, and an output layer. The input layer for the two
models contains pR, TR, γoTS, and liberated γg. The output layer for Model #1 contains Rso,
Bo, µo, and ρo. Whereas, the output layer for Model #2 contains undersaturated Co and
liberated γg. The accuracy of the RBF model was compared to all published correlations
for training and testing samples. The results show that the proposed model is far more
accurate than the correlations in predicting the oil properties. The behaviour of the model
in capturing the physical trend of the PVT data was also checked against experimentally
measured PVT properties of the test samples. The author concluded that although the
model was developed for a specific crude oil and gas system, the idea of using neural
73
network to model behaviour of reservoir fluid can be extended to other crude oil and gas
regression techniques.
Varotsis et al. (1999) presented a novel approach to predict a complete set of physical
properties describing the PVT behaviour of both oils and gas condensates using a non-
iterative approach. The ANN oil models are relative volume, Bo, µo, ρo, z-factor, GOR
and liberated γg. The input data for the oil ANN models consists of reservoir fluid
composition, Pb, Tr, ρob, µoD, flash molar ratio and flash liquid and gas densities. The
model was developed using a database containing full PVT data for 650 reservoir fluids
from around the world. Two hidden layers were used. The number of neurons in each
layer was varied and the networks were retrained until the optimum structure was
obtained.
Al-Shammasi (2001) presented two ANNs models to estimate Pb and Bob, based on
global data. Both models include four nodes in the input layer and two hidden layers, five
nodes in the first layer and three nodes in the second layer, [4-5-3-1]. He compared their
performances to numerical correlations. The ANNs models performed better, but suffer
Osman et al. (2001) presented an ANN model to predict the Bob. The model was
developed using 803 published data from the Middle East, Malaysia, Colombia, and the
Gulf of Mexico fields. One-half of the data was used to train the ANNs models, one
quarter to cross-validate the relationships established during the training process and the
remaining one quarter to test the models in order to evaluate their accuracy and trend
74
stability. With regard to the accuracy of the model, they reported an absolute percent
Al-Marhoun (2002) proposed an ANN models to predict Pb and Bob. The models were
developed using 283 datasets collected from Saudi reservoirs. He divided the datasets
into three groups of 142 data points for training, 71 data points for cross validation of the
models and 70 data points to evaluate the accuracy of the models. The Bob model
provides prediction values with an absolute average percent error of 0.5116% and a
standard division of 0.6625 with a correlation coefficient of 99.89% while the Pb model
provides prediction values with an absolute average percent error of 5.8915% and a
Goda et al. (2003) designed two models, one to predict Pb and the other to estimate
Bob. The Pb model based on Equations 1.4 has two hidden layers with each one having ten
neurons, [4-10-10-1]. The hidden layers are activated by the log sigmoid transfer function
while the output layer is activated by the line transfer function. The model was trained
with 160 data points and tested with 20 data sets. However, the Bob model has five inputs
connected with two hidden layers. The input nodes are Rsob, Tr, γo, γg and the estimated Pb
from the first model. Each one of the two hidden layers has eight neurons, [5-8-8-1] with
a log sigmoid transfer function. The output layer is joined with pure linear function.
Shokir et al. (2004) designed two models to estimate Pb and Bob. The Pb network
architecture was selected as a four-layer network. The input layer receives Rsob, TR, API
gravity and γg. Each of the two hidden layers has ten neurons activated by the log sigmoid
transfer function. The network was trained with 160 data points and tested with another
75
20 data points collected from Middle East oil systems. The Bob network has four layers.
The input layer has five neurons for five inputs which are TR, Rsob; API gravity, γg, and
Rasouli et al. (2008) presented two feedforward BP-NN models for the prediction of
Pb and Bob. The models trained with 106 experimental data points and were evaluated
with nine experimental data points. The models are based on Equations 1.4 and 1.5 has
three hidden layers. The Pb model, consists of six nodes in the first hidden layer, ten in
the second layer and six in the third layer, [4−6−10−6−1]. The Bob model consists of nine
nodes in the first hidden layer, twelve neurons in the second layer, and seven neurons in
the third layer, [4−9−12−7−1]. Input and output layers were activated by the logistic
activation function.
Dutta, and Gupta (2010) presented six models for Pb, Rsob, Bob and undersaturated Bo,
and saturated and undersaturated µo from Indian crude oil, based on samples from various
onshore oil fields in western India. The model is based upon MLP using Bayesian
Regularization coupled with a hybrid genetic algorithm (GA) as the error minimization
technique. The model architectures for the Pb, Rsob, Bob undersaturated Bo, saturated µo
[4−9−1], respectively. The transfer function in all the hidden layer neurons is hyperbolic
tangent and the output layer is linear. The performance of the new model was compared
to other existing correlations, and was found to outperform them in terms of average
absolute relative error, correlation coefficient, maximum and minimum absolute relative
76
Recently, Olatunji et al. (2010) proposed a new model to predict Pb and Bob. The
study used a Sensitivity Based Linear Learning Method as a learning technique for two-
layer feedforward NNs based on sensitivity analysis that uses a linear training algorithm
for each layer. The model inputs, based on Equations 1.4 and 1.5, two datasets (160
database and 283 database), were utilized in both internal validation (training) and
external validation (testing) by dividing each into a 70% training set and a 30% testing
set.
Most recently, Alimadadi and Fakhri (2011) predicted the Bo and ρo properties of Iran
crude oil using a committee machine (modular feedforward) with ANNs. The model was
trained using 170 PVT data points from 19 undersaturated oil reservoirs in south Iran.
Two additional datasets, each with 16 points, were used to validate and test the accuracy
of the designed model. The model processes its inputs using two parallel MLP networks,
and then recombines their results. The upper and lower MLP both share the same output
layer which corresponds to Bo and ρo. The upper MLP has one hidden layer with 15
nodes and five nodes in the input layer containing Pb, TRPr, API and Rso. The lower MLP
has 19 nodes in the input layer (with the input being the composition mole %), three
nodes in the hidden layer, and two nodes in the shared output layer.
• Others preferred to use radial-basis-function and the RBF neural network model.
77
• The majority of the PVT ANN models are designed to predict Pb and Bob and the
input data are based on the assumption that Pb and Bo are strong functions of Rsob,
• None of the previous studies deal with the prediction of the stock tank vent GOR,
RST
• All published PVT ANN models can be used to predict reservoir fluid properties
However, PVT ANN models require prior knowledge of some input parameters
that are rarely measured in the field. As a result, it is difficult to apply them in the
• Less attention is paid to the prediction of PVT properties in the absence of PVT
analysis. Until recently, not one specific ANN model has been published to
However, the main objective of this thesis is to build new ANN models to predict Pb,
Bob, Rsob, and RST using directly measured field parameters as nodes in the first network
layer. Alyuda Neurointelligent software (2006) was employed to construct the ANN
models using feedforward with the BP algorithm as a learning algorithm. The models are
shown to be robust, accurate and a quick prediction tool that can be used to estimate
reservoir fluid properties in the absence of a PVT laboratory analysis over the entire
range of operating conditions and fluid types. The main application of the models is
78
4.3 PVT Datasets
A complete description of PVT property data is not accessible with respect to some
their fluid properties using directly measured field data parameters as input variables. To
resolve this problem and develop accurate PVT models, the ANN must be exposed to
sufficient data points or in the training session, the network model learns and processes
only a few patterns and cannot provide a precise prediction and interpolation when it
encounters new PVT samples. The objectives of this chapter is to build new ANN models
to predict Pb, Bob, Rsob, and RST using directly measured field parameters as nodes in the
The PVT dataset used in this work contains a total of 118 reservoir fluid studies (PVT
analysis) which are the only available studies were collected from several oil fields
located in Sirte Based, Libya. Therefore, 493 data points were selected to design the
model. However, before the ANNs are carried out, the PVT data was subjected to an
investigation to find any outliers and data anomalies. Outliers are the observations that
investigator (Iglewicz, 1983). Analyzing the data detected eight observations that were
identified as outliers and rejected as unconvertible for use with the neural network.
Typically, the selection of the ANN model is made with the basic cross-validation
process. Moreover, to avoid overtraining, which results in poor model performance, the
remaining 485 data points were subdivided randomly into four sets, namely, the training
79
dataset, the validation dataset, the testing dataset, and the deploying dataset. The
validation dataset was used during training to see how the model would perform.
Training can be stopped when the performance of the model on the validation dataset
gives a minimum error. The testing dataset was used to fine-tune the models. However,
numerous ANNs were tried with various architectures and the accuracy of each one on
the testing dataset was tested in order to select the competing architecture. The lowest
error on the testing dataset is used to select the appropriate model and to indentify the
optimum network architecture. Finally, the deployment dataset was used to examine the
The overall data used to model Pb were split into four sets - 286 points (59.95%) to
the training set, 14.0% (67 points) to the validation set, 14.0% (67 points) to the testing
set and 57 points (12.95%) to the deployment set. The training dataset was used to train
Four inputs (Psp, Rsp, TR and γoST) are selected to be incorporated into the input layer.
In order to find an efficient network model and analyze the performance, the data must be
pre-processed in order to have a small variation of the output data. It is well known that,
in theory, the output data can have a large domain of variation. However, if the domain of
variation of the output data is large, the ANN model tends to be less stable. The
dimensionless output data ranges from 0 to 1, while the input data ranges from −1 to 1.
80
Table 4.1 − Ranges of Datasets used in Training, Validation, Testing and Deployment Pb
ANN Model.
81
To specify the number of hidden layers and the number of units in each layer, a trial and
error approach was carried out, beginning with one hidden layer and one hidden unit.
Hidden units were then gradually added. A total of 70 architectures were compared with
Table 4.2 shows the five best architectures. At this point, the selected architecture
with two hidden layers, nine nodes in the first hidden layer, and four nodes in the second
layer was trained using different algorithms. In each case, the architecture was retrained
five times with different initial weight randomization. The training algorithm with the
lowest average error and highest R2 was selected as a best training algorithm. Table 4.3
shows the average and maximum error during training, validation and testing of
It is evident that the Quasi-Newton training algorithm has a higher R2 of 99.12 %, and the
lowest training, validation and testing error of 77.535, 102.797 and 119.272 respectively.
82
4.4.1 Bubble Point Pressure ANN Model Evaluation
Both qualitative and graphical analysis of the model error terms was utilized to verify
the accuracy of the selected model. Table 4.4 summarizes the quantitative statistical error
analysis for the identified Pb ANN model. The analysis shows a small error results on
Furthermore, the graphical error analysis was carried out by plotting the predicted
values versus actual Pb observation values. Figures 4.1, 4.2, 4.3, and 4.4 show the scatter
diagram that compare the modeled Pb with the experimental value for training, validation,
testing and overall data points. It is evident that all the figures show the majority of points
tended to concentrate in the vicinity of the identity line y = x. This indicates excellent
agreement between the actual Pb values and the model predicted values.
83
Table 4.2 − Comparison of Pb Network Architectures - The 5 Best Architectures
Error, psia
Architecture R2
Training Validation Testing
[4-9-4-1] 79.142 116.597 119.693 99.11
[4-8-5-4] 72.462 132.506 158.292 99.19
[4-9-5-1] 85.493 124.296 181.009 98.83
[4-10-5-1] 59.905 121.312 168.052 99.43
[4-9-6-1] 77.231 104.626 172.360 99.16
84
Table 4.3 − Average Error during Training, Validation and Testing Pb Model
Training Training Error, psia Validation Error, psia Test Error, psia
R2 %
Algorithm Average Max Average Max Average Max
Q-N 99.12 77.532 456.831 102.797 575.879 119.272 938.42
CGD 98.46 104.549 624.501 131.09 570.42 126.823 974.400
L-M 97.75 135.28 490.169 159.560 581.692 162.040 970.161
LM Q-N 98.77 88.131 601.371 126.935 593.375 128.292 981.014
85
Table 4.4 − Ranges of Pb ANN Model Output with Statistical Error Analysis
Training output
Actual, psia ANN, psia AE, psia ARE
Average 1689.7 1686.1 130.7 0.126
Maximum 4244 3858.8 841.5 0.782
Minimum 121 172.5 1.195 0.0014
Validation output
Average 1731 1744 116.4 0.119
Maximum 4244 3922 495.7 1.023
Minimum 152 202 0.08 0.000
Testing output
Average 1707 1705 160 0.141
Maximum 3582 3569 755 0.669
Minimum 121 202 1.86 0.001
Overall output
Average 1699 1698 133 0.126
Maximum 4244 3922 841.6 1.023
Minimum 121 172 0.080 0.000
86
Scatter Diagram
Bubble Point Pressure ANN Model (Training Dataset)
5000
Modeled Bubble point Pressure
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Experimental PVT Bubble Point Pressure
Figure 4.1 – Cross Plot of Modeled Bubble Point Pressure on Training Dataset versus
Laboratory PVT Values
87
Scatter Diagram
Bubble Point Pressure ANN Model (Validation Error)
5000
Modeled Bubble Point Pressure
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Experimental PVT Bubble Point pressure
Figure 4.2 – Cross-plot of Modeled Bubble Point Pressure on Validation Dataset versus
Laboratory PVT Analysis
88
Scatter Diagram
Bubble Point Pressure ANN Model (Testing Dataset)
5000
Modeled Bubble Point Pressure
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Experimental PVT Bubble Point Pressure
Figure 4.3 - Cross-plot of Modeled Bubble Point Pressure on Testing Dataset versus
Laboratory PVT Analysis
89
Scatter Diagram
Bubble Point ANN Model (Overall Datapoints)
5000
Modeled Bubble Point Pressure
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Experimental PVT Bubble Point Pressure
Figure 4.4 - Cross-plot of Modeled Bubble Point pressure on Overall Points versus
Laboratory PVT Analysis
90
4.4.2 Contribution of Input Parameters
non-contributing features. The results of the predictive importance analysis are useful to
determine the percentage contribution of the input variables. As for ANNS, the
contribution factor is a rough measure of the importance of that variable in predicting the
network’s output relative to other input variables in the same network. Identifying the
the Rsp ranked first with an importance of 62.9% (the network prediction is influenced
mostly by Rsp) and TR ranked second with an importance of 21.4% while γoST and Psp
ranked third and forth with an importance of 8.5% and 7.2%, respectively.
As a final step in the modelling process, the Pb ANN model is deployed to a new
dataset. This is illustrated in Figure 4.5 where the estimated Pb, using ANN, was plotted
versus the PVT Pb values. The figure shows a tight cloud of points close to the identity
line indicating excellent agreement between the actual and model predicted values.
Table 4.5 summarizes the quantitative statistical error analysis for the deployment
test. The analysis shows a small error obtained on the deploy datasets. It shows the
91
Scatter Diagram
Deployment Test (Bubblepoint ANN Model)
5000
Estimated bubblepoint pressure, psia
4000
3000
2000
1000
0
0 1000 2000 3000 4000 5000
Actual bubblepoint pressure, psia
Figure 4.5 – Deployment Test: Cross-plot of Modeled Bubble Point versus Laboratory
PVT Analysis
92
Table 4.5- Statistical Error Analysis for Deployment Test
93
4.5 Bubble Point Oil FVF ANN Model
Again, the overall PVT dataset (the only available data, 485 points) used for
modeling Bob was split into four sets; the training set, the validation set, the testing set
and the deployment set. Each was separately used by the ANN for the tasks of training,
validation, testing and deploying. 289 points were used for training, 67 points were used
for validation, 67 points were used for testing and 57 points were used for the deployment
set. A range of each dataset is described in Table 4.6. As described in section 2.3, four
inputs (Psp, Rsp, TR and γoST) are selected to incorporate into the input layer and one node
into the output layer (Bob). Subsequently, the input data were pre-processed and
transforms the input data into a new representation before a network is trained.
With an eye to finding the optimum network design, a trial and error approach was
performed, starting with one hidden layer and a number of hidden nodes. A total of 149
architectures are compared to pick the best which displays excellent performance on the
94
Table 4.6 – Ranges of Datasets used in Training, Validation, Testing and Deployment of
the Bob ANN Model
95
The comparison indicates none of the networks offered an acceptable error with one
hidden layer. Table 4.7 presents the five best architectures of the compared 149
configuration of one layer of inputs with four nodes, two hidden layers with five nodes in
the first hidden layer and four in the second, and one output, [4−5−4−1]. An important
step before training the selection architecture model is to determine the optimal training
algorithm. The usual strategy is based solely upon a trial-and-error procedure, which may
be very time consuming and requires significant user experience. However, the selected
architecture was trained by the most commonly used training algorithms including Quasi-
functions. In each case, the architecture was retrained five times with different initial
were compared and tabulated in Table 4.8. The table presents the networks’ performance
in terms of average training error, average validation error and average testing error. The
best results of each training algorithm group are emphasized in bold. Thus, the results
indicate that the Levenberg-Marquardt algorithm gave the largest average errors and
96
Figure 4.6 – Bob BP-Model Architecture, [4-5-4-1]
97
Table 4.7 – Comparison of Bob Network Architectures - The 5 Best Architectures
Error, bbl/STB
Architecture Training Validation Testing R2
[4-5-2-1] 0.017 0.016 0.018 97.53
[4-3-5-1] 0.019 0.016 0.018 97.04
[4-5-4-1] 0.014 0.016 0.017 98.4
[4-8-6-1] 0.015 0.016 0.018 98.20
[4-9-6-1] 0.015 0.015 0.018 98.2
98
• With the use of the Q-N, LM Q-N and CGD training algorithms, the resulting
average errors are almost similar for all the used activation functions.
• When the hidden layers were activated by the hyperbolic activation function and
the output layer was activated by the logistic activation function and trained with
the Conjugate Gradient Descent BP algorithm, the model provided the smallest
• The smallest average error of 0.01409 on the validation dataset is achieved with
the network hidden layers being activated by the logistic activation function and
the hyperbolic activation function activated the output layer and in turn, was
• The smallest average testing error of 0.01688 was obtained when the network was
• The highest R2 of 98.76% is provided by training the network with the Conjugate
As the testing dataset is used neither in training nor in the validation phase, this
set is used to test the final model. Thus, the limited memory Quasi-Newton and
network Models M1, M5, M9 and M12, in Table 4.8, are selected as the best based on
model performance on the testing dataset and this can still be demonstrated by
99
determining a statistical error analysis to specify the best model among the four models
(M1, M5, M9 and M12). Comparisons of average, minimum and maximum of absolute
error and absolute percent relative error across the four models are presented in Table 4.9.
The table addresses the errors obtained from training, validation, testing and overall
datasets. However, the bottom line of comparing models and selecting the best is to
observe the absolute error and absolute relative error of the validation and testing dataset
in addition to the scatter diagram. The best results of each training model are emphasized
in bold. Via the results, the lowest errors in terms of Absolute Error (AE) and Absolute
Relative Error (ARE), based on validation and testing datasets can be summarized as:
layers are activated by the hyperbolic tangent function and the output layer
100
• The Testing Dataset,
with the hyperbolic function and the output layer was activated with the
layers are activated with the logistic activation function. The values are
As a result, from the above discussions, the networks M12 in Table 4.9 gave best
performance on validation dataset and overall dataset, while network M9 gave best
101
performance on testing dataset. Still there are challenges to select the best network
In this case, it is too difficult to select the best model from M9 and M12. The error
terms were analyzed graphically. y = x line cross-plot, error histogram and probability
plot for the error terms resulted from training, validation, testing and overall datasets
which are presented in Figures 4.2, 4.3, 4.4 and 4.5, respectively. To choose the best
model, the points must be observed and compared where the data points fall, which
roughly follows a straight line in the probability plot. The AD value provided with the
plot can be used. It measures how well the data follow a particular distribution. The better
the distribution fits the data, the smaller the AD value will be.
• Training set (Figure 4.7, c and d) shows both M9 and M12 are normally
with 0.01951 for M9. M12 has a smaller AD value. Therefore, M12 is the best
training dataset.
• Validation set (Figure 4.8, a - f): Both offer same shape and influence.
• Testing dataset (Figure 4.9): According to Figure 4.9 (e), Model M9 follows a
straight line better than M12 does. M9 has a low AD value and a low SD.
• Overall points (Figure 4.10): M12 offers a lower SD while M9 offers a lower AD
value.
102
Table 4.8 - Average Error during Training, Validation and Testing of Architecture [4-5-4-
1] with Various Training Algorithms and Input/Output Activation Functions. (Illustrates
the Effect of Choice of Training Algorithm and Activation Function)
Levenberg-Marquardt BP Algorithm
M13 Logistic Logistic 0.02044 0.01839 0.02189 97.22
M14 Hyperbolic Hyperbolic 0.01924 0.0191 0.02413 97.52
M15 Logistic Hyperbolic 0.01844 0.01809 0.02286 97.68
M16 Hyperbolic Logistic 0.02168 0.02 0.02272 96.86
103
Table 4.9 – Comparison between the Four Best Network Models
Model M1
Quasi-Newton BP Algorithm
Hidden activating Function = Logistic
Output activation function = Logistic
Training Set Validation Set Testing Set Overall Dataset
AE, ARE AE ARE AE ARE AE ARE
bbl/STB bbl/STB bbl/STB bbl/STB
SD 0.01655 0.01213 0.01497 0.01122 0.01915 0.01466 0.01677 0.01244
Min 0.00001 0.00001 0.00008 0.00007 0.00052 0.00039 0.00001 0.00001
Max 0.12892 0.09040 0.06832 0.04898 0.13192 0.09994 0.13192 0.09994
R2 % 97.18 97.59 95.74 97.07
Model M 5
Limited memory Quasi-Newton BP Algorithm
Hidden activating Function = Logistic
Output activation function = Logistic
Training Set Validation Set Testing Set Overall Dataset
AE ARE AE ARE AE ARE AE ARE
SD 0.01378 0.01039 0.0159 0.01184 0.01731 0.01331 0.01479 0.01118
Min 0.00028 0.00024 0.00011 0.00007 0.00056 0.00044 0.00011 0.00007
Max 0.11233 0.07877 0.08618 0.06178 0.09931 0.07523 0.11233 0.07877
R2 % 98.27 97.58 96.74 97.69
Model M 9
Limited memory Quasi-Newton BP Algorithm
Hidden activating Function = Logistic
Output activation function = Logistic
Training Set Validation Set Testing Set Overall Dataset
AE ARE AE ARE AE ARE AE ARE
SD 0.01368 0.01031 0.01550 0.01158 0.01719 0.01321 0.01465 0.01108
Min 0.00000 0.00000 0.00016 0.00013 0.00019 0.00014 0.00000 0.00000
Max 0.11129 0.07805 0.07972 0.05714 0.09691 0.07341 0.11129 0.07805
R2 % 98.29 97.65 97.76 98
Model M 12
Limited memory Quasi-Newton BP Algorithm
Hidden activating Function = Hyperbolic
Output activation function = Logistic
Training Set Validation Set Testing Set Overall Dataset
AE ARE AE ARE AE ARE AE ARE
SD 0.01145 0.00852 0.01445 0.01063 0.01852 0.0143 0.01347 0.0101
Min 0.00002 0.00002 0.00001 0.00001 0.00016 0.00014 0.00001 0.00001
Max 0.08624 0.06048 0.06679 0.0543 0.11195 0.08481 0.11195 0.08481
R2 98.74 97.95 96.27 98.3
104
Figure 4.7 (a, b) – Graphical Error Analysis on Training Dataset
1.7 1.7
Estimated Bob, bbl/STB
1.5 1.5
1.4 1.4
1.3 1.3
1.2 1.2
1.1 1.1
1.0 1.0
1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Actual Bib, bbl/STB Actual Bob, bbl/STB
(a) (b)
105
Figure 4.7 (c, d) – Graphical Error Analysis on Training Dataset
Frequency
50 50
40 40
30 30
20 20
10 10
0 0
-0.12 -0.08 -0.04 0.00 0.04 0.08 0.12 -0.12 -0.08 -0.04 0.00 0.04 0.08 0.12
Error Terms Error Terms
(c) (d)
106
Figure 4.7 (e, f) – Graphical Error Analysis on Training Dataset
Probability Plot of Error Terms (Training Dataset) Probability Plot of Error Terms (Training Dataset)
Bob ANN Model #12; Normal - 95% CI Bob ANN Model #12; Normal - 95% CI
99.9 99.9
Percent
60 60
50 50
40 40
30 30
20 20
10 10
5 5
1 1
0.1 0.1
-0.05 0.00 0.05 0.10 -0.050 -0.025 0.000 0.025 0.050 0.075 0.100
Error Terms Error Terms
(e) (f)
107
Figure 4.8 (a, b) – Graphical Error Analysis on Validation Dataset
1.7 1.7
1.5 1.5
1.4 1.4
1.3 1.3
1.2 1.2
1.1 1.1
1.0 1.0
1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Actual Bob, bbl/STB Actual Bob, bbl/STB
(a) (b)
108
Figure 4.8 (c, d) – Graphical Error Analysis on Validation Dataset
15
Frequency
Frequency
10
10
5
5
0 0
-0.09 -0.06 -0.03 0.00 0.03 0.06 0.09 -0.09 -0.06 -0.03 0.00 0.03 0.06 0.09
Error Terms Error Terms
(c) (d)
109
Figure 4.8 (e, f) – Graphical Error Analysis on Validation Dataset
Probability Plot of the Error Terms (Validation Dataset) Probability Plot of Error Terms (Validation Dataset)
Bob ANN Model # 9, Normal - 95% CI Bob ANN Model # 12, Normal - 95% CI
99.9 99.9
Mean 0.0003662
99 Mean 0.0005627 99 SD 0.02271
SD 0.02233
N 67
95 N 67 95 AD 1.184
90 AD 1.130 90 P-Value <0.005
P-Value 0.005
80 80
70 70
Percent
Percent
60 60
50 50
40 40
30 30
20 20
10 10
5 5
1 1
0.1 0.1
-0.10 -0.05 0.00 0.05 0.10 -0.10 -0.05 0.00 0.05 0.10
Error Terms Error Terms
(a) (b)
110
Figure 4.9 (a, b) – Graphical Error Analysis on Testing Dataset
1.7 1.7
Estimated Bob, bbl/STB
1.5 1.5
1.4 1.4
1.3 1.3
1.2 1.2
1.1 1.1
1.0 1.0
1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Actual Bob, bbl/STB Actual Bob, bbl/STB
(a) (b)
111
Figure 4.9 (c, d) – Graphical Error Analysis on Testing Dataset
Frequency
Frequency
15
20
15
10
10
5
5
0 0
-0.12 -0.08 -0.04 0.00 0.04 0.08 0.12 -0.13 -0.09 -0.05 -0.01 0.03 0.07 0.11
Error Terms Error Terms
(c) (d)
112
Figure 4.9 (e, f) – Graphical Error Analysis on Testing Dataset
Probability Plot of the Error Terms (Testing Dataset) Probability Plot of the Error Terms (Testing Dataset)
Bob ANN Model # 9, Normal - 95% CI Bob ANN Model # 12, Normal - 95% CI
99.9 99.9
Percent
60 60
50 50
40 40
30 30
20 20
10 10
5 5
1 1
0.1 0.1
-0.10 -0.05 0.00 0.05 0.10 -0.10 -0.05 0.00 0.05 0.10
Error Terms Error Terms
(e) (f)
113
Figure 4.10 (a, b) – Graphical Error Analysis on Overall Dataset
1.7 1.7
Estimated Bob, bbl/STB
1.5 1.5
1.4 1.4
1.3 1.3
1.2 1.2
1.1 1.1
1.0 1.0
1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Actual Bob, bbl/STB Actual Bob, bbl/STB
114
Figure 4.10 (c, d) – Graphical Error Analysis on Overall Dataset
150 150
Frequency
Frequency
100 100
50 50
0 0
-0.090 -0.045 0.000 0.045 0.090 -0.090 -0.045 0.000 0.045 0.090
Error Terms Error Terms
(c) (d)
115
Figure 4.10 (e, f) – Graphical Error Analysis on Overall Dataset
Probability Plot of Error Terms (Overall Dataset) Probability Plot of the Eror Terms (Overall Dataset)
Bob ANN Model # 9, Normal - 95% CI Bob ANN Model # 12, Normal - 95% CI
99.9 99.9
Percent
60 60
50 50
40 40
30 30
20 20
10 10
5 5
1 1
0.1 0.1
-0.05 0.00 0.05 0.10 -0.05 0.00 0.05 0.10
Error Terms Error Terms
(e) (f)
116
As result, the graphical analysis provides the same results for the validation set for
both models. M12 is the best on training and M9 is the best testing set. Therefore, the
final decision is taken based on the ability of a model to perform reliably on any novel
data in a given domain. Thus, M9 was considered the best BP-ANN model to predict oil
As a final step in the modelling process, the proposed Bob PB-ANN model is
deployed to a new dataset. This data consists of 62 data points which were introduced to
the trained BP-ANN model (M9) and the corresponding Bob’s were obtained. The test is
illustrated in Figure 4.11 where the model Bob is plotted versus the laboratory PVT
values. It shows a tight cloud of points near the identity line, indicating excellent
agreement between the actual Bob PVT data and model predicted values. Table 4.10
summarizes the quantitative statistical error analysis for the deployment test. The analysis
shows a small error obtained on the deploy datasets. It shows the superiority of the Bob
Prior to specifying the Rsob ANN model architecture, the overall collected PVT data
were split into four sets. Fifty-nine percent of the collected PVT data was used to train
the network models, 14% was used in the validation process to investigate the model’s
performance, another 14% was used to test the model and fine-tune it and the remaining
117
13% PVT data was used to deploy the selected model. The range of each dataset is
The development of ANN models starts with finding the best network structure to
starting from a low number of layers and nodes to a higher number of layers and nodes,
should be tested to select the best architecture. Successful architectures are those that
converge to the target error or reach the possible MSE and exhibit stable performance to
new data not included in the training. Topically, the best ANN architecture is the one that
provides the lowest training error and testing and validation errors with the highest R2.
118
Scatter Diagram
Deployment Test, Bob ANN Model
1.8
1.7
Estimated Bob, bbl/STB
1.6
1.5
1.4
1.3
1.2
1.1
1.0
1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Actual Bob, bbl/STB
119
Table 4.10- Statistical Error Analysis for Deployment Test
120
Table 4.11 − Ranges of Datasets used in Training, Validation, Testing and Deployment
Rsob ANN Model.
121
4.6.1 Model Design
A trial and error attempt was carried out to achieve a well-designed Rsob network
model architecture. An input vector and the corresponding desired output are considered
first. Three inputs (Rsp, Psp and γoST) are selected to incorporate into the input layer. Next,
a series of networks with single and multiple hidden layers with different numbers of
nodes in each hidden layer were trained with different activation functions and training
algorithms. A total of 48 architectures were compared, starting with one hidden layer and
one hidden node. Hidden nodes were then gradually added. The following network
In the training procedure, the inputs propagated forward through the network to the
output vector (Rsob). The output vector is compared with the desired output (actual Rsob),
and the errors are determined. The errors are then propagated back through the network
122
from the output to the input layer. This route is repeated until the errors begin to
minimize. The training results showed none of the networks with one or three hidden
layers gave an acceptable error. The 5 best network architectures are presented in Table
4.12 where the best results are emphasized in bold. It is evident that Architecture
[3−7−4−1] gives the lowest training and testing error, with the highest R2 of 99.27%.
Architecture [3−7−4−1], with the lowest average error and the highest R2, was
selected as the best. Following this, a selected architecture with two hidden layers, 7
nodes in the first hidden layer, and 4 nodes in the second was trained using different
algorithms and activation functions. In each case, the architecture was retrained five
times with different random initial weights with concerns over the possibility of
converging to a local minimum in weight space. However, a typical neural network might
have over two hundred weighs whose values must be found to produce an optimal
solution. Once a network settles on a minimum, whether local or global, the learning
stops. If a local minimum is reached, the error at the network outputs may still be
change in the activation function or in the learning algorithm will often fix the problem.
123
Table 4.12 − Comparison of Rsob Network Architectures - The 5 Best Architectures
124
Figure 4.12 – Rsob BP-ANN Model Architecture [3-7-4-1]
125
Table 4.13 shows the average and maximum error during training, validation and
Newton. It is evident that the Limited Memory Quasi-Newton training algorithm had the
higher R2, lowest average training, average validation and average testing error, as
emphasized in bold.
analysis of the model’s error terms was utilized to verify the accuracy of the selected
model. Table 4.14 summarizes the quantitative statistical error analysis for the identified
Rsob ANN model. The analysis shows a small AE and ARE on training, validation and
testing datasets. The model provided predicted Rsob values with an average absolute error
99.2%.
126
Table 4.13 - Average Error during Training, Validation and Testing Rsob Model with
Various Training Algorithms (Quick Propagation (QP), Quasi-Newton (Q-N), Conjugate
Gradient Descent (CGD), and Limited Memory Quasi-Newton (LM Q-N)
127
Table 4.14 – Statistical Error Analysis of Rsob ANN Model Output
Training Output
Actual ANN AE ARE
SCF/STB
Average 411.66 411.64 16.18 0.07
Maximum 17 51.312 0.01 0.00
Minimum 1256 1201.17 136.02 2.28
R2 99.09%
Validation Output
Average 413.01 413.10 15.66 0.07
Maximum 40 65.53 0.079 0.00
Minimum 1053 1039.19 59.8 0.64
R2 99.32%
Testing Output
Average 436.42 436.02 13.58 0.06
Maximum 42 69.54 0.14 0.00
Minimum 1135 1140.65 83.32 0.71
R2 99.47%
Overall Output
Average 415.81 415.75 15.68 0.07
Maximum 17 51.31 0.01 0.00
Minimum 1256 1201.17 136.01 2.28
R2 99.2%
128
Furthermore, the graphical error analysis was carried out by plotting the ANN model
predicted values versus laboratory PVT observation Rsob values. Figures 4.13, 4.14, 4.15
and 4.16 show the scatter diagram which compares the actual with the Rsob outputs for
training, validation, testing and overall datasets. It is evident that all the figures show the
majority of points tended to concentrate in the vicinity of the identity 1:1 line, indicating
excellent agreement between the actual and model predicted Rsob values. Moreover, the
histogram of the error terms resulting from training, validation, testing, and overall
datasets, are presented in Figures 4.17, 4.18, 4.19 and 4.20, respectively. All histograms
129
Scatter Diagram
Rsob ANN Model (Training Dataset)
1200
1000
Rsob ANN Model
800
600
400
200
0
0 200 400 600 800 1000 1200
Actual Rsob, scf/STB
Figure 4.13 – Cross-plot of Modeled Rsob on Training Dataset vs. Laboratory PVT Values
130
Scatter Daigram
Rsob ANN Model (Validation Dataset)
1200
1000
Estimated Rsob, scf/STB
800
600
400
200
0
0 200 400 600 800 1000 1200
Actual Rsob, scf/STB
Figure 4.14 – Cross-plot of Modeled Rsob on Validation Dataset vs. Laboratory PVT
Values
131
Scatter Diagram
Rsob ANN Model (Testing Dataset)
1200
1000
Estimated Rsob, scf/STB
800
600
400
200
0
0 200 400 600 800 1000 1200
Actual Rsob, scf/STB
Figure 4.15 – Cross-plot of Modeled Rsob on Testing Dataset vs. Laboratory PVT Values
132
Scatter Diagram
Rsob ANN Model (Overall Dataset)
1200
1000
ANN Model output
800
600
400
200
0
0 200 400 600 800 1000 1200
Actual Rsob, scf/STB
Figure 4.16 – Cross-plot of Modeled Rsob on Overall Data Points vs. Laboratory PVT
Values
133
Histogram of the Error Terms
Rsob ANN Model (Training Dataset)
90
Mean 0.02487
80 StDev 24.22
N 300
70
60
Frequency
50
40
30
20
10
0
-120 -80 -40 0 40 80 120
Error Terms
134
Histogram of the Error Terms
Rsob ANN Model (Validation Dataset)
18
Mean -0.09407
16 StDev 21.45
N 70
14
12
Frequency
10
0
-120 -80 -40 0 40 80 120
Error Terms
135
Histogram of the Error Terms
Rsob ANN Model (Testing Dataset)
25
Mean 0.4029
StDev 20.29
N 70
20
Frequency
15
10
0
-120 -80 -40 0 40 80 120
Error Terms
136
Histogram of the Error Terms
Rsob ANN Model (Overall Dataset)
140
Mean 0.06609
StDev 23.16
120 N 440
100
Frequency
80
60
40
20
0
-120 -80 -40 0 40 80 120
Error Terms
137
4.6.3 Model Deployment
Model deployment refers to the application of the model to a new dataset as a final
step in modeling development. This step involves the application of the selected model
into a new dataset in order to test the generalization of the model. Thus, after a
satisfactory Rsob, the ANN model has been identified as the best model and the model is
deployed to a new dataset. This is illustrated in Figure 4.21 where the estimated Rsob,
using the ANN model, was plotted against the actual Rsob values. The figure shows a tight
cloud of points near the identity line, indicating an excellent agreement between the
Table 4.15 summarizes the quantitative statistical error analysis for the deployment
test. The analysis shows a small error obtained on the deploy datasets. It shows the
Three inputs (PSP, TSP and γoST) are selected to incorporate into the input layer after a
forward-backward search algorithm is carried out to select the best subset of input
variables. Then, to specify the number of hidden layers and the number of units in each
layer, a trial and error attempt was carried out, starting with one hidden layer and one
hidden unit. Hidden units were then gradually added. A total of 672 architectures were
138
Scatter Diagram
Deployment Test, Rsob ANN Model
1200
1000
Estimated Rsob, scf/STB
800
600
400
200
0
0 200 400 600 800 1000 1200
Actual Rsob, scf/STB
139
Table 4.15- Statistical Error Analysis for Deployment Test
140
Hidden layer activation function = Logistic and hyperbolic tangent
Output activation function = Logistic and hyperbolic tangent
Error function = Sum-of-squares
Hidden layers search range = 1 to 3 HL
1st layer hidden units search range = 1 to 8 nodes
2nd hidden layer units search range = 1 to 5 nodes
3rd hidden layer units search range = 1 to 3 nodes
Fitness criteria = Inverse test error
The results show not one of the networks gave an acceptable error with one hidden
layer. However, the use of a three hidden layer network with an activation function of
hyperbolic tangent for both hidden and output layers gives better results. Table 4.16
presents the best architectures among the compared 672 architectures. It is evident that
the hyperbolic tangent activation function gives the lowest testing error.
One step prior to selecting the best architecture is to determine the optimal training
algorithm. The usual strategy is based solely upon a trial-and-error procedure, which
could be very time consuming and requires significant user experience. However,
in bold print in Table 4.15, are trained by the most commonly used training algorithms
Memory Quasi-Newton BP (LM Q-N) (Rahman et al. 2011 and Talib et al. 2009). In
each case, the architecture was retrained five times with different initial weights
4.17. The table presents the network’s performance in terms of average training error,
141
average validation error and average testing error. The best results of each training
As a result, Architecture [3-8-4-2-1] with three hidden layers, seven nodes in the first
hidden layer, four in the second and three nodes in the third was selected as a ANN
model to predict the RST. The model gave the best performance on training and testing
datasets.
Both qualitative and graphical analysis of the model error terms was utilized to verify
the accuracy of the selected model. Table 4.18 summarizes the quantitative statistical
error analysis for the identified RST BP-ANN model. The analysis shows a small error in
Furthermore, the graphical error analysis was carried out by plotting the predicted
values versus actual RST observation values. Figures 4.22, 4.23, 4.24 and 4.25 show the
scatter diagram that compare the actual Pb values with the outputs for training, validation,
testing and overall data points of the RST ANN model. It is evident that all the figures
show most of the points tended to concentrate in the vicinity of the identity line y = x.
This indicates excellent agreement between the laboratory PVT RST values and the
modeled values.
142
Table 4.16 - Comparison of RST Network Architectures - The Best Architectures
143
Table 4.17 - Average Error during Training, Validation and Testing RST Model with
Various Training Algorithms (Quasi-Newton (Q-N), Conjugate Gradient Descent (CGD),
and Limited Memory Quasi-Newton (LM Q-N)
144
Table 4.18 - Ranges of RST ANN Model Output with Statistical Error Analysis
145
Scatter Diagram
Stock-tank vent gas/oil raio ANN model (Training Dataset)
250
Estimated RST, SCF/STB
200
150
100
50
0
0 50 100 150 200 250
Actual RST, SCF/STB
Figure 4.22 – Cross-plot of Modeled RST on Training Dataset vs. PVT Values
146
Scatter Diagram
Stock-tank vent gas/oil raio ANN model (Validation Dataset)
250
Estimated RST, SCF/STB
200
150
100
50
0
0 50 100 150 200 250
Actual RST, SCF/STB
Figure 4.23 – Cross-plot of Modeled RST on Validation Dataset vs. PVT Values
147
Scatter Diagram
Stock-tank vent gas/oil raio ANN model (Testing Dataset)
250
Estimated RST, SCF/STB
200
150
100
50
0
0 50 100 150 200 250
Actual RST, SCF/STB
Figure 4.24 – Cross-plot of Modeled RST on Testing Dataset vs. PVT Values
148
Scatter Diagram
Stock-tank vent gas/oil raio ANN model (Overall Dataset)
250
Estimated RST, SCF/STB
200
150
100
50
0
0 50 100 150 200 250
Actual RST, SCF/STB
Figure 4.25 – Cross-plot of Modeled RST on Overall Data Points vs. PVT Values
149
Histogram of the Error Terms
Stock-tank gas/oil ratio ANN model (Training Dataset)
60
Mean -0.03540
StDev 12.59
50 N 235
40
Frequency
30
20
10
0
-60 -40 -20 0 20 40 60
Error Terms
150
Histogram of the Error Terms
Stock-tank gas/oil ratio ANN model (Validation Dataset)
20 Mean 1.185
StDev 18.72
N 55
15
Frequency
10
0
-60 -40 -20 0 20 40 60
Error Terms
151
Histogram of the Error Terms
Stock-tank gas/oil ratio ANN model (Testing Dataset)
20 Mean -1.178
StDev 15.27
N 55
15
Frequency
10
0
-60 -40 -20 0 20 40 60
Error Terms
152
Histogram of the Error Terms
Stock-tank gas/oil ratio ANN model (Overall Dataset)
80
Mean -0.02293
70 StDev 14.14
N 345
60
50
Frequency
40
30
20
10
0
-60 -40 -20 0 20 40 60
Error Terms
153
4.8 Conclusions
• Based on the results of this study, the following findings can be drawn:
four new models with which to estimate bubble point pressure, the bubble point
oil formation volume factor, bubble point solution gas/oil ratio and stock-tank
• The models were established based on 476 PVT flash separation tests collected
• Unlike existing PVT models, the models can be applied directly in the absence of
PVT analysis. They require on-hand production data with no need of additional
• The bubble point BP-ANN model has four inputs, two hidden layers, with nine
nodes in the first and four in the second. The model was trained with the Quasi-
Newton training algorithm. The entire output layer and the hidden nodes were
• The bubble point oil formation volume factor BP-ANN model was trained with
logistic activation function. It has four inputs, two hidden layers with five nodes
in the first hidden layer and four nodes in the second hidden layer.
• The bubble point solution gas/oil ratio BP-ANN model was trained with the
Limited Memory Quasi-Newton training algorithm. It has three inputs and two
154
hidden layers. Seven nodes are in the first hidden layer while four nodes are in the
• The stock-tank vent gas/oil ratio BP-ANN model was trained with the Quasi-
Newton training algorithm and activated by the logistic activation function. It has
three inputs, three hidden layers, eight nodes in the first layer, four nodes in the
second layer, and two nodes in the third layer. All the layers were activated by
• In order to build robust BP-ANN models and avoid the models from over fitting,
• The results show the trained models provided high accuracy on testing dataset.
This dataset was not viewed by the networks during the training process.
155
4.9 References
Alimadadi, F. and Fakhri, A., 2011. using a committee machine with artificial nural
networks to predict PVT properties of Iran crude oil. SPE Reservoir Evaluation &
Engineering.
Alyuda, 2006. NeuroIntelligence, www.alyuda.com
Amari, S et al., 1997. Asymptotic statistical theory of overtraining and cross-validation.
IEEE Trans Neural Networkers vol. 8, No. 5 pp. 985–996
Bishop, CM, 1995. Neural networks for pattern recognition. Oxford University Press
Freeman, James A. and Skapura, David M., 1999. Neural networks: Algorithms,
Applications and Programming Techniques, P
Bozorgmehry, R.B., F. Abdolahi and M.A. Moosavian, 2005. Characterization of basic
properties for pure properties for pure substances and petroleum fractions by neural
network. Fluid Phase Equilibr, 231: 188-196.
Freeman, J. and Skapura, D., 1991. Neural network algorithms, applications, and
programming techniques. Addison-Wesley, New York.
Gharbi, R. B. and Elsharkawy, A. M., 1997. “Neural-Network Model for Estimating the
PVT Properties of Middle East crude Oils,” SPE 37695, SPE Middle East Oil Show
Conf (Manamah, Bahrain, March15 –18) PROC V1, pp 151-166,
Gharbi, R. B. and Elsharkawy, A. M., 1997. “Universal Neural- Network Model for
Estimating the PVT Properties of Crude Oils,” SPE 38099, SPE Asia Pacific Oil &
Gas CONF (Kuala Lumpur, Malaysia, Apr. 14-16) PROC pp 619-628,
Hagan, M.T., H.B. Demuth and M.H. Beale, 1996. Neural Network Design. 1st Ed., PWS
Publishing Co., Boston, MA, USA.
Heaton, Jeff, 2005. Introduction to neural network with java, Heaton Research, Inc.
Iglewicz, B. and Hoaglin, D. C., 1993 How to Detect and Handle Outliers, American
Society for Quality Control, Milwaukee, WI
Krose, B. and Van der Smagt, P., 1996. An Introduction to Neural Networks, University
of Amsterdam.
156
Matignon, Randell, 2005. Neural network modeling using SAS enterprise miner, Auther
House.
Negnevitsky, M., 2005. Artificial Intelligence: A Guide to Intelligent Systems 2nd
Edition, Addison Wesley.
Rahman,A, Thulasiram, R. and Thulasiraman P. 2011. Performance analysis of
sequential and parallel neural network algorithm for stock price forecasting.
International Journal of Grid and High Performance Computing, 3(1), 45-68.
Richon, D., Laugier, S., 2003. Use of artificial neural networks for calculating derived
thermodynamic quantities from volumetric property data, Fluid phase equilibrium,
210, p. 247.
Sozen, A., Arcakilioglu, E., Ozalp, M., 2004. Investigation of thermodynamic properties
of refrigerant/absorbent couples using artificial neural networks, Chemical
engineering and processing, 43, p.1253.
StatSoft, Inc., 2011. Electronic Statistics Textbook (Electronic Version). Tulsa, OK: web:
http://www.statsoft.com/textbook/.
Talib, H. Saleh, J. Mokhtar, K and Ali, N. 2009. On the comparison of capacitance-based
tomography data normalization methods for multilayer perceptron recognition of
gas-oil flow patterns. Modern Applied Science Vol 3-1.
Varotsis, N., Gaganis, V., Nighswander, J., Guieze, P., 1999. A Novel Non-Iterative
Method for the Prediction of the PVT Behavior of Reservoir Fluids. SPE 56745.
Presented at the SPE ATCE, Houston, TX, October 3–6.
157
CHAPTER V
APPROACH
5.1 Introduction
The PVT properties of crude-oil such as bubble point pressure, the oil formation
volume factor, and dissolved gas-oil ratio play a key role in the calculation of reserves of
oil reservoirs as well as for identification of reservoir characteristics. The properties are
on samples collected from the bottom of the wellbore or after re-combining the liquid and
vapour samples collected from the separator at the surface. However, this process is
expensive and the costs are high. Therefore, researchers deduced many empirical
formulas and several ANN models which are used to calculate the PVT properties of
crude-oil.
provide a new method as an alternative technique with which to predict Pb, Bob, Rsob and
RST in the absence of PVT analysis. In this way, the dilemma that is associated with the
158
deficiencies resulting from these correlations are explained in detail in Chapters II and
III.
relatively new. In recent years, LS-SVM has rapidly developed into an important and
effective method of function approximation. However, the current literature review with
regard to the application of LS-SVM in the oil and gas industry is quite limited. In 2009,
flow. Peng and Yin (2010) applied LS-SVM to establish the void fraction measurement
models of oil-gas two-phase flow corresponding to different flow patterns. Cheng et al.
(2010) proposed an algorithm which combines Particle Swarm Optimization (PSO) with
Artificial Neural Networks (ANN) can learn the model data by self-adjusting their
parameters. Following training, ANN can accurately match with the expected data, and it
can quickly and accurately predict the unknown sample data. It can be said that one of the
most important features of ANN is its ability to discover the rules and patterns in the data
which general observation and many standard statistical methods cannot find especially,
when dealing with nonlinear problems. Many researchers recognized that through the use
of ANN, we can establish an appropriate model which can accurately predict the PVT
159
modeling schemes suffer from numerous drawbacks (Castillo et al., 2002 and Castillo et
hidden layers and the type of transfer function(s) for neurons in the
The Support Vector Machine (SVM) can be used for both regression and calcification
problems, just as with ANN. SVM is derived from statistical learning theory and was
found in 1995 by Vapnic et al. LS-SVM was proposed by Suykens et al. in 2002 and it is
the data space is transferred to a linear relationship in the feature space by means of
kernels. The model parameters, however, are given by the solution of a system of linear
defined as a trade-off between minimizing the training error and minimizing the
complexity of the LS-SVM model. The σ2 denotes the RBF kernel parameter and
represents the change in the RBF kernel function (width of kernel) (Pandit et al. 2011). It
160
is a positive real constant. Based on this advantage, the LS-SVM is introduced in this
chapter to investigate the capability of LS-SVR regression on modeling Pb, Bob, Rsob and
RST of crude oil systems in the absence of PVT analysis and to solve some of the above
ANNs limitations. The details of the LS-SVM algorithm can be found in Vapnik (1995),
Guo et al. (2006) and Chen et al. (2007). Appendix C provides a brief introduction as to
how to construct a LS-SVM model. The LS-SVM model can be expressed as:
= ∑ , + ! (5.1)
where the term K (x, xi) is the kernel function, xi is the input vector, αi is Lagrange
The commonly used kernel functions were linear kernel, polynomial kernel, sigmoid
kernel, Radial Basis Function (RBF) kernel. However, compared with other feasible
kernel functions, RBF can handle the nonlinear relationships between the spectra and
target attributes and is able to reduce the computational complexity of the training
procedure and give a good performance under general smoothness assumptions (Wang et
al. 2003). Thus, RBF kernel was recommended as the kernel function of the LS-SVM
This study employed the RBF as the nonlinear mapping kernel function. It is a
common function that is useful in nonlinear regression problems. However, the selecting
161
of γ and σ2 is the most significant topic when identifying a non-linear system with the LS-
SVM. The regularization parameter (γ ) determines the trade-off when minimizing the
training error. The parameter σ2 was the bandwidth, and implicitly defined the nonlinear
mapping input space to some high dimensional feature space. Consequently, the selection
can significantly affect model performance. A grid search method is used at first, and
then searching optimization is carried out employing a cross validation method until an
optimum parameter pair is obtained. The free LS-SVM toolbox (LSSVM V1.5, Suykens,
Leuven and Belgium) was applied to develop LS-SVM models. Appendix III introduces,
Several PVT empirical correlations and several ANN models are available in the
literature, but they suffer from a lack of application in the absence of laboratory analysis.
This shortcoming and the deficiencies resulting from the correlations are explained in
The main objective of this study is to overcome and then eliminate the dilemma by
building LS-SVM models using direct measured field parameters as input variables to
estimate Pb, Bob, Rosb, and RST. By incorporating the four proposed LS-SVM models,
engineers can estimate Pb, Bob, Rosb, and RST straightforwardly in the absence of PVT
analysis. The Pb and Bob LS-SVM models are proposed as a function of four directly
measured field parameters (Rsp, Psp, γoST and TR), whereas Rsob and RST LS-SVM models
162
are proposed as a function of only three direct measured field parameters (Rsp, Psp TSP and
γoST)
The LS-SVM parameters γ and σ2 must be appropriately selected for the RBF as the
kernel function. Immediately following the training process, the evaluation of the quality
and capability of the fitted model is applied. The prediction capability was mainly
assessed by correlation coefficients (R2) between the actual and the predicted outputs,
root mean-squared errors (RMSE), mean, minimum and maximum relative error (RE),
mean minimum and maximum Absolute Percent Relative Error (ARE), mean, minimum
and maximum of Absolute Error (AE), and Standard Deviation (SD). The best model is
the model with the highest R2 and the smallest other error statistical analysis. Appendix C
calibration model was developed based on the training dataset, validation dataset, and
testing dataset.
The PVT data used in this study were obtained from two-stage and single-stage flash
separation tests. A total of 118 reservoir fluid studies (439 data points) were collected
from various Libyan oil fields in the Sirte Basin. The majority of the data points are taken
from two-stage flash separation tests. In the single-stage separation test, the separator
pressure is atmospheric pressure and the stock-tank vent GOR value is equal to zero.
163
In order to build robust LS-SVM PVT models, the entire dataset was randomly
divided into three subsets; training set of 300 points, validation set of 70 points and
testing set of 69 points. Training and validation sets were used to build the LS-SVM
models, while testing sat was used to test the accuracy of the newly developed models
and not switched in the process of LS-SVM models derivation. The value ranges of
Four directly measured field parameters (Rsp, Psp, γoST and TR) are selected as inputs
to generate the Pb LS-SVM model. To obtain optimal LS-SVM model parameters, a grid
search is performed on the basis of leave-one-out cross-validation on the dataset. Grid search
techniques with 10-folds cross validation were employed to obtain the LS-SVM model
parameters (γ, σ2). The parameter combination (γ, σ2) was tuned simultaneously in a 10×10
grid. The contour plot of the optimization parameters γ and σ2 for the prediction of bubble
point pressure is shown in Figure 5.1. The grids ‘•’ in the first step are 10×10, and the
searching step in the first step is large. The natural logarithms of regularization parameter
γ and kernel parameter σ2 were tuned simultaneously. The optimal search area is
determined by the error contour line. The grids ‘×’ in the second step are 10 × 10, and the
searching steps in the second step are smaller. The optimal search area is determined
based on the first step. The optimal pair of (γ, σ2) for Pb model was found at the value of
277.88 [exp(5.627)] and 5.79 [exp(1.756)], respectively. The contour plot (Figure 5.1) of
the optimization error for LS-SVM regression model is Pb when optimizing the
parameters γ and σ2. The red square indicates the selected optimal settings.
164
After the LS-SVM model parameters are tuned, the model was developed and the
prediction performance was evaluated. The predicted results of the optimal Pb LS-SVM
regression ("#_#%&γ-./
''.)),σ* +.', ) model for training dataset are shown in Figure 5.2.
The figure shows all the points are scattered around the y = x line. It can be concluded
that the predicted bubble point values are in good agreement with the bubble point values
obtained from PVT analysis. The Root Mean Square Error (RMSE) is 148.4, Mean
Absolute Relative Error (MARE) is 8.09% and the Predicted Correlation Coefficient (R2)
is 98.27%. The statistical error analysis for the prediction capability, as obtained from the
LS-SVM model, is summarized in Table 5.2. As noted in Table 5.2, the proposed model
was statistically stable and fitted the data well. Moreover, the histogram of the error
distribution for the training dataset is presented in Figure 5.3. It confirms the error is
normally distributed.
165
Table 5.1 – Ranges of PVT Data
166
computing element [10 10] of [10 10]...
6
4
log(σ2)
0
0 1 2 3 4 5 6 7 8 9
log(γ)
Figure 5.1− Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Pb when Optimizing the Parameters γ [exp(5.627)] and σ2 [exp(1.756)]. The
Red Square Indicates the Selected Optimal Settings.
167
Figure 5.2− Performance of Training Dataset for Pb LS-SVM Model. Upper: Calculated
Pb vs. Data Observation Order. Lower: Calculated Pb vs. Experimental Values.
168
Table 5.2 − Statistical Error analysis of Pb LS-SVM Model
169
Histogram of the errors (Training dataset)
100
90 mean 0
SD 148.65
80
70
60
Frequency
50
40
30
20
10
0
-800 -600 -400 -200 0 200 400 600 800
Error
Figure 5.3− Error Distribution Resulting from the Training Dataset for Pb LS-SVM
Model
170
5.5.1 Accuracy of the Bubble Point Pressure LS-SVM Model
The final step in LS-SVM regression model building is to test its accuracy. A testing
dataset (69 data points) was used to scan the proposed Pb LS-SVM regression model’s
("#_#%&γ-./
''.)),σ* +.', ) model for testing the dataset is shown Figure 5.4. All data
points in the figure are scattered around the y = x line. The testing dataset provides Pb
with a RMSE of 275.55, a MARE of 14.214%, and a R2 of 93.84. The histogram of the
error term distribution confirms normal distribution, as clarified in Figure 5.5. Statistical
analysis of the error for the predicted capability on the testing dataset is summarized in
Table 5.2. Figure 5.6 shows the performance of the validation and testing datasets of the
LS-SVM model for the prediction of Pb. Both datasets have similar performances,
171
Figure 5.4− Performance of Testing Dataset for Pb LS-SVM Model. Upper: Calculated Pb
vs. Data Observation Order. Lower: Calculated Pb vs. Experimental Values.
172
Histogram of the errors (Testing dataset)
22
20 mean -19.77
SD 276.85
18
16
14
Frequency
12
10
0
-1000 -800 -600 -400 -200 0 200 400 600 800 1000
Error
Figure 5.5− Error Distribution Resulting From Testing Dataset for Pb LS-SVM Model
173
Plot of calculated pb vs. experement data
5000
4500
Bubblepoint pressure from PVT analysis
4000
3500
3000
2500
2000
Validation dataset
1500 y=x
1000
500
0
0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Bubblepoint Pressure from LS-SVM model
174
5.6 Bubble Point Oil Formation volume Factor, Bob LS-SVM Model
A LS-SVM model to calculate Bob in the absence of PVT analysis was developed
using 10×10 grid search is an error-surface spanned by the model parameters. In this grid
search, two steps are carried out. The first step was a search for crude with a large step
size presented in the form of ( • ) and the second step of the specified search is a small
step size, presented in the form of ( × ). The results of the two-step grid search are shown
in Figure 5.7. The optimal combination of (γ, σ2) was achieved with γ equalling 7142.508
and σ2 equalling 984.957. The Bob model was developed using the aforementioned
parameters and the prediction results for the training dataset are shown in Table 5.3. The
model provides Bob with a R2 of 97.11%, a MSE of 6.9*10-4, and a RMSE of 0.0263. The
in the training dataset is shown in Figure 5.8. The tightest cloud of points around the y =
x line indicates excellent agreement between the experimental and the calculated data
values. Moreover, the histogram of the error distribution for the training dataset in Figure
5.9 confirms the error is normally distributed with the mean being equal to zero.
Once the parameters of γ and σ2 were settled, there was no need to implement the grid
search step. The testing dataset can be used to inspect model accuracy. This inspection
illustrated in Table 5.3 and Figure 5.10. The figure shows the majority of the calculated
points fall too close to the y = x line. Furthermore, the histogram of the error terms in
175
Figure 5.11 show the errors are normally distributed with mean being equal to zero.
Figure 5.12 illustrates the scatter diagram of the modeled Bob using validation and testing
datasets versus experimental values. Again, this figure shows all the points of both
datasets scattered around the y = x line, indicating excellent agreement between the
When using LS-SVM, three crucial problems required solving, including the
determination of the optimal input feature subset, proper kernel function, and the optimal
kernel parameters. As mentioned previously, the RBF kernel was recommended as the
kernel function of the LS-SVM in this work. In addition, proper parameter setting played
a crucial role in building a good LS-SVM regression model with high prediction accuracy
and stability. Again, a two-step grid search technique with leave-one-out cross validation
was employed to obtain the optimal combination of (γ, σ2). Leave-one-out cross-
validation was used to avoid over fitting problems in the selection of optimal LS-SVM
limited range. In each iteration, 10 points are left out, and fits a model to the other data
points. The performance of the model is estimated based on the 10 points left out. This
176
computing element [10 10] of [10 10]...
9
X: 8.874
Y: 6.893
Z: 0.0008389
7
log(σ 2)
3
4 5 6 7 8 9 10 11 12 13
log(γ )
2 3 4 5 6 7
-3
x 10
Figure 5.7 - Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Bob when Optimizing the Parameters γ [exp(8.874)] and σ2 [exp(6.893)]. The
Black Square Indicates the Selected Optimal Settings.
177
Table 5.3 − Statistical Error Analysis of Bob LS-SVM Model
178
Function approximation using LS-SVMγRBF 2
=7142.5079,σ =984.9571
γ σ
1.9
1.8
1.7
Bubblepoint oil FVF, bbl/STB
1.6
1.5
1.4
1.3
1.2
1.1
1.8
Bubblepoint oil FVF from LS-SVM model
1.7
1.6
1.5
1.4
1.3
1.2
Training Data points
y=x line
1.1
1
1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
Bubblepoint oil FVF from PVT analysis
Figure 5.8− Performance of the Optimal Bob LS-SVM Model in Training Dataset
179
Histogram of the error (Training dataset)
60
Mean =0.0
50 SD =0.0263
40
Frequency
30
20
10
0
-0.1 -0.05 0 0.05 0.1
Error
Figure 5.9− Error Distribution Resulting from Training Dataset for Bob LS-SVM Model
180
Function approximation using LS-SVMγRBF 2
=7142.5079,σ =984.9571
γ σ
1.8
1.7
1.6
Bubblepoint oil FVF, bbl/STB
1.5
1.4
1.3
1.2
1.1
Testing Dataset
LS-SVM model
1
0 10 20 30 40 50 60 70
Obsewrvation order
1.7
Bubblepoint oil FVF from LS-SVM model
1.6
1.5
1.4
1.3
1.2
1.1
Figure 5.10− Performance of the Optimal Bob LS-SVM Model in Testing Dataset
181
Histogram of the errors (Testing dataset)
25
20
15
Frequency
10
0
-0.1 -0.08 -0.06 -0.04 -0.02 0 0.02 0.04 0.06 0.08 0.1
Error
Figure 5.11 –Error Distribution Resulting from Testing dataset for Bob LS-SVM Model
182
Plot of calculated Bob vs. experimental data
1.8
Testing dataset Validation dataset y = x line
1.7
1.6
from LS-SVM model
1.5
1.4
1.3
ob
B
1.2
1.1
1
1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8
B from PVT analysis
ob
183
The results of this two-step grid search are shown in Figure 5.13. The first step grid
search was for a crude search with a large step size, presented in the form of (•), and the
second step for the specified search with a small step size, is presented in the form of
(×).The optimal search area is determined by an error contour line. Following the grid
search process, the optimal combination of γ and σ2 would be achieved for the LS-SVM
models. Finally, all the different estimates of the performance are combined. For each
combination of γ and σ2 parameters, MSE was calculated and the optimum parameters
were selected which produced smaller MSE. The black square in Figure 5.13 indicates
the selected optimal combination of γ and σ2. The optimal pair of (γ, σ2) was found at the
values of 2936154.38 and 7865.36, respectively. In order to assess the relative robustness
of the proposed LS-SVM model, calculated Rsob was plotted against experimental values
in Figure 5.14. In the upper plot, the model estimating curve is very close to the actual
data, indicating that the LS-SVM model has a strong ability for regression analysis. The
lower plot illustrates the correlation between the LS-SVM model and the experimental
values. The correlation coefficient is very close to unity (R2 = 99.25%). Perfect accuracy
would result in the data points forming a straight line along the diagonal axis (y = x line).
5.7.1 Accuracy of the Bubble Point Solution Gas/Oil Ratio LS-SVM Model
The performance of the Rsob model was evaluated by 69 data points in the testing set.
The prediction results are summarized in Table 5.4. The R2 and RMSE of the testing set
were 98.73% and 28.7, respectively. Moreover, Figure 5.16 brings the calculated Rsob
using the proposed LS-SVM model against the experimental values. The sold line is the
184
regression line corresponding to the y = x. All the points are tight around this line
indicating excellent agreement between the experimental and the calculated Rsob values.
Notably, the errors are random and follow a normal distribution with zero mean, as
shown in Figure 5.17. However, if the mean is not zero, it might be that the model is not
the right choice for a particular data. Figure 5.18 shows the scatter points of both
validation and testing sets match the y = x line indicating no overfitting existed in the
model.
Usually, the RSP is a field measurement, while the RST is rarely measured in the field.
Thus, in order to obtain the Rsob indirectly, we must first estimate the RST, and add it to
the field measured RSP (Equation 2.1). In this work, a novel LSA-SVM model with
which to estimate RST was developed. PSP, TSP and γoST are the only independent
variables. A total of 364 two-stage separator tests were used to develop the RST LS-SVM
model. In order to build a robust RST LS-SVM model, 364 separator tests were randomly
split into three subsets, a training set which includes 246 separator tests while validation
and testing sets were 59 separator tests each. The value ranges of these sets are
185
computing element [10 10] of [10 10]...
10
9
X: 14.89
Y: 8.97
Z: 636.5
8
7
log(σ )
2
3
7 8 9 10 11 12 13 14 15 16
log(γ )
Figure 5.13 - Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Rsob when Optimizing the Parameters γ [exp(14.89)] and σ2 [exp(8.97)]. The
Black Square Indicates the Selected Optimal Settings.
186
Figure 5.14 − Performance of the Optimal Rsob LS-SVM Model in the Training Dataset
187
Table 5.4 − Statistical Error Analysis of Rsob LS-SVM Model
188
Histogram of the erro (Training dataset)
100
90
80
70
60
frequency
50
40
30
20
10
0
-150 -100 -50 0 50 100 150
Error
Figure 5.15 –Error Distribution Resulting from Training Dataset for Rsob LS-SVM Model
189
Figure 5.16 − Performance of the Optimal Rsob LS-SVM Model in the Testing Dataset
190
Histogram of the errors (Tisting dataset)
30
25
20
Frequency
15
10
0
-100 -80 -60 -40 -20 0 20 40 60 80 100
Error
Figure 5.17 – Error Distribution Resulting from Testing Dataset for Rsob LS-SVM Model
191
Plot of calculated Rsob vs. experimental data
1400
1000
Rsob from LS-SVM
800
600
400
200
0
0 200 400 600 800 1000 1200 1400
R sob from PVT analysis
192
A robust RST LS-SVM regression model with high prediction accuracy and stability is
built based on a proper parameter setting. The optimal LS-SVM parameters were
obtained by employing cross validation with a grid search. The optimal search area is
determined based on the first step. The optimal pair of (γ, σ2) was found at a value of γ =
47.6552 and σ2 = 1.3477, respectively. The optimizing process is shown in Figure 5.19.
The calculated RST, using the LS-SVM model, was plotted against the experimental
values in Figure 5.20. In the upper plot, the model estimating curve is very close to the
actual PVT data, indicating the LS-SVM has a strong ability for regression analysis. The
lower plot, illustrates the correlation between the LS-SVM model and the experimental
PVT values. Moreover, the histogram regarding the errors of the proposed model is
plotted in Figure 5.21. The error terms follow a normal distribution with a mean equal to
zero.
193
Table 5.5 – Ranges of PVT Data used in RST LS-SVM Model
194
computing element [10 10] of [10 10]...
5
2
log(σ 2)
1 X: 3.864
Y: 0.2984
Z: 484.8
-1
-2
-1 0 1 2 3 4 5 6 7 8
log(γ )
Figure 5.19 - Contour Plot of the Optimization Error of Cross Validation of RST when
Optimizing the Parameters γ [exp(3.864)] and σ2 [exp(0.298)]. The Black Square
Indicates the Selected Optimal Settings.
195
Function approximation using LS-SVMγRBF 2
=47.6552,σ =1.3477 γ σ
250
150
Calculated R
100
50
0
0 50 100 150 200 250
Observation order
150
ST
100
R
50
0
0 50 100 150 200 250
R from PVT analysis
ST
Figure 5.20 − Performance of the Optimal RST LS-SVM Model in the Training Dataset
196
Histogram of the error (Training dataset)
120
100
80
Frequency
60
40
20
0
-100 -80 -60 -40 -20 0 20 40 60 80 100
Error
197
5.8.1 Accuracy of the RST LS-SVM Model
In order to assess the relative robustness of the LS-SVM regression model for RST in
estimating Rsob, the calculated RST values are added to the field RSP (Equation 2.1) and
plotted against Rsob, as obtained from PVT analysis (Figure 5.22). The sold line, in Figure
5.21, is the regression line corresponding to the y = x. The figure shows all the points are
tight around this line indicating an excellent agreement between the experimental and the
calculated Rsob values. R2, RMSE and other error analysis are illustrated in Table 5.6. The
error analysis demonstrates a high R2 of 99.15% and a low RMSE of 15.76. Figure 5.23
shows the scatter points of the testing set match the y = x line indicating that no over
5.9 Conclusions
This chapter has described the LS-SVM model with which to develop a nonlinear
model to predict some PVT properties in the absence of experimental analysis. Four
robust LS-SVM models were built to calculate Pb, Bob, Rsob and RST. The models were
developed using the RBF kernel function. The model parameters (γ, σ2) were tuned by a
two-step grid search with a cross-validation process. The cross-validation process was
utilized to avoid the over fitting problem and to evaluate model performance during the
calibration stage. The models are obtained by selecting γ and σ2 parameters that give the
198
Table 5.6 − Error of Analysis of Rsob Obtained by Adding LS-SVM RST and Field RSP
199
Plot of calculated Rsob vs experimental data
1200
SP
+R
1000
γ =47.6552,σ =1.3477
Training dataset
2
800
using LS-SVM RBF
600
400
ST
=R
sob
200
R
0
0 200 400 600 800 1000 1200
R from PVT analysis
sob
200
Plot of calculated Rsob vs. experimental data
1000
900
SP
+R
800
γ =47.6552,σ =1.3477
700
2
600
using LS-SVM RBF
500
400
300
Tseting Dataset
ST
=R
200
sob
R
100
0
0 100 200 300 400 500 600 700 800 900 1000
R from PVT analysis
sob
Figure 5.23 − Performance of Rsob Calculated from RST LS-SVM Model using Testing
Dataset
201
5.10 References
An, S., Liu, W., Venkatesh, S.: Fast cross-validation algorithms for least squares support
vector machine and kernel ridge regression. Pattern Recogn. 40(8), 2154–2162 (2007)
Brabanter, J.D. et al.,2002. Robust cross-validation score function for LS-SVM non-
linear function estimation. International Conference on Artificial Neural Networks,
Madrid Spain, Aug. pp. 713-719.
Castillo, E. Fontenla-Romero, O. Betanzos, A. and Guijarro-Berdiñas, B. 2002. A global
optimum approach for one-layer neural networks. Neural Comput., 14 6, pp. 1429–
1449.
Castillo, E. Guijarro-Berdiñas, B. Fontenla-Romero, O. and Alonso-Betanzos, A., 2006.
A very fast learning method for neural networks based on sensitivity analysis. J.
Mach. Learn. Res., 7, pp. 1159–1182
Cheng, G. Guo, R. Wu, W., 2010. Petroleum lithology discrimination based on PSO-
LSSVM classification model. 2nd International Conference on Computer Modeling
and Simulation, ICCMS, Jan 22-24.
Chun-xiao Zhang, Taozhang 2009; LS-SVM model of oil holdup of oil-water two phase
flow using thermal method based on HHT, Proceedings of the Eighth International
Conference on Machine Learning and Cybernetics, Baoding, 12-15 July 2009
Deng, S. and T.H. Yeh, 2010. Applying least square vector machines to airframe wing-
box structural design cost estimation. Expert Systems with Applications, 37, 8417-
8423.
Duan, S.S. and Poo, 2003. A.N. Evaluation of simple performance measures for tuning
SVM hyperparameters. Neurocomputing, 50, pp. 41–59
Flecher, R., 1987 Practical methods of optimization, John Wiley & Sons.
Pandit,S., Mandal,C. Patra, A., 2011. A methodology for generation of performance
models for the sizing of analog high-level topologies. VLSI Design, Volume 2011,
Article ID
202
Patra, J. C. A. and Bos, V. D., 2000. Modeling of an intelligent pressure sensor using
functional link artificial neural networks. ISA Transactions, 39 15-27.
Pelckmans, K., et al., 2003. LS-SVMlab Toolbox User’s Guide,
<http://www.esat.kuleuven.ac.be/sista/lssvmlab/tutorial>
Peng, K.L. Wu, C.H. and Goo, Y.J., 2004. The development of a new statistical
technique for relating financial information to stock market returns. Int. J.
Manag., 21 4, pp. 492–505.
Peng, R.Z., 2007. Tool wear predictive model based on least squares support vector
machines, Zhejiang: Zhejiang University.
Peng, Zhenrui and Yin, Hong, 2010; ECT and LS-SVM Based Void Fraction
Measurement of Oil-Gas Two-Phase Flow Iran. J. Chem. Chem. Eng Vol. 29, No. 1
Scholkopf, B., Smola, A.J.: Learning with Kernels: Support Vector Machines,
Regularization, Optimization, and Beyond. MIT Press, Cambridge (2001)
Shi, D.F. and Nabil, N.G., 2007. Tool wear predictive model based on least squares
support vector machines. Mechanical Systems and Signal Processing, vol. 21, pp.
1799-1814.
Suykens, J.A.K. Gestel, T.V. Brabanter, J.D. and Vandewalle, J., 2002. Least square
support vector machines, World Scientific, Singapore.
Suykens, J.A.K., Van Gestel, T., De Brabanter, J., De Moor, B., Vandewalle, J.: Least
Squares Support Vector Machines. World Scientific, Singapore (2002)
Suykens, T.V. Gestel, J.D. Brabanter and Vandewalle, J., 2002).Least square support
vector machines, World Scientific, Singapore.
Van Gestel, T., Suykens, J.A.K., Baestaens, D.-E., Lambrechts, A., Lanckriet, G.,
Vandaele, B., De Moor, B., Vandewalle, J.: Financial time series prediction using
least squares support vector machines within the evidence framework. IEEE
Transactions on Neural Networks 12(4), 809–821 (2001)
Vapnic, V., 1995. The Nature of Statistical Learning Theory, Springer, New York.
203
CHAPTER VI
NETWORK APPROACH
6.1 Introduction
Insight into average reservoir pressure and its change in production (time), plays a
hydrocarbon in place, computing rock and fluid characteristics, reservoir material balance
calculations, pressure maintenance projects, surface and subsurface facility designs and
which measures long-term built-up pressure when the producing wells are shut in. The
test is started after the well is opened to a constant flow rate over an extended period of
time. This will cause the reservoir pressure to drawdown due to the withdrawing of fluids
from the well. Afterward, the well is shut-in to allow the pressure to build up. During the
period of time, the build-up pressure reaches average reservoir pressure. This time period
depends mainly upon the reservoir permeability and thus, may take a prolonged period of
204
time in low permeability reservoirs. The resulting pressure build-up is then analysed to
Currently, the only available method with which to obtain average reservoir pressure is to
conduct an extended build-up test. It must then be evaluated using Horner (Horner, 1951)
pressure changes as fluids (oil, gas and water) are released from the reservoir. However, a
significant economic impact is caused by shutting in the producing wells during the entire
build-up test.
ANNs are relatively new computational tools which are extensively utilized to solve
many complex engineering problems. ANNs can assist engineers and researchers by
problems that conventional computing has been unable to solve. Petroleum engineers
may benefit from ANNs on occasions when engineering data for designs and
interpretations are less than adequate, Mohaghegh (1995). However, a review of the
literature reveals that ANNs has been applied successfully to different branches of
Ardjmandpour et al. (2011), Popa et al. (2011), and Lu and Fleming (2011).
A basic ANNs has three layers of processors - an input layer, an output layer and one
hidden layer between them. Each layer has a number of nodes. ANNs and their learning
205
algorithms and backpropagation (BP) algorithms have been studied extensively in the
literature by Bulsari (1995), Veelenturf (1995), Gropal (1998), Mehrotra et al. (2000),
Basheer and Hajmeer (2000) and Rabunal and Dorado (2006). A brief overview of the
A neural network model has been established in this study to map relationships which
control reservoir oil, gas and water production performance in order to interpolate,
predict and estimate the current average reservoir pressure without closing the producing
wells. This method is suitable for constant and variable flow rates. After interpolating,
predicting and/or obtaining the current average reservoir pressure, it can be utilized in
problem in a way that will be addressable by neural networks. The selection of input
variables for the ANNs is the key problem in this approach. In other words, the first step
in any modeling design procedure consists of reducing the dimension of the input space.
Thus, our task is to find a set of inputs that are strongly correlated to the output.
The overall performance of oil reservoirs is largely determined by the nature of the
energy (reservoir pressure) available to move the oil to the surface. In petroleum
driving mechanism may be defined as natural forces in the reservoir that displace
reservoir fluids (oil, gas and water) from the reservoir to the producing wells and up to
206
the surface. Mainly, there are three common driving mechanism forces (water drive, gas
cap drive and solution gas drive) or a combination of these forces which provide the
natural energy necessary for oil recovery. Each reservoir is composed of a unique
drive mechanisms. Although no two reservoirs are identical in all aspects, they can be
grouped according to the drive mechanism by which they produce. It has been observed
that each drive mechanism has certain typical performance characteristics in terms of
Usually, in water drive reservoirs, the reservoir pressure declines very gradually
depend on the aquifer efficiency. The reason for the minimal decline in reservoir pressure
is that oil and gas withdrawals from the reservoir are replaced, almost volume for
volume, by water encroaching upon the aquifer into the oil zone. Many reservoirs are
bounded on a portion, or all of their peripheries, by water bearing rocks called aquifers.
The aquifers may appear to be larger than the reservoir they adjoin, and thus, appear
infinite for all practical purposes. The aquifers may diminish to the point of being
negligible in their effects on reservoir performance. However, in gas cap reservoirs, due
to the ability of the previously existing (above the oil) gas cap to expand, the reservoir
pressure falls slowly and continuously. The actual rate of pressure decrease is related to
the size of the gas cap when compared to the oil volume. On the other hand, the reservoir
pressure in the solution gas drive declined rapidly and continuously. This reservoir
pressure behavior is attributed to the fact that no external fluids or gas caps are available
to provide a replacement to the oil withdrawals. When oil is produced from the reservoir,
207
reservoir pressure decreases and dissolved gas bubbles emerge from the oil. This gas
expands and pushes the reservoir oil to the wellbore and up to the surface.
Accordingly, the decline of average reservoir pressure, at a given time, depends upon
reservoir pressure can be correlated as a function of the oil, gas and water production
rates (production history) and the number of producing wells. Figures 6.1, 6.2 and 6.3
illustrate the impact on the decline of an average reservoir pressure according to oil, gas
and water production rates. Figure 6.4 shows the impact on declining average reservoir
pressure according to the number of producing wells. Based on the above discussions,
four types of data (oil, gas and water production rates and the number of producing wells)
are selected to incorporate into the input layer of the network model.
A Libyan oil field, located in the Sirte Basin, was utilized in this study to develop a
model with which to interpolate, predict and estimate current average reservoir pressure.
The field, consisting of 60 naturally flowing oil production wells, began production in
January, 1970. The daily production rates were 15,884 BPD of oil, 30 BPD of water and
2,621 M SCF/D of gas. The initial reservoir pressure was 1188 psi and the saturation
pressure was 267 psi. The most recent average reservoir pressure of 832 psi occurred in
May 2009. As of May 2009, the daily field production rates were 11,877 BPD for oil,
81,915 BPD for water and 1,941M SCF/D for gas. A total of 49 average reservoir
pressure data points were recorded during more than 30 years of production (January
1979 to May 2009). The range of the dataset is illustrated in Table 6.1.
208
140,000 1250
1200
120,000
100,000
Average Reservoir Pressure
1100
80,000 1050
60,000 1000
950
40,000
900
20,000
850
0 800
0 10 20 30 40 50
Observation order
Figure 6.1 – Impact on the Decline of an Average Reservoir Pressure According to the
Oil Production Rate. (unpublished data, personal communication)
209
25,000 1250
1200
20,000
1150
1000
10,000
950
900
5,000
850
0 800
0 10 20 30 40 50
Observation order
Figure 6.2 – Impact on the Decline of an Average Reservoir Pressure According to the
Gas Production Rate (unpublished data, personal communication)
210
160000 1250
1150
1100
100000
1050
80000
1000
60000
950
40000
900
20000 850
0 800
0 10 20 30 40 50
Observation order
Figure 6.3 – Impact on the Decline of an Average Reservoir Pressure According to the
Water Production Rate (unpublished data, personal communication)
211
1250
60
1200 Average Reservoir Pressure
Producers
1150
Average reservoir pressure, psi
50
1100
40
Producers
1050
1000 30
950
20
900
10
850
800 0
0 10 20 30 40 50
Observation order
Figure 6.4 – Impact on the Decline of an Average Reservoir Pressure According to the
Number of Producers (unpublished data, personal communication)
212
6.3.2 Data Preparation
In order to evaluate the model prediction performance after it was properly trained,
the last two recorded average reservoir pressures (June 2007 and May 2009) were put
aside and not switched in the process of model derivation. The remainder of the data
points (47 points) was used to build the network. To avoid over fitting (overtraining), it is
while the training error steadily decreases, then a situation of over fitting may occur.
When the performance of the validation test ceases to improve, the algorithm halts.
213
Table 6.1- Range of Overall Data (unpublished data, personal communication)
Producers 60 12 40
214
Table 6.2 – Range of Data Used in Training Process
215
Accordingly, the training data points (47 points) are split randomly into three sets
which are: training set, 70%, validation set, 15% and testing set, 15%. The validation set
(7 points) is used to learn how to decide when to stop and the testing set (7 points) is used
to evaluate how well an ANN performs on previously unidentified data. The ranges of the
After identifying the model inputs and data preparation, the next step is to specify the
number of hidden layers and nodes in each layer. The usual strategy is based solely on
trial and error. The attempt began with one hidden layer and one hidden node. Hidden
nodes were gradually added. A total of 310 architectures were compared to the network
parameters, as presented in Table 6.3. Table 6.4 shows the five best architectures from
the compared 310 architectures. Before selecting the best architecture, the optimal
training algorithm should be determined. Once more, the usual strategy is based solely on
experience. However, the architectures presented in Table 6.4 are trained by the most
common three used training algorithms, namely, Conjugate Gradient Descent BP (CGD),
case, the architecture was retrained five times with different initial weight
randomizations. However, one method to avoid local minima is to retrain the network
with a different weight initialization (Freeman and Skapura, 1991). This will locate the
MSE function at different minima in the error surface, and then choose the acceptable
solution.
216
Table 6.3 - Network Architecture Parameters
217
Table 6.4 - Top Five Architectures - The best results are emphasized in bold
Error, psi
Architecture R2
Testing Training
[4-10-1] 11.664 3.1229 99.46
[4-8-3-1] 12.765 8.8507 97.92
[4-3-4-1-1] 13.769 13.739 96.30
[4-8-1-4-1] 11.272 16.468 92.1
[4-7-6-4-1] 14.187 14.313 94.07
218
As a result, a total of 15 models were compared and the results are tabulated in Tables
6.5 through 6.9. The tables present the network performances of the testing dataset in
terms of AE, ARE, and SE. The best results from each training algorithm group are
emphasized in bold. Consequently, the Architecture [4−10−1] provides the best results
The proposed reservoir pressure ANN model architecture is multilayered with four
input nodes and one hidden layer with ten nodes, [4−10−1]. The network was trained
algorithm. The neurons in the backpropagation used a logistic function as an input and an
The proposed model provides low minimum errors in training, validation, testing
and the overall dataset which is illustrated in Figures 6.5, 6.6, 6.7 and 6.8, respectively.
The Figures show all the points are scattered around the y = x line, indicating excellent
agreement between the actual and estimated average reservoir pressure points. The
statistical parameters for the prediction capability, as obtained from the network, are
summarized in Table 10. The proposed network provides prediction values of average
reservoir pressure with an average AE of 11.028 psi and an average ARE of 1.194%,
indicating the model describes the data very well. Moreover, the error distribution for the
training and overall datasets is presented in Figures 9 and 10 and both figures show the
219
Table 6.5 - Error Analysis of Architecture [4-10-1] under Various Training Algorithms
Training
Criterion AE, psi ARE% SE, (psi)2
Algorithms
Average 19.28 2.0168 693.4
SD 17.94 1.8705 880.2
CGD
Minimum 0.50 0.0596 0.2552
Maximum 44.62 5.0595 1991.4
Average 27.4 2.7605 1859.1
SD 33.29 3.2673 3643.9
QN
Minimum 0.070 0.0079 0.0049
Maximum 99.33 9.9929 9866.3
Average 34.46 3.6487 2465.2
SD 35.74 3.6823 4197.5
LM-QN
Minimum 3.34 0.3799 11.18
Maximum 107.54 10.8187 11564.3
220
Table 6.6 - Error analysis of Architecture [4-8-3-1] under Various Training Algorithms
Training
Criterion AE, psi ARE% SE, (psi)2
Algorithms
Average 27.18 2.864 1159.8
SD 20.52 2.1441 1510.7
CGD
Minimum 2.76 0.3263 7.6383
Maximum 59.66 6.2367 3559.3
Average 27.53 2.7476 2085.7
SD 36.44 3.581 4394.6
QN
Minimum 3.21 0.364 10.30
Maximum 108.78 10.9442 11834.36
Average 44.36 4.7589 3414.6
SD 38.03 3.9501 4846.6
LM-QN
Minimum 1.225 0.1446 1.5
Maximum 116.5 11.7204 13572
221
Table 6.7 - Error Analysis of Architecture [4-8-3-1] under Various Training Algorithms
Training
Criterion AE, psi ARE% SE, (psi)2
Algorithms
Average 38.83 4.1392 767622
SD 37.3 3.7969 217796
CGD
Minimum 7.90 0.9328 572074
Maximum 118.82 11.9539 1237794
Average 43.98 4.4257 3849
SD 43.75 4.1608 6425.6
QN
Minimum 3.66 0.4229 13.41
Maximum 129.73 13.0513 16830
Average 27.77 2.9179 1890.5
SD 33.45 3.342 4314.8
LM-QN
Minimum 2.656 0.3067 7.052
Maximum 108 10.8647 11663
222
Table 6.8 - Error Analysis of Architecture [4-8-1-4-1] under Various Training Algorithms
Training
Criterion AE, psi ARE% SE, (psi)2
Algorithms
Average 44.29 4.6882 4751.7
SD 52.8 5.3733 9418
CGM
Minimum 1.140 0.1346 1.3
Maximum 160.42 16.1389 25734.7
Average 33.61 3.4476 2566.5
SD 37.90 3.7487 5446.9
QN
Minimum 6.188 0.7146 38.29
Maximum 121.81 12.2546 14837.8
Average 50.44 4.9274 5963.8
SD 58.48 5.2843 10207
LM-QN
Minimum 5.82 0.6727 33.94
Maximum 159.71 14.0344 25508
223
Table 6.9 - Error Analysis of Architecture [4-7-6-4-1] under Various Training Algorithms
Training
Criterion AE, psi ARE% SE, (psi)2
Algorithms
Average 23.4 2.4445 1043.8
SD 22.28 2.2648 1794.7
CGD
Minimum 2.90 00.3426 8.42
Maximum 70.47 7.0899 4966.5
Average 34.68 3.6191 2100
SD 29.9 3.1814 2688
QN
Minimum 2.714 0.3205 7.367
Maximum 84.12 9.4099 7077
Average 28.26 2.9021 2068.4
SD 35.63 3.6052 4278
LMQN
Minimum 0.042 0.0048 0.0018
Maximum 107.92 10.8577 11648
224
Scatterplot of Model Output vs Actual Aaverage Reservoir Pressure
Training Dataset
1200
R2 = 97.49%
1120
960
880
800
800 880 960 1040 1120 1200
Actual average reservoir pressure, psi
Trainig Dataset
1200
Actual Average rese rvoir Pressure
Average rReservoir pressure, psi
Model Output
1100
1000
900
800
3 6 9 12 15 18 21 24 27 30 33
Observation order
Figure 6.5 – Actual Average Reservoir Pressure versus Modeled ANN Values for
Training Dataset
225
Scatterplot of Model Output vs Actual Average Reservoir Pressure
Validation Dataset
1100
1050 R2 = 96.16%
1000
Model output
950
900
850
800
800 850 900 950 1000 1050 1100
Actual average reservoir pressure, psi
Validation Dataset
1100
1000
950
900
850
1 2 3 4 5 6 7
Observation order
Figure 6.6 – Actual Average Reservoir Pressure versus Modeled ANN Values for
Validation Dataset
226
Scatterplot of Model Out vs Actual Average Reservoir Pressure
1200
R2 = 94.38%
1100
Model output
1000
900
800
800 900 1000 1100 1200
Actual Average reservoir pressure, psi
Testing Dataset
1200
Actual Average Reservoir Pressure
Average reservoir pressure, psi
Model Output
1100
1000
900
800
1 2 3 4 5 6 7
Observation order
Figure 6.7 – Actual Average Reservoir Pressure versus Model ANN Values for Testing
Dataset
227
Scatterplot of Model Output vs Actual Average Reservoir Pressure
Overall Dataset
1200
R2 = 96.67%
1100
Model output
1000
900
800
800 900 1000 1100 1200
Actual average reservoir pressure, psi
Overall Dataset
1200
Model Output
1100
1000
900
800
1 31 61 91 121 151 181 211 241 271
Observation order
Figure 6.8 – Actual Average Reservoir Pressure versus Modeled ANN Values for Overall
Dataset
228
Table 6.10 – Statistical Error Analysis of the Proposed Model
Training Dataset
AE, psi ARE%
Average 9.404 1.041
SD 10.063 1.140
Minimum 0.056 0.006
Maximum 45.705 5.067
Validation Dataset
AE, psi ARE%
Average 10.9 1.136
SD 11.686 1.135
Minimum 1.253 0.149
Maximum 38.972 3.8472
Testing Dataset
AE, psi ARE%
Average 19.278 2.017
SD 17.939 1.870
Minimum 0.505 0.06
Maximum 44.625 5.059
Overall Dataset
AE, psi ARE%
Average 11.028 1.194
SD 12.224 1.314
Minimum 0.0559 0.006
Maximum 45.705 5.067
229
Histogram of the Error
Training Dataset
12
Mean 1.023
SD 14.19
10 N 33
8
Frequency
0
-60 -40 -20 0 20 40 60
Error
230
Histogram of the Errors
Overall Dataset
14
Mean 0.3363
SD 16.88
12 N 47
10
Frequency
0
-60 -40 -20 0 20 40 60
Error
231
6.5 Estimation of Average Reservoir Pressure
analytical tool with respect to new data. In the majority of cases, the reservoir pressure is
not obtainable over long periods of time for several economic reasons. In this particular
case, during more than 30 years of production, there are just 49 reservoir pressure
readings available. In such a case, interpolation techniques are used to estimate the
data points. Figure 6.11 illustrates the results of the ANN interpolation technique. Figure
6.10 can be used to estimate the average reservoir pressure at any time. Furthermore, the
proposed model can be used to estimate current average reservoir pressure and to predict
average reservoir pressure. Two unknown points, (June 2007 and May 2009) which were
put aside and not switched in the process of model derivation, were used to evaluate the
proposed model performance. Table 6.11 shows the prediction performance of the
proposed model. The network provides average reservoir pressure values with an average
232
1250
1200
Average reservoir pressure, psi
1150
ANN Model Actual Forecast
1100
1050
1000
950
900
850
800
750
Jan-70 Jun-75 Dec-80 Jun-86 Nov-91 May-97 Nov-02 May-08
Time, month
233
Table 6.11 – Model Prediction Performance
234
6.6 Conclusions
• A significant economic effect occurs when shutting in producing wells during the
entire pressure build-up test. This study presents a new method with which to
predict and interpolate the average reservoir pressure without closing the
producing wells. This technique is suitable for constant and variable flow rates.
• The network consisted of one hidden layer with ten neurons. The input layer
received four parameters consisting of oil, gas and water production rates and a
• The results indicate average reservoir pressure can be predicted using neural
network models without the wells having to be closed. Nevertheless, the neural
network model has the ability to predict and interpolate average reservoir pressure
235
6.7 References
Alimadadi et al. 2011. Using a committee machine with artificial neural networks to
predict PVT properties of Iran crude oil. SPE Reservoir Evaluation & Engineering,
Vol. 14, No. 1
Ardjmandpour, N. et al. 2011. Artificial neural network forward modelling and inversion
of electrokinetic logging data. Geophysical Prospecting, Vol. 59, Issue 4, pages 721–
748, July 2011
Basheer, I.A. and Hajmeer, M., 2000. Artificial neural networks: fundamentals,
computing, design, and application. Journal of Microbiological Methods, 43:3–31.
Bulsari, A. B., 1995. Neural Networks for Chemical Engineers. Elsevier, Amsterdam.
Elmabrouk, S., Mayorga, R. and Shirif, E. 2010 “Artificial Neural Network Modeling for
Reservoir Oil Production Prediction.” Paper presented at 5th TOG, Tripoli, Libya
Nov.12-14
Gropal, Suchatita, 1998. Artificial Neural Networks for Spatial Data Analysis.
http://www.ncgia.ucsb.edu/giscc/units/u188/u188.html
Horner, D.R., "Pressure Build-up in Wells", Proc., 3rd World Petr. Cong., E.J. Brill,
Leiden (1951) Vol. II, p. 503.
Lu, Qin and Fleming, Graham C. 2011. Gas-Lift Optimization Using Proxy Functions in
Reservoir Simulation. SPE 140935 paper presented at SPE Reservoir Simulation
Symposium, 21-23 February 2011, The Woodlands, Texas, USA
Mehrotra, K. , Mohan, C. , and Ranka, S., 2000. Elements of Artificial Neural Networks.
Cambridge, MA: The MIT Press.
Miller, C.C., Dyes, A.B., Hutchinson, C.A., Jr., "The Estimation of Permeability and
Reservoir Pressure from Bottom Hole Pressure Buildup Characteristics", Trans.,
AIME (1950) 189, 91-104.
Popa, A. Cassidy, S. and Mercer, M. 2011. A data mining approach to unlock potential
from an old heavy oil field. SPE 144470 paper presented at SPE Western North
American Region Meeting, 7-11 May 2011, Anchorage, Alaska, USA
236
Rabunal, J.R. and Dorrado, J. 2006. Artificial Neural Networks in Real-life
Applications. Idea Group Publishing.
Veelenturf, L.P.J., 1995. Analysis and applications of artificial neural networks. Prentice
Hall International.
237
CHAPTER VII
RESERVOIR PRESSURE
7.1 Introduction
(BP-ANN) was established in Chapter VI to interpolate, predict and estimate the current
average reservoir pressure (Pavg). In this chapter a Genetic Algorithm (GA) was
employed for the evolution of connection weights in the BP- ANN model, as proposed in
Chapter VI, in order to improve upon model performance and support the simultaneous
weights of ANN using the GA instead of local search algorithms, including a gradient
descent algorithm. They suggested that global search techniques, including the GA, might
prevent ANN from falling into a local optimum (Gupta & Sexton, 1999). Thus, a
combination of Artificial Neural Network (ANN) and the Genetic Algorithm (GA) based
optimization, termed the “ANN-GA Model” was proposed to interpolate, predict and
238
7.2 Genetic Algorithm
GA is a search algorithm, based upon the survival of the fittest among string
structures, used to form a search algorithm (Goldberg, 1989). With respect to a solution
for optimization problems, GA has been investigated recently and shown to be effective
mechanisms of reproduction, crossover and mutation (Adeli and Hung, 1995). GA may
avoid falling into minimum when a local search method such as gradient descent is used
form of binary or real numbers. During the connection weights evolution, the architecture
The backpropagation (BP) algorithm and sigmoid function are used in this model as
weights to train the ANN and are optimized by the GA rather than the gradient descent
algorithm. The BP learning algorithm is a widely used algorithm but it has a drawback of
239
with respect to complex problem space and can be used as a powerful optimization tool.
2. Use the training and validation sets to design a suitable neural network
The construction and generic operation of the GA for FFNN is as follows: Given a
(chromosomes) is generated, each of which codifies a set of values for the weights of the
connections and biases of the neural network. Then the fitness of each individual is
evaluated, which entails allocating the values of the weights and biases codified by the
individual in order to assess the network connections and calculate the mean square error.
The set of training patterns are used to conduct this calculation. The error value is the
240
individual's fitness. Following the initialization stage, the genetic reproduction, crossover
and mutation operators are usually applied to output new generations until there is
convergence toward a population of individuals that encode the set of weights which
Training BP-ANN with GA is divided into two parts: the first part is to optimize the
weights with GA, and the second part is to train the BP-ANN with the result of GA. The
multilayered feedforward BP-ANN, designed in Chapter VI, is used in this study. Four
inputs (oil, gas and water production flow rates, and the number of producing wells) and
one hidden layer with ten nodes, [4−10−1] were chosen to interpolate, predict and
estimate the current average reservoir pressure. The number of hidden nodes is
determined using a trial and error approach. The main goal of this study is to establish a
more accurate ANN model to interpolate and estimate current average reservoir pressure
minimum when a local search method is used in the ANN model, as proposed in Chapter
VI. Learning in a neural network involves modifying the weights and biases of the
network in order to minimize the cost function. The cost function always includes an
error term (a measure of how close the network's predictions are to the training set).
The field data described in Table 6.2 was used to build the BP-ANN-GA model.
Again, the last two recorded average reservoir pressures (June 2007 and May 2009) were
put aside to use in the prediction process. However, they not switched in the process of
241
model derivation. The remaining data points were split randomly into three sets: training,
validation and testing datasets. The validation and testing sets consist of 7 points while
Based on trial and error, as explained in Chapter VI, the BP-ANN model architecture
is chosen to be multilayered with four input nodes and one hidden layer with ten nodes,
[4−10−1]. The neurons in the backpropagation used a logistic function as an input and
output activation function. The only difference between the model proposed in Chapter
VI and this model is that this model was trained using feedforward backpropagation with
a hybrid GA algorithm to locate the global minimum of the MSE rather than the
ANN is initialized with random weights using GA with each weight being between −1.0
to +1.0. This study requires two sets of vectors of weights. The first vector set consists of
connection weights between inputs to the hidden layer. The second vector set consists of
connection weights between the hidden layers to the output layer. This study uses 4 input
features and 10 hidden elements. The weights in the ANN are encoded in such a way that
each weight is between −1.0 to +1.0. At this point, weights are assigned to each link. The
network is trained with the training data to calculate the fitness of each chromosome and
the mean square error (MSE) was returned. In this method, GA searches several sets of
weight vectors simultaneously. Unlike the ANN model presented in Chapter VI, the
242
algorithm method. In this application, each string or chromosome in the population
represents the weight and bias values of the network. The initial population is randomly
crossover, probability of mutation, initial population, etc., for the GA, high efficiency and
performance can be achieved. The objective function is minimization of the MSE. The
considered fitness function is the minimum MSE and it is computed by recalling the
network. After achieving the fitness values of all chromosomes, they are ranked based on
the best fitness values. Half of the best ranked population is selected for the production of
off springs for the next generation. This half population undergo cross over with cross-
over probability. Again, this will be mutated to give a new offspring, with mutation
probability, which is combined with the selected best population to form a new
population for the next generation. Table 7.1 shows the PB-ANN-GA parameters used in
this study.
Basically, the performance of weight evolution using GA depended upon the number
of populations and generations. If the parameters are set too low, the evolution may
generations would require a longer computation time for convergence. In this study, the
number of individuals in the population was chosen as 100 and the number of generations
used to evolve the solution is 20. Meanwhile, the Generational GA evaluates the fitness
of the current population before creating an entirely new population to evaluate. The
probability of crossover on average is chosen (0.7) and the mutation rate is chosen
(0.033).
243
Eventually, the chromosome with the minimum error from the last generation of the
GA algorithm was chosen as the initial weight of BP-ANN to conduct BP-ANN training
according to the field data. The field average reservoir pressure and modeled values of
BP-ANN-GA are compared. It shows the mean AE is 7.517 psi and mean ARE is 0.817%
in the training dataset with a regression coefficient of 98.42%. The validation dataset
provides a regression coefficient of 96.7%, whereas the testing dataset provides 94.38%.
Table 7.2 shows the statistical error analysis of the modeled average reservoir pressure
for training validation and testing datasets. Figure 6.1 shows the field reservoir pressure
versus the modeled output for a given training, validation and testing data points.
The two unknown points, (June 2007 and May 2009, which were put aside and not
switched in the process of model derivation, were used to evaluate the proposed model
prediction performance. The oil, gas and water production rates and the number of
producing well values between June 2007 and May 2009 are used as the input vector to
the trained BP-ANN-GA model and then the output average reservoir pressure was
matched to the field values. The modeled values matched the field average reservoir
7.6 Conclusion
in Chapter VI), a new BP-ANN model, trained with GA (BP-ANN-GA), was established.
244
was introduced in this chapter. The algorithm was applied to the same oil field data used
BP-ANN (Chapter VI), this case study demonstrates its application with the results
similar results when testing and prediction data sets are used.
245
Table 7.1 – BP-ANN-GA Model Parameters
246
Figure 7.1 – Modeled Versus Field Average Reservoir Pressure
247
Table 7.2 – Statistical Error Analysis of the Proposed Model
Training Dataset
AE, psi ARE%
Mean 7.517 0.817
SD 7.942 0.869
Validation Dataset
AE, psi ARE%
Mean 10.4 1.122
SD 9.368 0.979
Minimum 0.923 0.108
Maximum 28.305 2.794
Testing Dataset
AE, psi ARE%
Mean 19.278 2.017
SD 17.94 1.87
Minimum 0.505 0.06
Maximum 44.625 5.059
248
Table 7.3 – Model Prediction Performance
Time Field Pavg, psi ANN-GA Pavg, psi AE, psi ARE %
June-07 820 830 10 1.219
May-09 832 832 0 0
249
7.7 References
Adeli, H. and S. Hung, 1995. Machine Learning: Neural Networks, Genetic Algorithms
and Fuzzy Systems. John Wiley and Sons Inc., New York, ISBN: 0-471-01633-0, pp:
211.
Davis, L., 1991. Handbook of Genetic Algorithms. 1st Edn., Van Nostrand Reinhold,
New York, ISBN-10: 0442001738
Goldberg, D.E., 1989. Genetic Algorithms in Search optimization and Machine Learning.
Addison Wesley and Longman Publishing Co., Inc., Boston, MA., USA.
Gupta, J. N. D., & Sexton, R. S., 1999. Comparing backpropagation with a genetic
algorithm for neural network training. Omega, 27(6), 679–684.
Shanthi,D. Sahoo,G. and Saravanana,N., 2009. Evolving Connection Weights of
Artificial Neural Networks Using Genetic Algorithm with Application to the
Prediction of Stroke Disease. International Journal of Soft Computing Vol.4 -2 pp.
95-102
250
CHAPTER VIII
APPROACH
8.1 Introduction
Insight into average reservoir pressure, and its change over time, plays a critical role
in reservoir development. In particular, almost every oil well intervention requires insight
with respect to average reservoir pressure. However, in order to determine the average
reservoir pressure, the well is shut-in, resulting in loss of production. This is regarded as a
pressure build-up test. In high permeability reservoirs, this may not be a significant issue,
but in medium to low permeability reservoirs, the shut-in period during the entire build-
up test may last several weeks before a reliable reservoir pressure can be estimated. This
loss of production opportunity, as well as the cost of monitoring the shut-in pressure, is
often unacceptable. It is of great practical value if the average reservoir pressure can be
obtained from the historical production and reservoir pressure data without having to
shut-in the well. It is clear that the production rate of a well is a function of many factors
such as permeability, viscosity, thickness etc. Also, the rate is directly related to the
driving force in the reservoir, i.e. the difference between the average reservoir pressure
251
Chapter V clarifies how the ANNs have the ability to estimate and interpolate average
reservoir pressure with good accuracy. The purpose of the work in this chapter, however,
Recently Support Vector Machine (SVM) has garnered support in replacing ANN
because of their predominance in global optimization but they are only one solution. LS-
SVM is a new type of SVM which was introduced by Suykens et al. (2002). LS-SVM
uses equality constraints instead of inequality constraints and a least squares error term in
order to obtain a linear set of equations in the dual space. However, they all face the same
underlying question - how to set the key parameters which may affect the performance of
the algorithm. In the LS-SVM average reservoir pressure model, a two-step grid search
with cross-validation on the training dataset was used to search for the best parameter γ
and σ2.
The historical monthly reservoir production data used in this study, from March 1997
to July 2010, is related to an oil reservoir located in the Sirte Basin, Libya. This oil
reservoir which has 60 naturally producing wells began production in January, 1970.
Generally, during the life of an oil reservoir, cumulative production is recorded, and
of average reservoir pressure were obtained during the time period from January 1970 to
May 2009, as illustrated in Table 8.1. In order to evaluate the LS-SVM model prediction
252
performance after it was properly trained, the last two recorded average reservoir
pressures (June 2007 and May 2009) were put aside and not switched in the process of
model derivation. The ranges of the dataset are illustrated in Table 6.1
253
Table 8.1 - Average Reservoir Pressure Data (unpublished data, personal
communication); Testing Dataset is emphasized in Bold
254
An issue of concern in evaluating the approximation ability of a predicting model is
to ensure that it has no over fitting on the training dataset and to obtain good model
performance. In general, as with any LS-SVM problem, if the model parameters (γ and
σ2) are not well selected, the predicted results will not be sufficient and the generalization
ability will be poor. Tuning the model parameters is usually completed by minimizing the
estimated generalization error. The generalization error is a function that measures the
generalization ability of the constructed models, i.e., the ability to predict correctly the
used to estimate the generalization error. Therefore, the 47 data points are separated into
two sets, called “the training set” (40 points) and “the testing set” (7 points). The testing
data points are emphasized in bold in Table 8.1. The LS-SVM model is constructed using
only the training set and is then tested using the test dataset. The test data are completely
unknown to the estimator. The errors it makes are accumulated to give the mean test set
To obtain satisfactory LS-SVM model performances, two model parameters (γ, σ2)
must be carefully chosen. The higher level parameters are usually referred to as “hyper
parameters”. In general, in any LS-SVM problem, if the hyper parameters of the model
are not well selected, the predicted results will not be adequate and the generalization
ability will be poor. Tuning of the hyper parameters is usually conducted by minimizing
the estimated generalization error. The generalization error is a function that measures the
generalization ability of the constructed models, i.e., the ability to predict correctly the
255
performance of an unknown sample. In the present methodology, the holdout technique is
The holdout method is the simplest method of cross validation. The data set is
separated into two sets, called the “training set” and the “testing set”. The LS-SVM is
constructed using only the training set. Then, the LS-SVM is asked to predict the output
values for the data in the testing set. This testing dataset is completely unknown to the
model (it has never observed the output values). The errors it makes are accumulated as
before in order to give the Mean Square Error (MSE), which is used to evaluate the
model.
As described in Chapter VI, the decline of average reservoir pressure in oil reservoirs,
problems. As a result, the average reservoir pressure can be correlated as a function of the
oil, gas and water production rates (production history) and the number of producers.
Accordingly, four inputs are selected to be the inputs to the LS-SVM average reservoir
pressure model. The LS-SVM regression model with RBF kernel function was
established. In this method, the two parameters of regularization parameter (γ) and the
kernel parameter (σ2) of RBF are crucial to the performance of the LS-SVM model. The
former determines the trade off between minimizing the training error and minimizing
model complexity and the latter is the bandwidth and implicitly defines the nonlinear
The input layer has four nodes (oil, gas and water production rates and the number of
producers) and the output layer has one node (average reservoir pressure). The RBF is
256
selected as a kernel function. A selected training sample set is used to train the LS-SVM.
Because of the variables in the collected data, each has different levels of magnitudes
from each other. Both the input and output data are normalized before the training (all
23 4567 23
′ =
589 23 4 567 23
(7.1)
where xi is the value of the ith dimension of the data, min(xi) and max(xi) are the
The main steps of the modeling process of the LS-SVM average reservoir pressure
are as follows:
Step 1: select proper data and split it into training and testing groups.
Step 4: set an initial guess for regularization parameter (γ) and kernel
parameter (σ2)
Step 6: apply the third level of Bayesian inference. This will estimate the
parameter σ2.
Step 7: tune the γ and σ2 obtained from Steps 5 and 6 using a two-grid search
with cross-validation on the training dataset. The parameter set (γ, σ2)
257
Step 8: Use a training dataset train with the LS-SVM having the tuned values
of γ and σ2
Step 10: test the LS-SVM model using the testing dataset. Calculate MSE,
RMSE and R2
Step 11: use the monthly production historical data from Mach 1970 to Jun
Step 12: use the monthly production data from July 2006 to July 2010 as inputs
Two-step grid searching and leave-one-out cross validation was applied to the
training dataset in order to realize the global optimization. The main idea behind this
method is to find the optimal parameters that minimize the prediction error of the
regression LS-SVM model. The prediction error can be estimated by leave-one-out cross-
validation. Cross validation was employed with respect to the training data of this work.
The LS-SVR model is carried out using the LS-SVMlab1.5 Toolbox, developed by
The four reservoir parameters (oil, gas and water production rates and a number of
producing wells) were employed as the input of the LS-SVM model. Figure 8.1 presents
the process of optimizing the LS-SVM method parameter γ and σ2 and in which the
symbol ‘•’ denotes the grid dot during the first search while the ‘×’ denotes the grid dots
258
in second time period. The optimal combination of γ and σ2 for average reservoir
pressure LS-SVM model was obtained at the values of 17.0586 and 4.1287, respectively.
The error term between calculated and measured average reservoir pressure values
was obtained and plotted as a histogram in Figure 8.2. The histogram shows the error
term is normally distributed with zero mean. Figure 8.3 shows the scatter plot of the
observation average reservoir pressure versus the output from training LS-
>?@
SVM;'.1+)<,=* 0.)' model. The model shows a correlation coefficient (R ) of 97.29%
2
259
computing element [10 10] of [10 10]...
5
3
log(σ )
2
2 X: 2.837
Y: 1.418
Z: 937.1
-1
0 1 2 3 4 5 6 7 8 9
log(γ )
Figure 8.1 - Contour Plot of the Optimization Error (MSE) for LS-SVM Regression
Model of Calculating Average Reservoir Pressure. The Black Square Indicates the
Selected Optimal Settings of γ [exp(2.837)] and σ2 [exp(1.418)].
260
8.4 Interpolating Average Reservoir Pressure
unknown values from data observed at known locations or times. After the LS-SVM
model is trained, interpolations are applied to estimate the missing average reservoir
pressure values on the basis of the remaining known sample points. In order to analyze
the interpolation quality, the errors between observed average reservoir pressure and the
interpolated values during the same period of time are calculated. The results show that
the MSE is 1510 psi, RMSE is 38.86 psi and the correlation coefficient is 83.6%. Figure
8.4 illustrates the accuracy of the LS-SVM model on the testing dataset. The plot shows
all the points remain close to the y = x line, indicating no over fit exists in the model.
Figure 8.5 can be used as a tool to interpolate the missing average reservoir pressure
values. The statistical error analysis for training testing and all data points are illustrated
in Table 8.2.
The two points of June 2007 and May 2009 (Table 8.1) are used as the input of the LS-
SVM model, and then forecasting is carried out. The model shows excellent forecasting
precision as illustrated in Table 8.3. The model provides forecasting average reservoir
261
Histogram of the Error
Training Dataset
18
16
14
12
Frequency
10
0
-50 -40 -30 -20 -10 0 10 20 30 40 50
Error
Figure 8.2 − Histogram of the Errors of the LS-SVM Average Reservoir Pressure Model
262
Scatterplot of LS-SVM model output vs. actual average reservoir pressure values
1200
1150
1100 R2 = 97.29%
1050
Model output
1000
950
900
850
Figure 8.3 − Performance of the Optimal LS-SVM Model in the Training Dataset
263
Performance of LS-SVM model
1200
1150
1100
Average reservoir pressure, psi
1050
1000
950
900
800
800 850 900 950 1000 1050 1100 1150 1200
Observation order
Figure 8.4 − Performance of the LS-SVM Model. Both training and Testing Dataset are
Scattered around the y = x Line indicating No Overfit Existing in the Model
264
1250
1200
Average reservoir pressur, psi
1150
1100
LS-SVM Actual Forecast
1050
1000
950
900
850
800
750
Mar-70 Aug-75 Feb-81 Aug-86 Jan-92 Jul-97 Jan-03 Jun-08 Dec-13
Time, month
Figure 8.5 − LS-SVM Model Performance. The Plot can be used to interpolate the
missing average reservoir pressure values.
265
Table 8.2 − Statistical Error Analysis of the Proposed Model
266
Table 8.3 – LS-SVM Model Forecasting Performance
267
8.6 Conclusions
Shutting in the wells during the entire pressure build-up test means a loss of income. The
predict and interpolate average reservoir pressure without shutting in the producing wells.
This technique is suitable for constant and variable flow rates. The study is based on
historical production and reservoir pressure data for an oil reservoir located in Sirte
Basin, Libya. The reservoir contained a set of 60 oil wells. There are just 49
measurements of average reservoir pressure from the period of March 1970 to July 2010
Oil, water and gas production rates and a number of producers are used as inputs for
the LS-SVM prediction model. The optimal model parameters (γ, σ2) are obtained by
applying a two-grid search with cross-validation on the training dataset only. The results
indicate the LS-SVM approach has the ability to interpolate and predict current average
268
8.7 References
269
CHAPTER IX
PRODUCTION PREDICTION
9.1 Introduction
management. Traditionally, numerical simulations and decline curve analysis have been
used to predict a reservoir’s future performance based on its current and past
performance.
under various operating strategies. A set of reservoir parameter values are placed in the
simulator to obtain a set of fluid flows. They are listed as a time-series over a specified
period of time. This time-series is compared with historical data to evaluate their
differences. If they are not in agreement, the input parameters are modified and the
between the simulated flow outputs and the historical production data. This history
matching is very time consuming and offers a non-unique solution with a high degree of
uncertainty.
Decline curve analysis fits the observed production rates of individual wells, groups
270
by extrapolating the declining function. The basis of decline curve analysis is to match
past production performance with a model assuming production continues to follow the
historical trend. Most decline curve analysis techniques are based on the empirical Arps
exponential, hyperbolic and harmonic equations (Arps 1954). In some cases, production
decline data do not follow a particular model but crossover the entire set of curves
(Camacho and Raghavan 1989). In addition, decline curve analysis does not take into
account opening or closing intervals and varying water or gas injection rates.
In recent years, there has been a steady increase in the application of Artificial Neural
Networks (ANNs) in engineering. ANNs have been used to address several fundamental
and specific problems that conventional computing problems have been unable to solve,
especially when the data for engineering design, interpretations and calculations are less
than adequate. Lately, neural network was found to be a very suitable technology to solve
problems in petroleum and gas engineering. Most recently, Elshafei et al. (2011)
presented two neural networks to predict the performance of the multi-stage gas/oil
separation plants in crude oil production. The neural networks accept the initial and final
pressures and temperatures of each stage and then attempt to utilize the oil composition
information to predict the stage gas/oil ratio. Recently, Elmabrouk et al. (2010)
established a neural network model that is able to map certain relationships that control
reservoir pressure without shutting in the wells. However, the current literature with
271
performance of oil wells with respect to spatial variation and time. The network contains
one hidden layer with 10 neurons, an input layer containing 11 neurons and one bias
neuron. The 11 inputs include the number of wells, the X and Y coordinates of the wells,
the cumulative production at time t−2, the cumulative production at time t−1, the
shut in switch, the average distance to surrounding wells, average cumulative production
of surrounding wells at time t and the cumulative production days at time t. The output
layer contains one neuron representing the cumulative production at time t+1. By
comparing the two cases, the results indicate an adequate capacity for short term
predictions.
Nguyen and Chen (2005) used Decision Support System (DSS) to model production
by incorporating curve fitting and Artificial Neural Network (ANN). DSS is a computer-
based system designed to model data and qualify decisions. According to the results,
Nguyen and Chen concluded no one model provides a consistently reliable prediction.
This is due to a weakness in the DSS and assumptions based on expert opinion. A major
system weakness is an assumption that the production rate is proportional to the initial
production.
General Knowledge, Optimal Search and Gols) to build a time series model to predict oil
production in petroleum reservoirs. They studied several models with which to forecast
production curves by applying historical data from Bonsucesso and Carmopolis Brazilian
oil fields. The study shows the Optimal Search approach worked very well and the Gols
272
approach gave reasonable results, while Specific Knowledge and General Knowledge
which to predict oil reservoir production performance taking into account the number of
production and injection wells in service and the varying water injection rates. Historical
production data from a Libyan oil field located in the Sirte Basin was used to train the
network. The training network is able to serve as a practical and robust reservoir
The most popular ANN model is the architecturally trained Multi-Layer Perceptron
weighted connections. The first layer consists of the input vector and the last layer
consists of the output vector. The intermediate layers, called “hidden layers”, represent
neural pathways and modify the input data through several weighted connections.
There are three major phases to network training with backpropagation. During the
initial phase, the input vector is presented to a network which is activated via the forward
pass. This generates a difference (error) between the input of the network and the desired
output. During the second phase, the computed output error propagates back through the
network (error backward pass). During the third phase, connection weights are corrected
by feeding the sum of the squared errors from the output layer back, through the hidden
273
layers, to the input layer. This process is repeated until the connection weights produce an
output which is within a predetermined tolerance of the desired output (Ripley 1996).
of trial-and-error (Heaton 2008). Thus, several networks with various (a) numbers of
hidden units, (b) training algorithms and (c) activation functions are attempted and a
generalization error is estimated for each. The network with the minimum estimated
An oil field in the Sirte Basin of Libya has 11 oil wells and 3 water injection wells.
The field has been in production since September, 1978 when it initially had an average
production rate of 3700 BOPD, an initial average reservoir pressure of 2960 psig and a
bubble point pressure of 2710 psig. Following 7 months of production, the average
reservoir pressure declined to 2665 psig. This rapid decline in reservoir pressure indicates
a limited aquifer size and very weak water influx. Pressure maintenance by water
injection was initiated in April, 1982 when 3 wells were converted from oil production
wells to injection wells and placed on the water injection system. As of December, 2009
the reservoir had produced a cumulative oil production of 38.5 MMSTB, and a
cumulative water production of 19.75 MMBW with an average oil production rate of
2316 BOPD and an average water cut of 38%. Figure 9.1 illustrates the reservoir’s oil
production performance.
274
Reservoir Oil Production, POPD
20000
18000
16000 Beginning
14000 of water
12000 flooding
Training data interval
10000
8000 Prediction
6000 interval
4000
2000
0
0 50 100 150 200 250 300 350 400
Time, month
Figure 9.1− Reservoir Oil Production Performance from Sept. 1978 to July 2010
(unpublished data, personal communication)
275
9.2.2 Data Pre-Processing
It is not possible to expect to achieve a good model from insufficient data. Therefore,
before training the network, the oil field production data are checked and pre-processed
to avoid peculiar answers from the trained network model. However, depending upon the
problem, there might be special features in the data that could be used in testing data
quality. One way to check for quality is to view graphical representations of the data in
areas. As presented in Figure 9.1, the oil field production data are time-dependent and
incorporate atypical patterns. Such patterns lead to the deterioration of the model
performance. Thus, the primary production period (the end of March 1982) in which
water injection was started and the period from April 1982 to January 1995 were
excluded as illustrated in Figure 9.1. The remainder of the production data points were
then split into two sets before starting to specify the model architecture. The first set
(Feb. 1995 to July 2009) was used to build the network model. The second set (Aug.
2009 to July 2010) was used to predict the average reservoir oil production rate.
To avoid over fitting and improve the network model generalization, the first dataset
was subdivided randomly into training, validation, and testing sets. The training set
(67%) is used to compute gradient and update the network weights and biases. The
validation set (16.5%) is used to check network performance during the training process.
Training can be stopped when the performance of the model on the validation dataset
gives a minimum error. The testing set (16.5%) is used to fine-tune models. It is not
276
architecture, select the appropriate model and assess the performance (generalization) of
a fully-specified model.
The key problem in our approach is the determination of how to select as few inputs
as possible. In other words, our task is to find an effective set of inputs to capture the
main features of our prediction problem. The proposed network was based on practical
experience and the necessity to map relationships that control past oil reservoir
production performance in order to predict production levels, while taking into account
the number of production and injection wells in service and the varying water injection
rates. The chosen network has five inputs and one output. The inputs are the average
reservoir oil production rate at time t, the average reservoir gas production rate at time t,
the water injection rate at time t+1, the number of oil wells in production at time t +1 and
the number of injection wells at time t+1. The output is the average reservoir oil
production rate at time t+1. The selection of an optimum network model can be achieved
using a trial-and-error approach. Accordingly, the best results were obtained by a network
model consisting of 2 hidden layers, 13 nodes in the first layer and 8 in the second layer.
The nodes in the hidden and output layers are activated with a logistic function then
trained by the Quasi-Newton algorithm which provides a backpropagation error with the
lowest sum-squared-error.
Figure 9.2 shows a comparison between the network outputs of the average reservoir oil
rate versus the actual average reservoir oil rate for each training data point. The average
277
Absolute Error (AE) of the network output is 381.2 BOPD and the average Absolute
Relative Error (ARE) is 10.41%. The network provides results which are very close to the
actual average reservoir oil rates. This indicates good agreement between the actual and
observed average reservoir oil rates. The performances of the training, validation and
testing datasets are presented in Figures 9.3, 9.4 and 9.5, respectively. The statistical
parameters used to measure the network’s prediction capability are obtained from the
training, validation and testing datasets and summarized in Table 9.1. The proposed
network provides prediction values for oil reservoir rates with average AEs of 391
BOPD, 324 BOPD and 392 for the training, validation and testing datasets, respectively.
Moreover, average AREs of 10.3%, 9.2% and 12% were obtained from the training,
validation and testing datasets, respectively indicating the network describes the data very
well.
The second dataset was used to predict reservoir oil production so the robustness and
accuracy of our network approach could be studied with respect to predicting oil
reservoir production as illustrated in Figure 9.6. The predicted reservoir oil rate values
agree with the historical values, indicating the training network can serve as a practical
robust reservoir production management tool. The network provides reservoir oil rates
278
9000
7000
6000
5000
4000
3000
2000
0
0 20 40 60 80 100 120 140 160 180
Time, month
Figure 9.2− ANN Model Output versus the Actual Reservoir Production (1st Dataset)
279
9000
7000
6000
5000
4000
3000
2000
0
0 20 40 60 80 100 120 140 160
Time, month
280
7000
Validation Dataset
6000
Reservoir Oil Production, BOPD
5000
4000
3000
2000
0
0 20 40 60 80 100 120 140 160
Time, month
281
8000
Reservoir Oil Production , BOPD
7000
Testing Dataset
6000
5000
4000
3000
2000
1000 Historical data Model
0
0 20 40 60 80 100 120 140 160
Time, month
282
Table 9.1 – Error Analysis of the Proposed Model
Training
Average 3922 3920 391 10
SD 1347 1224 351 8
Minimum 1832 2310 1.42 0.02
Maximum 8195 6644 1602 39.7
Validation
Average 3926 3960 325 9
SD 1224 1260 374 10
Minimum 1931 2374 1.84 0.04
Maximum 6250 6251 1680 39.4
Testing
Average 3556 3811 392 12
SD 1210 1263 303 9
Minimum 1985 2465 9.6 0.25
Maximum 7038 6759 1163 31.5
283
5000
Reservoir oil Production, BOPD
500
0
162 164 166 168 170 172 174 176
Time, month
Figure 9.6− Prediction of average reservoir oil production (one-year-ahead from Aug.
2009 to Jul. 2010)
284
9.4 Conclusions
This study provides a new, alternative method with which to predict oil reservoir
network technique can be used to predict oil reservoir performance, and can serve as a
285
9.5 References
Arps, J.J. (1945). Analysis of Decline Curves. Trans. AIME 160, 228-247.
Camacho, V. and Raghavan, R. (1989). Boundary-dominated flow in solution-gas drive
reservoirs. SPERE, 503- 512.
Da Silva, L.C.F., Portella, R.C.M. and Emerick, A.A. (2007). Predictive data-mining
technologies for oil-production prediction in petroleum reservoir. Paper SPE 107371
Presented at the SPE Latin American and Caribbean Petroleum Engineering
Conference, Buenos Aires, April 5-18.
Elmabrouk, S., Shirif, E. and Mayorga, R. (2010). A neural network approach to predict
average reservoir pressure. 5th Technology of Oil and Gas Forum. October 12-14,
Tripoli, Libya.
Elshafei, M., Khoukhi, A. and Abdulraheem, A. (2011). Neural network aided design of
oil production units. IEEE 10-13 pp. 638 – 641.
Heaton, Jeff (2008). Introduction to Neural Networks for Java, 2nd Edition. Heaton
Research, Inc. St. Louis, 143-170.
Nguhen, H.N. and Chan, C.W. (2005). Applications of data analysis techniques for oil
production prediction. JPSE 37, 153-169.
Ripley, B.D. (1996). Pattern Recognition and Neural Networks. Cambridge University
Press.
Zhong He, Yang, L., Yen, J. and Wu, C. (2001). Neural network approach to predict well
performance using available field data. Paper SPE 68801 Presented at SPE Western
Regional Meeting held in Bakersfield, California, March 26-30.
286
CHAPTER X
10.1 Introduction
The art of managing oil and gas reservoirs has progressed over the years. Water
injection has proven to be one of the most economical methods of reservoir management.
Water injection in reservoir engineering refers to the method wherein water is injected
into the reservoir. The technology can be valuable in helping to maintain a pressure that
continues to move oil to the surface. Reservoir simulation models are applied to forecast
geological model. Once the oil field enters into the production stage, many changes take
place in the reservoir and thus, the changes must be reflected in the reservoir model.
Usually, simulator programs are used to generate a model which simulates the reservoir.
In reservoir simulation, the simulator takes a set of reservoir parameter values as inputs
and returns a set of fluid flow information as outputs. The outputs are usually a time
series over a specific period of time. This time series is then compared with the historical
production data in order to evaluate their match. Experts modify the input model
parameters and rerun the simulation model. This process (history matching) continues
287
until the model output rates match the historical production rates. History matching is the
production data. The process of history matching is an inverse problem. In this problem,
a reservoir is a black box with unknown parameter values. Since inverse problems have
no unique solution, a large number of models, with good history matching, can be
achieved. In other words, in a reservoir simulation process, more than one combination of
inputs of reservoir parameter values gives the same output. As a result, history matching
input parameters
• Inverse problem (non-unique solution), where more than one model can
• Time consuming
the performance of oil production within water injection reservoirs. In this approach,
historical production and injection data are used as inputs. The approach can be
categorized as a new and rapid method with reasonable results. Another application of
this model is that it can be utilized to find the most economical scenario of water
injection to maximize ultimate oil recovery. This method could be a new window of fast
simulators. It has reasonable accuracy, requires little data and can forecast quickly.
288
10.2 Description of the Production and Injection Data
The historical production and injection data have been obtained from a small oil field
located in the Sirte Basin of Libya. The oil field initially had 14 oil wells. It has been in
production since September, 1978 when it initially had an oil production of 3700 BOPD,
an initial average reservoir pressure of 2960 psig and a bubble point pressure of 2710
psig. Following 7 months of production, the average reservoir pressure declined to 2665
psig. This rapid decline in reservoir pressure indicates a limited aquifer size and very
weak water influx. Pressure maintenance by water injection was initiated in April, 1982
when 3 wells were converted from oil production wells to injection wells and placed on
the water injection system to maintain the reservoir pressure that continues to move oil to
the surface. The field’s operations were shutdown in February 1999 and returned to
production in October 1999. Oil and gas production with water injection rates were used
from February 1995 to July 2010 in order to design the LS-SVM regression model. The
As a first step, the field production and injection dataset are split into two subsets,
namely, the “training dataset” and the “testing dataset”. The training set was enclosed the
first 162 months (February 1995 to July 2009). This set was used to train the model and
tune the model parameters γ and σ2. The testing set was enclosed the last 12 months
The selection of input features for the LS-SVM is the key problem in this approach.
In other words, the first step in any modeling design procedure is selecting the most
289
impact input parameters and at the same time reducing the dimension of this input space.
Thus, our task is to find a set of inputs that are strongly correlated to the output.
impact on the oil production are selected. Every reservoir, however, is unique because of
the myriad of geological and fluid dynamic characteristics and thus, the production can
vary drastically from one reservoir to another. These variations make it difficult to simply
predict the amount of oil it will require to produce from a particular reservoir. The second
minimizing the estimated generalization error. The generalization error is a function that
measures the generalization ability of the constructed models. The above mentioned field
data were used to create two LS-SVM models with which to predict a field oil production
rate under water injection. Thereinafter, the best forecasting performance will be chosen
to predict the performance of oil production under water injection reservoirs and to find
the most economical scenario of water injection to maximize ultimate oil recovery. The
two models employed Radial Basis Function (RBF) as a kernel function. The
regularization parameter (γ ) and the RBF kernel parameter (σ2) are tuned with two-step
A two-step grid search technique with 10-fold cross validation on the training dataset
was employed to obtain the optimal combination of (γ, σ2) for both LS-SVM models.
range. During each iteration of this algorithm, 10 points are left and fit a model on the
other data points. The performance of the model is estimated based on the 10 points left
290
out. This procedure is repeated for each 10 points. The first step grid search was for a
crude search with a large step size, presented in the form of (•), and the second step for
the specified search was a small step size, presented in the form of (×). The optimal
search area is determined by an error contour line. Following the grid search process, the
optimal combination of γ and σ2 would be achieved for the LS-SVM models. Finally, the
different estimates of the performance are combined. For each combination of γ and σ2
parameters, MSE was calculated and the optimum parameters were selected when they
This model has six inputs (oil and gas production rates at time t, water injection rate
at time t, producers at time t, producers at time t+1, water injection rate at time t+1) and
one output (oil production rate at time t+1). The training data set was used to tune the
model parameters while the testing dataset used to analyze the forecasting model
capability.
and σ2 are 34.1142 and 67.3738 respectively. The parameter combination (γ, σ2) was
tuned simultaneously in a 10×10 grid. Figure 10.1 shows the contour plot of optimization
the LS-SVM parameters. The grids ‘•’ in the first step is 10×10, and the searching step in
the first step is large. The natural logarithms of regularization parameter γ and RBF
kernel parameter σ2 were tuned simultaneously. The optimal search area is determined by
the error contour line. The grids ‘×’ in the second step is 10 × 10, and the searching steps
in the second step are smaller. The optimal search area is determined based on the first
291
step. The optimal pair of (γ, σ2) for LS-SVM oil production forecasting model is indicated
After the LS-SVM model parameters are tuned, the model was developed and the
regression for training dataset are shown in Figure 10.2 and 10.3. The history matching is
presented in Figure 10.2. It can be conclude that there is a good match between the
modeled and field oil flow rates. In addition, the cross plot (Figure 10.3) illustrates the
agreement between the field and modeled oil production rates. It shows that the
regression coefficient (R2) is 89.05% mean absolute relative error 8.93%. The statistical
error analysis for the reservoir oil forecasting capability, as obtained from the LS-SVM
model, is summarized in Table 10.1. As can be seen from the proposed model is
292
computing element [10 10] of [10 10]...
10
7
log(σ 2)
5 X: 3.53
Y: 4.21
Z: 2.675e+005
2
-1 0 1 2 3 4 5 6 7 8
log(γ )
4 5 6 7 8 9 10 11 12 13 14
5
x 10
Figure 10.1 - Contour Plot of the Optimization Error for LS-SVM Regression M1 when
Optimizing the Parameters. The red square indicates the selected optimal Settings of γ =
exp(3.53) & σ2 = exp(4.21)
293
function estimation using LS-SVMγRBF 2
=34.1142,σ =67.3738
γ σ
8000
7000
LS-SVM model Reservoir oil production
Reservoir oil production, BOPD
6000
5000
4000
3000
2000
1000
0 20 40 60 80 100 120 140 160 180
Time, months
294
Scatter plot of LS-SVM output vs. reservoir oil production
8000
7000
6000 LS-SVMRBF 2
γ =34.1142,σ =67.3738
LS-SVM model output
5000 R2 = 89.05%
4000
2000
1000
1000 2000 3000 4000 5000 6000 7000 8000
Reservoir oil production
Figure 10.3 - Cross Plot of Modeled Oil Production Rates vs. Field Values for M1
295
10.4.1 Prediction Performance of M1
forecasting tasks. The last 12 months of field production period from August 2009 to July
2010 (testing dataset) is used as the input of the trained LS-SVM regression model and
the perfection is carried out. The flowchart of load forecasting is shown in Figure 10.4.
The correlation between the predicted oil production rates and the field values is
illustrates in Figure 10.5. The plot explains the performance of the model prediction
process. It provides forecasting oil production rates with a RMSE of 404.78 BOPD and
mean ARE of 11.58.214%. The statistical analysis of the error for the predicted capability
features subset was selected, as a now prediction LS-SVM model (M2). The input
features of M2 are oil production rate at time t-1 and t, gas production rate at time t, water
injection rate at time t and t+1, producers at time t and t+1. The target is oil production at
time t+1. Thus, seven input features of LS-SVM with RBF kernel function used in
296
Figure 10.4 - Flow Chart of Load Forecasting
297
Cross plot of LS-SVM model output vs. reservoir oil production
4400
4200
4000
3800
LS-SVM model output
3600
LS-SVMRBF 2
γ =34.1142,σ =67.3738
3400
3200
R2 = 20.48%
3000
2600
2600 2800 3000 3200 3400 3600 3800 4000 4200 4400
Reservoir oil production
Figure 10.5 – Cross Plot of Predicted Oil Production Rates vs. Field Values for M1
298
A two-step grid search technique with 10-fold cross validation was employed to
obtain the model parameters (γ, σ2). The result of the optimal search area is determined
by error contour line in Figure 10.6. After the process of grid search, the optimal
combination of γ and σ2 would be achieved for the LS-SVM models. The red square in
the Figure 10.6 indicates the selected optimal combination of γ and σ2. The optimal pair
of (γ, σ2) was found at the value of 1341.4872 and 252.5918 respectively.
In order to assess the relative robustness of this model, the calculated oil production
rates of training dataset were plotted against the field values in Figure 10.7. It shows the
modeled oil flow rates are matching the field data. Figure 10.8 illustrated the correlation
between the LS-SVM model output and the field oil production rates. The model
provides oil rates with a correlation coefficient R2 of 90.56%, a RMSE of 402.08 BOPD
and a mean ARE of 8.22%. The statistical accuracy analysis of the proposed model is
summarized in Table 10.1. It shows stable model and fit the data well.
299
computing element [10 10] of [10 10]...
9
7
log(σ 2)
3
-1 0 1 2 3 4 5 6 7 8 9
log(γ )
4 5 6 7 8 9 10 11 12
5
x 10
Figure 10.6 - Contour Plot of the Optimization Error for M2 Model when Optimizing the
Parameters. The red square indicates the selected optimal settings of γ = exp(7.201) & σ2
= exp(5.532)
300
10.5.1 Prediction Performance of M2
The model’s prediction performance is evaluated by employed the testing dataset. This
dataset was used as input for the trained LS-SVM model then the predicted oil production
rates matched up to the field values. Figure 10.9 illustrated the forecasting accuracy
performance. The model provides forecasting values of oil production rates with a mean
301
function estimation using LS-SVMγRBF 2
=1341.4872,σ =252.5918
γ σ
8000
6000
5000
4000
3000
2000
1000
0 20 40 60 80 100 120 140 160 180
Observation order
302
Cross plot of LS-SVM model output vs. reservoir oil production
8000
7000
RBF
6000 LS-SVM =1341.4872, 2
γ σ =252.5918
2
R = 90.56%
LS-SVM model
5000
4000
2000
1000
1000 2000 3000 4000 5000 6000 7000 8000
Reservoir oil production
Figure 10.8 – Correlation between Modeled and Field Oil Flow Rates for M2
303
Cross plot of LS-SVM model output vs. reservoir oil production
4400
4200
4000
LS-SVM model prediction
3800
3400
3200 RBF
LS-SVMγ =1341.4872,σ2=252.5918
γ σ
2
R = 29.85%
3000
2800
2600
2600 2800 3000 3200 3400 3600 3800 4000 4200 4400
Reservoir oil production
Figure 10.9 – Forecasting Oil Production against Field Values for M2 Model
304
Table 10.1 − Statistical Error Analysis of M1 and M2 LS-SVM Models
LS-SVM M1 LS-SVM M2
Statistical Error
6 inputs 7 inputs
Analysis
Training Forecasting Training Forecasting
R2 , % 89.05 20.48 90.59 29.85
MSE, (BOPD)2 188150 163520 161670 144530
RMSE, BOPD 433.76 404.78 402.08 380.17
SD, BOPD 435.00 355.03 403.34 347.36
Mean AE, BOPD 326.76 358.28 299.3 333.43
Max AE, BOPD 1543 631.57 1601 635.2
Min AE, BOPD 3.582 49.8 1.38 20.29
Mean ARE, % 8.93 11.58 8.22 10.79
Max ARE, % 49.25 23.06 43.72 23.19
Min ARE, % 0.078 1.255 0.031 0.511
305
Absolute relative error %
30 Model 1 Model 2
25
20
15
10
0
1 2 3 4 5 6 7 8 9 10 11 12
Months
Figure 10.10 − Illustrates the Comparison of ARE between the Two Models in
Forecasting Data Domain
306
10.6 Results and Discussions
Figure 10.10 illustrates the comparison of ARE between the two models in
forecasting data domain. The final results of forecasting precision among these models
can be judged based on MSE, RMSE and MARE. The lowest MSE, RMSE and MARE
are found with M2 in training and forecasting stages. The model M2 provides prediction
oil flow rates with a MSE of 144530 BOPD, a RMSE of 380 BOPD and a MARE of
10.79 %. Table 10.1 summarizes the assessment results for both models.
10.7 Conclusions
This study applied a novel approach to predict oil production in oil fields under water
injection. Because the selection of input features for the LS-SVM is the key problem in
this approach, two LS-SVM models with different input features were compared. The
first model had six inputs while there were seven inputs in the second model. The case
study involving a small oil field under water injection located in the Sirte Basin of Libya
demonstrated that the input features of Model 2 are superior to those of the Model 1. The
LS-SVM with the inputs of oil production rate at time t-1 and t, gas production rate at
time t, water injection rate at time t and t+1, and producers at time t and t+1 is robust in
of oilfield under water injection with reasonable accuracy. The approach can be
employable to find the most economical injection rates by applying various water
307
injection scenarios. Meanwhile, it could be a new window for fast simulators. It has
308
CHAPTER XI
11.1 Conclusions
This thesis puts foreword there different function approximation tools with different
engineering.
pressure.
The results of the three used functions approximations provided predicted values with
acceptable magnitude of errors by the oil industry. Based on the results of this study, the
• A review of the literature shows little attention has been paid to the prediction of
reservoir fluid properties in the absence of PVT analysis. Currently, not one
specific model has been published to directly estimate the bubble point pressure in
the absence of PVT analysis and, presently, there is just one published empirical
correlation (Labedi 1990) to estimate bubble point oil FVF and one correlation
(Valko and McCain 2002) to estimate stock-tank vent gas oil ratio directly in the
309
• The majority of the available PVT Models are used to estimate the reservoir oil
However, due to the fact the models require prior knowledge of parameters such
as solution GOR and specific gas gravity, it is difficult to apply them in the
absence of PVT analysis. Since these parameters are rarely measured in the field,
absence of PVT analysis. Thus, there is no need for additional correlations. The
only required input parameters are separator GOR, separator pressure, stock-tank
• Four novel empirical correlations, four ANN models and four LS-SVM models to
estimate Pb, Bob, Rsob and RST in the absence of PVT analysis, were proposed as a
function of readily available field data. All the proposed PVT models can be
• The proposed PVT MRA, ANN and LS-SVM models were developed based on
118 reservoir fluid studies collected from various Libyan oil fields in the Sirte
Basin.
• The estimated Rsob and RST values from the proposed models can be used as a
basic input variable in many PVT imperial correlations in order to estimate other
• The bubble point and oil FVF models have four inputs while bubble point GOR
310
• The results show the proposed PVT models provided high accuracy on testing
dataset. This dataset was not viewed by the models during the building process.
conditions, the first and most important operation, before running a complete
reservoir fluid study, is to check the validity of the samples. The bottom-hole
sample, used for the PVT study, is selected according to the results obtained
bubble point pressure models can be used to select the most representative sample
• Since the average reservoir pressure is measured during a pressure build up test
while the well is shut-in, a significant economic effect occurs in producing wells
during the entire build up test. Moreover, average reservoir pressure is measured
average reservoir pressure without shutting in the producing wells. BP-ANN, BP-
ANN-GA and LS-SVM models are performed to interpolate the missing values,
311
• This case study is based on historical production and reservoir pressure data for
an oilfield located in Sirte Basin, Libya. The reservoir contained a set of 60 oil
wells. There are just 49 measurements of average reservoir pressure from the
• The three proposed models have the ability to predict and interpolate average
reservoir pressure accurately by employing oil, water and gas production rates and
a number of producers are used as inputs for the three proposed models.
• This study shows the ANN and LS-SVM approximation techniques provide a new
• A case study involving a small oil field under water injection located in the Sirte
Basin of Libya demonstrates the proposed approach can be used to predict oil
reservoir performance, and can serve as a practical and robust tool with regard to
reservoir management.
• Since the selection of input features is the key problem in this approach, the
following input features of oil production rate at time t-1 and t, gas production
rate at time t, water injection rate at time t and t+1, and producers at time t and
• This approach is suitably employable to find the most economical injection rates
312
window for fast simulators. It has reasonable accuracy, requires little data and is
In general, ANN and LS-SVM approximation techniques can serve as a practical, cost-
11.2 Recommendations
• Since the proposed PVT models based on two-stage separation test, these
• Because the proposed PVT models were developed specifically for Libyan
oil reservoirs located in Sirte Basin, there testing accuracy are needed to
consider for use in other Libyan Basins and/or other reservoirs around the
world.
313
• The developed models for oil production prediction were successfully
applied. Further studies are needed for other oilfield to test the robustness of
engineers find out an optimal water injection rates and/or to modify the
reservoir damage
reservoir.
314
Appendix A
A.1 Introduction
In any system in which variable quantities change, it is of interest to examine the effects
that some variables exert or appear to exert on others. Often there exists a functional
relationship which is too complicated to describe in simple terms. In this case, it may
approximate to this functional relationship by some simple mathematical function, which
contains the appropriate variables and which approximates to the true function over some
limited ranges of the variables involved. Variables may be related by some sort of
mathematical equation, while the equation might be physically meaningless, and it might
be a linear or nonlinear model. This model can be generated by what is called “regression
analysis”.
Y = β 0+ β 1 X + ε (A.1)
where β0 and β1, are constants called “the model regression coefficients” (or regression
parameters), and ε is the random error or residual term. The random error εi is the term
which accounts for the variation of the ith response variable Yi away from the linear
315
predictor β0 + β1 Xi at the point Xi. In other words, residual term is an estimate of
experimental errors obtained by subtracting the observed responses from the predicted
responses. The predicted response is calculated from the model. That is,
ε i = Yi − β 0 − β 1 xi , i = 1, ……, n (A.2)
Before discussing the multiple regression models in detail, a short description of a simple
linear regression model is given.
A.3 Correlation and Regression Modeling in the Oil and Gas Industry
Multiple regression analysis is an essential tool to analysis data. In the oil industry,
regression analysis is one of the most widely used statistical tools because it provides a
conceptually simple method with which to investigate functional relationships among
variables. A very long list of examples of regression analysis applications exists in the
petroleum industry. A partial list would include its application to reservoir & production
engineering, drilling operations and exploration.
316
2002). Moreover, Salama (2000) and Yi et al. (2005) used regression analysis to produce
models for sand production management.
However, a very long list of regression models to develop PVT properties of reservoir
fluids have been offered in the petroleum engineering literature. A comprehensive
literature review of PVT correlation will be addressed in Chapters 2 and 3.
317
formal statistical tests and probability calculations. It should be used as an informal
analysis directed towards uncovering patterns in the data. Before discussing the
procedure of regression model building, regression assumptions will be addressed.
How to detect: nonlinearity is usually most evident in a plot of observed versus predicted
values or a plot of residuals versus predicted values, which are a part of standard
regression output. The points should be symmetrically distributed around a diagonal line
in the former plot or a horizontal line in the latter plot. Careful observation will disclose
possible evidence of a "bowed" pattern, indicating that the model makes systematic errors
whenever it is making unusually large or small predictions.
318
produced by transformations of the independent variables, or by transformation of the
independent variable, or by both. There are many possible transformations, and different
transformations could be undertaken within the same model.
1) Power Model: y = a x1b x2c x3d ε By taking the logarithms of both sides, the
above equation converts into a linear form of Equation A.1.
How to detect: plots of residual versus dependent and versus all independent variables are
observed. Note evidence of residuals that are becoming larger (i.e., more spread-out). If
the residuals seem to increase or decrease in average magnitude with the fitted values, it
is an indication that the variance of the residuals is not constant. If the points in the plot
lie on a curve around zero, rather than fluctuating randomly, it is an indication that the
assumption (the random errors εI have zero mean) is broken. If assumption (ii) is
satisfied, the residuals can be expected to vary randomly around zero and the spread of
the residuals would be expected to be approximately the same throughout the plot. Figure
A.1 addresses different features in the residual plots.
319
Figure A.1 - Different Features in Residual Plots, Where the Residuals (y-axis) Plotted
versus Fitted Values (x-axis). After Draper and Smith (1998)
320
Figure A.1 (a) shows a residual plot with no systematic pattern. It appears the assumption
(ii) is satisfied for the data associated with this residual plot. In Figure A.1 (b), there is a
clear curved pattern: Assumption (2) may be broken. In Figure A.1 (c), the random
variation of the residuals increases as the fitted values increase. This pattern indicates the
variance is not constant. Finally, in Figure A.1 (d), most of the residuals are randomly
scattered around 0, but one observation has produced a residual which is much larger
than any of the other residuals. The point may be an outlier.
How to fix: if the plot of the residuals against dependent variable shows a conical pattern
as Figure A.1 (c), a transformation is attempted upon the dependent variable before
performing a regression analysis. If it is a trend pattern, as in Figure A.2 (a, c and d),
there is a means error in the analysis caused by wrongly omitting a constant term, b0 in
the model. If the plot takes a curve pattern as in Figure A.1 (b) or Figure A.2 (b), the
model may require an extra term or there may be a need for a transformation on Y.
To obtain the average deviation from the regression line, the square root of the SSE
divided by n − p is used, where p is the number of predictors and n is the sample size.
The square root is used because the residuals were squared and p was subtracted from n
because p degrees of freedom were lost from using p independent variables. Thus, it is
desirable to have the standard error of estimate as small as possible. However, standard
error of estimate has an inverse relationship of R2. As R2 increases, standard error of
estimate decreases, and vice versa. The standard error of estimate is usually computed by
the majority of statistical packages.
321
Figure A.2 - Further Different Features in Residual Plots
322
A.5.3 Normality - Assumption (3)
To satisfy this assumption, we must answer the question “are the model residuals
normally distributed”? The histogram and the normal probability plot can answer this
question. The histogram and the normal probability plot are used to check whether or not
it is reasonable to assume that the random errors, inherent in the process, have been
drawn from a normal distribution. Sometimes the error distribution is “skewed” by the
presence of a few large outliers. Figure A.3 illustrates the histogram plot of the residuals.
It shows the residual term is almost normally distributed. Figure A.4 demonstrates some
shapes of the normal probability plot.
How to detect: the best test for normally distributed errors is a normal probability plot of
the residuals. If the distribution is normal, the points on this plot should fall close to the
diagonal line. A bow-shaped pattern of deviations from the diagonal, as in Figure A.4 (a
and d) indicates that the residuals have excessive skewness (i.e., they are not
symmetrically distributed, with too many large errors in the same direction). An S-
shaped pattern of deviations indicates that the residuals have excessive kurtosis (i.e.,
there are either too many or too few large errors in both directions) as in Figure A.4 (b).
Figure A.4 (e) shows most points fall close to the diagonal line, concluding that the
residual appears to be homoscedastic and more closely follows a normal distribution. In
statistics, a sequence or a vector of random variables is homoscedastic if all random
variables are in the sequence.
323
Figure A.3 - Histogram of the Residuals. The Residuals Close Normal Distribution
324
(a) (b)
(c)
(d) (e)
Figure A.4 - Tests of the Normality. The Histogram and the Normal Probability Plot in (c
and e) Show that the Error Terms are Normally Distributed.
325
A.5.3.1 Outliers, Leverage and Influential Points in Regression
Another use for residual analysis is to see if there are any outlier or leverage values in the
data set. An outlier among the residuals is one that is greater than the rest in the absolute
value. Leverage basically means that a single data point located well outside the bulk of
the data (an outlier), has an over proportional effect on the resulting regression curve.
Observations with large leverage values may exert a disproportionate influence upon a
model and produce misleading results. However, the problem with the residual
distribution is mainly due to one or two very large errors. To obtain a satisfactory model,
it is important to take into account these particular points. Automatic rejection of outliers
is not always a very wise procedure. Sometimes, the outlier is providing information
which other data points cannot provide due the fact that it arises from an unusual
combination of circumstances which may be of vital interest and requires further
investigation rather than rejection. Such values should be scrutinized closely. Are they
genuine, (i.e., not the result of data entry errors)? Are they explainable? Are similar
events likely to occur again in the future, and how influential are they in the model-fitting
results?
To determine the extent of influence, you can fit the model with and without those
observations and compare the coefficients, p-values, R2, and residual plots. If the model
changes significantly when the influential observation is removed, then first determine
whether the observation is a data entry or measurement error. If not, examine the model
further to determine if an important term (e.g., interaction term) or variable has been
omitted, or has incorrectly specified the model. More data may be required to resolve the
issue.
However, the majority of statistical packages have a method with which to see how the
data spread and scatter. The box plot, otherwise known as a “box-and-whisker plot” is a
graphical representation of data that shows how the data are spread. The plot shows how
the data are scattered within those ranges. The advantage of using the box plot is that
when it is used for multiple variables, not only does it graphically show the variation
between the variables but it also shows the data spread, data variations within the ranges
and shows the obviousness of the presence of outliers.. Figure A.5 illustrates the
components of a box plot.
To determine the presence of outliers, the interquartile range (IQR) must be found first.
The IQR measures the vertical distance of the box; it is the difference between the third
quartile, Q3 and the first quartile, Q1 values (IQR = Q3 - Q1). An outlier is defined as any
observation away from the closest quartile by more than 1.5 IQR, (Q3 +1.5(Q3-Q1)). An
outlier is considered extreme when it is away from the closest quartile by more than 3
IQR. This is illustrated in Figure A.5
326
Figure A.5 - Box Plot
327
A.5.4 Multicollinearity - Assumption (4)
Multicollinearity is a high degree of correlation (linear dependency) among several
independent variables. It commonly occurs when a large number of independent variables
are incorporated into a regression model. It is because some of them may measure the
same concepts or phenomena. If the model fits the data well, the overall R2 value will be
high, and the corresponding P value will be low (a great fit is unlikely to be a
coincidence). In addition to the overall P value, multiple regression also reports an
individual P value for each independent variable. A low P value herein means that this
particular independent variable significantly improves the fit of the model. It is calculated
by comparing the goodness-of-fit of the entire model to the goodness-of-fit when that
independent variable is omitted. If the fit is much worse when that variable is omitted
from the model, the P value will be low, thus stating that the variable has a significant
impact upon the model. In some cases, multiple regression results may seem paradoxical.
Even though the overall P value is very low, all the individual P values are high. This
means that the model fits the data well, even though none of the X variables has a
statistically significant impact on predicting Y. How is this possible? When two X
variables are highly correlated, they both convey essentially the same information. In this
case, neither may contribute significantly to the model after the other is included. But
together they contribute significantly. If both variables were removed from the model, the
fit would be much worse. So, the overall model fits the data well, but neither X variable
makes a significant contribution when it is added to the last model. When this happens,
the X variables are “collinear” and the results show “multicollinearity”.
328
correlations among the variables. If high bivariate correlations are present, one can
be deleted from the two variables. However, this may not always be sufficient.
2. Regression coefficients will change dramatically according to whether other
variables are included or excluded from the model. Various approaches can be
taken by adding and then removing variables from the regression model. This can
be easily accomplished by running a stepwise regression.
3. Examine the Variance Inflation Factor (VIF). VIF measures the impact of
multicollinearity among the variables in a regression model. VIF is always greater
than or equal to 1. There is no formal VIF value to determine the presence of
multicollinearity. Values of VIF that exceed 10 are often regarded as indicating
multicollinearity.
In most statistics packages, the results are shown for both the R2 value and the Variance
Inflation Factor (VIF).
How to fix: The best solution is to understand the cause of multicollinearity and remove
it. Multicollinearity occurs because two (or more) variables are related (they basically
measure the same). If one of the variables does not appear to be logically essential to the
model, removing it may reduce or eliminate multicollinearity. Or, perhaps a method can
be found to combine the variables. The two most common ways to display the
multicollinearity are correlation matrix and stepwise regression.
A.5.4.1 Coefficient of Determination R2
Coefficient of determination, R2 is the proportion of the variation in the dependent
variable explained by the regression model, and is a measure of the goodness of fit of the
model. It is calculated as follows:
SSR SSE
R2 = = 1− (A.5)
SST SST
A fundamental equality, SST = SSR + SSE, where SST stands for the total sum of squared
deviations in Y from its mean, SSR denotes the sum of squares due to regression, and SSE
represents the sum of squared residuals. Those quantities may be computed from
2
−
SST = ∑ y i − y
2
^ −
SSR = ∑ y i − y (A.6)
2
^
SSE = ∑ y i − y i
329
Usually, the values are computed and supplied as part of the regression output by
statistical packages named “the ANOVA table”. ANOVA stands for analysis of variance.
However, because SSE is less than or equal to SST, the value of R2 will always lie
between 0.0 and 1.0 (or, in percentage, between 0% and 100%). To achieve R2 near
100 %, we must minimize the sum of squares of the errors. The equation is often asked,
“What is a good value of R2?” That depends upon the area of application. In some
applications, variables tend to be more weakly correlated and there considerable noise.
Lower values for R2 could be expected in these cases. In another area of application, a
value of R2 of 60% might be considered good. In physics and engineering, where most
data comes from closely controlled experiments, a much higher R2’s is expected and a
value of 60% would be considered low. Generally, some experience with the particular
area is necessary to judge the R2’s well. Table 1.1 shows the range of R2 in various PVT
correlations to estimate a solution gas-oil ratio, oil formation volume factor and
undersaturated oil compressibility.
A.5.4.2 Adjusted R2
2
However, R tends to somewhat over-estimate the success of the model when applied to
the real world, so an adjusted R2 value is calculated which takes into account the number
of variables in the model and the number of observations (participants) the model is
based upon. This adjusted R2 value gives the most useful measure regarding the success
of our model. If, for example an adjusted R2 value of 85% is given, it can be said that the
model has accounted for 85% of the variance in the criterion variable.
Does a large R2 mean a good model? R2 automatically increases as more variables are
added to a multiple regression model, even if the additions are insignificant. The
prediction always appears to become superior as more independent variables are added.
This is a false indication and does not actually increase confidence in the predictions.
When there are similar numbers of variables and data samples, multiple R will always be
high. Adjusted R2 is a modified R2 where the addition of independent variables does not
automatically increase R. It is an improved measure of the overall fit of a multiple
regression model.
n −1
adjR 2 = 1 − (1 − R 2 )
n − p − 1 (A.7)
where,
n = number of samples
p = number of independent variables
330
Figure A.6 - ANOVA Table
331
A.5.4.3 Correlation Coefficient among Variables, Cor(X1,X2)
The correlation coefficient is one of the most widely used coefficients to study the
relation among variables. In simple linear regression, R2 is equal to the square of the
correlation coefficient between the response variable Y and the predictor X or to the
square of the correlation coefficient between the response variable Y and the fitted values
Y. The correlation coefficient between Y and X, Cor(Y,X) is given by
− −
∑ ( y i − y ) ( xi − x )
Cor (Y , X ) = (A.8)
− −
∑(y i − y) 2
∑ (x x − x) 2
Thus, the correlation coefficient is symmetric, that is Cor(Y, X) = Cor(X, Y). Furthermore,
Cor(Y, X) satisfies 0.1 ≤ Cor (Y , X ) ≤ 1.0 . Values near zero mean no linear correlation and
values near ± 1 mean a very strong linear correlation. The negative sign means that the
two variables are inversely related. That is, as one variable increases the other variable
decreases. The properties make the correlation coefficient a useful quality with which to
measure both the direction and the strength of the relation between X1, X2 and Xn. Figure
A.7 supplies different types of correlation coefficients.
332
Figure A.7 - Different Features in Correlation Coefficient
333
In forward selection, the variables enter into the model one at a time in an order
determined by the strength of their correlation with the criterion variable. The effect of
adding each is assessed as it is entered, and the variables that do not significantly add to
the success of the model are excluded. However, in backward selection, all the predictor
variables are entered into the model. The weakest predictor variable is then removed and
the regression re-calculated. If this significantly weakens the model, then the predictor
variable is re-entered – otherwise it is deleted. This procedure is then repeated until only
useful predictor variables remain in the model.
As explained earlier, that regression analysis is used to produce an equation that will
predict a dependent variable using one or more independent variables (Equation 1.1).
However, when running the regression, the statistical packages provide very useful
information. Beside the model coefficients (estimated parameters), it presents the
standard deviation of each estimated parameters, T-statistics and P-value. This is
illustrated in Figure A.8. First of all, the purpose is to discover whether the coefficients
on the independent variables are really different from zero, and if the independent
variables are having a genuine effect on the dependent variable, or if any apparent
differences from zero are just due to random chance. The null (default) hypothesis is
always that each independent variable is having absolutely no effect (has a coefficient of
zero) and a reason to reject this theory is sought.
A P-value of 5% or less (significant level 95%) is the generally accepted point at which
to reject the null hypothesis. With a P-value of 5% (0.05), there is only a 5% chance that
the estimated parameters would have come up in a random distribution, and thus, it can
be said that with a 95% probability of being correct that the variable is having some
effect, assuming the model is specified correctly. The test is carried out by comparing this
observed value with the appropriate critical value obtained from statistical tables. Critical
value can be read directly from the statistical tables at n-p and 0.95/2, where n denotes an
observation number. Note, the number was divided by 2 because it has a two-sided
alternative hypothesis. Also, n-p was used as a degree of freedom because p is the
334
estimated parameters (including the constant, b0). The P-value is the sum of the two
shaded areas under the curve in Figure A.8.
Furthermore, while model completeness is one of the most important aspects of model
adequacy, this type of test does not address other important aspects of model quality. In
statistical jargon, this type of test for model adequacy is usually called a "lack-of-fit" test.
The most common strategy used to test for model adequacy is to compare the amount of
random variation in the residuals from the data used to fit the model with an estimate of
the random variation in the process using data that are independent of the model. If these
two estimates of the random variation are similar, that indicates that no significant terms
are likely to be missing from the model. If the model-dependent estimate of the random
variation is larger than the model-independent estimate, then the model miss-specified, (a
model that provides an incorrect description of the data).
The ANOVA table contains the degrees of freedom, the sum of squares, the mean
squares, an F-statistic and a corresponding P-value. For a multiple linear regression
model with p independent variables, the model has p degrees of freedom (i.e. the number
of independent variables). Since the total degrees of freedom is unchanged at n−1, it
follows that the degrees of freedom for the error is n − p − 1. A summary of the first
table is given by
335
Figure A.8 - Probability Density Function Graph of a T-distribution. The P-value is the
sum of the two shaded areas under the curve.
336
In a multiple linear regression model, the hypothesis of overall regression is the
hypothesis that the independent variables (as a whole) provide some useful information
about the response variable. Thus, the hypothesis of overall regression is accepted, if the
null hypothesis that none of the explanatory variables influence the response is rejected.
That is, H0: b1=b2=………=bp=0 against the alternative hypothesis H1: bj ≠ 0 for at least
one j = 1, 2, . . . , p. If H0 is false, that is, if the dependent variable depends upon at least
one of the independent variables, then model = b0+b1x1 +…+ bpxp should explain a
substantial amount of the variation in the data, relative to the residual variation. Thus, if
H0 is false, the model variation (SSR) should be large relative to the residual variation
(SSE). Then, reject H0, if the value of the ratio (SSR/p) / (SSE/(n-p-1)) is large. The F-
value in the ANOVA table is exactly this ratio. However, the hypothesis can be rejected
if the P-value is less than 0.05 (5% significance level) as explained earlier.
Why is the H0 tested? If the full model explains the variation in the data better than the
reduced model (Model #2), it must be because the p−m extraneous variables provide
more useful information regarding the response variable. Hence, H0 should be rejected
and the extraneous variables in the model should be retained. If, on the other hand, the
full model (Model #1) and the reduced model (Model #2) explain (almost) the same
amount of the variation, the simpler, reduced, model may as well be used as the full one.
Thus, we do not reject H0. More specifically, we reject H0 if the difference between the
model variation for the full model and the model variation for the reduced model is large,
relative to the residual variation under the full model.
If the models are nested, two distinct approaches can be used to compare models. The
first method is based on statistical hypothesis and ANOVA. It is based on analysis of the
difference between the sum of squares of the two models. It also takes into account the
number of parameters of each model. The ANOVA calculations compute an F-ratio and
337
P-value. If the P-value is lower, it can be concluded that there are sufficient data to
convince and to reject the reduced model (the null hypothesis) in favour of the alternative
hypothesis if the full model is correct. If the P-value is small, there are two possibilities:
The P-value is able to tell how frequently this could be expected to happen. So, which
model is correct? Which model fits better? The full model is correct but this would be
expected to be by chance and thus, that is the wrong question to ask!! Which model
should be accepted? This is answered by statistical hypothesis testing. This method is
only valid when comparing nested models. It cannot be used for non-nested models.
The second method for comparing models is not based on hypothesis testing at all. After
fitting a number of different models to a given data set, they can be compared by:
To compare the models and before running the regression analysis, the data set must be
divided into two sets. One set for fitting the regression model and the other set (about
10% of the whole data) to validate and compare the models. Average relative error,
absolute average relative error, minimum and maximum of both averages give a good
advice. See Appendix C for more details.
338
A.10 References
A.10.1 Regression Analysis
Cohen, J., Cohen, P., West, S. G., & Aiken, L. S., 2003. Applied multiple
regression/correlation analysis for the behavioral sciences, 3rd Ed. Mahwah, NJ:
Lawrence Erlbaum Associates.
Draper, N.R. and Smith, H., 1998. Applied Regression Analysis, third addition, John
Wiley and Sons, Inc. New York.
Greene, William H., 2000. Econometric Analysis (Fourth edition). Upper Saddle
River, NJ: Prentice-Hall.
Guthrie, R.K. and Greenberger, M.H., 1955. The Use of Multiple Correlation
Analysis for Interpreting Petroleum Engineering Data. Drilling and Production
Practice, API, 1955, 130.
Havlicek, L., & Peterson, N., 1977. Effects of the violation of assumptions upon
significance levels of the Pearson r. Psychological Bulletin, 84, 373-377.
Stevens, J. P., 2002. Applied multivariate statistics for the social sciences, fourth
addition. Mahwah, NJ: Lawrence Erlbaum Associates.
Tabachnick, B. G., & Fidell, L. S., 2001. Using multivariate statistic, fourth addition.
Needham Heights, MA: Allyn & Bacon
339
Jensen, J. L., Hinkley, D.V., and Lake, L.W., 1987. A statistical study of reservoir
permeability: distributions, correlations, and averages. SPE Formation Evaluation,
2 (4), 461-468.
Jian, F. X. et al., 1994. A Genetic Approach to the Prediction of Petrophysical
Properties. Journal of Petroleum Geology, Vol. 17-1.
Sahin, A. and Saner, S., 2001. Statistical distributions and correlations of
petrophysical parameters in the eastern Saudi Arabia Arab-d reservoir, Abqazq
oilfield. Journal of Petroleum Geology, vo1. 24-1.
Sahin, A. Ali, A. Z., Saner, S. and Menouar, H., 2007. Permeability anisotropy
distributions in an upper Jurassic carbonate reservoir, eastern Saudi Arabia.
Journal of Petroleum Geology, Vol. 30-2
340
Colonna, J., Brissaud, F., and Millet, J.L., 1972. Evolution of Capillarity and Relative
Permeability Hysteresis,” SPEJ 28; Trans., AIME, 253.
Eigestad, G.T. and Larsen, J.A., 2000. Numerical Modeling of Capillary Transition
Zones,” paper SPE 64374 presented at the SPE Asia Pacific Oil and Gas
Conference and Exhibition, Brisbane, Australia, Oct. 16–18.
Eikje, E. et al., 1992. Relative Permeability Hysteresis in Micellar Flooding,” J. Pet.
Sci. Eng. 7, No. 1–2, 91.
Evrenos, A.I. and Comer, A.G.1969. Numerical Simulation of Hysteretic Flow in
Porous Media. Paper SPE 2693 presented at the SPE Annual Fall Meeting,
Denver, Sep 28.– Oct. 1.
Geffen, T.M. et al.1951. Experimental Investigation of Factors Affecting Laboratory
Relative Permeability Measurements. Trans., AIME 192, 99.
Guzman, R.E. and Fayers, F.J., 1997. Mathematical Properties of Three-Phase Flow
Equations. SPEJ 291.
Hagoort, J., 1984. Measurements of Relative Permeability for Computer
Modeling/Reservoir Simulation. Oil & Gas J. Feb. 62.
Hammervold, W.L. et al., 1998, Capillary Pressure Scanning Curves by the
Micropore Membrane Technique. J. Pet. Sci. & Eng. 20, 253.
Honarpour, M., Koederitz, L., and Harvey, A.H., 1986. Relative Permeability of
Petroleum Reservoirs. CRC Press, Inc., New York City.
Honarpour, M.M., Huang, D.D., and Al-Hussainy, R., 1996. Simultaneous
Measurements of Relative Permeability, Capillary Pressure, and Electrical
Resistivity With Microwave System for Saturation Monitoring. SPEJ 283.
Huang, D.D., Honarpour, M.M., and Al-Hussainy, R., 1997. An Improved Model for
Relative Permeability and Capillary Pressure Incorporating Wettability. Paper
presented at the 1997 Society of Core Analysts International Symposium,
Calgary, Sep. 7–10
Kjosavik, A. Ringen, J.K. and Skjaeveland, S.M., 2002. Relative Permeability
Correlation for Mixed-Wet Reservoirs” SPE 77328, March SPE Journal
Killough, J.E., 1976. Reservoir Simulation with History-Dependent Saturation
Functions. SPEJ 37; Trans., AIME, 261.
Kriebernegg, M. and Heinemann, Z., 1996. A New Model for History Dependent
Saturation Functions in Reservoir Simulation. Paper presented at the European
Conference on the Mathematics of Oil Recovery, Leoben, Austria, Sep. 3–6.
Land, C.S., 1971. Comparison of Calculated With Experimental Imbibition Relative
Permeability. SPEJ Trans., AIME, 251.
Land, C.S., 1968. Calculation of Imbibition Relative Permeability for Twoand Three
Phase Flow From Rock Properties. SPEJ 149; Trans., AIME, 243.
Larsen, J.A. and Skauge, A., 1999. Simulation of the Immiscible WAG Process Using
Cycle-Dependent Three-Phase Relative Permeabilities,” paper SPE 56475
presented at the SPE Annual Technical Conference and Exhibition, Houston, 3–6
October.
Laroche, C., Vizika, O., and Kalaydjian, F., 1999. Network Modeling To Predict the
Effect of Wettability Heterogeneities on Multiphase Flow. Paper SPE 56674
presented at the SPE Annual Technical Conference and Exhibition, Houston, 3–6
October.
341
Lenhard, R.J. and Oostrom M., 1998. A Parametric Model for Predicting Relative
Permeability-Saturation-Capillary Pressure Relationships of Oil–Water Systems
in Porous Media with Mixed Wettability. Transport in Porous Media 31, 109.
Lomeland, Frode et al., 2005. A new versatile relative permeability correlation.
Symposium of the Society of Core Analysts held in Toronto, Canada, Aug. 21-25.
McDougall, S.R. and Sorbie, K.S. 1995. The Impact of Wettability on Waterflooding
Pore-Scale Simulation. SPERE 208.
Morrow, N.R. and Harris, C.C., 1965. Capillary Equilibrium in Porous Materials,”
SPEJ 15.
Moulu, J.-C. et al., 1999. A New Three-Phase Relative Permeability Model for
Various Wettability Conditions. Paper SPE 56477 presented at the SPE Annual
Technical Conference and Exhibition, Houston, 3–6 October.
Nordvedt, J.E., Mejia, G., Yang, P., Watson, A.T., 1993. Estimation of Capillary
Pressure and Relative Permeability Functions from Centrifuge Experiments. SPE
Reservoir Engineering, Vol. 8, No. 4, pp. 292-298.
Schembre, J. M. and Kovsce, A. R., 2006. Estimation of Dynamic Relative
Permeability and Capillary Pressure from Countercurrent Imbibition Experiments.
Transport in Porous Media Vol. 65-1.
Skjaeveland, S.M. et al., 2000. Capillary Pressure Correlation for Mixed-Wet
Reservoirs. SPEREE 2-60.
Standing, M.B., 1975. Notes on Relative Permeability Relationships, Proc.,U. of
Trondheim, Norway.
Topp, G.C. and Miller, E.E., 1966, Hysteretic Moisture Characteristics and Hydraulic
Conductivities for Glass-Bead Media. Soil Sci. Soc. Amer. Proc. 30, 156.
Wardlaw, N.C. and Taylor, R.P., 1976. Mercury Capillary Pressure Curves and the
Interpretation of Pore Structure and Capillary Behaviour in Reservoir Rocks.
Bull. Can. Pet. Geo. 24, No. 2, 225.
342
A.10.8 Regression analysis in Flow in Porous Media
Morshed, J. and Powers, S.E., 2000. Regression and Dimensional Analysis for
Modeling Two‐Phase Flow. Transport in Porous Media 38: 205–221.
343
APPENDIX B
B.1 Introduction
In order to find relationship between the input and output data derived from experimental
work, a more powerful method than the traditional ones are necessary. For some time,
ANN methods have been used in many diverse fields as a function approximation tools,
producing results comparable with (or better than) regression models and other statistical
methods, see for example Negnevitsky (2005). However, one of their key advantages is
their ability to easily model complex, non-linear systems. ANN is an especially efficient
algorithm to approximate any function with finite number of discontinuities by learning
the relationships between input and output vectors (bozorgmehry et al., 2005; Haganet et
al., 1990). These algorithms can learn from the experiments, and also are fault tolerant in
the sense that they are able to handle noisy and incomplete data. ANN methods are a
powerful tool and widely used technique for supervised neural networks learning in many
application areas, ranging from pattern recognition, and function approximation, to
speech recognition and robot control. Thus, ANNs are able to deal with non-linear
problems, and once trained can perform prediction and generalization rapidly (Sozen et
al., 2004). They have been used to solve complex problems that are difficult to be solved
if not impossible by the traditional methods and provide more accurate and robust
solutions for problems. Examples include; regression, classification, control,
optimization, pattern recognition, and so on. Especially it is desired to have the minimum
difference between the predicted and actual outputs (Richon and Laugier, 2003).
The BP algorithm consists of two paths; forward path and backward bath. The backward
bath is described for the backpropagation training algorithm. Forward bath contain
creating a feedforward network, initializing weight, simulation and training the network.
The network weights biases are updated in backward path. Once the network weights
have been initialized, the network is ready for training. The training process requires a set
of proper inputs and targets as outputs. During training, the weights are interactively
adjusted to minimize the network performance function (mean square error, MSE). MSE
is the default performance function for forward networks which is the average squared
error between the network outputs and the target output. The basic procedure for training
the network is embodied in the following description:
344
i. Apply an input vector from training dataset to the network and calculate the
corresponding output value,
ii. Compare the actual output with the computed output value and determine the
difference (error).
iii. Determine in which direction (increase or decrease, + or −) to change each weight
in order to reduce the error,
iv. Determine the amount by which to change each weight,
v. Apply the corrections to the weights, and
vi. Repeat items a through e with all the training vectors until the error for all vectors
in the training set is reduced to an acceptable value.
1. Apply the N-dimensional input vector, xp = (xp1, xp2, …, xPN)t to the input layer of
the network, where the ‘t’ superscript means transpose. The input units distribute
the values to the hidden-layer units.
2. Calculate the net-input values to the hidden layer units. The net-input to the jth
hidden unit is
N
nethpj = ∑ wjih xpi + θ hj (B.1)
i =1
The weight on the connection from the jth node to the ith node is denoted wji,
where whji is the weight on the connection from the ith input units, the “h”
superscript refer to quantities on the hidden layer and θ jh is the bias term.
345
Hidden Layer
Input Layer Output Layer
hn11
HL
xp1 hn22
HL
xp2 HL
hn33
y
xp3
hn44
HL
xp4
hn55
HL
346
3. Calculate the outputs from the hidden layer. The function used for this calculation
is referred to as the activation function as illustrated in Figure B.2. The activation
of this node is equal to the net-input; then, the output of this node is
where f jh is the input activation function. The activation function can be linear so
that we have a linear network, or nonlinear. Activation functions for the hidden
units are needed to introduce nonlinearity into the network. Without nonlinearity,
hidden units would not make nets more powerful than just plain perceptrons
(which do not have any hidden units, just input and output units). However, it is
the nonlinearity that makes multilayer networks so powerful.
4. Move to the output layer. Calculate the net-input values to each unit:
L
netopk = ∑ wkjo ipj + θko
i =1
(B.3)
where y pk are the desired output values and opk is the model output values from
Equation (3.4).
Note that, δ pih is calculated before the connection weights to the output-layer unit
have been updated.
347
8. Update weights on the output layer:
Every weight update on the hidden layer depends on all the error terms on the
output layer, δ pk
o
. This result is where the notion of BP arises. The known errors
on the output layer are propagated back to the hidden layer to determine the
appropriate weight changes on that layer. The order of the weight updates on an
individual layer is not important. Be sure to calculate the error term, Ep
M
Ep = 1
2 ∑δ
k =1
2
pk (B.9)
since Ep is the measure of how well the network is learning. When the error is
acceptably small for each of the training-vector pairs, training can be
discontinued. Thus, our goal is to minimize the network performance function
(Mean Squares Errors, MSE) Eq. (3.10).
N
MSE = 1
N ∑ (E )
i =1
p
2 (B.10)
In a neural network model the function y = g(x) is modeled as a nonlinear function, and
the neural network model can be rewrite as
Where, β, α1, . . . , αm−1 are unknown parameter vectors, Xi is a vector of known constants
and εi are residuals (error terms).
348
NeuroIntelligence software (Alyuda 2006) was used for the artificial neural networks
calculations. This proved to be a very powerful tool for carrying out the artificial neural
network design and analysis. Usually, trial-and-error is required to achieve appropriate
network architecture. In this chapter, a multiple feedforward backpropagation network
was used to build the following six models:
One of the most important problems that neural network designers face today is choosing
an appropriate network size for a given application. Network size involves in the case of
layered neural network architectures,
i. Training Dataset
The training dataset is used to train and build a network model. In other words, this
dataset is used to obtain the network weights.
349
Weights Net-input
Output
xp1
w1
xp2 w2 nethpj f jh(net hpj) y
w3
xp3
Activation
Inputs Function
350
ii. Validation Dataset
However, during the training process the accuracy of the model is computed with a
dataset that was not used in the training process which is validation dataset. The task of
this dataset is to check the performance of the network model. In order to avoid
overtraining which result poor network model performance, it is necessary to use an early
stopping technique as illustrated in Figure B.3. Training can be stopped when the
performance of the model on the validation dataset gives minimum error. The training set
error will naturally decrease with increasing numbers of epochs of training. However, the
error on the unseen validation dataset will start off decreasing as the under-fitting is
reduced, but then it will eventually begin to increase again as over-fitting occurs. Indeed,
the solution to get the best generalization and lowest error, the validation dataset is used
for early stopping (Bishop 1995 and Amari et al. 1997).
i. Testing Dataset
The testing dataset is used to fine-tune models. This dataset is used neither in training nor
in validation; it is used to pick the best model architecture. Typically, several network
models are tried out with various architectures (as will be explained in detail in Section
B.5) the accuracy of each one on the testing dataset is tested to choose the competing
architecture. The lowest error on the testing dataset is used for selecting the appropriate
model and to indentify the optimum network architecture.
The training and testing dataset error will naturally decrease with increasing numbers of
training epochs after the number of epochs of training the testing error begin to increase
as over-fitting takes place. In this iteration, the network should track.
351
Figure B.3- Illustration of the Early-stopping Rule. The trained network is saved at the
point defined by the early-stopping criterion, and training continues exclusively to check
whether the error will fall again. This will ensure that the increase in the validation error
is not a temporary event.
352
B.4.4 Number of Hidden Layers and Hidden Nodes per Hidden Layer
A major problem in designing a neural network is establishing the optimal number of
layers and hidden nodes per hidden layer. The hidden layer is responsible for internal
representation of the data and the information transformations input and output layers. A
search of the available literature suggested many rules of thumb for determining the
number of neurons in the hidden layer. It became apparent that there is no general rule by
which a suitable number of hidden layer neurons can be chosen and that the optimum
number is very specific to the problem being solved (Krose and Van der Smaget, 1996)
and (Heaton, 2005). Usually, trial-and-error is required to achieve a better network model
architecture. It is arbitrary to the selection of the various neural network architectures and
configuration settings. Hence, it is advised that building a well-designed neural network
model takes patience and involves a large amount of time spent in network training and
training a series of networks with single and multiple hidden layers with different number
of nodes in each hidden layer.
n
y = f ∑ wj x j + b (B.11)
i =0
where wi represents the weight vector, xi is the input vector (i = 1, 2..., n), b is the bias, f
is the activation function, and y is the output.
Many types of activation functions have been proposed and tried, but the surviving,
most used are the logistic and hyperbolic tangent activation function.
f (s) = 1 (B.12)
1 + e− s
353
It is an S-shaped (sigmoid) curve and their outputs take values in the ranges of
[0,1].
f (s) = 2 −1 (B.13)
1 + e−2s
It a sigmoid curve, like the logistic function, except that output lies in the
range [−1, +1]. Often performs better than the logistic function because of its
symmetry.
Again, the activation function plays a critical role in NN modeling. However, the
selection of the correct activation function to the NN design is not that critical. Since the
hyperbolic tangent function produces both positive and negative target values, therefore,
it tend to yield faster network training than logistic function, (Matignon, 2005).
i. Quasi-Newton
Quasi-Newton is an enhancement of the BP training algorithm. Since,
back propagation adjusts the network weights after each case, Quasi-
Newton works out the average gradient of the error surface across all cases
before updating the weights once at the end of the epoch. There is also no
need to select learning or momentum rates for Quasi-Newton, and thus, it
can be much easier to use than back propagation. However, Quasi-Newton
is the most popular algorithm in nonlinear optimization, with a reputation
for fast convergence (StatSoft, 2011).
ii. Levenberg-Marquardt
A Levenberg-Marquardt training algorithm can only be used on networks
with a single output unit.
354
iv. Quick Propagation
Quick propagation works by making the assumption that the error surface
is locally quadratic, with the axes of the hyper-ellipsoid error surface
aligned with the weight. If this is true, then the minimum of the error
surface can be found after only a couple of epochs. Based on this
assumption, quick propagation needs to select a learning rate and
algorithm coefficient (training algorithm’s parameter), (StatSoft, 2011).
In fact, the selection of a training algorithm is based upon the characteristics of the
problem itself. Essentially, training a neural network model means selecting one model
from the set of allowed models that minimize the cost function (minimum square error,
Equation B.10). As a result, the selection of the neural network architecture will come
down to a trial-and error procedure. If the model architecture and learning algorithm are
selected appropriately, the resulting ANN can be extremely robust.
Finally, there are concerns regarding the possibility of converging to a local minimum in
weight space. A typical neural network might have a couple of hundred weights whose
values must be found to produce an optimal solution. Since the output of a neural
network, as a function of the inputs, is often highly nonlinear, this makes the optimization
process complex. If an error was plotted as a function of the weights, a rough surface
with many local minima would likely be observed, as shown in Figure B.4a. A final topic
concerns the possibility of converging to a local minimum in weight space. Figure B.4b
illustrates the idea. It shows a cross-section of a hypothetical error surface in weight
space. The point, zmin, is called a “global minimum”. Notice, however, that there are other
minimum points called “a local minimum”. A search for the global minimum might
accidentally find one of these local minima instead of the global minimum.
Once a network settles on a minimum, whether local or global, learning stops. If a local
minimum is reached, the error at the network outputs may still be unacceptably high.
Fortunately, this problem does not appear to cause much difficulty in practice. If a
network stops learning before reaching an acceptable solution, a change in the number of
hidden nodes or in the learning parameters will often fix the problem or simply start over
with a different set of initial weights (retrain the network), (Freeman and Skapura, 1991).
Actually, the retraining approach was adopted in the modeling process in this chapter.
However, when a network reaches an acceptable solution, there is no guarantee that it has
reached the global minimum rather than a local one. If the solution is acceptable from an
error standpoint, it does not matter whether the minimum is global or local, or whether
the training was halted at some point before a true minimum was reached. Thus, the
algorithm does not always step in the direction of the global minimum of the error
surface. As a result, to avoid local minima, simply try a number of ANN models with
different optimization algorithms with random starting points and use the one with the
best value. In this way, randomly jumps in the hope that the location is getting closer to
the global minimum.
355
Figure B.4- (a) Hypothetical Surface in Weight Space with Many Local Minimums. (b)
Cross-section of a Hypothetical Error Surface in Weight Space.
356
B.5 Selections of the Best Architecture
However, the development of any ANN model is based on a thorough understanding of
the research problem. Designing an ANN model, however, can be summarized in the
following eight-step procedure:
Step 2: Variable selection (number of inputs and outputs, if applicable). Generally, Steps
1 and 2 would be considered simultaneously prior to developing a model.
Step 5: Neural network training parameters and configuration values (number of input
neurons, percentage iterations for constant initial learning rate, initial weights,
learning rate increment, and radius decrement).
Step 6: Neural network training (presentation of records and the number of iterations).
The development of ANN models starts with finding the best network structure that will
represent the complicated relationship between variables. Possible ANN architectures,
starting from a low number of layers and nodes to a higher number of layers and nodes,
should be tested to select the best architecture. Successful architectures are those that
converge to the target error or reach the possible MSE and exhibit stable performance to
new data not included in the training. Topically, the best ANN architecture is the one that
simultaneously fulfils all the following criteria;
357
are used to evaluate the model. Quantitative analysis is determined in terms of correlation
coefficient (R2), standard deviation (SD), average, maximum and minimum absolute
error (AE), average, and maximum and minimum absolute relative error (ARE). One
common distinction between the absolute error and the relative error is the absolute error
is the magnitude of the difference between the exact value and the approximation. The
relative error is the absolute error divided by the magnitude of the exact value. The
percent error is the relative error expressed in terms of per 100. Average absolute error is
a quantity used to measure how close forecasts or predictions are to the actual values.
Standard deviation shows how much variation or dispersion there is from the average.
Correlation coefficient R2 provides a measure of how well future outcomes are likely to
be predicted by the model.
Graphical analysis is determined in terms of scatter diagrams that compare the predicted
ANN values versus the actual observation values. From this plot, it can be observed as to
how the two comparable datasets agree with each other. In this case, an identity line (y=x
line, or a 1:1 line) is often drawn as a reference. The more the two datasets agree, the
more the scatters tend to concentrate in the vicinity of the identity line. This indicates an
excellent agreement between the actual data and network model predicted data. If the two
datasets are numerically identical, the scatters fall on the identity line exactly.
358
APPENDIX C
C.1 Introduction
The Support Vector Machine (SVM) can be used for both regression and calcification
problems, as with ANN. A SVM is derived from statistical learning theory and was
found in 1995 by Vapnic et al. SVM is one of the effective algorithms in machine
learning and data mining with the advantages of simple model structure selection, fast
processing speed, and high learning precision. It is widely used in handling regression
and classification problems based on statistical learning theory. The method adheres to
the principle of structural risk minimization, seeking to minimize an upper bound of the
generalization error, rather than minimize the training error. Suykens et al. (2002)
proposed the LS-SVM method, in which the least squares loss function was modified and
introduced to the SVM. LS-SVM is based on equality constraints and a sum square error
cost function as opposed to earlier approaches that utilize inequality constraints and solve
complex convex optimization problems. The LS-SVM reformulation simplifies the
problem and solution by adopting a linear system rather than quadratic problems that are
difficult to compute. This principle is illustrated in Figure C.1. Consider a nonlinear
relationship in the input space (original space, Figure C.1, left panel). Then, the inputs (x)
are mapped into a high dimensional space by means of Φ (Figure C.1, right panel). In this
space, a linear model is fitted. There are, however, only two parameters to be tuned, γ and
σ2, which are less than that for standard SVM (three parameters). When adopting LS-
SVM with the RBF kernel function, the parameter combination (γ, σ2) should be
established, where γ denotes the regularization parameter, namely, the trade-off between
minimizing the training error and minimizing the complexity of the LS-SVM model; σ2
denotes the RBF kernel parameter and represents the change in the RBF kernel function
(width of kernel) and it is a positive real constant.
SVM and LS-SVM is a kernel-base algorithm. A kernel is a function that transfers the
input data to a high-dimensional space where the problem is solved. Kernel function can
be linear or nonlinear. The model parameters of an LS-SVM are given by the solution of
a system of linear equations. The generalization performance of the LS-SVM is however,
strongly dependent upon the model selection process, in this case the careful selection of
an appropriate kernel function and the good values of the model parameters (Suykens et
al., 2002).
359
Figure C.1 - Illustration of the Key Ingredient of LS-SVM: Transformation Φ of the input
data to the feature space.
360
C.2 LS-SVM for Function Approximation
The following paragraph is a brief introduction to the LS-SVM regression model. The
main objective of function approximation is to approximate a function f (x) from a given
training set of N samples B , CD
in which xi ∈ R (ND vector space) is the input data
N
and yi ∈ r (1D vector space) is the corresponding output value. The least square method
learns in a linear relation between a set of indicators fi (x), i = 1…..N, and the output is
= ! + ! + ⋯ + ! (C.1)
In this equation, the b's are the n parameters that receive a value during the training by
minimizing the summed squared approximation error over all examples. The indicators
can be non-linear functions of the input vectors. Therefore, the training data X and the
output Y can be defined as
⋯
E=F ⋮ ⋱ ⋮ I, J = F ⋮ I
(C.2)
D ⋯
The goal of function approximation is to train the underlying relation between the inputs
and the output values. Then, the LS-SVM model is constructed in this high dimensional
feature space and takes the form
= K L ϕ + ! (C.3)
where the nonlinear mapping ϕ (.) maps the input data into a higher dimensional feature
space. b is the bias and w is a weight vector having the same dimension as the feature
space. In LS-SVM, the weight vector (w) is that will give the smallest summed squared
approximation error over all samples is located. After training, a function approximation
implements a mapping x → that can be evaluated for any x. The performance of
function approximation is measured by means of the approximation error ei, which is
defined as:
ei = yi − (C.4)
Thus, nonlinear function is estimated in its original space because the linear function
estimation is in the feature space. Then, the optimization problem of LS-SVM is
formulated wherein the regularization parameter (γ) is used to avoid overfitting. As a
result, the regularization parameter needs to be tuned while in the training process.
= ∑ M , + ! (C.5)
where k (x, xi) is the kernel function, xi is the input vector, αi is Lagrange multiplier called
the “support value”, b is the bias term. However, αi and b are the results from the
optimization. The details of the LS-SVM algorithm could be found in Vapnok, (1995),
Gue et al., (2006) and Chen et al., (2007).
361
There are several common kernel functions such as linear, polynomial, radial basis
function (RBF) kernel, and Multilayer Perceptron (MLP). Theoretically speaking, the
linear kernel type is the simplest and most efficient kernel to perform similarity
calculation. RBF kernel is a nonlinear function and a more compact supported kernel, and
could reduce the computational complexity of the training procedure while giving good
performance under general smoothness assumptions (Wu et al., 2008). Moreover, in
comparison with some other feasible kernel functions, the RBF is a more compactly
supported kernel function and able to improve generalization performance of the LSSVM
(Brabanter et al., 2002 and Shi and Nabil, 2007). As a result, RBF kernel (Equation C.6)
was recommended as the kernel function for the LS-SVM regression models in this
thesis.
where, σ2 is a positive real constant denoting the kernel width parameter. As a result of
kernel regression, the nonlinear relationship in the data space is transferred to a linear
relationship in the feature space by means of kernels.
Consequently, from the training LS-SVM problem, it can be observed that there are two
free parameters. Regularization parameter (γ) and kernel width parameter (σ2), may affect
the LS-SVM generalization performance. The regularization parameter γ determined the
trade-off between minimizing the training error and minimizing model complexity. The
parameter σ2 was the bandwidth and implicitly defined the nonlinear mapping from the
input space to a high dimensional feature space. The small value of σ2 can eventually lead
to overfitting. On the contrary, a large value can make the model more parsimonious and
not accurate (Espinoza et al., 2003).
• Divide the entire dataset into training, validation and testing datasets. The training
and validation datasets are used to build the LS-SVM model and the testing
dataset is used to verify the model’s performance.
• Initialize the hyper parameters γi, σ2i using the training dataset.
• Use the grid search algorithm with the cross-validation method to perform fine-
tuning of the optimal parameter combination (γopt, σ2opt), as obtained in Step 3.
The present research applied a ten-fold cross-validation method, dividing the
training into ten aliquot parts. The grid training data comprises nine aliquot parts
and the other part is the grid validation dataset. Train the LS-SVM model with the
grid training data and the optimized parameter combination (γopt, σ2opt). Test the
362
LS-SVM model with the validation data and iterate the process ten times. The
average training error is collected and calculated afterwards. Insert the new
parameter combination and repeat the process until the stop criteria are
approached. This process can successfully obtain the optimal parameter
combination (γ, σ2) with the minimized error.
• Adopt the tuned parameter combination (γ, σ2) to build the LS-SVM prediction
model. Substitute the testing dataset into the LS-SVM model and the prediction
data can now be obtained successfully.
The above steps are also illustrated in the flowchart presented in Figure C.2.
• To estimate performance of the learned model from available data using one
algorithm. In other words, to gauge the generalizability of an algorithm.
• To compare the performance of two or more different algorithms and find the best
algorithm for the available data, or alternatively, to compare the performance of
two or more variants of a parameterized model.
In order to avoid the possibility that the LS-SVM model over fits the data it learns from,
it is important to measure its performance when approximating data which have not been
used during the training process. One of the most common methods of obtaining a model
with good generalization capabilities consists in minimizing not just the training error but
the cross-validation error. The l-fold cross-validation error of a model is obtained by
dividing the available data in l sub-sets and, alternately using one of the sub-sets as test
data and the rest as training data. Therefore, a total of l models are trained and cross-
validated.
L-fold cross validation is an abstract pattern used for the estimation of the classification
accuracy of learning algorithms. It is used to determine the accuracy with which a
learning algorithm is able to predict inferences using unused sample data during the
training phase. However, 1-iteration of the L-fold cross-validation is performed in the
following way:
363
Figure C.2 − LS-SVM Modeling Procedure
364
• First, a random permutation of the validation dataset is generated and partitioned
into L folds of approximately equal size. Of the L subsets, a single subset is
retained as the validation data to estimate the performance of a predictive model,
and the remaining L – 1subsets, together, are used as training data.
• Then a model is trained on the training set and its accuracy is evaluated on the L-
fold. Model training and evaluation is repeated L times, with each of the L subsets
used exactly once as a testing set. Figure C.2 shows 1-iteration of a 3-fold cross
validation algorithmic pattern using 30 samples of data.
• initial value assignments of γ and σ2 were given and their search ranges were
determined
• according to the initial value assignments, the position of the first grid point of the
cross validation was determined
• the ranges of parameters were respectively divided into n equal spaces, and the n-
cross validation was used as the objective function of the grid calculation to
calculate all the grid points, and finally,
• the contour line was drawn based on the results of the cross validation. Under the
conditions of the contour line, the parameters γ and σ2 were sought. If the
obtained γ and σ2 do not meet the accuracy requirements, the range of the
parameters should be narrowed down and the above steps repeated until they meet
the accuracy requirements. Sequentially, the optimal parameter values were
obtained.
365
Figure C.2 – One-iteration of a 3-fold Cross Validation Algorithmic Pattern
366
C.7 References
Brabanter, J.D. et al., 2002. Robust cross-validation score function for LS-SVM non-
linear function estimation. International Conference on Artificial Neural Networks,
Madrid Spain, pp. 713-719.
Chen, Q.S. Zhao, J.W. Fang, C.H. and Wang, D.M., (2007). Spectroc. Acta Pt. A, 66 , p.
568.
Deng, S. and T.H. Yeh, 2010. Applying least square vector machines to airframe wing-
box structural design cost estimation. Expert Systems with Applications, 37: 8417-
8423.
Espinoza, M. Suykens, J.A.K. and De Moor, B., 2003. Least squares support vector
machines and primal space estimation Proceedings of the 42nd IEEE Conference on
Decision and Control Maui Hawaii, USA, pp. 3451–3456.
Guo, H. Liu, H.P. and Wang, L. (2006) J. Syst. Simul., 18 p. 2033.
Pelckmans, K., et al., 2003. LS-SVMlab Toolbox User’s Guide.
http://www.esat.kuleuven.ac.be/sista/lssvmlab/tutorial
Shi, D.F. Nabil, N.G., 2007. Tool wear predictive model based on least squares support
vector machines. Mechanical Systems and Signal Processing, vol. 21, pp. 1799-1814.
Suykens, J.A.K., Van Gestel, T., De Brabanter, J., De Moor, B., Vandewalle, J. (2002).
Vapnik, V.N. (1995), The Nature of Statistical Learning Theory, Springer-Verlag,
New York.
Vapnic, V., 1995. The Nature of Statistical Learning Theory, Springer, New York.
Wu, Di et al., 2008. Application of image texture for the sorting of tea categories using
multi-spectral imaging technique and support vector machine. JFE, vol. 88
367
APPENDIX D
ERROR ANALYSIS
D.1 Introduction
Statistical error analysis is used to analysis errors (N ) in order to check the performance
and the accuracy of the developed models. The difference (error) occurs due to
randomness or because the estimator does not account for information that could produce
a more accurate estimate. In trying to ascertain whether the error measures in the training
process are reliable, it should be considered whether the model under consideration is
likely to have over fitted the data. The overall data must split into training, validation and
testing subsets. The error measures in the training, validation and testing datasets are very
important. Indeed, the model's performance in the testing dataset is the best guide to its
ability to predict the future. However, it should be expect that the errors made in
predicting the future could be larger than those made in fitting the past.
For the purpose of communicating model results to others, it is usually best to report the
Mean Squared Error (MSE), Root Mean Squared Error (RMSE) and Absolute (AE) and
Absolute Relative Percent Error (ARE), Error Standard Deviation (SD), multiple
regression coefficients (R2) and minimum, maximum and average values of both relative
and absolute relative error.
The error term (N ) can be defined as the deviation of the calculated value from the true
value.
N = − (D.1)
where, is an observation and is a corresponding fitted value obtained by use of the
fitted regression model.
T*
&#S =
(D.2)
368
D.2 Root Mean Square Error (RMSE)
RMSE is frequently used to measure the differences between values predicted by a model
and the values actually observed. It also known as a quadratic mean. The RMSE can be
compared to an observed variation in errors of a typical model.
RMSE = (N VU0.5
(D.3)
#^ = _ ∑ − ̅ (D.6)
where, is the average value of the n observations of . A low SD indicates that the
data points tend to be very close to the average. Whereas high SD indicates that the data
are spread out over a large range of values.
∑ X3 4X3 *
S = a1 − ∑ X3 4X*
b× 100 (D.7)
369