Predicting Production Using A Neural Network (AI) 1995
Predicting Production Using A Neural Network (AI) 1995
SPEMember
This paper was prepared for presentation at the Petroleum Computer Conferenoe held In Houston, TX, U,S.A., 11-14 June 1995.
This paperwes selected for presentation by an SPE Program Committee following review of Information contained In an abstraot submitted by the author(s). Contents of the paper, as presented,
have not been reviewed by the Sooietyof Petroleum Engineers and are sUbjeoted to correction by the author(S). The material, as presented, does not neoessarily refleot any position of the
Sooiety of Petroleum Engineers, Its officers, or members. Papers presented at SPE meetings are sUbject to publication review by Editorial Committees of the Sooiety of Petroleum Engineers.
Permission to oopyls restrioted to an ebstraot of not more than 300 words. Illustrations may not be oopled. The abstraot should oontaln oonspiouous aoknowledgment of where and by whom
the paper Is presented. Write librarian, SPE, P,O. Box 833836, Rlohardson, TX 75083·3836, U,S,A. (Facsimile 214-952-9435).
196
SPE 30202 ROBERT 1. BOOMER 3
197
4 PREDICTING PRODUCTION USING A NEURAL NETWORK SPE 30202
(ARTIFICIAL INTELLIGENCE BEATS HUMAN INTELLIGENCE)
geologically similar locations, top of structure versus day (BOPD), at the one-, three-, SiX·, and twelve-
down dip. month time intervals.
198
SPE 30202 ROBERT J. BOOMER 5
type of data considered from each well in their A Neural Network initially assumes a random
prediction process. It is very difficult for humans to relationship between all the inputs and the desired
consider more than a handful of variables at anyone outputs. By comparing its first attempt at an answer
time. If a methodology could consider more offset to the desired output, it self-modifies this initial
wells and parameters, such as water production, random relationship into a relationship that best fits
injection withdrawal ratio, etc., then more accurate the outputs. Figure 5 is a graphical representation of a
predictions could be made. What is needed is a tool typical Neural Network.
that provides the professional the means to consider a
larger number of variables. Neural Networks are
ideally suited to this type of problem.
t Weight 11-H1
Input 1 \. Weight 11-H2
Hidden Node 1
• Weight H1..Q1
• Weight H1-02
neurons (a brain).5 Mathematically, Neural Networks l'Pm N \ W~1g1It IN-H' • W~ight HM·02
199
6 PREDICTING PRODUCTION USING A NEURAL NETWORK SPE 30202
(ARTIFICIAL INTELLIGENCE BEATS lIDMAN INTELLIGENCE)
network. It is called feed forward because the inputs The choice of which transform function and the
are presented to the network on one side and they number of hidden nodes is left to the Network
flow through the network to the outputs on the other designer, usually several combinations are tried and
side. the one yielding the smallest error is finally chosen.
The signals from the hidden nodes proceed in a similar
Every connection in a network is associated with a fashion to the output nodes where the final signal is
weight factor. These weight factors are independent scaled-up to yield the network's answer. The first pass
of each other. The mathematical workings of a of data through the network produces an answer that
network are simple. First the possible range in values is usually way off the mark since all the connection
for every input and output is determined. Then all weights are initially random numbers and have not yet
inputs and outputs are scaled for their range with 0 for learned anything. The networks answer is compared
the smallest possible value to 1 for the maximum to the correct answer and the network makes self-
possible value. At every node, any input to that node adjustments to all the connection weights to minimize
is multiplied by its weight factor and then summed this error.
with the product of the other inputs and their
This self-adjustment is the reason that another name
individual weight factors (see Figure 5). This resulting
for this type of network is IIback propagation ll
sum is processed by a transform function within the
network. The error is back propagated through the
node.
network by adjusting the connection weights in a
proportional fashion so the error is minimized. This
Two of the more commonly used transform functions whole process, feed forward and back propagate, is
are the hyperbolic tangent and logistic. They are repeated over and over again for every example the
IIsquashing functions;" that is, they pass along a signal network is supposed to learn. Learning is finished
with a value between -1 to 1 and 0 to 1, respectively. when the average error for all the examples is
This transform function provides non-linearity to the minimized. This iterative learning may take as many
network and constrains the node's signal within a fixed as several million passes of each example through the
range. Regardless of how lI excited ll an individual node network before an acceptable average error is reached.
gets, it has a fixed maximum signal that it passes on
through the network, just like their biological neuron Neural Networks are so good at learning that they can
counterparts. Figure 6 shows a single node and its use actually memorize all of the training examples. This
of the logistic function. usually produces a network that does not
IIunderstand, II it will be poor at applying its newfound
knowledge. It is just like the student who had a copy
of the answer sheet and just memorized the sequence
of answers. He may get a good grade on the test but
x 1 he does poorly when having to apply this knowledge.
200
SPE 30202 ROBERT J. BOOMER 7
The true test of the validity of a network is done by The professional typically looks only at the nearest
giving the network a "final exam." A third set of offset wells to a prospective drilling location. This
examples is used for this testing. The network error usually is the nearest four to eight wells. Since a
on this data set is the true measure of how good the Neural Network can easily handle more wells than
Neural Network performs. The selection of data for this, a Data Mask is utilized to capture data for wells
the Learning, Testing, and VerifYing data sets is not a that are not direct offset wells. The Data Mask is
matter to be taken lightly. To have a Neural Network shown in Figure 7. The Data Mask is a 5x5 grid
that you have confidence in requires that the data in covering 1,000 acres. Each square grid (1,320 feet
each of these three data sets be representative of the per side) equals 40 acres. This Data Mask is placed
problem you are trying to solve. In terms of Artificial over the drilling location's 40-acre tract and the data
Intelligence, the data must cover the Domain of the listed in Table 1 is collected for every well within the
problem. boundaries of the Data Mask. Since the number of
inputs in a Neural Network is a fixed number, the
It is not unusual to try various combinations of number of data values collected by the Data Mask
transform functions with a varying number of hidden have to be constant for every drilling location. This is
nodes and selected inputs before deciding upon the impossible as the number of wells within this Data
final network architecture. This entire process-feed Mask varied between less than 10 to greater than 50.
forward, back propagate, stop and test, and then The solution is to divide the Data Mask into three
continue training-is easily accomplished with concentric "rings." The "rings," labeled X, 10, and 9,
commercially available software. form a bullis-eye target with the drilling location
located at the center in the X-Ring. Data collected
Production Prediction Neural Network within each ring is then averaged so the data collected
in Table 1 is used with the Neural Network on a per-
The first step in building any Neural Network is to ring basis instead of a per-well basis. The point in
develop a data model that captures the data that a time that this current rate and cumulative data
professional would use to solve the problem. What represent is the calendar month preceding the month
inputs would an expert use to generate an output? that oil production for the newly drilled well is first
The review of the past methods of predicting reported to the state of New Mexico, i.e., initial
production is the starting point. Data used most often completion.
is either offset production data or reservoir
characterization data. Since production data is readily
available for all wells in the Vacuum Field and DATA MASK
reservoir characterization data is not readily available
for all the wells in the field, only production data is
. ------ ..... ----
selected as inputs. The specific production data
gathered for every offset well is shown in Table 1.
Oil Oil
Water Water
Percent Oil Percent Oil
Injection Injection 1120'
201
8 PREDICTING PRODUCTION USING A NEURAL NETWORK SPE30202
(ARTIFICIAL INTELLIGENCE BEATS HUMAN INTELLIGENCE)
In summary, the data input items, data an expert might This is the verify data set, data the network had never
use to make a production forecast, for the Neural seen while training. It shows that the Neural Network
Network are the current rates and cumulatives for all slightly underestimates production 38 percent of the
the wells within a %-mile radius of any proposed time. It has a mean error of -19 BOPD and a mean
drilling location expressed as an average for three absolute error of 27 BOPD. This network is over 3.5
distinct locations from the drilling location as of the times (a mean absolute error of 27 BOPD for the
month before the drilling well is initially completed. Neural Network versus and a mean absolute error of
The outputs for the Neural Network are the one-, 99 BOPD for the Professionals) more accurate than
three-, six-, and twelve-month cumulative production the Professionals.
volumes for the proposed drilling location. All this
well data was collected for each of the 250 producing
wells completed in the Vacuum Field since 1970. To Initial, 3, 6, & 12 Month
gather the data for these 250 wells required that data Average Cumulative Oil (SOPD)
for 7883 offset wells be captured. A schematic of the T PREDICTED
network is shown as Figure 8. 1000 r--~-~-------~-....."...,
1st Month
i.,2~~~"-I~k2~~::::::;;:;~
1~---l--+-f ............t-H+-+--I---+--lH-H_--t----l-+--H-+-t+l
X-Rlng Cumulative
Cumulative -: OIL 1 10 100 1000
011 3rd Month
ACTUAL
X-Rlng
Water Rate
~~~~~~~§§;~~~~ Cumulative
011
6th Month
• Verify's -Perfect Predictor
Oumulatlve
011
41st Year Figure 9 Neural Network's Performance
Oumulatlve
9-Ring 011
Cumulative Conclusions
IIW
202
SPE 30202 ROBERT J. BOOMER 9
'f SOPD
250 I I I i I ! i
200
i
······+..·
I
·1·· ·..··.. !..· ·..·.. t · I · ,.· ·
! i!
·'I'··..· ·r ·r I!
r····..·.·
150 1 j + .L + ++.
100 . · ··j···..· +· ··f ·I.. ··· ···· . ·..· ·.. j ·t· ·..·..t.. · .
50 ·..·..·!·· ··.. ·j· · i..· . · t·..· ····I ·..,..······ r.. .· i· ··.. ··t····..······[·..·..·..·..
o I 1 I ! i i II 1
1st Year
1 2 3 4 5 6 7 B 9 10 11 12 Average Produotion
Month tm50-100 BOPD ~100-150 BOPD _150-200 BOPD
-Actual .... Professional "'P- Neural
203
10 PREDICTING PRODUCTION USING A NEURAL NETWORK SPE 30202
(ARTIFICIAL INTELLIGENCE BEATS HUMAN INTELLIGENCE)
References
204