45
1 ESTIMATING PVT PROPERTIES OF CRUDE OIL 1 SYSTEM BASED ON FUNCTIONAL NETWORKS 2 Emad A. El-Sebakhy and S. Y. Al-Bokhitan 3 Information & Computer Science Department, College of Computer Sciences and Engineering, 4 King Fahd University of Petroleum & Minerals, Dhahran 31261, Saudi Arabia 5 [email protected] , [email protected] 6 Abstract 7 PVT properties are very important in the reservoir engineering computations. Numerous approaches have been proposed to estimate these PVT 8 properties, such as, empirical correlations, statistical regression, and neural networks modeling schemes. Unfortunately, the developed neural 9 networks correlations have some limitations as they were originally developed for certain ranges of reservoir fluid characteristics and 10 geographical area with similar fluid compositions. Accuracy of such correlations is often limited and global correlations are usually less 11 accurate compared to local correlations. Recently, functional networks have been proposed as a novel modeling scheme for both prediction and 12 classification based on selecting families of linearly independent functions to approximate each neuron functions under the threshold of minimum 13 description length. This new framework dealt with general form of neuron functions instead of sigmoid-like ones, which allows numerous of 14 choice for solving more complex nonlinear problems. It has featured in a wide range of medical, science and business applications, often with 15 promising results. The objective of this research is to investigate the capability of functional networks in modeling PVT properties of crude oil 16 systems and assess in building decision making tools in the field of oil and gas industry. 17 To demonstrate the usefulness of the functional networks modeling scheme in oil and gas industry, we briefly describe the required steps and the 18 learning algorithm of functional networks technique for predicting the PVT properties of crude oil systems. Comparative studies will be carried 19 out to compare their performance with the performance of neural networks, nonlinear regression, and the common empirical correlations 20 algorithms. Results show that the performance of functional networks is accurate, reliable, and outperform most of the existing approaches. 21 Future work can be achieved by using this new framework as a modeling approach for modeling other oil and gas industry problems, such as, 22 permeability and porosity prediction, identify liquid-holdup flow regimes, fracture corridor, identification of lithofacies types, seismic pattern 23 recognition, estimating pressure drop in pipes and wells, and optimization of well production, and other reservoir characterization. 24 Keywords ; Functional Networks, Feedforward Neural Networks PVT Properties, Empirical Correlations, Description Length 25 Feedforward neural networks, Empirical correlations, PVT properties; Formation volume factor; Bobble point pressure. 26

ESTIMATING PVT PROPERTIES OF CRUDE OIL …mlearn/FunctionalNetworks-EstPVT-Properties...1 ESTIMATING PVT PROPERTIES OF CRUDE OIL ... 23 permeability and porosity ... 2 Reservoir fluid

  • Upload
    lamnhan

  • View
    232

  • Download
    1

Embed Size (px)

Citation preview

1

ESTIMATING PVT PROPERTIES OF CRUDE OIL 1

SYSTEM BASED ON FUNCTIONAL NETWORKS 2

Emad A. El-Sebakhy and S. Y. Al-Bokhitan 3

Information & Computer Science Department, College of Computer Sciences and Engineering, 4 King Fahd University of Petroleum & Minerals, Dhahran 31261, Saudi Arabia 5

[email protected], [email protected] 6

Abstract 7

PVT properties are very important in the reservoir engineering computations. Numerous approaches have been proposed to estimate these PVT 8

properties, such as, empirical correlations, statistical regression, and neural networks modeling schemes. Unfortunately, the developed neural 9

networks correlations have some limitations as they were originally developed for certain ranges of reservoir fluid characteristics and 10

geographical area with similar fluid compositions. Accuracy of such correlations is often limited and global correlations are usually less 11

accurate compared to local correlations. Recently, functional networks have been proposed as a novel modeling scheme for both prediction and 12

classification based on selecting families of linearly independent functions to approximate each neuron functions under the threshold of minimum 13

description length. This new framework dealt with general form of neuron functions instead of sigmoid-like ones, which allows numerous of 14

choice for solving more complex nonlinear problems. It has featured in a wide range of medical, science and business applications, often with 15

promising results. The objective of this research is to investigate the capability of functional networks in modeling PVT properties of crude oil 16

systems and assess in building decision making tools in the field of oil and gas industry. 17

To demonstrate the usefulness of the functional networks modeling scheme in oil and gas industry, we briefly describe the required steps and the 18

learning algorithm of functional networks technique for predicting the PVT properties of crude oil systems. Comparative studies will be carried 19

out to compare their performance with the performance of neural networks, nonlinear regression, and the common empirical correlations 20

algorithms. Results show that the performance of functional networks is accurate, reliable, and outperform most of the existing approaches. 21

Future work can be achieved by using this new framework as a modeling approach for modeling other oil and gas industry problems, such as, 22

permeability and porosity prediction, identify liquid-holdup flow regimes, fracture corridor, identification of lithofacies types, seismic pattern 23

recognition, estimating pressure drop in pipes and wells, and optimization of well production, and other reservoir characterization. 24

Keywords –; Functional Networks, Feedforward Neural Networks PVT Properties, Empirical Correlations, Description Length 25

Feedforward neural networks, Empirical correlations, PVT properties; Formation volume factor; Bobble point pressure. 26

2

1. Introduction 1

Reservoir fluid properties are very important in petroleum engineering computations, such as, material balance 2

calculations, well test analysis, reserve estimates, inflow performance calculations, and numerical reservoir 3

simulations. Ideally, these properties are determined from laboratory studies on samples collected from the bottom 4

of the wellbore or at the surface. Such experimental data are, however, very costly to obtain. Therefore, the 5

solution is to use the empirically derived correlations to predict PVT properties, Osman et al.38. There are many 6

empirical correlations for predicting PVT properties, most of them were developed using equations of state (EOS) 7

or linear/nonlinear multiple regression or graphical techniques or feedforward neural networks (ANN or FFN) or 8

multilayer perceptron (MLP). However, they often do not perform very accurate and suffer from a number of 9

drawbacks. Each correlation was developed for a certain range of reservoir fluid characteristics and geographical 10

area with similar fluid compositions and API oil gravity. Thus, the accuracy of such correlations is critical and not 11

often known in advance. Among those PVT properties is the bubble point pressure (Pb), Oil Formation Volume 12

Factor (Bob), which is defined as the volume of reservoir oil that would be occupied by one stock tank barrel oil 13

plus any dissolved gas at the bubble point pressure and reservoir temperature. Precise prediction of Bob is very 14

important in reservoir and production computations. The objective of this study is to develop a new functional 15

networks prediction model for both Pb and Bob based on the kernel function scheme using worldwide experimental 16

PVT data. 17

18 Artificial neural networks have been proposed for solving many problems in the oil and gas industry, including 19

permeability and porosity prediction, identification of lithofacies types, seismic pattern recognition, prediction of 20

PVT properties, estimating pressure drop in pipes and wells, and optimization of well production. However, the 21

technique suffers from a number of numerous drawbacks. The main objective of this study is to investigate the 22

capability of functional networks on modeling PVT properties of crude oil systems and solve some of the neural 23

networks limitations, such as, the limited ability to explicitly identify possible causal relationships and the time-24

consumer in the development of back-propagation algorithm, which lead to an overfitting problem and gets stuck at 25

a local optimum of the cost function, Hinton39 and White40. To demonstrate the usefulness of this new 26

computational intelligence framework, the developed support vector machines regression prediction model was 27

3

developed using a database with 782 observations (after deleting the redundant 21 observations from the actual 803 1

observations in Osman et al.38 and Goda et al.41 published data gathered from Malaysia, Middle East, Gulf of 2

Mexico, and Colombia. That algorithm was found to be faster and more stable than other schemes reported in the 3

petroleum engineering literatures. The results show that the new support vector machines regression modeling 4

scheme outperforms both the standard neural networks and all the most common existing correlations models in 5

terms of absolute average percent error, standard deviation, and correlation coefficient. 6

2.0 Literature review and Related work 7

For the last 60 years engineers realized the importance of developing and using empirical correlations for PVT 8

properties. Studies carried out in this field resulted in the development of new correlations. In the following 9

subsections, we propose the most existing PVT properties estimation techniques in both petroleum engineering and 10

computer science communities. 11

2.1 Empirical Correlations PVT Models 12

For the last 60 years engineers realized the importance of developing and using empirical correlations for PVT 13

properties. Studies carried out in this field resulted in the development of new correlations. Standing1,3 presented 14

correlations for bubble point pressure and for oil formation volume factor. Standing’s correlations were based on 15

laboratory experiments carried out on 105 samples from 22 different crude oils in California. Katz2 presented five 16

methods for predicting the reservoir oil shrinkage. Vazquez and Beggs4 presented correlations for oil formation 17

volume factor. They divided oil mixtures into two groups, above and below thirty degrees API gravity. More than 18

6000 data points from 600 laboratory measurements were used in developing the correlations. Glaso5 developed 19

correlation for formation volume factor using 45 oil samples from North Sea hydrocarbon mixtures. Al-Marhoun6 20

published correlations for estimating bubble point pressure and oil formation volume factor for the Middle East oils. 21

He used 160 data sets from 69 Middle Eastern reservoirs to develop the correlation. Abdul-Majeed and Salman7 22

published an oil formation volume factor correlation based on 420 data sets. Their model is similar to that of Al-23

Marhoun6 oil formation volume factor correlation with new calculated coefficients. 24

25

4

Labedi8 presented correlations for oil formation volume factor for African crude oils. He used 97 data sets from 1

Libya, 28 sets from Nigeria, and 4 sets from Angola to develop his correlations. Dokla and Osman9 published set of 2

correlations for estimating bubble point pressure and oil formation volume factor for UAE crudes. They used 51 3

data sets to calculate new coefficients for Al-Marhoun6 Middle East models. Al-Yousef and Al-Marhoun10 pointed 4

out that the Dokla and Osman9,11 bubble point pressure correlation was found to contradict the physical laws. In 5

1992, Al-Marhoun12 published a second correlation for oil formation volume factor. The correlation was developed 6

with 11,728 experimentally obtained formation volume factors at, above, and below bubble point pressure. The data 7

set represents samples from more than 700 reservoirs from all over the world, mostly from Middle East and North 8

America. 9

10 Macary and El-Batanoney13 presented correlations for bubble point pressure and oil formation volume factor. They 11

used 90 data sets from 30 independent reservoirs in the Gulf of Suez to develop the correlations. The new 12

correlations were tested against other Egyptian data of Saleh et al.14, and showed improvement over published 13

correlations. Omar and Todd15 presented oil formation volume factor correlation, based on Standing1 model. Their 14

correlation was based on 93 data sets from Malaysian oil reservoirs. In 1993, Petrosky and Farshad16 developed new 15

correlations for Gulf of Mexico crude oils. Standing1 correlations for bubble point pressure, solution gas oil ratio, 16

and oil formation volume factor were taken as a basis for developing their new correlation coefficients. Ninety data 17

sets from Gulf of Mexico were used in developing these correlations. 18

19 Kartoatmodjo and Schmidt17 used a global data bank to develop new correlations for all PVT properties. Data from 20

740 different crude oil samples gathered from all over the world provided 5392 data sets for the correlation 21

development. Al-Mehaideb18 published a new set of correlations for UAE crudes using 62 data sets from UAE 22

reservoirs. These correlations were developed for bubble point pressure and oil formation volume factor. The 23

bubble point pressure correlation like Omar and Todd15 uses the oil formation volume factor as input in addition to 24

oil gravity, gas gravity, solution gas oil ratio, and reservoir temperature. 25

26 Saleh et al.14 evaluated the empirical correlations for Egyptian oils. They reported that Standing1 correlation was the 27

best for oil formation volume factor. Sutton and Farshad19, 20 published an evaluation for Gulf of Mexico crude oils. 28

They used 285 data sets for gas-saturated oil and 134 data sets for under saturated oil representing 31 different 29

5

crude oils and natural gas systems. The results show that Glaso5 correlation for oil formation volume factor perform 1

the best for most of the data of the study. Later, Petrosky and Farshad16 published a new correlation based on Gulf 2

of Mexico crudes. They reported that the best performing published correlation for oil formation volume is Al-3

Marhoun6 correlation. McCain21 published an evaluation of all reservoir properties correlations based on a large 4

global database. He recommended Standing1 correlations for formation volume factor at and below bubble point 5

pressure. 6

7 Ghetto et al.22 performed a comprehensive study on PVT properties correlation based on 195 global data sets 8

collected from the Mediterranean Basin, Africa, Middle East, and the North Sea reservoirs. They recommended 9

Vazquez and Beggs4 correlation for the oil formation volume factor. Elsharkawy et al.23 evaluated PVT correlations 10

for Kuwaiti crude oils using 44 samples. Standing1 correlation gave the best results for bubble point pressure while 11

Al-Marhoun6 oil formation volume factor correlation performed satisfactory. Mahmood and Al-Marhoun24 12

presented an evaluation of PVT correlations for Pakistani crude oils. They used 166 data sets from 22 different 13

crude samples for the evaluation. Al-Marhoun12 oil formation volume factor correlation gave the best results. The 14

bubble point pressure errors reported in this study, for all correlations, are among the highest reported in the 15

literature. Hanafy et al.25 published a study to evaluate the most accurate correlation to apply to Egyptian crude oils. 16

For formation volume factor Macary and El-Batanoney13 correlation showed an average absolute error of 4.9% 17

while Dokla and Osman9 showed 3.9%. The study strongly supports the approach of developing a local correlation 18

versus a global correlation. 19

20 Al-Fattah and Al-Marhoun26 published an evaluation of all available oil formation volume factor correlations. They 21

used 674 data sets from published literature. They found that Al-Marhoun12 correlation has the least error for global 22

data set. Also, they performed trend tests to evaluate the models’ physical behavior. Finally, Al-Shammasi27 23

evaluated the published correlations and neural network models for bubble point pressure and oil formation volume 24

factor for accuracy and flexibility to represent hydrocarbon mixtures from different geographical locations 25

worldwide. He presented a new correlation for bubble point pressure based on global data of 1661 published and 48 26

unpublished data sets. Also, he presented neural network models and compared their performance to numerical 27

correlations. He concluded that statistical and trend performance analysis showed that some of the correlations 28

6

violate the physical behavior of hydrocarbon fluid properties. Published neural network models for predicting the 1

missing major parameters are reproduced. 2

2.2 PVT Modeling Based on Neural Networks 3

Artificial neural networks are parallel-distributed information processing models that can recognize highly complex 4

patterns within available data. In recent years, neural network have gained popularity in petroleum applications. 5

Many authors discussed the applications of neural network in petroleum engineering28-32. Recently, it is shown in 6

both machine learning and data mining communities that artificial neural networks have the capacity to learn 7

complex linear/nonlinear relationships amongst input and output data. There are many different types of the neural 8

network. The most common widely used neural network in literature is known as the feedforward neural networks 9

with backpropagation training algorithm, Osman et al.38, Ali42, and Duda et al.43. This type of neural networks is 10

excellent computational intelligence modeling scheme in both prediction and classification tasks. Few studies were 11

carried out to model PVT properties using neural networks. Recently, feedforward neural network serves the 12

petroleum industry to predict the PVT correlations, Osman et al.38, Elsharkawy33-34. 13

14 The author in Al-Shammasi46,47 presented neural network models and compared their performance to numerical 15

correlations. He concluded that statistical and trend performance analysis showed that some of the correlations 16

violate the physical behavior of hydrocarbon fluid properties. In addition, he pointed out that the published neural 17

network models missed major model parameters to be reproduced. He uses (2HL) neural networks (4-5-3-1) 18

structure for predicting both properties: bubble point pressure and oil formation volume factor. He evaluates 19

published correlations and neural-network models for bubble point pressure (pb) and oil formation volume factor 20

(Bo) for their accuracy and flexibility in representing hydrocarbon mixtures from different locations worldwide. The 21

study presents a new, improved correlation for pb based on global data. It also presents new neural-network models 22

and compares their performances to numerical correlations. The evaluation examines the performance of 23

correlations with their original published coefficients and with new coefficients calculated based on global data, 24

data from specific geographical locations, and data for a limited oil-gravity range. The evaluation of each coefficient 25

class includes geographical and oil-gravity grouping analysis. The results show that the classification of correlation 26

models as most accurate for a specific geographical area is not valid for use with these two fluid properties. 27

7

Statistical and trend performance analysis shows that some published correlations violate the physical behavior of 1

hydrocarbon fluid properties. Published neural-network models need more details to be reproduced. New developed 2

models perform better but suffer from stability and trend problems. 3

The authors in Varotsis et al.36 introduced a novel approach for predicting the complete PVT behavior of reservoir 4

oils and gas condensates using neural network. The method uses key measurements that can be performed rapidly 5

either in the lab or at the well site as input to a neural network. This network was trained by a PVT studies database 6

of over 650 reservoir fluids originating from all parts of the world. Tests of the trained ANN architecture utilizing a 7

validation set of PVT studies indicate that, for all fluid types, most PVT property estimates can be obtained with a 8

very low mean relative error of 0.5-2.5%, with no data set having a relative error in excess of 5%. This level of error 9

is considered better than that provided by tuned Equation of State (EOS) models, which are currently in common 10

use for the estimation of reservoir fluid properties. In addition to improved accuracy, the proposed neural network 11

architecture avoids the ambiguity and numerical difficulties inherent to equations of state (EOS) models and 12

provides for continuous improvements by the enrichment of the neural network training database with additional 13

data. 14

The authors in Osman et al.38, Elsharkawy33-34 were carried comparative studies between the feedforward neural 15

networks performance and the four empirical correlations: Standing correlation, Al-Mahroun correlation, Glaso 16

correlation, and Vasquez and Beggs Correlation, see Osman et al.38, Goda et al.41, Danesh44, Smith et al. 45, and Al-17

Marhoun6 for more details about both the performance and the carried comparative studies results. In 1996, Gharbi 18

and Elsharkawy33-34 published neural network models for estimating bubble point pressure and oil formation volume 19

factor for Middle East crude oils. Gharbi and Elsharkawy33 use the neural system with log sigmoid activation 20

function to estimate the PVT data for Middle East crude oil reservoirs, while in Gharbi and Elsharkawy34 they 21

developed a universal neural network for predicting PVT properties for any oil reservoir. In Gharbi and 22

Elsharkawy33, two neural networks are trained separately to estimate the bubble point pressure (Pb) and oil 23

formation volume factor (Bob), respectively. The input data were solution gas-oil ratio, reservoir temperature, oil 24

gravity, and gas relative density. They used two hidden layers (2HL) neural networks: The first neural network, (4-25

8-4-2) to predict the bubble point pressure and the second neural network, (4-6-6-2) to predict the oil formation 26

volume factor. Both neural networks were built using a data set of size 520 observations from Middle East area. The 27

8

input data set is divided into a training set of 498 observations and a testing set of 22 observations. The authors in 1

Gharbi and Elsharkawy34 follow the same criterion of Gharbi and Elsharkawy33, but on large scale covering 2

additional area: North and South America, North Sea, South East Asia, with the Middle East region. They 3

developed a one hidden layer neural network using a database of size 5434 representing around 350 different crude 4

oil systems. This database was divided into a training set with 5200 observations and a testing set with other 234 5

observations. The results of their comparative studies were shown that the FFN outperforms the conventional 6

empirical correlation schemes in the prediction of PVT properties with reduction in the average absolute error and 7

increasing in correlation coefficients. 8

9 The author in Elsharkawy35 presented a new technique to model the behavior of crude oil and natural gas systems 10

using a radial basis function neural network (RBFN) model. The model can predict oil formation volume factor, 11

solution gas-oil ratio, oil viscosity, saturated oil density, under saturated oil compressibility, and evolved gas 12

gravity. He used differential PVT data of ninety samples for training and another ten novel samples for testing the 13

model. Input data to the RBFN model were reservoir pressure, temperature, stock tank oil gravity, and separator 14

gas gravity. Accuracy of the model in predicting the solution gas oil ratio, oil formation volume factor, oil 15

viscosity, oil density, under saturated oil compressibility and evolved gas gravity was compared for training and 16

testing samples to all published correlations. The comparison shows that the proposed model is much more accurate 17

than these correlations in predicting the properties of the oils. The behavior of the model in capturing the physical 18

trend of the PVT data was also checked against experimentally measured PVT properties of the test samples. He 19

concluded that although, the RBFN model was developed for specific crude oil and gas system, the idea of using 20

neural network to model behavior of reservoir fluid can be extended to other crude oil and gas systems as a 21

substitute to PVT correlations that were developed by conventional regression techniques. 22

23

The authors in Osman et al.38 use the feedforward learning scheme with log sigmoid transfer function in order to 24

estimate the formation volume factor at the bubble point pressure. The developed neural networks model was 25

developed using 782 (after deleting the redundant 21 observations from the actual 803 data published in Osman et 26

al.38, Goda et al.41 published data gathered from Malaysia, Middle East, Gulf of Mexico, and Colombia. They 27

designed a one hidden layer (1HL) feedforward neural network (4-5-1) with the backpropagation learning 28

9

algorithm: The input layer has four neurons covering the input data of gas-oil ratio, API oil gravity, relative gas 1

density, and reservoir temperature, one hidden layer with five neurons, and single neuron for the formation volume 2

factor in the output layer. The results of the developed calibration neural network model outperform the most 3

common empirical correlations techniques with absolute average error of 1.789%, and correlation coefficient of 4

0.988. 5

The authors in Al-Marhoun48 developed two new models to predict the bubble point pressure, and the oil formation 6

volume factor at the bubble-point pressure for Saudi crude oils. The models were based on artificial neural 7

networks, and developed using 283 unpublished data sets collected from different Saudi fields. Of the 283 data sets, 8

142 were used to train the Bob and Pb Artificial Neural Network models, 71 to cross-validate the relationships 9

established during the training process and adjust the calculated weights, and the remaining 70 to test the model to 10

evaluate its accuracy. The results show that the developed Bob model provides better predictions and higher 11

accuracy than the published empirical correlations. The neural networks model provides predictions of the 12

formation volume factor at the bubble point pressure with an absolute average percent error of 0.5116%, standard 13

deviation of 0.6626 and correlation coefficient of 0.9989. In addition, the developed Pb model outperforms the 14

published empirical correlations. It provides predictions of the bubble point pressure with an absolute average 15

percent error of 5.8915%, standard deviation of 8.6781 and correlation coefficient of 0.9965. 16

The authors in Osman and Abdel-Aal49 introduced the abductive Network as an alternative modeling tool, which 17

avoids many of the neural networks drawbacks. Unlike neural network, the abductive network uses various types of 18

more powerful polynomial functional elements based on prediction performance, which is based on the self-19

organizing group method of data handling (GMDH), this technique uses well-proven optimization criteria for 20

automatically determining the network size and connectivity, and element types and coefficients for the optimum 21

model, thus reducing the modeling effort and the need for user intervention. The abductive network model 22

automatically selects influential input parameters and the input-output relationship can be expressed in polynomial 23

form. In this research they used abductive network to predict both Bubble point pressure (Pb) and Bubble point Oil 24

Formation Volume Factor (Bob) for Saudi crude oils. The abductive networks models were implemented based on 25

283 observations published in Al-Marhoun et al.48. Of the 283 data sets, 198 were used to train the Bob and Pb 26

abductive network models and 85 to test the model to evaluate its accuracy. They obtained an acceptable agreement 27

10

between the measured and predicted values of Pb, with correlation coefficient is 0.9898 and the average absolute 1

percentage error is 5.62%. This indicates that the abductive network model outperforms all other empirical 2

correlations, where average absolute percentage error in predicting Pb values using empirical equations is around 3

13% as it is reported in McCain et al.51 (1998). Moreover, the Bob was predicted using abductive networks using 4

only solution gas-oil ration, Rs and reservoir temperature, Tf with correlation coefficient is 0.9959 and the average 5

absolute percentage error is 0.86%, which is better than the published empirical correlations. 6

The authors in Goda et al.41 used the feedforward neural networks with both log-sigmoid and pure linear activation 7

functions and backpropagation training algorithm to estimate both bubble point pressure and oil formation volume 8

factor through two linked feedforward neural networks from distinct inputs: gas-oil ratio, API oil gravity, relative 9

gas density, and reservoir temperature. They use a single neuron in the output layer that is, joint with pure linear 10

activation function. The first network architecture was selected as a two hidden layers (2HL) neural network (4-10-11

10-1) to predict the bubble point pressure from data set of 180 observations collected from Middle East oil systems. 12

The provided data set is divided to a training set of size 160 and a testing set of size 20. The second chosen neural 13

network was two hidden layers (2HL) neural network (5-8-8-1) to predict the oil formation volume factor. The input 14

layer has five neurons for five inputs which are: gas-oil ratio, API oil gravity, relative gas density, and reservoir 15

temperature, and the estimated bubble point pressure from the first network using the same data set used in the first 16

neural network. The main difference between the work in Goda et al.41 and other work carried out on the prediction 17

of PVT data by feedforward neural network is that the authors use one neural network to estimate bubble point 18

pressure using the four input variables: gas-oil ratio, API oil gravity, relative gas density, and reservoir 19

temperature. Next, they used a second neural network with five input parameters (the same four input variable plus 20

the estimated bubble point pressure from the first neural network) in order to determine the new output of oil 21

formation volume factor. 22

The authors in Osman and Al-Marhoun50 developed two new models to predict different brine properties. The first 23

model predicts brine density, formation volume factor (FVF), and isothermal compressibility as a function of 24

pressure, temperature and salinity. The second model is developed to predict brine viscosity as a function of 25

temperature and salinity only. An attempt was made to develop a comprehensive model to predict all properties in 26

terms of pressure, temperature, and salinity. The results were satisfactory for all other properties except for 27

11

viscosity. This was attributed to the fact that viscosity depends only on temperature and salinity. The models were 1

developed using 1040 published data sets. These data were divided into three groups: training, cross-validation and 2

testing. Radial Basis Functions (RBF) and Multi-layer Preceptor (MLP) neural networks were utilized in the study. 3

Trend tests were performed to ensure that the developed model would follow the physical laws. Results show that 4

the developed models outperform the published correlations in terms of absolute average percent relative error, 5

correlation coefficients, and standard deviation. 6

2.3 The Shortcoming of Neural Networks 7

Experience with neural networks has revealed a number of limitations for the technique. One such limitation is the 8

complexity of the design space15. With no analytical guidance on the choice of many design parameters, the 9

developer often follows an ad hoc, trial-and-error approach of manual exploration that naturally focuses on just a 10

small region of the potential search space. Architectural parameters that have to be guessed a priori include the 11

number and size of hidden layers and the type of transfer function(s) for neurons in the various layers. Learning 12

algorithm parameters to be determined include initial weights, learning rate, and momentum. Although acceptable 13

results may be obtained with effort, it is obvious that potentially superior models can be overlooked. The 14

considerable amount of user intervention not only slows down model development, but also works against the 15

principle of ‘letting the data speak’. To automate the design process, external optimization criteria, e.g. in the form 16

of genetic algorithms, have been proposed16. Over-fitting or poor network generalization with new data during 17

actual use is another problem17. As training continues, fitting of the training data improves but performance of the 18

network with new data previously unseen during training may deteriorate due to over-learning. A separate part of 19

the training data is often reserved for monitoring such performance in order to determine when training should be 20

stopped prior to complete convergence. However, this reduces the effective amount of data used for actual training 21

and would be disadvantageous in many situations where good training data are often scarce. Network pruning 22

algorithms18 have also been described to improve generalization. The commonly used back-propagation training 23

algorithm with a gradient descent approach to minimizing the error during training suffers from the local minima 24

problem, which may prevent the synthesis of an optimum model19. Another problem is the opacity or black-box 25

nature of neural network models. The associated lack of explanation capabilities is a handicap in many decision 26

12

support applications such as medical diagnostics, where the user would usually like to know how the model came to 1

a certain conclusion. Additional analysis is required to derive explanation facilities from neural network models; 2

e.g. through rule extraction20. Model parameters are buried in large weight matrices, making it difficult to gain 3

insight into the modeled phenomenon or compare the model with available empirical or theoretical models. 4

Information on the relative importance of the various inputs to the model is not readily available, which hampers 5

efforts for model reduction by discarding less significant inputs. Additional processing using techniques such as the 6

principal component analysis21 may be required for this purpose. 7

3.0 Functional Networks 8

Recently, functional networks have been introduced by [57, 58] as a generalization of the standard neural networks. 9

It dealt with general functional models instead of sigmoid-like ones. This new framework have not been used in the 10

field of software engineering and none of the above authors, however, used and/or evaluated the capability of 11

functional networks in predicting defect-prone classes. The main goal of this paper is to investigate the capability of 12

this new computational intelligence modeling scheme in spam filtering and evaluate its performance against the 13

most common data mining schemes, such as, two neural networks techniques: (i) feedforward neural networks 14

(FFN) and (ii) radial basis function (RBF). The results and the detailed performance are shown in details at the end 15

of the paper. 16

In Functional Networks, the functions associated with the neurons are not fixed but are learnt from the available 17

data. There is no need, hence, to include weights associated with links, since the neuron functions subsume the 18

effect of weights. Functional networks allow neurons to be multi-argument, multivariate, and different learnable 19

functions, instead of fixed functions. Functional networks allow converging neuron outputs, forcing them to be 20

coincident. This leads to functional equations or systems of functional equations, which require some compatibility 21

conditions on the neuron functions. Functional networks have the possibility of dealing with functional constraints 22

that are determined by the functional properties of the network model. 23

24 Functional networks as a new modeling scheme has been used in solving both prediction and classification 25

problems. It is a general framework useful for solving a wide range of problems in probability, statistics, signal 26

processing, pattern recognition, functions approximations, real-time flood forecasting, science, bioinformatics, 27

13

medicine, structure engineering, and other business and engineering applications, see [57-60], and the references 1

therein for more details. The performance of functional networks has shown bright outputs for future applications in 2

both industry and academic research of science and engineering based on its reliable and efficient results. Several 3

comparative studies have been carried to compare its performance with the performance of the most popular 4

prediction/classification modeling data mining, machine learning schemes in literature [12, 13]. The results show 5

that functional networks performance outperforms most of these popular modeling schemes in machine learning, 6

data mining, and statistics communities. Dealing with functional networks in prediction/classification required some 7

concepts and definitions, which can be briefly discussed as follows: 8

We use the set 1{ }px … x= , ,X to be the set of nodes, such that each node ix is associated with a variable iX . 9

The neuron (neural) function over a set of nodes X is a tuple U x f=< , , >z , where x is a set of the input 10

nodes, f is a processing function and z is the output nodes, such that ( )z f x= , where x and Z are two non–11

empty subsets of X . We illustrate the use of functional networks in problem by an example: 12

13

Figure 1 Functional network architecture: An example 14

15

As it can be seen in Fig.1, a functional network consists of: a) several layers of storing units, one layer for 16

containing the input data (xi; i = 1,2,3, 4), another for containing the output data (x7) and none, one or several layers 17

to store intermediate information (x5 and x6); b) one or several layers of processing units that evaluate a set of input 18

values and delivers a set of output values (fi); and c) a set of directed links. Generally, functional networks extend 19

the standard neural networks by allowing neuron functions fi to be not only true multiargument and multivariate 20

functions, but to be different and learnable, instead of fixed functions. In addition, the neuron functions in 21

14

functional networks are unknown functions from a given family, such as, polynomial, exponential, Fourier...etc, to 1

be estimated during the learning process. Furthermore, functional networks allow connecting neuron outputs, 2

forcing them to be coincident [12, 18, and 19]. 3

4 The functional network uses two types of learning: a) structural learning b) parametric learning. In structural 5

learning, the initial topology of the network, based on some properties available to the designer is arrived at and 6

finally a simplification is made using functional equation to a simpler architecture. In parametric learning, usually 7

activation functions by considering the combination of ‘‘basis’’ functions are estimated by using the least square, 8

steepest descent and mini-max methods [29]. In this paper we use the least square method for estimating activation 9

functions in both classification and prediction. 10

Generally, functional network is a problem driven, which means that the initial architecture is designed based on a 11

problem in hand. In addition, to the data domain, information about the other properties of the function, such as 12

associativity, commutativity, and invariance, are used in selecting the final network. In the functional network, 13

neuron functions are arbitrary known or unknown to be learned from the provided data, but in neural networks they 14

are a sigmoid, linear or radial basis and other functions. In functional networks, neuron functions in which weights 15

are incorporated are learned, and in neural networks, weights are learned. Neural networks work well if both input 16

and output data are normalized in a specific range, say between 0 and 1, but in functional networks there is no such 17

restriction. Furthermore, it can be pointed out that neural networks are special cases of functional networks [57]. 18

Dealing with functional networks required the following six-steps through the functional networks implementations 19

and learning process: 20

- Statement of the problem. 21

- Specify the initial topology based on the domain of the problem in hand. 22

- Simplify the initial architecture using functional equations and the equivalence concept and check the 23

uniqueness condition of the desired architecture, see [18 and 19] for more details. 24

- Gathering the required data and handle multicollinearity problem with the implementation of the required 25

quality control check before the functional networks implementation. 26

15

- The learning procedures and training algorithm based on either Structure or Parametric learning by 1

considering the combinations of linear independent functions, 1 2{ }ss s sm…ψ ψ ψΨ = , , , , for all s to 2

approximate the neuron functions, that is, 3

1

( ) ( ) for allsm

s si sii

g x a x sψ=

,∑ (1) 4

where the coefficients sia are the parameters of the functional networks. The most popular linearly 5

independent functions in literature are: 6

• {1 }mX … XΨ = , , , , or 7

• {1 ( ) ( ) ( )}l lCos X … Cos X Sin XΨ = , , , , , where 2m l= , or 8

• {1 }X X mX mXe e … e e− −Ψ = , , , , , , 9

where m is the number of elements in the combination of sets of linearly independent function. The 10

parameters in (1) can be learned using one of the known optimization (loss criterion) techniques, such as, 11

least squares, conjugate gradient, iterative least squares, minmax, or maximum likelihood estimation. 12

- Model selection and validation. The best functional networks model is chosen based on the minimum 13

description length and some other quality measurements, such as, the correlation coefficients and the root-14

mean-squared errors. The selection is achieved using one of the well known selection schemes, such as, 15

exhaustive selection, forward selection, backward elimination, backward forward selection, and forward 16

backward elimination. 17

- Finally, if the fifth step is satisfactory, then the functional networks model is ready to be used in predicting 18

unseen new unseen data sets from real-world industry applications. 19

20 The learning method of a functional network consists of obtaining the neuron functions based on a set of data 21

{ }D , Y= ,X where R p∈X represents the matrix of size ( n p× ) of p input feature variables and the column 22

Y R∈ (Forecasting) and R Y ⊆ (Classification) is an output vector. 23

24 In the case of pattern classification, we assumed that the observation vector is of fixed length, say p , and that Y is a 25

set of class labels kA , for all 1k … c= , , . Therefore, we assume that we have given a training set 26

16

1{ }i i ipD y x … x= , , , for all 1i … n= , , of feature (predictor) variables1 pX … X, , , where

1

T

j j njX x … x⎛ ⎞⎜ ⎟⎝ ⎠

= for all 1

1j … p= , , drawn from c classes, and the categorical response variable ( )1T

nY y … y= , where {0 }iy … c∈ , , 2

represents a group number, for all 1i … n= , , . We use the lower case letters 1i ipx … x, , for all 1i … n= , , to refer 3

to the values of each observation of the predictor variables, and y k= to the response variable Y to refer to class 4

kA for all 0 1k … c= , , , , where 2c ≥ . We use ikπ for the probability that observation i falls in class kA , that 5

is, 6

1 2 1 20

( | ) ( | ) where 0 0 1 1 and 1c

ik i k i i ip i i i ip ik ikk

P A x x … x P y k x x … x k … c i … nπ π π.=

= ∈ , , , = = , , , , ≥ , ∀ = , , , , = , , , = ,∑x (2) 7

Where 1( )i i ipx … x. ≡ , ,x represents the ith observation of input matrix, x . The goal in both forecasting and 8

classification is to determine the linear/nonlinear relationship among the output and the input variables using one of 9

the following equations: 10

( ) ( )1 2 1 2( ) or ( ) ; k=1,2,...,c.i i i ip i ik i i ip ig y f x x … x g f x x … xε π ε= , , , + = , , , + (3) 11

or 12

( ) ( )1 2 1 2 or ; k=1,2,...,c.i i i ip i ik i i ip iy f x x … x f x x … xε π ε= , , , + = , , , + (4) 13

In this paper we propose two distinct functional networks called Generalized Associativity Model and Separable 14

Functional Networks Model to approximate( )1i ipf x … x, ,

. We briefly expressed these models as follows: 15

16 1. The Generalized Associativity Model which leads to the additive model [57,58]: 17

( ) ( )1 21

.k

k k kj

f x x … x h x=

, , , = ∑ (5) 18

The corresponding functional network is shown in Figure 2. 19

20

17

1

Figure 2 Functional network representing additive model 2

3

2. The Separable Functional Networks Model which considers a more general form for ( )1i ipf x … x, , : 4

( ) ( ) ( ) ( )

1 2

1 2 1 1 2 2

1 2

1 2 ...1 1 1

... ... , (6)k

k k k

k

qq q

k r r r r r r r r rr r r

f x x … x C h x h x h x= = =

, , , = ∑∑ ∑ 5

where 1 2 ... kr r rC are unknown parameters and the sets of functions { }: 1,..., ,

ss r s sr qφΦ = = s =1,2,...,k, 6

are linearly independent. An example of this functional network for k =2 and q1 = q2 = q is shown in Figure 7

3. Equations (5) and (6) are functional equations since their unknowns are functions. Their corresponding 8

functional networks are the graphical representations of these functional equations. 9

10

Figure 3. The functional network for the separable model with p =2 and 1 2 .q q q= = 11

18

We note that the graphical structure is very similar to a neural network, but the neuron functions are unknown. Our 1

problem consists of learning 1 2, ,..., kh h h in (5) and 1 2 ... kr r rC in (6). In order to obtain 1 2, ,..., kh h h in (5), we 2

approximate each ( ) ,j jh x for j=1,2,…,k by a linear combination of sets of linearly independent functions jsψ 3

defined above ,that is, 4

( ) ( )1

; 1,..., , (7)jq

j j js js js

h x a x j pψ=

= =∑ 5

and the problem is reduced to estimate the parameters jsa , for all j and s, see [57,58] for more details. 6

Generally, by setting appropriate input and applying system identification to study the defect prone classes 7

identification, one can customize the characteristics of input value according to wishful output. In this paper, we 8

follow the same procedures in both [57,58] and choose the least squares criterion to learn the parameters, but the 9

additive model requires add some constrains to guarantee uniqueness. Alternatively, one can choose different 10

optimization criterion based on his interest. The main advantage in choosing the least squares method is that the 11

least squares criterion leads to solve a linear system of equations in both prediction and classification problems. 12

4.0 Data Acquisition and Implementation Process 13

Data sets used for this work are the three different data sets used by. These data sets are collected from distinct 14

published sources: Al-Marhoun6, (Al-Marhoun48 or Osman and Abdel-Aal49), and Osman et al.38. We have done the 15

quality control check on all these data sets and clean it from redundant data and un-useful observations, for instance, 16

the data set used in Osman et al.38 after dropping the repeated data sets, we ended up with 782 not 803 data sets: 17

Katz2 (53), Vazquez and Beggs4 (254), Glaso5 (41), Ghetto et al.22 (173), Omar and Todd15 (93), Gharbi and 18

Elsharkawy33 (22), and Farshad et al.37 (146). Each data set contains reservoir temperature, oil gravity, total 19

solution gas oil ratio, and average gas gravity, bubble point pressure and oil formation volume factor at the bubble 20

point pressure. The repeated data sets were reported by other investigators26, 27. 21

To evaluate the performance of each modeling scheme, the entire database is divided using the stratified criterion. 22

Therefore, we use 70% of the data for building the support vector machines learning model (internal validation) and 23

30% of the data for testing/ validation (external validation or cross-validation criterion). We repeat both internal and 24

external validation processes for 1000 times. Therefore, the data were divided into two or three groups for training 25

19

and cross validation check. Therefore, Of the 782 data points, 382 were used to train the neural network models, the 1

remaining 200 to cross-validate the relationships established during the training process and 200 to test the model to 2

evaluate its accuracy and trend stability. For the testing data, the statistical summary of the investigated quality 3

measures corresponding to FunNet modeling scheme, neural networks system, Standing1, 3 correlation, Al-Mahroun6 4

correlation, and Glaso5 correlation using three different data sets (Al-Marhoun6, (Al-Marhoun48 or Osman and Abdel-5

Aal49), and Osman et al.38) to predict both bubble point pressure (Pb) and Oil Formation Volume Factor (Bob) were 6

given in Tables 1 through 6, respectively. 7

We observe that this cross-validation criterion gives the ability to monitor the generalization performance of the 8

support vector machine regression scheme and prevent the kernel network to over fit the training data. In this 9

implementation process, we used three distinct kernel functions, namely, polynomial, sigmoid kernel, and Gaussian 10

Bell kernel. In designing the support vector machine regression, the important parameters that will control its 11

overall performance were initialized, such as, kernel='poly'; kernel option = 5; epsilon = 0.01; lambda = 12

.0000001; verbose=0, and the constant C either 1 or 10 for simplicity. The cross-validation method used in this 13

study utilized as a checking mechanism in the training algorithm to prevent both over fitting and complexity 14

criterion based on the root-mean-squared errors threshold. 15

4.1 Quality Control and Data Domain 16

During the implementation, the user should be aware of the input domain values to make sure that the input data 17

values fall in a natural domain. This step called the quality control and it is really an important step to have very 18

accurate and reliable results at the end. The following is the most common domains for the input/output variables, 19

gas-oil ratio, API oil gravity, relative gas density, reservoir temperature; bubble point pressure, and oil formation 20

volume factor that are used in the both input and output layers of modeling schemes for PVT analysis: 21

22 • Gas oil ratio range between 26 and 1602, scf/stb. 23

• Oil formation volume factor varies from 1.032 to 1.997, bbl/stb. 24

• Bubble point pressure, starting from 130 psia, ending with 3573, psia. 25

• Reservoir temperature range is 74° F to 240° F. 26

• API gravity which changes between 19.4 to 44.6. 27

20

• Gas relative density, change from 0.744 to 1.367. 1

4.2 Statistical Quality Measures 2

To compare the performance and accuracy of the new model to other empirical correlations, statistical error analysis 3

is performed. The statistical parameters used for comparison are: average percent relative error (Er), average 4

absolute percent relative error (Ea), minimum and maximum absolute percent error (Emin and (Ermax)) root mean 5

square errors (Erms), standard deviation (SD), and correlation coefficient (r), see Osman et al.38 for more details 6

about the corresponding equations of these statistical parameters. The statistical quality measures for the 7

investigated data in this research are given in Tables 1, 2, and 3, respectively. 8

9

Table 1 STATISTICAL DESCRIPTION OF THE DATA USED FOR PVT MODELS (160 RECORDS) 10

Parameter Min Max Average Stdev Temperature (Tf ), ºF 74 240 144.43 39.036 Gas-Oil Ratio (Rs) 26 1602 557.66 403.12 Gas Relative Density (γg) 0.69 1.367 0.96417 0.1711 API Oil Gravity (API) 19.4 44.6 32.388 5.7444 Bubble point Pressure (Pb), psi 130 3573 1731.1 1084.6 Oil FVF at Pb, RB/STB 1.032 1.997 1.3036 0.206

11

Table 2 STATISTICAL DESCRIPTION OF THE DATA USED FOR PVT MODELS (283 RECORDS) 12

Parameter Min Max Average Stdev Temperature (Tf ), ºF 75 240 147.35 47.772 Gas-Oil Ratio (Rs) 24 1453 432.46 303.58 Gas Relative Density (γg) 0.7527 1.8195 1.008 0.1483 API Oil Gravity (API) 17.5 44.6 31.622 5.2518 Bubble point Pressure 90 3331 1390.2 860.66 Oil FVF at Pb, RB/STB 1.0308 1.889 1.2504 0.1582

13

Table 3 STATISTICAL DESCRIPTION OF THE DATA USED FOR PVT MODELS (782 RECORDS) 14

Parameter Min Max Average Stdev Temperature (Tf ), ºF 58 341.6 181.9 51.984 Gas-Oil Ratio (Rs) 8.61 3617.3 541.75 483.68 Gas Relative Density (γg) 0.511 1.789 0.88825 0.1856 API Oil Gravity (API) 11.4 63.7 34.588 8.7286 Bubble point Pressure 107.33 7127 2006.1 1291.2 Oil FVF at Pb, RB/STB 1.028 2.887 1.3362 0.2721

15

21

1

5.0 results And Discussions 2

The paper provides the estimation of PVT properties using functional networks as a novel paradigm modeling 3

scheme. After training the support vector machines regression, the model becomes ready for testing and evaluation 4

using the cross-validation criterion. Comparative studies were carried out to compare the performance and accuracy 5

of the new FunNet model versus both the standard neural networks and the three common published empirical 6

correlations: Standing1,3, Al-Mahroun6, and Glaso5 correlations. One can investigate other common empirical 7

correlations, such as, Vazquez and Beggs (1980)4 and Al-Marhoun (1988)6 besides these chosen empirical 8

correlations, see Osman et al.38 and Al-Shammasi46,47 for more details about these empirical correlations 9

mathematical formulas. The results of comparisons in the testing (external validation check were summarized in 10

Tables 1 through 6, respectively. We observe form these results that the new computational intelligence (FunNet) 11

modeling scheme outperforms both neural and abductive networks, and the most common published empirical 12

correlations. The proposed model shown a high accuracy in predicting the Bob values with a stable performance, 13

and achieved the lowest absolute percent relative error, lowest minimum error, lowest maximum error, lowest root 14

mean square error, and the highest correlation coefficient among other correlations for the used three distinct data 15

sets. 16

We draw a scatter plot of the absolute percent relative error (EA) versus the correlation coefficient for all 17

computational intelligence forecasting schemes and the most common empirical correlations. Each modeling 18

scheme is represented by a symbol; the good forecasting scheme should appear in the upper left corner of the graph. 19

Figure 3 shows a scatter plot of (EA) versus (r) for all modeling schemes that are used to determine Bo based on the 20

data set used in Osman et al.38. We observe that the symbol corresponding to FunNet scheme falls in the upper left 21

corner with EA = 1.368% and r = 0.9884, while neural network is below FunNet with EA = 1.7886% and r = 0.9878, 22

and other correlations indicates higher error values with lower correlation coefficients, for instance, Al-Marhoun 23

(1992) has EA = 2.2053% and r = 0.9806; Standing1 has EA = 2.7238% and r = 0.9742; and Glaso5 Correlation with 24

EA = 3.3743% and r = 0.9715. Figure 4 shows the same graph type, but with Pb instead, based on the data set used 25

in Osman et al.38 using the same modeling schemes. We observe that the symbol corresponding to FunNet falls in 26

the upper left corner with EA = 1.368% and r = 0.9884, while neural network is below FunNet with EA = 1.7886% 27

22

and r = 0.9878, and other correlations indicates higher error values with lower correlation coefficients, for instance, 1

Al-Marhoun (1992) has EA = 2.2053% and r = 0.9806; Standing1 has EA = 2.7238% and r = 0.9742; and Glaso5 2

Correlation with EA = 3.3743% and r = 0.9715. The same implementations process were repeated for the other data 3

sets used in Al-Marhoun6, (Al-Marhoun48 or Osman and Abdel-Aal49), but for the sake of simplicity, we did not 4

include it in this context. 5

6 Figures 4-15 illustrate six scatter plots of the predicted results versus the experimental data for both Pb and Bob 7

values using the same three distinct data sets. These cross plots indicates the degree of agreement between the 8

experimental and the predicted values based on the high quality performance of the FunNet modeling scheme. The 9

reader can compare theses patterns with the corresponding ones of the published neural networks modeling and 10

common empirical correlations in Al-Marhoun6, Al-Marhoun48, Osman and Abdel-Aal49, and Osman et al.38. Finally, 11

we conclude that developed FunNet modeling scheme has better and reliable performance compared to the most 12

published modeling schemes and empirical correlations. 13

The bottom line is that, the developed FunNet modeling scheme outperforms both the standard feedforward neural 14

networks and the most common published empirical correlations in predicting both Pb and Bob using the four input 15

variables: solution gas-oil ratio, reservoir temperature, oil gravity, and gas relative density. 16

17

18 Figure 4 Pb Performance of 160 data set 19

20

23

0 500 1000 1500 2000 2500 3000 3500 40000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 611.518

Actual output

R2 = 0.865

1 Figure 5 Standing model performance of Pb for 160 data set 2

3

0 500 1000 1500 2000 2500 3000 3500 40000

1000

2000

3000

4000

5000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 513.6229

Actual output

R2 = 0.976

4 Figure 6 Glaso model performance of Pb for 160 data 5

6

0 500 1000 1500 2000 2500 3000 3500 40000

1000

2000

3000

4000

5000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 226.2146

Actual output

R2 = 0.982

7 Figure 7 Marhoun 92 model performance of Pb for 160 data 8

9

24

1 1.2 1.4 1.6 1.8 21

1.5

2

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.014595

Actual output

R2 = 0.997

1 1.2 1.4 1.6 1.8 21

1.5

2P

redi

cted

out

put

Testing Performance: RMSE = 0.011748

Actual Output

R2 = 0.999

1 Figure 8 ANN model performance of Pb for 160 data set 2

3

0 500 1000 1500 2000 2500 3000 3500 40000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 67.7134 and MDL = 502.7821

Actual output

R2 = 0.998

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 47.7293 and MDL = 502.7821

Actual Output

R2 = 0.999

4 Figure 9 FunNet model performance of Pb for 160 data 5

6

7 Table 4 Pb performance of 160 data set 8

Er Ea Emin Emax RMSE SD R Standing 12.811 24.684 0.62334 59.038 611.52 25.159 0.86537

Glaso -18.887 26.551 0.28067 98.78 513.62 25.171 0.97569

25

Marhoun 5.1023 8.9416 0.13115 87.989 226.21 12.839 0.98234 NN training -1.3049 5.2929 0.009235 41.471 56.712 9.6324 0.99854 NN testing -3.1499 7.3483 0.22508 69.015 59.172 14.871 0.9988

FunNet training -0.34191 5.3787 0.039086 45.405 67.713 9.2256 0.99802 FunNet testing -0.53716 5.6194 0.086737 46.107 47.729 10.097 0.99914

1

Table 5 Bo performance of 160 data set 2

Er Ea Emin Emax RMSE SD R Standing -2.628 2.7202 0.016741 13.292 0.057954 2.5823 0.9953 Glaso -0.45136 2.0084 0.03218 11.075 0.041129 2.673 0.99352 Marhoun -0.55292 0.9821 0.008597 6.5123 0.020783 1.2842 0.99586 NN training 0.12729 0.84748 0.036196 4.6519 0.014595 1.109 0.99748 NN testing -0.0709 0.66818 0.085465 2.3676 0.011748 0.89835 0.99897 FunNet training -0.00932 0.66562 0.001801 3.9941 0.011498 0.9197 0.99839 FunNet testing 0.20413 0.4922 0.010289 1.3852 0.007869 0.58667 0.99943

3

0.5 1 1.5 2 2.50.8

0.85

0.9

0.95

1

1.05

Absolute Percent Relative Error (EA)

Cor

rela

tion

Coe

ffici

ent (

Rho

)

Standing (1947)Glaso (1980)Al-Marhoun (1992)ANN TrANN TsFunNet TrFunNet Ts

4 Figure 10 Bo performance of 160 data 5

6

26

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 21

1.5

2

2.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.057954

Actual output

R2 = 0.995

1 Figure 11 Standing model performance of Bo for 160 data 2

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 21

1.5

2

2.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.041129

Actual output

R2 = 0.994

3

Figure 12 Glaso model performance of Bo for 160 data 4

5

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 21

1.5

2

2.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.020783

Actual output

R2 = 0.996

6 Figure 13 Marhoun model performance of Bo for 160 data 7

27

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 21

1.2

1.4

1.6

1.8

2

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.014595

Actual output

R2 = 0.997

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 21

1.2

1.4

1.6

1.8

2

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 0.011748

Actual Output

R2 = 0.999

1 Figure 14 ANN model performance of Bo for 160 data 2

1 1.2 1.4 1.6 1.8 21

1.5

2

2.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.011498 and MDL = -481.2753

Actual output

R2 = 0.998

1 1.2 1.4 1.6 1.8 21

1.5

2

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 0.0078693 and MDL = -481.2753

Actual Output

R2 = 0.999

3 Figure 15 FunNet model performance of Bo for 160 data 4

5.0 conclusion and Recommendation 5

In this study, three distinct published data sets were used in investigating the capability of the FunNets modeling 6

scheme as a new framework for predicting the PVT properties of oil crude systems. Based on the obtained results 7

and comparative studies, the conclusions and recommendations can be drawn as follows: 8

A new computational intelligence modeling scheme based on the FunNets scheme to predict both bubble point 9

pressure and oil formation volume factor using the four input variables: solution gas-oil ratio, reservoir temperature, 10

28

oil gravity, and gas relative density. As it is shown in the petroleum engineering communities, these two predicted 1

properties were considered the most important PVT properties of oil crude systems. 2

3 - The developed FunNets modeling scheme outperforms both the standard feedforward neural networks and the 4

most common published empirical correlations. Thus, the developed FunNets modeling scheme has better, 5

efficient, and reliable performance compared to the most published correlations. 6

- The developed FunNets modeling scheme shown a high accuracy in predicting the Bob values with a stable 7

performance, and achieved the lowest absolute percent relative error, lowest minimum error, lowest maximum 8

error, lowest root mean square error, and the highest correlation coefficient among other correlations for the 9

used three distinct data sets. 10

11 The functional networks modeling scheme is flexible, reliable, and shows bright future in implementing it for the oil 12

and gas industry, especially permeability and porosity prediction, history matching, predicting the rock mechanics 13

properties, and seismic litho-faceis classification. 14

NOMENCLATURE 15

Bob = OFVF at the bubble- point pressure, RB/STB 16

Rs = on solution gas oil ratio, SCF/STB 17

T = reservoir temperature, degrees Fahrenheit 18

γo = oil relative density (water=1.0) 19

γg = gas relative density (air=1.0) 20

Er = average percent relative error 21

Ei = percent relative error 22

Ea = average absolute percent relative error 23

Emax = Maximum absolute percent relative error 24

Emin = Minimum absolute percent relative error 25

RMS = Root Mean Square error 26

29

Acknowledgement 1

The authors wish to thank King Fahd University of Petroleum and Minerals and Mansoura University for the 2

facilities utilized to perform the present work and for their support. 3

References 4

1. Standing M.B.: “A Pressure-Volume-Temperature Correlation for Mixtures of California Oils and Gases,” 5

Drill&Prod. Pract., API (1947), pp 275-87. 6

2. Katz, D. L.: “Prediction of Shrinkage of Crude Oils,” Drill&Prod. Pract., API (1942), pp 137-147. 7

3. Standing, M. B.: Volumetric and Phase Behavior of Oil Field Hydrocarbon System. Millet Print Inc., 8

Dallas, TX (1977) 124. 9

4. Vazuquez, M. and Beggs, H.D.: “Correlation for Fluid Physical Property Prediction,” JPT (June 1980) 10

968. 11

5. Glaso, O. “Generalized Pressure-Volume Temperature Correlations,” JPT (May 1980), 785. 12

6. Al-Marhoun, M., “PVT Correlations for Middle East Crude Oils,” Journal of Petroleum Technology, pp 13

650-666, 1988. 14

7. Abdul-Majeed, G.H.A. and Salman, N.H.: “An Empirical Correlation for FVF Prediction,” JCPT, (July-15

August 1988) 118. 16

8. Labedi, R.: “Use of Production Data to Estimate Volume Factor Density and Compressibility of Reservoir 17

Fluids.” J. Pet. Sci. & Eng., 4(1990) 357. 18

9. Dokla, M. and Osman, M.: “Correlation of PVT Properties for UAE Crudes,” SPEFE (March 1992) 41. 19

10. Al-Yousef H. Y., Al-Marhoun, M. A.: “Discussion of Correlation of PVT Properties for UAE Crudes,” 20

SPEFE (March 1993) 80. 21

11. Dokla, M. and Osman, M.: “Authors’ Reply to Discussion of Correlation of PVT Properties for UAE 22

Crudes,” SPEFE (March 1993) 82. 23

12. Al-Marhoun, M. A.: “New Correlation for formation Volume Factor of oil and gas Mixtures,” JCPT 24

(March 1992) 22. 25

30

13. Macary, S. M. & El-Batanoney, M. H.: “Derivation of PVT Correlations for the Gulf of Suez Crude Oils,” 1

Paper presented at the EGPC 11th Petroleum Exploration & Production Conference, Cairo, Egypt (1992). 2

14. Saleh, A. M., Maggoub, I. S. and Asaad, Y.: “Evaluation of Empirically Derived PVT Properties for 3

Egyptian Oils,” paper SPE 15721 , presented at the 1987 Middle East Oil Show & Conference, Bahrain, 4

March 7-10. 5

15. Omar, M.I. and Todd, A.C.: “Development of New Modified Black oil Correlation for Malaysian Crudes,” 6

paper SPE 25338 presented at the 1993 SPE Asia Pacific Asia Pacific Oil & Gas Conference and 7

Exhibition, Singapore, Feb. 8-10. 8

16. Petrosky, J. and Farshad, F.: “Pressure Volume Temperature Correlation for the Gulf of Mexico.” paper 9

SPE 26644 presented at the 1993 SPE Annual Technical Conference and Exhibition, Houston, TX, Oct 3-10

6. 11

17. Kartoatmodjo, T. and Schmidt, Z.: “Large data bank improves crude physical property correlations,” Oil 12

and Gas Journal (July 4, 1994) 51. 13

18. Almehaideb, R.A.: “Improved PVT Correlations For UAE Crude Oils,” paper SPE 37691 presented at the 14

1997 SPE Middle East Oil Show and Conference, Bahrain, March 15–18. 15

19. Sutton, R. P. and Farshad, F.: ” Evaluation of Empirically Derived PVT Properties for Gulf of Mexico 16

Crude Oils” SPRE (Feb. 1990) 79. 17

20. Sutton, Roberts P. and Farshad, F.: ”Supplement to SPE 1372, Evaluation of Empirically Derived PVT 18

Properties for Gulf of Mexico Crude Oils” SPE 20277, Available from SPE book Order Dep., Richardson, 19

TX. 20

21. McCain, W. D.: “ Reservoir fluid property correlations-State of the Art,” SPERE, (May 1991) 266. 21

22. Ghetto, G. De, Paone, F. and Villa, M.: “Reliability Analysis on PVT correlation,” paper SPE 28904 22

presented at the 1994 SPE European Petroleum Conference, London, UK, October 25-27. 23

23. Elsharkawy, A. M. Elgibaly, A. and Alikhan, A. A.: “Assessment of the PVT Correlations for Predicting 24

the Properties of the Kuwaiti Crude Oils,” paper presented at the 6th Abu Dhabi International Petroleum 25

Exhibition & Conference, Oct. 16-19, 1994. 26

31

24. Mahmood, M. M. and Al-Marhoun, M. A.: “Evaluation of empirically derived PVT properties for 1

Pakistani crude oils,” J. Pet. Sci. & Engg. 16 (1996) 275. 2

25. Hanafy, H.H., Macary, S.A., Elnady, Y. M., Bayomi, A.A. and El-Batanoney, M.H.: “ Empirical PVT 3

Correlation Applied to Egyptian Crude Oils Exemplify Significance of Using Regional Correlations,” 4

paper SPE 37295 presented at the SPE Oilfield Chemistry International Symposium, Houston, Feb. 18–21, 5

1997. 6

26. Al-Fattah, S. M. and Al-Marhoun, M. A.: “Evaluation of empirical correlation for bubble point oil 7

formation volume factor,” J. Pet. Sci. & Engg. 11(1994) 341. 8

27. Al-Shammasi, A. A.,: “Bubble Point Pressure and Oil Formation Volume Factor Correlations,” paper SPE 9

53185 presented at the 1997 SPE Middle East Oil Show and Conference, Bahrain, March 15–18. 10

28. Kumoluyi, A.O. and Daltaban, T.S.: “High-Order Neural Networks in Petroleum Engineering,” paper SPE 11

27905 presented at the 1994 SPE Western Regional Meeting, Longbeach, California, USA, March 23-25. 12

29. Ali, J. K.: “Neural Networks: A New Tool for the Petroleum Industry,” paper SPE 27561 presented at the 13

1994 European Petroleum Computer Conference, Aberdeen, U.K., March 15-17. 14

30. Mohaghegh, S. and Ameri, S.,:" A Artificial Neural Network As A Valuable Tool For Petroleum 15

Engineers," SPE 29220, unsolicited paper for Society of Petroleum Engineers, 1994. 16

31. Mohaghegh, S.:" Neural Networks: What it Can do for Petroleum Engineers," JPT, (Jan. 1995) 42. 17

32. Mohaghegh, S.:" Virtual Intelligence Applications in Petroleum Engineering: Part 1 - Artificial Neural 18

Networks,” JPT (September 2000). 19

33. Gharbi, R.B. and Elsharkawy, A.M.: “Neural-Network Model for Estimating the PVT Properties of Middle 20

East Crude Oils,” paper SPE 37695 presented at the 1997 SPE Middle East Oil Show and Conference, 21

Bahrain, March 15–18. 22

34. Gharbi, R.B. and Elsharkawy, A.M.: “Universal Neural-Network Model for Estimating the PVT Properties 23

of Crude Oils,” paper SPE 38099 presented at the 1997 SPE Asia Pacific Oil & Gas Conference, Kuala 24

Lumpur, Malaysia, April 14-16. 25

35. Elsharkawy, A.M.: “Modeling the Properties of Crude Oil and Gas Systems Using RBF Network,” paper 26

SPE 49961 presented at the 1998 SPE Asia Pacific Oil & Gas Conference, Perth, Australia, October 12-14. 27

32

36. Varotsis N., Gaganis V., Nighswander J., and Guieze P.,: “A Novel Non-Iterative Method for the 1

Prediction of the PVT Behavior of Reservoir Fluids,” paper SPE 56745 presented at the 1999 SPE Annual 2

Technical Conference and Exhibition, Houston, Texas, October 3–6. 3

37. Farshad, F.F, Leblance, J.L, Garber, J.D. and Osorio, J.G.: “Empirical Correlation for Colombian Crude 4

Oils,” paper SEP 24538, Unsolicited (1992), Available from SPE book Order Dep., Richardson, TX. 5

38. Osman, E. A., Abdel-Wahhab, O. A., and Al-Marhoun, M. A., “Prediction of Oil Properties Using Neural 6

Networks,” Paper SPE 68233, 2001. 7

39. Hinton G. E., “Connectionist learning procedures”. Artificial Intel., 40:185–234, 1989. 8

40. White H., “Learning in artificial neural networks: A statistical perspective”. Neural Comput., 1:425–464, 9

1989. 10

41. Hussam M. Goda, Eissa M. El-M Shokir, Khaled A. Fattah, and Mohamed H. Sayyouh "Prediction of the 11

PVT Data using Neural Network Computing Theory". The 27th Annual SPE International Technical 12

Conference and Exhibition in Abuja, Nigeria, August 4-6, 2003: SPE85650. 13

42. Ali, J. K., “Neural Networks: A New Tool for the Petroleum Industry,” Paper SPE 27561, 1994. 14

43. Duda R. O., Hart P. E., and Stock D. G., “Pattern Classification, Second Edition, John Wiley and Sons, 15

New York, 2001. 16

44. Danesh, A., “PVT and Phase Behavior of Petroleum Reservoir Fluid,” Elsevier Publishing Company, 17

1998. 18

45. Smith, C., Tracy, G., and Farrar, R., “Applied Reservoir Engineering, Volume 1,” Oil & Gas Consultants 19

International, Inc., 1992. 20

46. Al-Shammasi, A. A.: “Bubble Point Pressure and Oil Formation Volume Factor Correlations,” SPE Middle 21

East Oil Show, Bahrain, 20–23 February 1999. SPE53185. 22

47. Al-Shammasi, A. A.: “A Review of Bubble point Pressure and Oil Formation Volume Factor 23

Correlations,” SPE Reservoir Evaluation & Engineering (April 2001) 146-160. This paper (SPE 71302) 24

was revised for publication from paper SPE 53185. 25

33

48. Al-Marhoun, M.A., and Osman, E. A., “Using Artificial Neural Networks to Develop New PVT 1

Correlations for Saudi Crude Oils,” Paper SPE 78592, presented at the 10th Abu Dhabi International 2

Petroleum Exhibition and Conference (ADIPEC), Abu Dhabi, UAE, October 8-11, 2002. 3

49. Osman E. A. and Abdel-Aal, "Abductive Netwoks: A New Modeling Tool for the Oil and Gas Industry". 4

SPE 77882 Asia Pacific Oil and Gas Conference and Exhibition Melbourne, Australia, 8–10 October 2002. 5

50. Osman E., Al-Marhoun M., “Artificial Neural Networks Models for Predicting PVT Properties of Oil Field 6

Brines", paper SPE 93765, 14th SPE Middle East Oil & Gas Show and Conference in Bahrain, March 7

2005. 8

51. McCain W. D. Jr., R. Soto B., Valko, P.P., and Blasingame, T. A. “Correlation Of Bubble point Pressures 9

For Reservoir Oils - A Comparative Study,” paper SPE 51086 presented at the 1998 SPE Eastern Regional 10

Conference and Exhibition held in Pittsburgh, PA, 9–11 November. 11

52. Castillo E., Cobo A., Gutiérrez J. M. and Pruneda E., (1998), Introduction to Functional Networks with 12

Applications, A Neural Based Paradigm, Kluwer Academic Publishers: New York. 13

53. E. El-Sebakhy, “Constrained Estimation of Functional Networks for Statistical Pattern Recognition 14

Problems”, Journal of AI and Patt. Rec. In Press, 2007. 15

54. E. Castello, Functional networks. Neural Proc. Lett. 7 (1998), pp. 151¯159. 16

55. S. Rajasekaran, V. Shalu Mohan, and O. Khamis, (2004), "The optimisation of space structures using 17

evolution strategies with functional networks", Engineering with Computers (2004) 20: 75–87. Springer-18

Verlag London. 19

56. M. Bruen and J.C.I. Dooge, An efficient and robust method for estimating unit hydrograph ordinates, J 20

Hydrol 70 (1984), pp. 1–24. 21

57. E. Castillo, J. M. GutiCrrez, (1998), “Nonlinear time series modeling and prediction using functional 22

networks. Extracting information masked by chaos”. Physics Letters A (244) 71-84. 23

58. El-Sebakhy A. E., Functional Networks Training Algorithm for Statistical Pattern Recognition; IEEE 24

Symposium on Computers and Communications, ISCC 2004. Ninth International Symposium on Volume, 25

vol. 1, pp. 92 –97, 28 June–1 July, 2004. 26

34

59. El-Sebakhy A. E., Hadi S. A., and Faisal A. K., Iterative Least Squares Functional Networks Classifier; 1

IEEE Transactions Neural Networks, vol. 18, no. 3, pp. 844 –850, March 2007. 2

60. El-Sebakhy E., Faisal K., El-Bassuny T., Azzedin F., and Al-Suhaim A., (2006), “Evaluation of Breast 3

Cancer Tumor Classification with Unconstrained Functional Networks Classifier”; the 4th ACS/IEEE 4

International Conf. on Computer Systems and Applications. 281-287. 5

61. El-Sebakhy A. E., “Relative Solvent Accessibility of Residues in Proteins Using Unconstrained Functional 6

Networks”, International Journal of Bioinformatics (2007). (In press). 7

62. El-Sebakhy A. E. and Albokhitan Y. S., “Forecasting PVT Properties of Crude Oil Systems based on 8

Support Vector Machines Modeling Scheme”. International journal of Mathematics and Computers In 9

Simulation, (2007). (In press). 10

63. Li C, Liao X, Wu Z, Yu J. Complex functional networks. Math Comput Simul 2001;57:355–65. 11

64. Castillo E., Functional networks. Neural Process. Lett., 7(3):151–159, 1998. 12

65. Castillo E, Cobo A, Gutierrez JM, Pruneda E., Functional networks with applications-a neural-based 13

paradigm. Boston/ Dordrecht/London: Kluwer Academic Publishers; New York, 1998. 14

66. Castillo E, Cobo A, Gutierrez JM, and Hadi A. S., A general framework for functional networks. 15

Networks, 35:70–82, 1999. 16

67. Castillo E. and Gutierrez JM,, A comparison of functional networks and neural networks. In Proceedings 17

of the IASTED International Conference on Artificial Intelligence and Soft Computing, pages 439–442, 18

IASTED ACTA press 1998. 19

68. Castillo E, Cobo A, Gutierrez JM, Pruneda E.,, Working with differential, functional and di erence 20

equations using functional networks. Applied Mathematical Modelling 23 (1999) 89 – 107. 21

69. Castillo E, Cobo A, Gutierrez JM, Nonlinear time series modeling and prediction using functional 22

networks. Extracting information masked by chaos. Physics Letters A 244 (1998) 71-84. 23

70. Castillo, E., Hadi, A., and Lacruz, B. (2001), "Optimal Transformations in Multiple Linear Regression 24

Using Functional Networks," Proceedings of the International Work-Conference on Artificial and natural 25

Neural Networks. IWANN 2001, in Lecture Notes in Computer Science 2084, Part I, 316–324. 26

35

71. Castillo, E., Gutiérrez, J. M., Hadi, A. S., and Lacruz, B. (2001), "Some Applications of Functional 1

Networks in Statistics and Engineering," Technometrics, 43, 10–24. 2

72. Pruneda, B. Lacruz, and C. Solares, A First Approach to Solve Classification Problems Based on 3

Functional Networks. ICANN 2005, LNCS 3697, pp. 313–318. Springer-Verlag Berlin Heidelberg 2005. 4

73. Solares C., Vieira E. W., and Minguez R., Functional Networks and the Lagrange Polynomial 5

Interpolation. E. Corchado et al. (Eds.): IDEAL 2006, LNCS 4224, pp. 394–401, 2006. Springer-Verlag 6

Berlin Heidelberg 2006. 7

74. Rajasekaran S., Mohan V. S., and Khamis O., The optimization of space structures using evolution 8

strategies with functional networks. Engineering with Computers (2004) 20: 75–87. 9

75. [28] Bruen M. and Yang J., Functional networks in real-time flood forecasting- a novel application. 10

Advances in Water Resources 28 (2005) 899–909 11

12

13

1 14

15

36

Appendices 1

Table 6 Performance of Pb for 283 data set 2   Er  Ea  Emin  Emax  RMSE  SD  R 

Standing  10.198  19.81  0.25511  93.672  425.19  22.252  0.89739 Glaso  ‐17.262  21.418  1.4363  70.407  377.72  18.349  0.98371 

Marhoun  8.8175  10.56  0.053296  57.593  193.63  9.9281  0.99054 NN training  ‐1.015  4.4734  0.015138  129.62  52.162  12.183  0.99804 NN testing  ‐3.3969  5.9478  0.024248  98.647  58.741  13.662  0.99843 

FunNet training  ‐0.12266  5.2877  0.013004  49.056  62.256  8.3431  0.99717 FunNet testing  2.8209  4.4175  0.045593  33.382  55.77  6.3978  0.99852 

 3

5 10 15 200.7

0.75

0.8

0.85

0.9

0.95

1

1.05

Absolute Percent Relative Error (EA)

Cor

rela

tion

Coe

ffici

ent (

Rho

)

Standing (1947)Glaso (1980)Al-Marhoun (1992)ANN TrANN TsFunNet TrFunNet Ts

 4 Figure 16  Pb Performance for 283 data set 5

 6  7

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 425.1929

Actual output

R2 = 0.897

 8 Figure 17  Standing Pb Performance for 283 data set 9

37

 1

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

5000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 377.7177

Actual output

R2 = 0.984

 2 Figure 18 Glaso Pb Performance for 283 data set 3

 4

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 193.6304

Actual output

R2 = 0.991

 5 Figure 19 Marhoun Pb Performance for 283 data set 6

 7

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 52.1623

Actual output

R2 = 0.998

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 58.7409

Actual Output

R2 = 0.998

 8 Figure 20 ANN  Pb Performance for 283 data set 9

38

 1  2

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 62.2558 and MDL = 853.879

Actual output

R2 = 0.997

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 55.77 and MDL = 853.879

Actual Output

R2 = 0.999

 3 Figure 21 FunNet Pb Performance for 283 data set 4

 5

2 Table 7 Performance of Bo for 283 data set 6

  Er  Ea  Emin  Emax  RMSE  SD  R Standing  ‐1.9738  2.1334  0.013382  13.337  0.044046  2.2257  0.99465 Glaso  0.2157  1.7939  0.002994  11.539  0.032632  2.3297  0.99197 Marhoun  ‐0.74564  1.0037  0.001112  5.0515  0.019915  1.1431  0.99722 NN training  0.35333  0.64064  0.002366  3.1729  0.01134  0.77523  0.99765 NN testing  0.26255  0.65798  0.002105  2.8004  0.010851  0.80728  0.99843 FunNet training  ‐0.00472  0.53251  0.006154  2.8419  0.008861  0.68708  0.99839 

FunNet testing 0.00654

4  0.34755  0.002623  2.6395  0.00624  0.50494  0.99926 

 7

Table 8 Performance of Pb for 782 data set 8

  Er  Ea  Emin  Emax  RMSE  SD  R Standing  67.602  67.73  0.16209  102.08  1609.3  13.696  0.86661 Glaso  ‐1.6162  18.523  0.10563  138.96  473.3  25.27  0.94471 

Marhoun  8.0076  20.011  0.025427  109.12  593.81  25.015  0.90583 NN training  ‐6.2973  15.098  0.023293  157.58  326.97  21.73  0.96786 NN testing  ‐10.4  16.345  0.13776  117.43  282.54  20.821  0.97887 

FunNet training  ‐5.2249  15.751  0.084898  173.29  344.69  24.745  0.9618 FunNet testing  ‐2.1471  13.361  0.068983  176.07  303.75  20.792  0.97558 

 9

39

 1 Figure 22 Performance of Bo for 283 data set 2

 3

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.91

1.5

2

2.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.044046

Actual output

R2 = 0.995

 4 Figure 23  Standing Bo Performance for 283 data set 5

 6

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.91

1.2

1.4

1.6

1.8

2

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.032632

Actual output

R2 = 0.992

 7 Figure 24 Glaso Bo Performance for 283 data set 8

 9

40

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.91

1.2

1.4

1.6

1.8

2

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.019915

Actual output

R2 = 0.997

 1 Figure 25 Marhoun Bo Performance for 283 data set 2

 3

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.91

1.5

2

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.01134

Actual output

R2 = 0.998

1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.91

1.5

2

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 0.010851

Actual Output

R2 = 0.998

 4 Figure 26 ANN Bo Performance for 283 data set 5

 6  7

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 62.2558 and MDL = 853.879

Actual output

R2 = 0.997

0 500 1000 1500 2000 2500 3000 35000

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 55.77 and MDL = 853.879

Actual Output

R2 = 0.999

 8 Figure 27 FunNet Bo Performance for 283 data set 9

41

 1

15 20 25 30 35 40 45 50 55 60 65

0.7

0.75

0.8

0.85

0.9

0.95

1

1.05

Absolute Percent Relative Error (EA)

Cor

rela

tion

Coe

ffici

ent (

Rho

)

Standing (1947)Glaso (1980)Al-Marhoun (1992)ANN TrANN TsFunNet TrFunNet Ts

 2 Figure 28 Performance of Pb for 782 data set 3

 4

0 1000 2000 3000 4000 5000 6000 7000 8000-1000

0

1000

2000

3000

4000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 1609.2527

Actual output

R2 = 0.867

 5 Figure 29 Standing Pb Performance for 782 data set 6

 7

0 1000 2000 3000 4000 5000 6000 7000 80000

2000

4000

6000

8000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 473.299

Actual output

R2 = 0.945

 8 Figure 30 Glaso Pb Performance for 782 data set 9

42

0 1000 2000 3000 4000 5000 6000 7000 80000

2000

4000

6000

8000

10000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 593.8129

Actual output

R2 = 0.906

 1 Figure 31 Marhoun Performance for 782 data set 2

 3

0 1000 2000 3000 4000 5000 6000 70000

2000

4000

6000

8000

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 326.9711

Actual output

R2 = 0.968

0 1000 2000 3000 4000 5000 6000 7000 80000

2000

4000

6000

8000

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 282.541

Actual Output

R2 = 0.979

 4 Figure 32 ANN Pb Performance for 782 data set 5

 6

 7 Figure 33 FunNet Pb Performance for 782 data set 8

43

Table 9 Performance of Bo for 782 data set 1

  Er  Ea  Emin  Emax  RMSE  SD  R Standing  ‐1.6353  3.1083  0.010503  89.786  0.099927  5.7655  0.95621 Glaso  1.0911  3.2591  0.011654  65.773  0.083353  5.2274  0.96009 

Marhoun  ‐0.78385  2.5738  0.009682  60.354  0.082717  5.0028  0.95986 NN training  ‐0.18716  2.2458  0.009277  35.814  0.066396  4.0402  0.97106 

NN testing 0.01351

6  1.9103  0.002308  10.128  0.04241  2.5873  0.98763 FunNet training  ‐0.13669  2.0913  0.002445  38.728  0.056962  3.8214  0.97848 FunNet testing  ‐0.22852  1.7134  0.001953  8.1564  0.033451  2.3508  0.99272 

 2  3

1.8 2 2.2 2.4 2.6 2.8 3 3.2

0.8

0.85

0.9

0.95

1

1.05

Absolute Percent Relative Error (EA)

Cor

rela

tion

Coe

ffici

ent (

Rho

)

Standing (1947)Glaso (1980)Al-Marhoun (1992)ANN TrANN TsFunNet TrFunNet Ts

 4 Figure 34 Performance of Bo for 782 data set 5

 6

1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 31

2

3

4

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.099927

Actual output

R2 = 0.956

 7 Figure 35 Standing Bo Performance for 782 data set 8

 9

44

1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 31

1.5

2

2.5

3

3.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.083353

Actual output

R2 = 0.960

 1 Figure 36 Glaso Bo Performance for 782 data set 2

 3

1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 31

1.5

2

2.5

3

3.5

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.082717

Actual output

R2 = 0.960

 4 Figure 37 Marhoun Bo Performance for 782 data set 5

 6

1 1.5 2 2.5 31

1.5

2

2.5

3

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.066396

Actual output

R2 = 0.971

1 1.5 2 2.5 31

1.5

2

2.5

3

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 0.04241

Actual Output

R2 = 0.988

 7 Figure 38  ANN Bo Performance for 782 data set 8

 9

45

1 1.5 2 2.5 31

2

3

4

Pre

dict

ed o

utpu

t

Training Performance: RMSE = 0.056962 and MDL = -1544.9958

Actual output

R2 = 0.978

1 1.5 2 2.5 31

1.5

2

2.5

3

Pre

dict

ed o

utpu

t

Testing Performance: RMSE = 0.033451 and MDL = -1544.9958

Actual Output

R2 = 0.993

 1 Figure 39 FunNet Bo Performance for 782 data set 2