- Open Access
- Total Downloads : 234
- Authors : M. Ravi Kumar, Ermias Girma Aklilu
- Paper ID : IJERTV6IS020172
- Volume & Issue : Volume 06, Issue 02 (February 2017)
- DOI : http://dx.doi.org/10.17577/IJERTV6IS020172
- Published (First Online): 14-02-2017
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License: This work is licensed under a Creative Commons Attribution 4.0 International License
A Comparative Study on the Prediction of Compresibility Factor for Generalized Equation of State using Artificial Neural Network
M. Ravi Kumar1, Ermias Girma Aklilu2 Department of Chemical Engineering, College of Engineering and Technology, Samara University, Samara, Afar, Ethiopia.
Abstract – The functional relationship among reduced pressure, compressibility, acentric factor, and reduced temperature of a gas was known as generalized equation of state. Though several generalized equations of state were available in literature each equation has its inherent disadvantages. There is no single generalized equation of state, which was mathematically friendly and can satisfy the experimental compressibility data with engineering accuracy.
Hence a two parameter zTr , Pr of generalized EOS model
was developed using Artificial Neural Network in the present work. A feed forward network with back propagation of errors algorithm was used in training the network. In the supervised learning, sum squared error was used as lapnov function. The model was found to be in good agreement with the experimental results. In this work Comparison of various neurons were used to find the compressibility factor.
Keywords: Artificial Neural Network, Compressibility Factor, Sum Squared Error, Acentric Factor.
INTRODUCTION
The development of Artificial Neural Network began approximately 58 years ago (Mc Culloch and Pitts, 1943), inspired by a desire to understand the human brain and emulate its functioning. Within the last decade, it has experienced a huge resurgence due to the development of more sophisticated algorithms and emergence of more powerful computational tools. Extensive research has been devoted to investigate the potential of ANN as computational tools that acquire, represent and compute a mapping from one multivariate input space to another. Mathematically an ANN is often viewed as a Universal Approximator. The ability to identify a relationship from given patterns make it possible for ANNs to solve large scale complex problems such as pattern recognition, nonlinear modeling, classification, association and control. Although the idea of ANNs was proposed over fifty years ago, a tremendous growth in the interest of this computational mechanism has occurred since Rumelhart et al (1986) rediscovered a mathematically rigorous theoretical framework for neural networks, i.e., back- propagation algorithm. Consequently ANNs have found applications in diverse areas such as
-
Neuro-Psychology
-
Physics
-
Various Engineering Fields
-
Computer Science
-
Acoustics
-
Robotics
-
Image Processing
-
Financing and Others.
A neural network is characterized by its architecture that represents the pattern of connection between the nodes, its method of determining connection weights and the activation function. A typical ANN consists of a number of nodes that are organized according to a particular arrangement. One way of classifying a neural network is by the number of layers.
-
Single layer Hop field et al
-
Bilayer Grosberg adaptive resonance network
-
Multi-layer most back propagation networks.
Input layer: The input layer receives the input variables for the problem at hand. This consists of all the quantities that can influence the output. The input layer is thus transparent and is a means of providing information to the network.
Hidden layer: The layers processing information are in between the input and output layers, which has any number of nodes.
Output layer: This layer gives the output of the network to external receptor.
Each layer consists of certain number of nodes, which are connected by links. No two nodes in the same layer will be having any links. A synaptic weight is assigned to each link to represent the relative connection strength of two nodes at both ends in predicting the input- output relationship. In this present study comparison of neurons used to find the compressibility factor.
Compressibility factor (Z) Theory: Virial Equation: –
A useful auxiliary thermodynamic property is defined by the equation.
(1)
factor.
use.
This dimensionless ratio is called compressibility
(2)
An alternative expression for Z is also in common
(3)
Both of these equations are known as virial
and virial equation is truncated to three terms, the appropriate form is:
(7)
Pitzer Correlations for the Compressibility Factor:
The correlation for Z takes the form:
Z = Z0 + Z1 (8)
expansions, and the parameter B, C, D, etc., and B, C, D, etc., are called virial coefficients. Parameters B and B are second virial coefficients; C and C are third virial coefficients; etc, For a given gas the virial coefficients are functions of temperature only.
The two sets of coefficients is 2 and 3 are related as follows;
etc.
Ideal gas:-
Since, the terms B/V, C/V2, etc., of the virial expansion [Eq. (3)] arise on account of molecular interactions, the virial coefficient B, C, etc., would be zero if no such interactions existed. The virial expansion would then reduce to :
Z = 1 or PV = RT
For a real gas, molecular interactions do exist, and exert an influence on the observed behavior of the gas. As the pressure of a real gas is reduced at constant temperature, V increases and the contributions of the terms B/V, C/V2, etc., decrease. For a pressure approaching zero, Z approaches unity, not because of any change in the Virial coefficient, but because V becomes infinite. Thus in the limit as the pressure approaches zero, the equation of state assumes the same simple form as for the hypothetical case of B = C = … = 0; i.e.,
Z = 1 or PV = RT
Differentiation of Eq.(2) for a given temperature gives:
(4)
From which, Thus the equation of tangent line is:Z = 1 + BP
a result also given by truncating the Eq. 2 to two terms. A
more common form of this equation results from the substitution, B = B/RT;
(5)
Equation (3) may also be truncated to two terms for application at low pressure:
(6)
Where Z0 and Z1 are functions of both Tr and Pr. When = 0, as is the case for the simple fluids, the second term disappears, and Z0 becomes identical with Z. Thus a generalized correlation for Z as a function of Tr and Pr based on data for just argon, krypton, and xenon provides the relationship Z0 = F0(Tr, Pr). By itself, this represents a two-parameter corresponding states correlation for Z. Since the second term of Eq. (8) is a relatively small correlation to this correlation, its omission does not introduce large errors, and a correlation for Z0 may be used alone for quick but less accurate estimates of Z than are obtained from a three parameters correlation.
Equation (8) is simple linear relation between Z and for given values of Tr and Pr do indeed yield approximately straight line, and their slopes provide values for Z1 from which the generalized function Z1 = F1(Tr, Pr) can be constructed.
The simplest form of the Virial equation has validity only to low to moderate pressures where Z is linear in pressure. The generalized Virial-coefficient correlation is therefore useful only where Z0 and Z1 are at least approximately linear functions of reduced pressure.
Levenberg Marquardt Method: –
This ANN design model uses Levenberg Marquardt method as multi variable optimization technique to predict the sum of squared errors or the designed target value.
Neural Network Architecture: –
A neural network is characterized by its architecture that represents the pattern of connection between the nodes, its method of determining connection weights and the activation function. A typical ANN consists of a number of nodes that are organized according to a particular arrangement. One way of classifying a neural network is by the number of layers.
-
Single layer Hop field et al.
-
Bilayer Grosberg adaptive resonance network.
-
Multi-layer most back propagation networks.
Input layer: The input layer receives the input variables for the problem at hand. This consists of all the quantities that can influence the output. The input layer is thus transparent and is a means of providing information to the network.
Hidden layer: The layers processing information are in between the input and output layers, which has any number of nodes.
Output layer: This layer gives the output of the network to external receptor.
Each layer consists of certain number of nodes, which are connected by links. No two nodes in the same layer will be
having any links. A synaptic weight is assigned to each link to represent the relative connection strength of two nodes at both ends in predicting the input-output relationship.
Input Hidden Output
Layer Layer Layer
U1 Y1
Outputs
U2
Y2
U K
Desired output Bias1() Bias2(m) Error
Fundamentals of Neural Computations: –
A node/neuron/processing element is an abstraction for a computational procedure, consists of three steps.
-
Calculating the weighted sum of input (net input) signals.
-
Applying a threshold (the value which excites or inhibits the nodes firing) on the weighted sum of input signals.
-
Calculating the output according to particular output function called transfer function or activation function.
Weight factor is nothing but the strength of an interconnection between two neurons expressed numerically.
Transfer function is nothing but the mapping relation between the output of node and its threshold applied weighted sum of input signals.
Examples of threshold functions are
-
Radial basis transfer function: f(x) =
-
Logarithmic sigmoid transfer function: ( log sig ) :
f(x) =
Normalizing the input-output data: The normalization of input and output data is very critical for successful training of network. If input and output variables do not have same order of magnitude some variables may appear to have more significance than they actually do. The training algorithm has to compensate for this and this is not very effective in many training algorithms. And also transfer functions like sigmoid and hyperbolic tangent are insensitive as argument (threshold input) is greater than 2. Hence three commonly used normalizations are given here.
X1 norm = ; Xi norm = ;
weight factors that accelerates the rate that the weight factors are adjusted.
The equation for adjusting the weight factors is:
New weight factor = [old weight factor] + [learning rate] x [gradient descendent correction term] + [momentum coefficient] x previous weight change]
When the learning rate is very slow, the training requires more number of iterations and the effect is very pronounced for complex networks and the training may be trapped at local minima (of obj. fn.) on the contrary if learning rate is very high the minima is never reached. Hence initially reasonably high and safe learning rate should be taken which goes on decreasing with iterations. General suggestion for learning rate and momentum coefficient is 0.3 and 0.4 respectively.
Selecting proper transfer function: Sigmoid and hyperbolic tangent functions are recommended for prediction/modeling and radial basis for fault diagnosis and feature categorization applications.
Selecting training data (Representative data): A representative data set prepared for training should contain information of all peaks, troughs and middle points also. Total number of data points available is divided into 4:1 ratio of training to testing data sets. Testing data sets should also contain some data points that belongs peaks, toughs and middle points.
Back propagation of error algorithm: –
The back propagation or errors learning algorithm is nothing but changing the weight factors based between layer of interest and its previous layers based or error back propagated from the layer preceded by the layer of interest.
The weight factors in the network between any two layers are calculated as follows for minimizing the sum-squared error using steepest descent method.
Let wij be the weight factor between ith hidden
Zero mean Normalization method: X
i norm = ;
neurons and jth output neuron. Then the change in wij i.e.,
wij is given by wij = 0 yi ej. where 0 is learning rate
Where Ri max= Max (Xi max Xi avg , Xi avg – Xi min ); Xi avg =
;
Initializing the weight distribution: – One broad suggestion for initializing weights is take random weight factors in between 1/n to 1/n, where n is number of neurons in input layer.
Setting learning rate and momentum coefficient: – Learning rate and momentum coefficient are two very important parameters that control how effectively the back propagation algorithm trains the neural network. The learning rate is a positive constant and that controls the rate at which the new weight factors are adjusted based on the calculated gradient descent correction term. The momentum coefficient is extra weight added on to the
for output layer yi is output of ith neuron in the hidden layer previous to output layer. Ej is error of jth neuron in output layer given by the difference between its actual and network outputs.
Wij(new) = wij(old) + wij
If Tj is the bias for jth neuron in output layer, Tj = 0 ej If Th is the bias for jth neuron in hidden layer, Th = 0 eh When momentum correction is applied,
w[i][j]t = w[i][j]t + w[i][j](t-1)
where t indicates iteration number and is the momentum coefficient.
E2 = t2 y2 ; w12 = 0 x1 e2 ; ep = y2 (1-y2) [w21 e1 + w22 e21];
w21 = 0 x2 e2 ; w12 = h w12 e2
The back propagation learning algorithm: Step 0 : Initialize weights.
Step 1: While stopping condition is false. For each pair of set:
-
Calculate outputs of all neurons in the network
-
Calculate new weight factors and biases
-
Step 2: While stopping condition is false (go to step 1) else break
RESULTS AND DISCUSSIONS
The Program code is run for a target goal of 0.002 and simulated for different number of Neurons as follows: Target Error = 0.002 No of Neurons =1
A Three dimensional plot of Tr, Pr and Z values for the first run is drawn as shown below.
Fig. 1. 3-D Plot between (Tr, Pr, Z)
Sum-squared error goal (SSE) = 2.92067.
TRAINLM: Network error did not reach the error goal. Further training may be necessary, or try different initial weights and biases and/or more hidden neurons.
Fig. 2 ANN Prediction Vs Actual data points in testing for one number of neurons with an error of 0.002
Fig. 3 Experimental Vs Model Predicted for testing Data points for one number of neurons with an error of 0.002
Expt. values |
Model values |
Expt. values |
Model values |
Expt. values |
Model values |
Expt. values |
Model values |
0.0029 |
0.1159 |
0.8206 |
0.8991 |
0.1158 0.1139 |
0.9946 |
0.9476 |
|
0.0130 |
0.1138 |
0.6635 |
0.5256 |
0.1564 0.1138 |
0.9768 |
0.9525 |
|
0.0239 |
0.1187 |
0.1321 |
0.1703 |
0.1904 0.1138 |
0.9573 |
0.9531 |
|
0.0442 |
0.1188 |
0.1614 |
0.1192 |
0.2200 0.1138 |
0.9174 |
0.9489 |
|
0.9804 |
0.2214 |
0.8561 |
0.9389 |
0.0778 0.1189 |
0.9965 |
0.9585 |
|
0.0093 |
0.2903 |
0.7574 |
0.8111 |
0.1109 0.1162 |
0.9826 |
0.9586 |
|
0.0178 |
0.3757 |
0.6138 |
0.3573 |
0.1415 0.1149 |
0.9659 |
0.9577 |
|
0.0344 |
0.3786 |
0.1844 |
0.1557 |
0.1703 0.1143 |
0.9322 |
0.9558 |
|
0.9935 |
0.9313 |
0.0021 |
0.1612 |
0.0611 0.3844 |
0.0882 |
0.1149 |
|
0.9725 |
0.9429 |
0.0110 |
0.1294 |
0.0983 0.2393 |
0.1429 |
0.1139 |
|
0.9528 |
0.9487 |
0.0239 |
0.1187 |
0.1321 0.1703 |
0.2084 |
0.1138 |
|
0.9115 |
0.9445 |
0.0522 |
0.1149 |
0.1664 0.1327 |
0.2892 |
0.1138 |
|
0.9963 |
0.9582 |
0.9922 |
0.8897 |
0.8338 0.9176 |
0.0687 |
0.1677 |
|
0.9821 |
0.9579 |
0.9505 |
0.7136 |
0.7360 0.7481 |
0.1063 |
0.1190 |
|
0.9648 |
0.9573 |
0.0178 |
0.3757 |
0.6138 0.3573 |
0.1476 |
0.1143 |
|
0.9300 |
0.9550 |
0.0371 |
0.1664 |
0.2901 0.1717 |
0.1939 |
0.1138 |
|
0.0825 |
0.1161 |
0.9961 |
0.9575 |
0.0026 0.1183 |
0.8059 |
0.8720 |
|
0.1322 |
0.1140 |
0.9790 |
0.9553 |
0.0119 0.1210 |
0.1006 |
0.3903 |
|
0.1904 |
0.1138 |
0.9528 |
0.9487 |
0.0221 0.1244 |
0.1301 |
0.1391 |
|
0.2604 |
0.1138 |
0.8810 |
0.8975 |
0.0413 0.1246 |
0.1626 |
0.1163 |
|
0.0670 |
0.2363 |
0.9968 |
0.9591 |
0.9849 0.3543 |
0.8509 |
0.9349 |
|
0.1027 |
0.1251 |
0.9832 |
0.9585 |
0.9377 0.4825 |
0.7471 |
0.7819 |
|
0.1415 |
0.1149 |
0.9648 |
0.9573 |
0.8958 0.6043 |
0.5887 |
0.3224 |
|
0.1842 |
0.1139 |
0.9523 |
0.9532 |
0.0336 0.6078 |
0.1779 |
0.1494 |
Table 1 Experimental Vs Model Values for one number of Neurons with an Error of 0.002
Fig. 4 Experimental Data points Vs Model Predictions (All data sets) for one number of neurons with an error of 0.002
Target Error = 0.002 No of Neurons =10 SSE = 0.0182752.
TRAINLM: Network error did not reach the error goal. Further training may be necessary, or try different initial weights and biases and/or more hidden neurons.
Fig 5 ANN Prediction Vs Actual data points in testing for 10 number of neurons with an error of 0.002
Fig 6 Experimental Vs. Model Predicted for testing Data points for 10 number of neurons with an error of 0.002
Fig 7Experimental Data points Vs Model Predictions (All data sets) for 10 numbers of neurons with an error of 0.002
Expt. values |
Model values |
Expt. values |
Model values |
Expt. values |
Model values |
Expt. values |
Model values |
0.0029 |
0.0000 |
0.8206 |
0.8154 |
0.1158 0.1151 |
0.9946 |
1.0000 |
|
0.0130 |
0.0000 |
0.6635 |
0.6640 |
0.1564 0.1577 |
0.9768 |
1.0000 |
|
0.0239 |
0.0000 |
0.1321 |
0.1318 |
0.1904 0.1891 |
0.9573 |
1.0000 |
|
0.0442 |
0.0449 |
0.1614 |
0.1596 |
0.2200 0.2195 |
0.9174 |
0.9243 |
|
0.9804 |
0.9803 |
0.8561 |
0.8567 |
0.0778 0.0784 |
0.9965 |
1.0000 |
|
0.0093 |
0.0003 |
0.7574 |
0.7616 |
0.1109 0.1113 |
0.9826 |
1.0000 |
|
0.0178 |
0.0179 |
0.6138 |
0.6137 |
0.1415 0.1419 |
0.9659 |
1.0000 |
|
0.0344 |
0.0364 |
0.1844 |
0.1865 |
0.1703 0.1727 |
0.9322 |
0.9369 |
|
0.9935 |
1.0000 |
0.0021 |
0.0026 |
0.0611 0.0620 |
0.0882 |
0.0885 |
|
0.9725 |
1.0000 |
0.0110 |
0.0000 |
0.0983 0.0958 |
0.1429 |
0.1436 |
|
0.9528 |
1.0000 |
0.0239 |
0.0000 |
0.1321 0.1318 |
0.2084 |
0.2075 |
|
0.9115 |
0.9167 |
0.0522 |
0.0520 |
0.1664 0.1666 |
0.2892 |
0.2893 |
|
0.9963 |
1.0000 |
0.9922 |
1.0000 |
0.8338 0.8312 |
0.0687 |
0.0672 |
|
0.9821 |
1.0000 |
0.9505 |
1.0000 |
0.7360 0.7342 |
0.1063 |
0.1064 |
|
0.9648 |
1.0000 |
0.0178 |
0.0179 |
0.6138 0.6137 |
0.1476 |
0.1472 |
|
0.9300 |
0.9353 |
0.0371 |
0.0354 |
0.2901 0.2888 |
0.1939 |
0.1930 |
|
0.0825 |
0.0829 |
0.9961 |
1.0000 |
0.0026 0.0000 |
0.8059 |
0.8107 |
|
0.1322 |
0.1322 |
0.9790 |
1.0000 |
0.0119 0.0000 |
0.1006 |
0.1008 |
|
0.1904 |
0.1891 |
0.9528 |
1.0000 |
0.0221 0.0000 |
0.1301 |
0.1328 |
|
0.2604 |
0.2619 |
0.8810 |
0.8789 |
0.0413 0.0411 |
0.1626 |
0.1614 |
|
0.0670 |
0.0685 |
0.9968 |
1.0000 |
0.9849 1.0000 |
0.8509 |
0.8527 |
|
0.1027 |
0.1013 |
0.9832 |
1.0000 |
0.9377 0.9377 |
0.7471 |
0.7449 |
|
0.1415 |
0.1419 |
0.9648 |
1.0000 |
0.8958 0.8958 |
0.5887 |
0.5895 |
|
0.1842 |
0.1845 |
0.9523 |
0.9318 |
0.0336 0.0341 |
0.1779 |
0.1769 |
Table 2 Experimental vs. Model Values for 10 number of Neurons with an Error of 0.002
CONCLUSION:
In the present work, as the functional relationship is extremely complex an ANN can be suggested as best method to solve with regard to its accuracy. We suggest an ANN with 37 hidden neurons, 2 input neurons and one
output neuron to find compressibility factor from given Tr and Pr using Levenberg Marquardt second order optimization technique. The network proposed with above parameters is found to be very accurate and better than R-K or Vander Waal two parameter model.
Number of Neurons |
Sum-squared error (SSE) |
Is it a desired SSE |
1 |
2.92067 |
No |
10 |
0.0182752 |
No |
20 |
0.0178607 |
No |
30 |
0.0218734 |
No |
40 |
0.0019945 |
Yes |
Table3 SSE values for different number of neurons with target goal of 0.002
As shown in table 2 the predicted SSE is matches with desired target value for forty (40) number of neurons which
is the optimum number of neurons for the designed ANN for equation of state model.
Number of Neurons |
Sum-squared error (SSE) |
Is it a desired SSE |
1 |
2.92067 |
No |
15 |
0.0258809 |
No |
30 |
0.00528496 |
No |
45 |
0.000951105 |
Yes |
Table4 SSE values for different number of neurons with target goal of 0.001
As shown in table 3 the predicted SSE is maches with desired target value for forty five (45) numbers of neurons which is the optimum number of neurons for the designed ANN for equation of state model.
REFERENCES
-
A neural network weight determination model designed uniquely for small data set learning.
-
Expert Systems with Applications, Volume 36, Issue 6, August 2009, Pages 9853-9858, Der- Chiang Li, Chiao-Wen Liu.
-
-
Neural Networks in Bioprocessing and Chemical Engineering by D.R. Baughman and Y.A.Liu, Academic Press, (1995).
-
An ordinal optimization theory-based algorithm for a class of simulation optimization problems and application. Expert Systems with Applications, Volume 36, Issue 5, July 2009, Pages 9340-9349. Shih-Cheng Horng, Shieh-Shing Lin.
-
An artificial neural network approach to compressor performance prediction. Applied Energy, Volume 86, Issues 7-8, July-August 2009, Pages 1210-1221, K. Ghorbanian, M.
Gholamrezaei
-
Artificial neural networks to predict daylight illuminance in office buildings. Building and Environment, Volume 44, Issue 8, August 2009, Pages 1751-1757, Tuçe Kazanasmaz, Murat Günaydin, Selcen Binol.
-
Adaptive neuro-fuzzy based inferential sensor model for estimating the average air temperature in space heating systems, Building and Environment, Volume 44, Issue 8,
August 2009, Pages 1609-1616, S. Jassar, Z. Liao, L. Zhao
-
Flow forecast by SWAT model and ANN in Pracana basin, Portugal. Advances in Engineering Software, Volume 40, Issue 7, July 2009, Pages 467-473, Mehmet C. Demirel,
Anabela Venancio, Ercan Kahya
-
Heat transfer of a helical double-pipe vertical evaporator: Theoretical analysis and experimental validation. Applied Energy, Volume 86, Issues 7-8, July-August 2009, Pages 1144-1153. D. Colorado-Garrido, E. Santoyo-Castelazo, J.A. Hernández, O. GarcÃa-Valladares, J. Siqueiros, D. Juarez-
Romero
-
Modeling the characteristics of turbo compressors for fuel cell systems using hybrid method based on moving least squares. Applied Energy, Volume 86, Issues 7-8, July- August 2009, Pages 1283-1289. R. Tirnovan, S. Giurgea, A.
Miraoui, M. Cirrincione
-
Artificial intelligence approaches to achieve strategic control over project cash flows. Automation in Construction, Volume 18, Issue 4, July 2009, Pages 386-393. Min-Yuan Cheng, Hsing-Chih Tsai, Chih-Lung Liu.
-
A particle swarm optimization-aided fuzzy cloud classifier applied for plant numerical taxonomy based on attribute similarity. Expert Systems with Applications, Volume 36, Issue 5, July 2009, Pages 9388-9397. Hongfei Lu, Erxu Pi, Qiufa Peng, Lanlan Wang, Changjiang Zhang.
-
Neural Computing: A Technology Handbook for Professional II/PLUS and Neural Works Explorer.
-
Himmelblau, D.M., Use of Artificial Neural Networks to Monitor Faults and for Troubleshooting in the Process Industries
-
Hoskins,J.C. Artificial Neural Networks for knowledge Representation in Chemical Engineering.
-
Zhang, J. and A.J.Morris, On-line process fault Diagnosis Using Fuzzy Neural Networks
-
Introduction to Artificial Neural Systems, by Jacek M. Zurada. West Publishing company.