Behera ANN 2013 2
Behera ANN 2013 2
Behera ANN 2013 2
net/publication/272251584
CITATIONS READS
13 121
4 authors, including:
Nagaratnam Sivakugan
James Cook University
190 PUBLICATIONS 1,910 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Principles of Foundation Engineering 9th Edition with Braja M Das View project
Cemented Paste Backfill Modification Using Different Types of Binders and Insight into its Mechanical, Micro-Structure, and Flow Properties View project
All content following this page was uploaded by Nagaratnam Sivakugan on 24 November 2016.
trained weights of the neural network model. Finally, the are presented in Table 2. They are computed from the
predicted RF values were compared with those derived database. The schematic diagram of ANN architecture is
empirically by Patra et al. (2012b). shown in Fig. 2. The number of hidden layer neurons is
varied with mean square error (MSE). The minimum
Database and preprocessing MSE is found to be 0?001 for four neurons in the hidden
layer (Fig. 3). Therefore, the final ANN architecture used
The database available in Patra et al. (2012b) has been in this study will be 3–4–1, i.e. 3 (input)–4 (hidden layer
used in the present study. Load tests were carried out on neuron)–1 (output), as shown in Fig. 4. MSE is defined as
model strip footings subjected to eccentrically inclined
n
P 2
loads in the manner as shown in Fig. 1b that were RFi {RFp
increased to failure. The details of the tests and the
MSE~ i~1 (2)
procedure have been described in detail by Patra et al. n
(2012a, b). The data consists of parameters like load Coefficient of efficiency, R2 is expressed as
eccentricity (e), load inclination (a), embedment ratio (Df /
B), friction angle (w), and ultimate bearing capacity (qu). E1 {E2
R2 ~ (3)
Seventy-eight laboratory model tests were conducted. The E1
input parameters for the ANN model are e/B, a/w and Df /
where
B and the output is the RF. The RF is given by
quðDf =B,e=B,a=wÞ X
n 2
RF ~ (1) E1 ~ RFi {RF (3a)
quðDf =B,e=B~0,a=w~0Þ i~1
Table 1 Dataset used for training and testing of ANN model (Patra et al., 2012b)
Calculated
Data Expt no. e/B Df/B a/w Experimental qu Experimental RF (equation
type (1) (2) (3) (4) (5) (kN/m2) (6) RF (equation (1)) (7) (19)) (8)
Table 1 Continued
Calculated
Data Expt no. e/B Df/B a/w Experimental qu Experimental RF (equation
type (1) (2) (3) (4) (5) (kN/m2) (6) RF (equation (1)) (7) (19)) (8)
experimental reduction factor and predicted reduction sums the products across all the hidden neurons, which is
factor for training data set. Residual (er) can be defined as defined as Si. The relative inputs are corresponding to
the difference between the experimental and predicted RF absolute Si values, with the most important input
value and is given by corresponding to the highest Si value. The details of
connection weight approach are presented by Olden et al.
er ~RFi {RFp (4) (2004).
The residuals are plotted with the experiment number as The sensitivity analysis based on Pearson correlation
shown in Fig. 7. It is observed that the residuals are coefficient is presented in Table 4 which shows the cross
distributed evenly along the horizontal axis of the plot. correlation of inputs with the reduction factor. It is seen
Therefore, it can be said that the network is well trained that the parameters (e/B) and (a/w) are inter-related with a
and can be used for prediction with reasonable accuracy. cross-correlation value of 0?3. This is possibly due to the
reinforcing effect. From the table, it is observed that RF is
Sensitivity analysis highly correlated to a/w with a cross-correlation values of
0?87, followed by e/B (50?59) and Df /B (50?28). The
Sensitivity analysis is of utmost concern for selection of
sensitivity analysis for the model as per Garson’s
important input variables. Different approaches have been
algorithm is presented in Table 5. The a/w is found to be
suggested to select the important input variables. The
the most important input parameter with the relative
Pearson correlation coefficient is defined as one of the
importance value being 37?3% followed by 34?1% for Df /B
variable ranking criteria in selecting proper inputs for the
and 28?6% for e/B. The relative importance of the input
ANN (Guyon and Elisseeff, 2003; Wilby et al., 2003). Goh
variables as calculated following connection weight
(1994) and Shahin et al. (2002) have used Garson’s approach (Olden et al., 2004) is also presented in
algorithm (Garson, 1991) in which the input-hidden and Table 5. As per connection weight approach method, a/w
hidden-output weights of the trained ANN model are is found to be the most important input parameter (Si
partitioned and the absolute values of the weights are used value: 22?13) followed by e/B (Si value: 21?69) and Df /B
in selecting the important input variables, the details of (Si value: 0?41). The Si values being negative imply that
which with example have been presented by Goh (1994). It both a/w and e/B are indirectly and Df /B is directly related
does not provide information on the effect of input to RF values.
variables in terms of direct or inverse relation to the
output. Olden et al. (2004) proposed a connection weights Neural Interpretation Diagram (NID)
approach based on the NID, in which the actual values of Ozesmi and Ozesmi (1999) proposed neural interpretation
input-hidden and hidden-output weights are considered. It diagram for visual interpretation of the connection weight
2 ANN architecture
4 Neural Interpretation Diagram (NID) showing lines representing connection weights and effects of inputs on reduction fac-
tor (RF)
5 Correlation between predicted reduction factor with 6 Correlation between predicted reduction factor with
experimental reduction factor for training data experimental reduction factor for testing data
Step 1
The input parameters were normalized in the range [21, 1]
ANN model equation for RF based on trained by the following expressions
neural network
A model equation is developed using the weights obtained X1 {Xmin
Xn ~2 (6)
from trained neural network model (Goh et al., 2005). The Xmax {Xmin
mathematical equation relating input parameters (e/B, Df / where Xn is the normalized value of input parameter X1,
B, a/w) to output (RF) can be given by and Xmax and Xmin are maximum and minimum values of
( " !#)
X h X
m input parameter X1 in the data set.
RFn ~fn b0 z wk fn bhk z wik Xi (5)
k~1 i~1 Step 2
where RFn is the normalized value of RF in the range [21, Calculate the normalized value of reduction factor (RFn)
1], fn is the transfer function, h is the no. of neurons in the using the following expressions
hidden layer, Xi is the normalized value of inputs in the e
: : Df : a
range [21, 1], m is the no. of input variables, wik is the A1 ~{0 4188 z0 3928 {0 8511 {0:005
B n B n w n
connection weight between ith layer of input and kth (7)
neuron of hidden layer, wk is the connection weight
e
between kth neuron of hidden layer and single output Df a
neuron, bhk is the bias at the kth neuron of hidden layer, A2 ~0:17 z5:2462 z3:1071 z3:874 (8)
B n B n w n
and bo is the bias at the output layer.
The model equation for the RF of a shallow strip e
Df a
footing subjected to eccentrically inclined load as shown in A3 ~4:0168 z8:1273 {4:8781 z7:3396
B n B n w n
Fig. 1b was formulated using the values of the weights and (9)
biases shown in Table 3 as per the following steps.
Table 3 Values of connection weights and biases
Weight
wik wk Bias
e
Df a
A4 ~6:2788 {0:3513 z3:247 z4:8671 8 Comparison of ANN results with experimental RF and
B n B n w n Patra et al. (2012b) for training data
(10)
A1
e {e{A1
B1 ~0:7337 (11)
eA1 ze{A1
A2 RFn ~C1 (16)
e {e{A2
B2 ~{0:1013 A (12)
e 2 ze{A2
A3 Step – 3
: e {e{A3 Denormalize the RFn value obtained from equation (16)
B3 ~0 0693 A (13)
e 3 ze{A3 to actual RF as
A3 RF ~0:5(RFn z1)(RFmax {RFmin )zRFmin (17)
: e {e{A3
B4 ~{0 2621 A (14)
e 3 ze{A3
RF ~0:5(RFn z1)(1{0:291)z0:291 (18)
C1 ~0:0972zB1 zB2 zB3 zB4 (15)
Comparison
Table 4 Cross-correlation of input and output for reduction Patra et al. (2012b) proposed an empirical equation for
factor reduction factor (RF) and can be expressed as
Parameters e/B Df /B a/w RF quðDf =B,e=B,a=wÞ h e i
a ð1:5{0:7Df =BÞ
RF ~ ~ 1{2 1{
e/B 1 0 0.3 20.59 quðDf =B,e=B~0,a=w~0Þ B w
Df /B 1 0 0.28 (19)
a/w 1 20.87 Figures 8 and 9 show the comparison of RF obtained from
RF 1
equation 17 with equations (19) and (1). It can be seen
Table 5 Relative importance of different inputs as per Garson’s algorithm and connection weight approach
References
Das, S. K. and Basudhar, P. K. 2006. Undrained lateral load capacity of
piles in clay using artificial neural network, Comput. Geotech., 33,
(8), 454–459.
Garson, G. D. 1991. Interpreting neural-network connection weights,
Artif. Intell. Exp., 6, (7), 47–51.
Goh, A. T. C. 1994. Seismic liquefaction potential assessed by neural
network, J. Geotech. Eng., ASCE, 120, (90), 1467–1480.
Guyon, I. and Elisseeff, A. 2003. An introduction to variable and feature
9 Comparison of ANN results with experimental RF and
selection, J. Mach. Learn. Res., 3, 1157–1182.
Patra et al. (2012b) for testing data Goh, A. T. C., Kulhawy, F. H. and Chua, C. G. 2005. Bayesian neural
network analysis of undrained side resistance of drilled shafts,
that ANN results are much closer to line of equality than J. Geotech. Geoenv. Eng., ASCE, 131, (1), 84–93.
the empirical ones from Patra et al. (2012b). Ozesmi, S. L. and Ozesmi, U. 1999. An artificial neural network approach
to spatial modeling with inter specific interactions, Ecol. Model.,
116, 15–31.
Conclusion Olden, J. D., Joy, M. K. and Death, R. G. 2004. An accurate comparison
Based on the neural network model developed herein, the of methods for quantifying variable importance in artificial neural
networks using simulated data, Ecol. Model., 178, (3), 389–397.
following conclusions may be drawn. Patra, C. R., Behera, R. N., Sivakugan, N. and Das, B. M. 2012a.
1. As per residual analysis, the errors are distributed Ultimate bearing capacity of shallow strip foundation under
evenly along the horizontal axis (Fig. 7). It can be eccentrically inclined load: part I, Int. J. Geotech. Eng., 6, (3),
concluded that the network was well trained and can 343–352.
predict the result with reasonable accuracy. Patra, C. R., Behera, R. N., Sivakugan, N. and Das, B. M. 2012b.
Ultimate bearing capacity of shallow strip foundation under
2. Using correlation coefficient it was observed that a/w
eccentrically inclined load: part II, Int. J. Geotech. Eng., 6, (4),
is the most important input parameter followed by e/B and 507–514.
Df /B. Perloff, W. H. and Barron, W. 1976. Soil mechanics: principles and
3. Similarly, using Garson’s algorithm, a/w is found to applications; New York, Ronald Press.
be the most important input parameter followed by Df /B Shahin, M. A., Maier, H. R. and Jaksa, M. B. 2002. Predicting settlement
of shallow foundations using neural network, J. Geotech. Geoenv.
and e/B.
Eng., ASCE, 128, (9), 785–793.
4. Based on connection weight approach method, and Wilby, R. L., Abrahart, R. J. and Dawson, C. W. 2003. Detection of
Pearson correlation coefficient gave similar results. Hence, conceptual model rainfall-runoff processes inside an artificial neural
it may be concluded that sensitivity analysis using network, Hydrol. Sci., 48, (2), 163–181.