Skip to content
BY-NC-ND 3.0 license Open Access Published by De Gruyter December 18, 2013

Soft computing techniques for compressive strength prediction of concrete cylinders strengthened by CFRP composites

  • Mostafa Jalal EMAIL logo

Abstract

This study presents the application of soft computing techniques, namely, as multiple regressions (MRs), neural networks (NNs), genetic programming (GP), and adaptive neuro-fuzzy inference system (ANFIS) for modeling of compressive strength of carbon fiber-reinforced polymer (CFRP) confined concrete cylinders. The proposed soft computing models are based on experimental results collected from literature. They represent the ultimate strength of concrete cylinders after confinement with CFRP composites, which is in terms of diameter and height of the cylindrical specimen, ultimate circumferential strain in the CFRP jacket, elastic modulus of CFRP, unconfined concrete strength, and total thickness of CFRP layer used. The accuracy of the proposed soft computing models is very satisfactory compared to experimental results. Moreover, the results of proposed soft computing models are compared with five models existing in the literature proposed by various researchers so far and are found to be, by far, more accurate.

1 Introduction

One of the most important composite materials accepted by civil engineers is fiber-reinforced-polymer (FRP) composites, which have been utilized in different construction applications such as repair and rehabilitation of existing structures as well as in new construction applications. One of the successful and most popular structural applications of FRP composites is the external strengthening, repair, and ductility enhancement of reinforced concrete (RC) columns in both seismic and corrosive environments [1, 2]. The main types of FRP composites used in external strengthening and repair of RC columns are glass fiber-reinforced polymers (GFRP), carbon fiber-reinforced polymers (CFRP), and aramid fiber-reinforced polymers (AFRP). Types of FRP confinement can be spiral, wrapped, and tube. FRP composites offer several advantages due to the extremely high strength-to-weight ratio, good corrosion behavior, and electromagnetic neutrality. Thus, the effect of FRP confinement on the strength and deformation capacity of concrete columns has been extensively studied, and several empirical and theoretical models have been proposed [3–7].

Meanwhile in recent years, soft computing and artificial intelligence techniques like artificial neural networks (ANN), adaptive neuro-fuzzy inference system (ANFIS), and genetic programming (GP) have been of interest to researchers in the modeling of various civil engineering applications such as the prediction of compressive strength of unconfined and confined concrete and its stress-strain behavior [8–10].

In this study, a large test database built from an extensive survey of existing tests on CFRP-confined concrete cylinders is carefully examined to establish the effect of various variables. Then, a comparative study of soft computing methods such as multiple regressions (MR), ANN, genetic programming (GP), and ANFIS models for strength estimation of CFRP confined concrete cylinders is carried out, and finally, the best models’ results are also compared with five important empirical models so that the suitability and accuracy of the proposed models are confirmed.

2 Behavior of FRP-confined concrete

In recent years, advanced fiber-reinforced composites have been increasingly used especially in urgent rehabilitation and strengthening of structures located in densely populated earthquake zones. The gain in axial strength and ductility of reinforced concrete columns due to FRP wrapping is one of the most important reasons for increasing attention. A variety of experimental studies [11–30], performed for the description of the general behavior of FRP-confined concrete, point out a certain increase in the compressive strength of test specimens. Actually, the axial behavior of confined concrete was primarily researched by Richart et al. [31], and the following well-known relation was proposed for expressing the increase in the compressive strength based on their test results:

(1)fcc=fc+k1fl (1)

where fcc is the compressive strength of the confined concrete, fc is the compressive strength of the unconfined concrete, fl is the effective lateral confining stress, and k1 is an experimental constant. In the literature, there are several important models [3–7] for the description of the FRP-confined concrete. The majority of such models have focused on the analytical representation of the behavior of concrete specimens with circular cross-section adopting the classical approach proposed by Richart et al. [31]. The maximum confining pressure, fl, can be found in the usual way as:

(2)fl=2ntEfrpεrupd (2)

in which, EFRP represents the tensile modulus of FRP composite, t is the thickness of the composite jacket, εrup is the ultimate circumferential strain in the composite jacket, and d is the diameter of the concrete core.

After this approach, many researchers investigated specifically the FRP-confined concrete, and consequently, a considerable number of models developed. All of the proposed models were developed empirically by either doing regression analysis using existing test data or by a development based on the theory of plasticity with four or five parameters to be determined using available experimental data. Table 1 presents five important existing empirical models to predict the compressive strength of FRP-confined concrete [11, 13, 32–34].

Table 1

Some of the important strength models for FRP-confined concrete.

ModelAuthorYearTypeFormula
1Lam and Teng [32]2002Linearfcc/fc=1+2flfc
2Xiao and Wu [13]2000Second orderfcc/fc=1.1+(4.1-0.75fc2El)flfc
3Saafi et al. [11]1999Nonlinearfcc/fc=1+2.2(flfc)0.84
4Samaan et al. [33]1998Nonlinearfcc/fc=1+6.0fl0.7fc
5Saadatmanesh et al. [34]1994Second orderfcc/fc=2.254(1+7.94flfc)0.5-2flfc-1.254

3 Soft computing techniques, aims, and methodology

The definition of soft computing is not precise. Lotfi A. Zadeh, the inventor of the term soft computing, describes it as follows [35]: “Soft computing is a collection of methodologies that aim to exploit the tolerance for imprecision and uncertainty to achieve tractability, robustness, and low solution cost. Its principal constituents are fuzzy logic, neurocomputing, and probabilistic reasoning. Soft computing is likely to play an increasingly important role in many application areas, including software engineering. The role model for soft computing is the human mind”.

As demonstrated in Figure 1, this study encompasses the application of different soft computing techniques such as MRs, neural networks (NN), genetic programming (GP), and ANFIS models to predict the compressive strength of CFRP-confined concrete cylinders. The structure of this paper is as follows: First, the concepts and formulation of the models are explained. Then, the details of the models and their performance are described. Four MR models, one optimal NN, one GP model, and five ANFIS models are trained and tested by using the gathered database. The prediction performances of the models are evaluated by mean absolute percentage error (MAPE), root mean squared error (RMSE), and correlation coefficient (R). Finally, these models are compared with each other and with five important empirical models presented by researchers to discover the most suitable one.

Figure 1 Overview of the study.
Figure 1

Overview of the study.

4 Experimental database

In this study, an extensive literature review on experimental studies related to compressive strength of CFRP-confined concrete cylinders has been carried out, and an experimental database has been gathered. A total number of 128 specimens from 20 references with the ranges of variables included in the database [11–30] for which the statistical properties for scaling the input parameters are presented in Table 2.

Table 2

Statistical properties of experimental data.

Input parametersd (mm)h (mm)t (mm)Efrp (MPa)εrup (mm)fc (MPafcc (MPa)
Mean131.19294.810.41211,221.600.009139.6877.51
Max2006102611,6000.0207171303
Min511020.089019,9000.001717.3931.40
Standard deviation34.98120.860.36108,110.510.003325.6143.54
Coefficient of variation0.260.410.880.510.360.640.56

5 Preprocessing of data and performance criteria

To prevent the saturation problem and, consequently, the low rate of the training of the models [36], a simple linear mapping was applied on the rough input data for ANN and ANFIS. The following mapping function converts the real input values to the corresponding values in the interval of [0.1, 0.9]:

(3)iM=0.1+(0.90.1)×(iRimin)/(imaximin) (3)

where iM is the mapped input, iR is the rough input, imax and imin are the maximum and minimum rough input values, respectively. Clearly, an inverse mapping would be applied at the output layer to gain the real results.

The predicted and experimental values are compared using three criteria including RMSE, MAPE, and R. RMSE, MAPE, and R are calculated by using Eqs. (4), (5), and (6), respectively.

(4)RMSE=1N(ti-oi)2 (4)
(5)MAPE=1N(|ti-oioi|×100) (5)
(6)R=1-((ti-oi)2(oi)2) (6)

where ti is the target value, oi is the output value, and N is the number of samples.

6 Numerical modeling

6.1 MR (linear/nonlinear) models

MRs generally consist of linear and nonlinear regressions. Linear regression is a form of regression analysis in which the relationship between one or more independent and dependent variables is modeled by a linear regression equation, while in nonlinear regression, the aim is to find an appropriate nonlinear equation for this relationship. The general form of a MR model can be stated as follows:

(7)y=f(ai×xi) (7)

where y, f, ai, and xi are the dependent variable, linear or nonlinear function, constant coefficients, and the dependent variables of the concerned problem, respectively. The major issue is to find an appropriate function f with statistically well-adjusted coefficients ai. This is accomplished through iterative estimation algorithms, which are usually performed by statistical methods. In this paper, four regression models presented in Table 3 are considered to predict the compressive strength of CFRP-confined concrete cylinders. The evaluated coefficients are presented in Table 4.

Table 3

Multiple regression models for strength estimation of CFRP-confined concrete cylinders.

ModelReg. typeLinear/nonlinear regression models
1Lineara0+a1d+a2h+a3t+a4ε+a5E+a6fcc
2Logarithmica0+a1lnd+a2lnh+a3lnt+a4lnε+a5lnE+a6lnfcc
3Powera0+a1da2+a3ha4+a5ta6+a7εa8+a9Ea10+a11fcca12
4Quadratica0+a1d+a2h+a3t+a4ε+a5E+a6fcc+a7d2+a8h2+a9t2+a10ε2+a11E2+a12fcc2
Table 4

Evaluated coefficients of regression models.

Modela0a1a2a3a4a5a6
127.377-0.221-0.02828.8979.625E-051284.0991.092
2-125.244-49.653-2.33935.43240.22229.72942.97
315.79-0.2041.131-2.745-0.39828.0230.805
475.268-2.1060.019155.5944.369E-046010.2800.218
Modela7a8a9a10a11a12
1
2
31.4030.284-22.7675.2344.8540.676
40.007-6.187E-05-54.619-5.820E-10-304616.90.002

6.2 Neural network modeling

6.2.1 Brief overview of NNs

Artificial neural networks (NNs) are computer models that mimic the biological nervous system. An NN can be defined as a massively parallel distributed processor that has a natural propensity for storing experiential knowledge and making it available for use [37]. The basic element of an NN is the artificial neuron as shown in Figure 2, which consists of three main components, namely as weights, bias, and an activation function.

Figure 2 Basic elements of an artificial neuron.
Figure 2

Basic elements of an artificial neuron.

Each neuron receives inputs x1; x2;...; xn, attached with a weight wi, which shows the connection strength for that input for each connection. Each input is then multiplied by the corresponding weight of the neuron connection. A bias bi can be defined as a type of connection weight with a constant nonzero value added to the summation of inputs and corresponding weights ui, given by

(8)ui=j=1Hwijxj+bi (8)

The summation ui is transformed using a scalar-to-scalar function called an “activation or transfer function”, F(ui), yielding a value called the unit’s “activation”, given by

(9)Yi=f(ui) (9)

Activation functions serve to introduce nonlinearity into NNs. which makes NNs so powerful. NNs are commonly classified by their network topology (i.e., feedback, feed forward) and learning or training algorithms (i.e., supervised, unsupervised). For example, a multilayer feed forward NN with back propagation indicates the architecture and learning algorithm of the NN.

6.2.2 Optimal NN model selection

The performance of an NN model mainly depends on the network architecture and parameter settings. One of the most difficult tasks in NN studies is to find this optimal network architecture, which is based on determination of numbers of optimal layers and neurons in the hidden layers by trial and error approach. The assignment of initial weights and other related parameters may also influence the performance of the NN in a great extent. However, there is no well-defined rule or procedure to have optimal network architecture and parameter settings where trial and error method still remains valid. This process is very time consuming. In this study, Matlab NN toolbox is used for NN applications. Matlab NN toolbox randomly assigns the initial weights for each run each time, which considerably changes the performance of the trained NN even if all parameters and NN architecture are kept constant. This leads to extra difficulties in the selection of optimal network architecture and parameter settings. To overcome this difficulty, a program has been developed in Matlab, which handles the trial and error process automatically. The program tries various numbers of layers and neurons in the hidden layers both for first and second hidden layers for a constant epoch for several times and selects the best NN architecture with the minimum MAPE or RMSE of the testing set, and maximum R value. For instance, an NN architecture with one hidden layer with five nodes is tested 10 times, and the average error of 10 trials is stored where in the second cycle the number of hidden nodes is increased to six, and the process is repeated. All of the errors for different networks are stored and compared to select the best network. This process is repeated N times where N denotes the number of hidden nodes for the first hidden layer. This whole process is repeated for a changing number of nodes in the second hidden layer. Moreover, this selection process can be performed for different back propagation training algorithms such as trainlm, trainscg, and trainbfg, wherein the best one here selected is trainlm, which is a Levenberg-Marquardt algorithm. The Levenberg-Marquardt (LM) algorithm randomly divides input vectors and target vectors into three sets including training, validation, and testing. Changing the relative percentages of these three sets could slightly improve the generalization process. In this study, as a preliminary step, some data sets with different relative percentages were examined in which the training data varies from 50% to 90%, and the values of 60%, 20%, and 20% were selected for training, validation, and testing, respectively, in order to obtain the most efficient distribution of data sets. So, 60% of the whole data was specified as the training data in which the network would be adjusted according to its error. Similarly, 20% of the database was considered as the validating data, which was used to measure network generalization and to halt training when generalization stops improving. Finally, the remaining 20% of the whole data was specified as the testing data, which has no effect on training and so provides an independent measure of network performance during and after training. The network architecture used in this study was called NN 6-n-1, where the first digit is the number of input nodes, n is the number of nodes in the hidden layer, and the last digit is the number of output nodes as shown in Figure 3.

Figure 3 Schematic diagram of ANN model.
Figure 3

Schematic diagram of ANN model.

In this program, up to 20 neurons in the hidden layer were tested, and the optimal network was selected based on the minimum error and maximum correlation. The flowchart of the whole process is shown in Figure 4, in which the minimum error has occurred with six neurons in the hidden layer. So, the optimal NN architecture was obtained as 6-6-1. The results for training, validation, and testing of the NN 6-6-1 are summarized in Figures 57.

Figure 4 Flowchart of optimal NN selection.
Figure 4

Flowchart of optimal NN selection.

Figure 5 Performance of NN 6-6-1.
Figure 5

Performance of NN 6-6-1.

Figure 6 Training state of NN 6-6-1.
Figure 6

Training state of NN 6-6-1.

Figure 7 Regressions of training, validation, testing, and all data simulated by NN 6-6-1.
Figure 7

Regressions of training, validation, testing, and all data simulated by NN 6-6-1.

6.3 Genetic programming (GP)

Genetic algorithm (GA) is an optimization and search technique based on the principles of genetics and natural selection. A GA allows a population composed of many individuals to evolve under specified selection rules to a state that maximizes the “fitness” (i.e., minimizes the cost function). The method was developed by John Holland [38] and finally popularized by one of his students, David Goldberg [39], solved a difficult problem involving the control of gas-pipeline transmission for his dissertation [40]. The fitness of each individual in a genetic algorithm is the measure the individual has been adapted to the problem that is solved employing this individual. It means that fitness is the measure of optimality of the solution offered, as represented by an individual from the genetic algorithm. The basis of genetic algorithms is the selection of individuals in accordance with their fitness; thus, fitness is obviously a critical criterion for optimization [41]. GP is an extension to GA proposed by Koza [42]. Koza defines GP as a domain-independent problem-solving approach in which computer programs are evolved to solve, or approximately solve, problems based on the Darwinian principle of reproduction and survival of the fittest and analogs of naturally occurring genetic operations such as crossover (sexual recombination) and mutation. GP reproduces computer programs to solve problems by executing the following steps:

  1. Generate an initial population of random compositions of the functions and terminals of the problem (computer programs).

  2. Execute each program in the population and assign it a fitness value according to how well it solves the problem.

  3. Create a new population of computer programs. Copy the best existing programs (reproduction).

    1. Create new computer programs by mutation.

    2. Create new computer programs by crossover (sexual reproduction).

    3. Select an architecture-altering operation from the programs stored so far.

  4. The best computer program that appeared in any generation, the best-so-far solution, is designated as the result of genetic programming [42].

6.3.1 Brief overview of GEP

Gene expression programming (GEP) software, which is used in this study is an extension to GP that evolves computer programs of different sizes and shapes encoded in linear chromosomes of fixed length. The chromosomes are composed of multiple genes, each gene encoding a smaller subprogram. Furthermore, the structural and functional organization of the linear chromosomes allows the unconstrained operation of important genetic operators such as mutation, transposition, and recombination. A significance of the GEP approach is that the creation of genetic diversity is extremely simplified as genetic operators work at the chromosome level. Another significance of GEP consists of its unique, multigenic nature which allows the evolution of more complex programs composed of several subprograms. As a result, GEP surpasses the old GP system in 100–10,000 times [43–45]. APS 3.0 (http://www.gepsoft.com/), GEP software developed by Candida Ferreira, is used in this study. The fundamental difference between GA, GP, and GEP is due to the nature of the individuals: in GAs, the individuals are linear strings of fixed length (chromosomes); in GP, the individuals are nonlinear entities of different sizes and shapes (parse trees); and in GEP, the individuals are encoded as linear strings of fixed length (the genome or chromosomes), which are afterwards expressed as nonlinear entities of different sizes and shapes (i.e., simple diagram representations or expression trees). Thus, the two main parameters of GEP are the chromosomes and expression trees (ETs). The process of information decoding (from the chromosomes to the ETs) is called translation, which is based on a set of rules. The genetic code is very simple where there exist one-to-one relationships between the symbols of the chromosome and the functions or terminals they represent. The rules, which are also very simple, determine the spatial organization of the functions and terminals in the ETs and the type of interaction between sub-ETs [43–45]. That is why two languages are utilized in GEP: the language of the genes and the language of ETs. A significant advantage of GEP is that it enables us to infer exactly the phenotype given the sequence of a gene, and vice versa, which is termed as Karva language. Consider, for example, the algebraic expression d4d3-d0+d1d4-d4, which can be represented by a diagram (Figure 8), which is the expression tree.

Figure 8 Expression tree (ET).
Figure 8

Expression tree (ET).

6.3.2 Solving a simple problem with GEP

For each problem, the type of linking function, as well as the number of genes and the length of each gene, are a priori chosen for each problem. While attempting to solve a problem, one can always start by using a single-gene chromosome and then proceed by increasing the length of the head. If it becomes very large, one can increase the number of genes and obviously choose a function to link the sub-ETs. One can start with addition for algebraic expressions or OR for Boolean expressions, but in some cases, another linking function might be more appropriate (like multiplication or IF, for instance). The idea, of course, is to find a good solution, and GEP provides the means of finding one very efficiently [44]. As an illustrative example, consider the following case where the objective is to show how GEP can be used to model complex realities with high accuracy. So, suppose one is given a sampling of the numerical values from the curve (remember, however, that in real-world problems, the function is obviously unknown):

(10)y=3a2+2a+1 (10)

There are over 10 randomly chosen points in the real interval [-10, +10] and the aim is to find a function fitting those values within a certain error. In this case, a sample of data in the form of 10 pairs (ai, yi) is given, where ai is the value of the independent variable in the given interval, and yi is the respective value of the dependent variable (ai values: -4.2605, -2.0437, -9.8317, -8.6491, 0.7328, -3.6101, 2.7429, -1.8999, -4.8852, 7.3998; the corresponding yi values can be easily evaluated). These 10 pairs are the fitness cases (the input) that will be used as the adaptation environment. The fitness of a particular program will depend on how well it performs in this environment [44]. There are five major steps in preparing to use GEP. The first is to choose the fitness function. For this problem, one could measure the fitness fi of an individual program i by the following expression:

(11)fi=j=1Ct(M-|C(i,j)-Tj|) (11)

where M is the range of selection, C(i,j) the value returned by the individual chromosome i for fitness case j (out of Ct fitness cases), and Tj is the target value for fitness case j. If, for all j, ∣C(i,j)Tj∣ (the precision) is less than or equal to 0.01, then the precision is equal to zero, and fi=fmax=Ct*M. For this problem, use an M=100, and therefore, fmax=1000. The advantage of this kind of fitness function is that the system can find the optimal solution for itself. However, there are other fitness functions available, which can be appropriate for different problem types [44]. The second step is choosing the set of terminals T and the set of functions F to create the chromosomes. In this problem, the terminal set consists obviously of the independent variable, i.e., T={a}. The choice of the appropriate function set is not so obvious, but a good guess can always be done in order to include all the necessary functions. In this case, to make things simple, use the four basic arithmetic operators. Thus, F={+, -, *, /}. It should be noted that there are many other functions that can be used. The third step is to choose the chromosomal architecture, i.e., the length of the head and the number of genes.

The fourth major step in preparing to use GEP is to choose the linking function. In this case, we will link the sub-ETs by addition. Other linking functions are also available such as subtraction, multiplication, and division. Finally, the fifth step is to choose the set of genetic operators that cause variation and their rates. In this case, one can use a combination of all genetic operators (mutation at pm=0.051; IS and RIS transposition at rates of 0.1 and three transpositions of lengths 1, 2, and 3; one-point and two-point recombination at rates of 0.3; gene transposition and gene recombination both at rates of 0.1). To solve this problem, let us choose an evolutionary time of 50 generations and a small population of 20 individuals in order to simplify the analysis of the evolutionary process and not fill this text with pages of encoded individuals. However, one of the advantages of GEP is that it is capable of solving relatively complex problems using small population sizes, and thanks to the compact Karva notation, it is possible to fully analyze the evolutionary history of a run. A perfect solution can be found in generation 3, which has the maximum value 1000 of fitness. The sub-ETs codified by each gene are given in Figure 9. Note that it corresponds exactly to the same test function given above in Eq. (10) [44]. Thus, expressions for each corresponding sub-ET can be given as follows:

(12)y=(a2+a)+(a+1)+(2a2)=3a2+2a+1 (12)
Figure 9 ET for the problem of Eq. (10).
Figure 9

ET for the problem of Eq. (10).

6.3.3 Numerical application of GP

The main focus of the GP in this paper is to model the compressive strength of CFRP-confined concrete cylinders based on experimental database gathered from the literature. Among these experimental data, 96 sets were used for GP training and 32 sets for GP testing. As mentioned earlier, the compressive strength of confined concrete is considered as a function of geometric and mechanical properties given as follows:

fcc=f(fc,d,h,t,εrup,EFRP)

Related parameters for the training of the GP model are given in Table 5. Detailed information on the values given in Table 3 can be found in Section 6.3.2.

Table 5

Parameters of the GP model.

P1Function set+, -, *, /, √, ex, ln(x), power
P2Chromosomes55
P3Head size11
P4Number of genes5
P5Linking functionAddition
P6Fitness function error typeMean absolute error (MAE)
P7Mutation rate0.044
P8Inversion rate0.1
P9IS transposition rate0.1
P10RIS transposition rate0.1
P11One-point recombination rate0.3
P12Two-point recombination rate0.3
P13Gene recombination rate0.1
P14Gene transposition rate0.1

6.4 ANFIS model

ANFIS is the famous hybrid neuro-fuzzy network for modeling the complex systems [46, 47]. ANFIS incorporates the human-like reasoning style of fuzzy systems through the use of fuzzy sets and a linguistic model consisting of a set of IF-THEN fuzzy rules. The main strength of ANFIS models is that they are universal approximators [46] with the ability to solicit interpretable IF-THEN rules.

6.4.1 Architecture of ANFIS

The architecture of an ANFIS model with two input variables is shown in Figure 10. Suppose that the rule base of ANFIS contains two fuzzy IF-THEN rules of Takagi and Sugeno type as follows:

Figure 10 The reasoning scheme of ANFIS [44].
Figure 10

The reasoning scheme of ANFIS [44].

Rule 1: IF x is A1 and y is B1, THEN f1=p1x+q1y+r1.

Rule 2: IF x is A2 and y is B2, THEN f2=p2x+q2y+r2.

where A1, A2 and B1, B2 are membership values of input variables x and y, respectively; p1, q1, r1 and p2, q2, r2 are the parameters of the output functions f1 and f2, respectively. A basic Takagi-Sugeno inference system, which produces an output function f from input variables x and y by applying triangular membership functions is illustrated schematically in Figure 10, and also, the corresponding equivalent ANFIS architecture is shown in Figure 11.

Figure 11 Schematic of ANFIS architecture [44].
Figure 11

Schematic of ANFIS architecture [44].

The functions of each layer are described as follows [46–48]:

Layer 1 – Every node i in this layer is a square node with a node function:

(13)Oi1=μAi(x) (13)

where x is the input to node i, and Ai is the linguistic label (fuzzy sets: small, large,...) associated with this node function.

Layer 2 – Every node in this layer is a circle node labeled Π, which multiplies the incoming signals and sends the product out. For instance,

(14)w=iμAi(y)×μBi(y),i=1,2 (14)

Each node output represents the firing weight of a rule.

Layer 3 – Every node in this layer is a circle node labeled N. The ith node calculates the ratio of the ith rule’s firing weight to the sum of all rule’s firing weights:

(15)w¯i=wi/(w1+w2),i=1,2 (15)

Layer 4 – Every node in this layer is a square node with a node function:

(16)Oi4=w¯i(pix+qi+ri) (16)

where w¯i is the output of layer 3, and {pi, qi, ri} is the parameter set.

Layer 5 – The signal node in this layer is a circle node labeled Σ that computes the overall output as the summation of all incoming signals, i.e.,

(17)Oi5=iw¯ifi=iwifi/iwi (17)

The basic learning rule of ANFIS is the back-propagation gradient descent, which calculates error signals recursively from the output layer backward to the input nodes. This learning rule is exactly the same as the back-propagation learning rule used in the common feed-forward neural networks [47, 48]. Recently, ANFIS adopted a rapid learning method named as hybrid-learning method, which utilizes the gradient descent and the least-squares method to find a feasible set of antecedent and consequent parameters [47, 48]. Thus, in this paper, the latter method is used for constructing the proposed models.

The structure of the proposed ANFIS networks was consisted of six input variables (i.e., d,h,t,E,ε,andfc) and one output variable (fcc). In this paper, for comparison purposes, five types of membership functions (MFs) including the triangular, trapezoidal, bell-shape, Gaussian, and π-shaped were utilized to construct the suggested models. The ANFIS models were trained by 96 data sets and tested by 32 data sets. Moreover, to monitor the performance of the training process, 32 data sets were randomly selected as checking sets from both training and testing pairs. Parameter types and their values used in the ANFIS models can be seen in Table 6.

Table 6

Different parameter types and their values used for ANFIS models.

Different parameter of ANFIS models
TypeValue
MF type1) Triangular
2) Trapezoidal
3) Bell-shape
4) Gaussian
5) π-shaped
Number of MFs12
Output MFLinear
Number of nodes161
Number of linear parameters448
Number of nonlinear parameters48
Total number of parameters496
Number of training data pairs96
Number of checking data pairs32
Number of fuzzy rules64
ANDProduct
ORMaximum
ImplicationProduct
AggregationMaximum
DefuzzificationWeighted average

7 Results and discussion

7.1 MR models

Generally, four regression models were proposed for the strength estimation of CFRP-confined concrete cylinders. The first model is linear, and the rest are nonlinear. Reported in Table 7 are the performance criteria of the models, i.e., RMSE, MAPE, and R values, which are calculated from Eqs. (12), (13), and (14), respectively. The results show that model 4 (quadratic) has the lowest error and highest R. Also, the goodness of predictions with model 4 might be confirmed. Thus, the model 4 is proven to have better predictions than the other proposed MR models here.

Table 7

Performance criteria of MR models.

Multiple regression models
TypeLinearLogarithmicPowerQuadratic
MAPE23.7234.8424.4812.79
RMSE25.0818.6225.0313.85
R0.95570.97670.95580.9875

Moreover, in Figure 12, a comparison is made between real experimental observations and the predictions of model 4, which is the best of MR models here. In this figure, which is hereafter called X-Y plot, the horizontal axis represents experimental observations, and the vertical one is related to the model’s results. In this plot, the more points congregated about the diagonal line, the better performance for the model could be judged. As can be seen in this figure, the model has a relatively good correlation with the experimental results.

Figure 12 Predictions of the best MR model against experimental data.
Figure 12

Predictions of the best MR model against experimental data.

7.2 ANN model

As mentioned before, the optimal neural network obtained by a program, with its general flowchart presented in Figure 4. The ANN performance criteria are also presented in Table 8. Again, the X-Y plot was utilized to verify the robustness of this model as shown in Figure 13. It can be observed that more points were well gathered around the diagonal line.

Table 8

Performance criteria of ANN model.

Optimal ANN model
MAPERMSER
10.4511.990.9904
Figure 13 Predictions of the ANN model against experimental data.
Figure 13

Predictions of the ANN model against experimental data.

7.3 GP model

The performance of the proposed GP predictions vs. experimental results are given in Figure 14, and the performance criteria of the model are also presented in Table 9. As can be observed, the predictions are quite satisfactory with correlation coefficient of about R=0.94.

Table 9

Performance criteria of GP model.

GP model
MAPERMSER
10.1710.770.9925
Figure 14 Predictions of the GP model against experimental data.
Figure 14

Predictions of the GP model against experimental data.

7.4 ANFIS models

The performance of ANFIS models was also examined by MAPE, RMSE, and R, and the results are summarized in Table 10. As seen in this table, all of the ANFIS models have acceptable prediction performance. Among these models, model 2, which is constructed with trapezoidal MFs, exhibited the best performance. In addition, aggregation of the representative points around the X-Y plot shown in Figure 15, confirms that the ANFIS model with trapezoidal-type MFs is the best model for strength estimation of CFRP-confined concrete cylinders.

Table 10

Performance criteria of ANFIS models.

ANFIS models
TypeTriangularTrapezoidalGaussianBell-shapedπ-shaped
MAPE2.301.852.082.002.14
RMSE2.372.032.192.122.28
R0.999640.999740.999690.999710.99967
Figure 15 Predictions of the best ANFIS model against experimental data.
Figure 15

Predictions of the best ANFIS model against experimental data.

7.5 Empirical models

Five important empirical models for verification of the proposed models in this study were selected [11, 13, 32–34] including the strength models proposed by Lam and Teng (2002), Xiao and Wu (2000), Saafi et al. (1999), Samaan et al. (1998), and Saadatmanesh et al. (1994). The models’ performance criteria are also presented in Table 11 so that the best one among these five empirical models could be selected. It can be inferred from the table that the best model should be selected by considering the lowest RMSE/R or MAPE/R ratio, as the final conclusion on the best model cannot be made by separately considering the error or R values. Therefore, the best empirical model among these five models is proven to be the model proposed by Xiao and Wu. The X-Y plot of this model is shown in Figure 16.

Table 11

Performance criteria of empirical models.

Empirical models
TypeLam and Teng [32]Xiao and Wu [13]Saafi et al. [11]Samaan et al. [33]Saadatmanesh et al. [34]
MAPE35.9313.5324.6320.1216.54
RMSE26.0514.7220.8920.7917.76
R0.92440.98740.98720.99360.9977
Figure 16 Predictions of the best empirical model against experimental data.
Figure 16

Predictions of the best empirical model against experimental data.

7.6 Comparison of the models

The performance criteria of the different soft computing techniques used in this study such as the optimal neural network model, GP model, the best regression model (i.e., model 4), ANFIS model (i.e., model 2), and empirical model (i.e., model 2) are compared in Table 12. Furthermore, the predictions of the five models along with experimental results are plotted comparatively in Figure 17. The relative errors (RE) of the five predictive models are also plotted for all samples in Figure 18. By comparing the results, it might be concluded that MR and empirical models have, to some extent, similar accuracy. After that, the optimal ANN model presents better predictions, and the GP model gives even more accurate results. Finally, it can be obviously seen that the ANFIS models are far more powerful and accurate and can recognize the relationships very well because of their distributed and parallel computing nature.

Table 12

Comparison of performance criteria values for MR, ANN, ANFIS, and empirical models.

Type of models
Multiple regression (model 4)Optimal ANNGPANFIS (model 2)Empirical model (model 2) [13]
MAPE12.7910.4510.171.8513.53
RMSE13.8511.9910.772.0314.72
R0.98750.99040.99250.999740.9874
Figure 17 Experimental results along with predictions of the models.
Figure 17

Experimental results along with predictions of the models.

Figure 18 Relative errors (RE) of MR, ANN, GP, ANFIS, and empirical models.
Figure 18

Relative errors (RE) of MR, ANN, GP, ANFIS, and empirical models.

8 Conclusion

In this paper, various soft computing methods, namely, MR (linear and nonlinear), ANN, GP, and ANFIS models were utilized for strength estimation of CFRP-confined concrete cylinders. For strength estimation of CFRP-confined concrete cylinders, the constitutive parameters as input included the diameter and height of the cylindrical specimen, ultimate circumferential strain in the CFRP jacket, elastic modulus of CFRP, unconfined concrete strength, and total thickness of the CFRP layer used. Totally, 128 data sets were gathered from the literature based on which the models were developed. The performance criteria used to evaluate the suitability and robustness of each model included MAPE, RMSE, and R. Multiple regression models included one linear and three nonlinear models among which model 4 (quadratic) was proven to be the best one based on the performance criteria. The optimal ANN architecture was obtained by a program that automatically finds the best ANN model based on performance criteria. A GP model was developed, which yielded precise and satisfactory results. Five ANFIS models were also presented in which model 2 with trapezoidal MFs had the best performance. Comparing the MR, ANN, GP, ANFIS, and empirical models, it was considered that the mean absolute% errors (MAPEs) of the best empirical model, the best MR, optimal ANN, GP, and ANFIS models were 13.53%, 12.79%, 10.45%, 10.17%, 1.85%, respectively. Furthermore, as the errors of five ANFIS models were around 2%, all of them could be suitable for strength estimation. In general, the regression model 4 and empirical model 2 could be proposed for rough estimate of strength, while the ANN, GP, and particularly ANFIS models could be proposed in the case of high accuracy requirements.


Corresponding author: Mostafa Jalal, Young Researchers Club and Elites, Science and Research Branch, Islamic Azad University, Tehran, Iran, e-mail: ;

References

[1] Hollaway LC. In Advanced Polymer Composites for Structural Applications in Construction: Proceedings of the 2nd International Conference, held at the University of Surrey, Guildford, UK on 20–22 April, 2004.Search in Google Scholar

[2] Leung CKY, Ng MYM, Luk HCY. J. Compos. Constr. (ASCE) 2006, 10, 125–138.10.1061/(ASCE)1090-0268(2006)10:2(125)Search in Google Scholar

[3] Matthys S, Toutanji H, Audenaert K, Taerwe L. ACI Struct. J. 2005, 102, 258–267.Search in Google Scholar

[4] Lam L, Teng JG, Cheng CH, Xiao Y. Cem. Concr. Res. 2006, 28, 949–958.Search in Google Scholar

[5] Deniaud C, Neale KW. Compos. Struct. 2006, 73, 318–330.Search in Google Scholar

[6] Teng JG, Yu T, Wong YL, Dong SL. Constr. Build. Mater. 2007, 21, 846–854.Search in Google Scholar

[7] Lee C, Hegemier GA. J. Compos. Constr. (ASCE) 2009, 13, 442–454.10.1061/(ASCE)CC.1943-5614.0000029Search in Google Scholar

[8] Ashrafi HR, Jalal M, Garmsiri K. Expert Syst. Appl. 2010, 37, 7663–7668.Search in Google Scholar

[9] Jalal M, Ramezanianpour AA, Pouladkhan AR, Tedro P. Neural Comput. Applic. 2012, 23, 455–470.Search in Google Scholar

[10] Jalal M, Ramezanianpour AA. Compos. Part B 2012, 43, 2990–3000.10.1016/j.compositesb.2012.05.044Search in Google Scholar

[11] Saafi M, Toutanji HA, Li Z. ACI Mater. J. 1999, 96, 500–509.Search in Google Scholar

[12] Rochette P, Labossiere P. J. Compos. Constr. (ASCE) 2000, 4, 129–136.10.1061/(ASCE)1090-0268(2000)4:3(129)Search in Google Scholar

[13] Xiao Y, Wu H. J. Mater. Civ. Eng. 2000, 12, 139–146.Search in Google Scholar

[14] Harmon TG, Slattery KT. In First International Conference on Advanced Composite Materials in Bridges and Structures, Sherbrooke, Québec, Canada, 1992, pp. 299–306.Search in Google Scholar

[15] Picher F, Rochette P, Labossiére P. In Proceedings of the First Composites in Infrastructure ICCI’96, 15–17 January, Tucson, Arizona, 1996, pp. 829–841.Search in Google Scholar

[16] Watanabe K, Nakamura H, Honda Y, Toyoshima M, Iso M, Fujinaki T, Kaneto M. In Proceedings of the Third International Symposium on Non-Metallic FRP for Concrete Structures, Japan, 1997, pp. 233–240.Search in Google Scholar

[17] Kono S, Inazumi M, Kaku T. In Proceedings of the Second International Conference on Composites in Infrastructure ICCI’98, Tucson, Arizona, 5–7 January, 1998, pp. 343–355.Search in Google Scholar

[18] Toutanji HA. ACI Mater J. 1999, 96, 397–404.Search in Google Scholar

[19] Matthys S, Taerwe L, Audenaert K. In Fourth International Symposium on Fiber Reinforced Polymer Reinforcement for Reinforced Concrete Structures, SP-188, Farmington, Michigan, USA, American Concrete Institute, 1999, pp. 217–229.Search in Google Scholar

[20] Shahawy M, Mirmiran A, Beitelman A. Compos. Part B Eng. ASCE 2000, 31, 471–480.10.1016/S1359-8368(00)00021-4Search in Google Scholar

[21] Micelli F, Myers JJ, Murthy S. In Proceedings of CCC2001 International Conference on Composites in Construction, Porto, Portugal, 2001.Search in Google Scholar

[22] Mirmiran A, Shahawy M, Samaan M. J. Struct. Eng. ASCE 1999, 125, 1085–1093.10.1061/(ASCE)0733-9445(1999)125:10(1085)Search in Google Scholar

[23] De Lorenzis L, Micelli F, La Tegola A. In Advanced Polymer Composites for Structural Applications in Construction, Proceedings of the First International Conference, Shenoi, RA, Moy, SSJ, Hollaway, LC, Eds., Thomas Telford: London, UK, 2002, pp. 231–239.10.1680/apcfsaic.31227.0024Search in Google Scholar

[24] Dias da Silva V, Santos JMC. In Composites in Constructions, Proceedings of the International Conference, Figueiras, J, Juvandes, L, Faria, R, Marques, AT, Ferreira, A, Barros, J, Appleton, J, Eds., Lisse, A.A. Balkema Publishers: The Netherlands, 2001, pp. 257–262.Search in Google Scholar

[25] Pessiki S, Harries KA, Kestner JT, Sause R, Ricles JM. J. Compos. Constr. ASCE 1997, 5, 237–245.10.1061/(ASCE)1090-0268(2001)5:4(237)Search in Google Scholar

[26] Wang P, Cheong KK. In FRP Composites in Civil Engineering, Proceedings of the International Conference, Teng JG, Ed., Elsevier Science Ltd: Oxford, UK, 2001, pp. 327–334.Search in Google Scholar

[27] Shehata IAEM, Carneiro LAV, Shehata LCD. Mater. Struct. 2002, 35, 50–58.Search in Google Scholar

[28] Kshirsagar S, Lopez-Anido RA, Gupta RK. ACI Mater. J. 2000, 97, 703–712.Search in Google Scholar

[29] Berthet JF, Ferrier E, Hamelin P. Constr. Build. Mater. 2005, 19, 223–232.Search in Google Scholar

[30] Lin CT, Li YF. Can. J. Civil Eng. 2003, 30, 882–889.Search in Google Scholar

[31] Richart FE, Brandtzaeg A, Brown RL. A Study of the Failure of Concrete Under Combined Compressive Stress. University of Illinois, Engineering Experimental Station: Illinois, USA, 1928.Search in Google Scholar

[32] Lam L, Teng J-G. J. Struct. Eng. ASCE 2002, 128, 612–623.10.1061/(ASCE)0733-9445(2002)128:5(612)Search in Google Scholar

[33] Samaan M, Mirmiran A, Shahawy M. J. Struct. Eng. ASCE 1998, 124, 1025–1031.10.1061/(ASCE)0733-9445(1998)124:9(1025)Search in Google Scholar

[34] Saadatmanesh H, Ehsani MR, Li MW. ACI Struct. J. 1994, 91, 434–447.Search in Google Scholar

[35] Zadeh LA. IEEE Software, 1994, 11, 48–56.10.1109/52.329401Search in Google Scholar

[36] Fausett LV. Fundamentals of Neural Networks: Architectures, Algorithms, and Applications, 1st ed., Prentice Hall, 1994. ASIN: B008VR8XAS.Search in Google Scholar

[37] Haykin S. Neural Networks – a Comprehensive Foundation, Macmillan College Publishing Company: New York, 2000.Search in Google Scholar

[38] Holland JH. Adaptation in Natural and Artificial Systems, University of Michigan Press: Ann Arbor, 1975.Search in Google Scholar

[39] Goldberg DE. Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley: Reading, MA, USA, 1989.Search in Google Scholar

[40] Haupt RL, Haupt SE. Practical Genetic Algorithms, 2nd ed., Wiley- Interscience, 2004, pp. 272. ISBN 0471455652.10.1002/0471671746Search in Google Scholar

[41] Chambers L. The Practical Handbook of Genetic Algorithms Applications, 2nd ed., Chapman & Hall/CRC, 2001.10.1201/9781420035568Search in Google Scholar

[42] Koza JR. Genetic Programming: on the Programming of Computers by Means of Natural Selection, MIT Press: Cambridge, MA, USA, 1992.Search in Google Scholar

[43] Ferreira C. Gene Expression Programming: Mathematical Modelling by an Artificial Intelligence, 2nd ed., Springer-Verlag: Germany, 2002.Search in Google Scholar

[44] Ferreira C. In Invited Tutorial of the 6th Online World Conference on Soft Computing in Industrial Applications, 2001, pp. 10–24.Search in Google Scholar

[45] Ferreira C. Complex Syst. 2001, 13, 87–129.Search in Google Scholar

[46] Jang JSR. IEEE Trans. Syst. Man. Cyber. 1993, 23, 665–685.Search in Google Scholar

[47] Jang JSR, Sun CT. Proc. IEEE 1995, 83, 378–406.10.1109/5.364486Search in Google Scholar

[48] Ramezanianpour AA, Sobhani J, Sobhani M. In Proceedings of the Fourth International Conference on Engineering Computational Technology, Civil-Comp Press: Lisbon, Portugal, 2004, p. 138.Search in Google Scholar

Received: 2013-10-2
Accepted: 2013-10-26
Published Online: 2013-12-18
Published in Print: 2015-1-1

©2015 by De Gruyter

This article is distributed under the terms of the Creative Commons Attribution Non-Commercial License, which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Downloaded on 5.6.2024 from https://www.degruyter.com/document/doi/10.1515/secm-2013-0240/html
Scroll to top button