innovative seismic design optimization …...two types of random variables are considered: those...
TRANSCRIPT
![Page 1: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/1.jpg)
1
INNOVATIVE SEISMIC DESIGN OPTIMIZATION WITH RELIABILITY CONSTRAINTS
ANARGYRI TH. GARAVELAS, NIKOS D. LAGAROS, MANOLIS PAPADRAKAKIS*
Institute of Structural Analysis & Seismic Research, National Technical University of Athens,
9, Iroon Polytechniou Str., Zografou Campus, 157 80 Athens, Greece e-mail: {cv01053,nlagaros, mpapadra)@central.ntua.gr
ABSTRACT: Performance-Based Design (PBD) methodologies is the contemporary trend in designing better
and more economic earthquake-resistant structures where the main objective is to achieve more predictable and
reliable levels of safety and operability against natural hazards. On the other hand Reliability-Based Optimization
(RBO) methods directly account for the variability of the design parameters into the formulation of the
optimization problem. The objective of this work is to incorporate PBD methodologies under seismic loading into
the framework of RBO in conjunction with innovative tools for treating computational intensive problems of
real-world structural systems. Two types of random variables are considered: Those which influence the level of
seismic demand and those that affect the structural capacity. Reliability analysis is required for the assessment of
the probabilistic constraints within the RBO formulation. The Monte Carlo Simulation (MCS) method is
considered as the most reliable method for estimating the probabilities of exceedance or other statistical quantities
albeit with excessive, in many cases, computational cost. First or Second Order Reliability Methods (FORM,
SORM) constitute alternative approaches which require an explicit limit-state function. This type of limit-state
function is not available for complex problems. In this study, in order to find the most efficient methodology for
performing reliability analysis in conjunction with performance-based optimum design under seismic loading, a
Neural Network approximation of the limit-state function is proposed and is combined with either MCS or with
FORM approaches for handling the uncertainties. These two methodologies are applied in RBO problems with
sizing and topology design variables resulting in two orders of magnitude reduction of the computational effort.
KEY WORDS: Performance-based design, reliability-based optimization, response surface, Monte Carlo, neural
networks, sizing-topology design variables, earthquake resistant structures
*Corresponding Author
![Page 2: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/2.jpg)
2
1. INTRODUCTION
In deterministic optimization problems randomness and uncertainty are considered through
several safety factors introduced into the constraints of the optimization problem. On the other
hand, non-deterministic performance measures are increasingly being taken into consideration
in many contemporary engineering applications that involve various reliability requirements.
In structural optimization, non-deterministic performance measures can be taken into account
using two distinct formulations: Robust Design Optimization (RDO) [1,2] and
Reliability-Based design Optimization (RBO) [3-5].
The modern conceptual approach to structural design under seismic loading is based on
the principal that a structure should meet performance-based objectives for a number of
different hazard levels, ranging from earthquakes of small intensity with small return period, to
more destructive events with large return period. This approach constitutes the
Performance-Based Design (PBD) concept which has been introduced in order to increase the
safety against natural hazards. According to PBD the structures should be able to resist
earthquakes in a quantifiable manner and to preset levels of desired possible damage. The
current state of practice in performance-based engineering can be found in US guidelines such
as FEMA-350 [6] and FEMA-356 [7].
The structural performance during an earthquake depends highly on a number of
structural parameters which are inherently uncertain. Such parameters are, among others, the
material properties, the workmanship, the hysteretic behaviour of structural members and
joints. The intensity and the earthquake ground motion characteristics are also random.
Furthermore, uncertainty is also involved in the analysis procedure that would be adopted and
in the numerical simulation of the structure. In order to account for as many as possible of the
above uncertainties, a reliability-based in conjunction with a performance-oriented approach
should be considered. For earthquake engineering applications, the reliability problem can be
defined as a problem of two separate types of variables, representing the demand and the
capacity. This concept is adopted in the reliability-based design procedure suggested in the
SAC/FEMA project [8]. The SAC/FEMA methodology allows the consideration of
uncertainties on both capacity and demand. However, for the calculation of the limit-state
probabilities, prior knowledge of the distribution and the variance of the capacity and the
demand is necessary.
A limited number of studies have been published in the past, where the PBD concept is
implemented in a structural optimization problem considering uncertainties. These studies are
![Page 3: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/3.jpg)
3
restricted to relatively small-scale structures due to the increased computational cost. Αleatory
and epistemic uncertainties are introduced in a structural optimization environment by Beck et
al. [9]. In the work by Wen [10], the issue of the proper consideration of the uncertainty in the
demand and capacity and the balance of reliability against cost is investigated based on the
minimization of the expected lifecycle cost. The concept of performance-based design within
the context of robust design optimization for the design of structures is examined by Lagaros
and Fragiadakis [11]. In the work by Foley et al. [12] a state-of-the-art model code
performance-based design methodology is proposed. This design methodology is applied to
multiple-objective optimization problems for single storey and multi-storey structural
frameworks with fully and partially restrained connections.
Structural reliability analysis can be performed either with simulation methods, such as
the Monte Carlo Simulation (MCS) method, or with other approximation methods. First and
second order reliability methods (FORM, SORM) require prior knowledge of the mean and the
variance of each random variable. Furthermore, these methods require a differentiable failure
function. On the other hand, although the major advantage of MCS is that accurate solutions
can be obtained for almost every problem, yet it requires excessive computational cost in many
cases. Variance reduction techniques, such as Importance Sampling, Directional Simulation,
Antithetic Variates or Adaptive Sampling, have been proposed in order to reduce the
computational effort of MCS. The disadvantage of these methods is that they require prior
knowledge of the behaviour of the structure in order to determine the most effective sampling
region, which for many practical problems is not clearly identifiable. Recent results [13] reveal
that variance reduction techniques still require significant number of the system response
evaluations to estimate failure probabilities of the order less than 10−3. Other recently proposed
simulation methods, such as Line Sampling [13] and Subset Simulation [14], were proved to be
very efficient in reducing the required sample size and the computational cost; however, their
performance and ergodicity are sensitive to the values of certain parameters which are not
known a priori.
The computational cost is the main barrier that prohibited the application of uncertain
optimization methodologies considering earthquake loading. For the application of a
computationally efficient MCS method to complex structural models it would be necessary to
have an approximate knowledge of the limit-state function g(x). Such an approximation
contributes in reducing the excessive number of repeated finite element analyses required for
the MCS. On the other hand, since complex structural reliability problems are characterized by
![Page 4: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/4.jpg)
4
the implicit nature of the limit-state functions, the implementation of FORM or SORM requires
an explicit approximation of either the entire limit-state function g(x) or of its limit-state
surface g(x)=0 in the space of the random variables x. The Response Surface (RS) method is
customarily used for approximating the limit-state function. It was found, though, that the
performance of the RS method is dependent on the experimental points required to define the
limit-state function approximation due to the rigid and non-adaptive structure of the function
implemented by the RS method [15,16]. Several attempts have been presented in the past where
the limit-state function is estimated with a Neural Network (NN) approximation. These
attempts, however, were limited to static loads [17,18]. In the present work innovative and
efficient procedures for performing RBO are implemented, within the performance-based
design framework of FEMA-356 under seismic loading, for the design of real-world structural
systems. Randomness and uncertainty are taken into consideration with two methodologies for
reducing the computational cost. The methodologies are based on the Monte Carlo and the First
Order Reliability methods and exploit the Neural Network predictions of the limit state
function. Furthermore, sizing and topology design variables are incorporated into deterministic
and probabilistic formulations of the optimization problem.
2. PERFORMANCE-BASED DESIGN
The majority of the seismic design codes belong to the category of the prescriptive design
codes, which take into consideration site selection and development of conceptual, preliminary
and final design stages. According to a prescriptive design code the strength of the structure is
evaluated at one limit state, between life-safety and near collapse, using a response spectrum
based loading corresponding to one design earthquake [19]. In addition, the serviceability limit
state is usually checked in order to ensure that the structure will not deflect or vibrate
excessively. On the other hand, PBD is a different approach for the seismic design of structural
systems which includes, apart from the site selection and the consideration of the design stages,
the performance of the building after construction in order to ensure reliable and predictable
seismic performance over its life.
Prescriptive building codes do not provide acceptable levels of a building life-cycle
performance, since they only include provisions aiming at ensuring adequate strength of
structural members and, indirectly or implicitly, of the overall structural strength. The basic
philosophy of a PBD procedure is to allow engineers to determine explicitly the seismic
demands at preset performance levels by introducing design checks of a higher level.
![Page 5: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/5.jpg)
5
Performance-based seismic design has the following distinctive features with respect to a
prescriptive design requirements: (i) Allows the owner in cooperation with the structural
engineer to define the appropriate level of seismic hazard and the corresponding performance
level where the seismic demand will be evaluated. (ii) The structure is designed to meet the
requirements corresponding to a number of levels of seismic intensity.
2.1. Performance Levels and Objectives
The main part in a performance-based seismic design procedure is the definition of the
performance objectives. A performance objective is defined as a given level of performance for
a specific hazard level. In this study, three performance objectives corresponding to the
“Enhanced Rehabilitation Objectives” of FEMA-356 [7] for buildings have been considered.
The first part in the definition of a performance objective is the selection of the level of
structural performance. The following performance levels have been considered:
i) Operational level: the overall damage is characterized as very light. No permanent drift
is encountered, while the structure essentially retains original strength and stiffness.
ii) Life Safety level: the overall damage is characterized as moderate. Permanent drift is
encountered but partial or total structural collapse is avoided. Gravity-load bearing
elements continue to function while there is no out-of plane failure of the infill walls.
The overall risk of life-threatening injury as a result of structural damage is expected to
be low. It should be possible to repair the structure; however, for economic reasons this
may not be practical.
iii) Collapse Prevention level: the overall damage is characterized as severe. Substantial
damage has occurred to the structure, including significant degradation in the stiffness
and strength of the lateral-force resisting system. Large permanent lateral deformation
of the structure and degradation in the vertical load bearing capacity is encountered.
However, all significant components of the gravity load resisting system continue to
carry their gravity load demands. The structure may not be technically repairable and is
not safe for reoccupancy, since aftershock activity could induce collapse.
The second part in the definition of a performance objective is to determine the
corresponding seismic hazard level. Earthquake hazard, according to FEMA-350 [6], includes
parameters such as direct ground fault rupture, ground shaking, liquefaction, lateral spreading
and land sliding. Ground shaking is the only earthquake hazard that structural design
provisions of building codes directly address. Ground shaking hazard is defined by means of a
![Page 6: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/6.jpg)
hazard curve, which indicates the probability that a measure of seismic intensity (e.g peak
ground acceleration or 1st mode spectral acceleration) will be exceeded over a certain period of
time. The three levels of seismic hazard considered can be defined as follows:
i) Occasional Earthquake Hazard level: with probability of exceedance 50% in 50 years
and interval of recurrence 72 years.
ii) Rare Earthquake Hazard level: with probability of exceedance 10% in 50 years and
interval of recurrence 475 years.
iii) Maximum Considered Event Earthquake Hazard level: with probability of exceedance
2% in 50 years and interval of recurrence 2475 years.
According to the Enhanced Objectives of FEMA-350 the following three performance
objectives are considered: (i) Operational level corresponding to Occasional Earthquake
Hazard level, (ii) Life Safety level in connection to Rare Earthquake Hazard level, (iii)
Collapse Prevention level correlated to the Maximum Considered Event Earthquake Hazard
level.
2.2. Structural analysis phase – Evaluation of structural capacity
According to FEMA-356 four alternative analytical procedures, based on linear and nonlinear
static and dynamic structural response, are recommended for the structural analysis of
buildings under earthquake loading. In this study the Nonlinear Static analysis Procedure
(NSP) of FEMA, also known as pushover analysis, is used to assess the structural capacity. In
order to assess the capacity of the structure, for every performance level considered, a lateral
load distribution that follows the fundamental mode is adopted. For the PBD procedure, prior to
any structural analysis, the column to beam strength ratio is calculated and is examined whether
the sections chosen are of class 1 , as Eurocode 3 (EC3) [20] suggests. Class 1 cross-sections
are those which can form a plastic hinge with the rotation capacity required from plastic
analysis without reduction of the resistance. The column to beam strength ratio is calculated as:
1 20M pl ,column
M pl ,beam.
∑
∑≥ (1)
where ΣMpl,column and ΣMpl,beam is the sum of the design values of plastic moment resistances of
the structural members at each joint.
6
The PBD procedure consists of the following steps: (i) All EC3 checks must be satisfied
for the gravity loads; (ii) if the checks of Step (i) are satisfied then NSP is performed in order to
explicitly calculate the demand for the defined intensity levels. The structural capacity is
![Page 7: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/7.jpg)
associated to the maximum interstorey drift values θ, and the acceptance criteria of Step (ii) are
confirmed if satisfied or not in order to accept or revise the design. NSP is terminated as soon as
a target displacement that corresponds to 1.5 times the 2% in 50 years (2/50) probability of
exceedance earthquake is achieved [7]. A flowchart for the PBD procedure employed in this
study is shown schematically in Figure 1.
For every tentative design the capacity is assessed at three performance levels using the
displacement coefficient method [6]. The structure is “pushed” with a triangular distribution of
the lateral loads until the target displacement is reached. The target displacement can be
obtained from the FEMA-356 formula:
2
0 1 2 3 24e
tTC C C C Sa gδπ
= (2)
where C0, C1, C2, C3, are modification factors. C0 relates the spectral displacement to the likely
building roof displacement. C1 relates the expected maximum inelastic displacements to the
displacements calculated for linear elastic response. C2 represents the effect of the hysteresis
shape on the maximum displacement response and C3 accounts for P-Δ effects. Te is the
effective fundamental period of the building in the direction under consideration. Sa is the
response spectrum acceleration corresponding to the Te period.
The applicability of NSP has limitations; it provides very good estimates of global as well
as local inelastic deformation demands in the case of structures that vibrate primarily in their
fundamental mode. NSP is capable to expose the design weaknesses that might remain hidden
when an elastic analysis is considered. Such weaknesses include story mechanisms, excessive
deformation demands, strength irregularities, and overloads on potentially brittle elements,
such as columns and connections. However, as higher mode contributions become more
significant, NSP overestimates the maximum displacements computed during the analysis
procedure. High-rise buildings or buildings with irregular distribution of strength and/or mass
are not suitable for the application of the NSP. On the contrary, low to mid-rise or regular
buildings vibrate primarily in the fundamental mode and their response is reasonably predicted
by NSP. This type of building is considered in the present study. Furthermore, to take into
account the effect of simultaneous ground shaking in two orthogonal directions, the
recommendation of FEMA-350 is employed, where multidirectional excitation effects are
accounted for by combining 100% of the response due to loading in the longitudinal direction
with 30% of the response due to loading in the transverse direction, and vice versa. The most
7
![Page 8: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/8.jpg)
severe maximum interstorey drift value obtained when both of these load combinations are
applied is used to obtain the seismic demand.
3. RELIABILITY-BASED STRUCTURAL DESIGN OPTIMIZATION
Reliability theory is introduced in structural engineering and structural optimization in order to
consider, in a more rational way, all existing uncertainties which are not known with the
desirable degree of precision and can influence structural response. Reliability is recognized as
a safety constraint in structural engineering, therefore any optimum design under reliability
constraints should balance both cost and safety. In the case of deterministic optimum design,
stress and displacement constraints are considered in accordance with the design code safety
factors, while in reliability based optimum design additional probabilistic constraints related to
the behaviour of the structure are considered.
The main goal of Reliability-Based design Optimization (RBO) is to perform a safe and
economic design with respect to extreme loading events. In RBO problems, probabilistic
constraints are incorporated into the optimization procedure leading to unbiased estimates of
the structural performance and, subsequently, to the determination of designs that are located
within a range of target failure probabilities. In the case of the incorporation of PBD approach
into the RBO formulation, the objective function is assumed to be the weight of the structure,
while the deterministic constraints correspond to serviceability and ultimate limit state checks
and the probabilistic constraints to the probabilities of violation of the ultimate limit state
checks. In particular, the probabilities of exceeding the limit states, which correspond to each
hazard level, are taken as the probabilistic constraints. The probabilistic constraints enforce the
condition that the probability violation is smaller than a certain threshold value.
The RBO problem can be formulated in the following form:
(3)
3
a
min ( , )subject to g ( , ) 0 (serviceability checks) g ( , ) 0 (ultimate limit state checks) P(g ( , ) 0) P (probabilistic checks)
where
IN x
EC x
PBD x
PBD
C∈
≤≤
> ≤
s s μs μs μ
s x
F
n
2
R N( , )∼ x x
∈sx μ σ
The vectors s, x and μr represent the design, the random variable and the mean value vectors,
respectively. F is the feasible region, where both the serviceability gEC3 and the ultimate limit
state constraint functions gPBD are satisfied, Pa is the allowable probability of violating the
8
![Page 9: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/9.jpg)
ultimate limit state constraints, while CIN corresponds to the objective function representing the
initial construction cost.
4. NEURAL NETWORKS
Artificial Neural Networks (NN) are biologically inspired, since they are composed by
elements that perform in a manner analogous to the elementary functions of a biological
neuron. These elements are known as artificial neurons. NN are organized in a way that is
related to the anatomy of the brain and they exhibit a surprising number of the brain’s
characteristics such as: learning from experience, generalizing from previous examples and
abstracting essential characteristics from sets of inputs containing irrelevant data. The use of
hidden layers and nonlinear activation functions enhance the ability of the NN to “learn” the
complicated relationship between a set of input and a set of output data.
A feed-forward NN attempts to create a desired mapping between the inputs and the
targets of a training set. The training set is composed by m input-target pairs D=[xm, tm]. A
Neural Network architecture A consists of a specific number of layers, a number of units in
each layer and a type of activation function. If a set of weight parameters w is assigned to the
connections of the network, a mapping y(xm; w, A) is defined between the input vector xm and
the output vector y. The quality of this mapping is measured using the following error function:
( 2
D1( | , ) = ( ; , )2
m
m−∑D w )my x w tE A A (4)
A learning algorithm tries to determine the weight parameters w in order to minimize the error
function ED. Iterative minimization algorithms are therefore used to obtain the optimum weight
parameters w. For the solution of the minimization problem the operator O is applied, resulting
to the following iterative formula:
(5) ( +1) ( ) ( ) ( ) = ( ) = +Δt t tw w wO tw
t
Most of the numerical minimization methods are based on the above expression, while to
initiate the algorithm a starting vector of weight parameters w(0) is necessary. The changing part
of the algorithm Δw(t) can be further decomposed in:
(6) ( ) ( )Δ att=w d
where d(t) specifies the direction of search and at is the corresponding step size.
Learning algorithms are classified to local or global algorithms. Global algorithms make
use of the knowledge about the state of the entire network, such as the direction of the overall 9
![Page 10: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/10.jpg)
weight update vector. In the widely used back-propagation global learning algorithm the
gradient descent algorithm is used. In contrast, local adaptation strategies are based on specific
information of the weight values such as the temporal behaviour of the partial derivative of the
weights. The local approach is better related to the NN concept of distributed processing, where
the computations are performed independently. Moreover, it appears that for many applications
local strategies achieve faster and more reliable predictions than global techniques. The
Resilient backpropagation learning algorithm, abbreviated as Rprop [21], is adopted in this
study. Rprop is a local algorithm, based on an adaptive version of the Manhattan-learning rule
that has been proved very efficient in the past [22].
5. SOLVING THE OPTIMIZATION PROBLEM
Evolutionary Algorithms (EA) are population based, probabilistic, direct search optimization
algorithms gleaned from principles of Darwinian evolution. Starting with an initial population
of μ candidate designs, an offspring population of λ designs is created from the parents using
variation operators. Depending on the manner in which the variation and selection operators are
designed and the spaces in which they act, different classes of EA have been proposed. In the
EA algorithm employed in this study, each member of the population is equipped with a set of
parameters:
10
c
ncRn
(7) γd
c σ
d c d
nnd
n n nc
[( , , ( , , )] (Ι ,Ι )
Ι =D R
Ι =R R [ π,π] a
+
+
= ) ∈
×
× × −
a s γ s σ α
where sd and sc are the vectors of discrete and continuous design variables defined in the
discrete and continuous design sets , respectively. Vectors γ, σ and α are the
distribution parameter vectors taking values in
n and dDn n, and [ , ] ,aR Rγ σ+ + −π π respectively. Vector γ
corresponds to the variances of the Poisson distribution. Vector corresponds to the
standard deviations (1 ≤ n
σnR +∈σ
σ ≤ nc) of the normal distribution. Vector is related to the
inclination angles (n
n[ π,π] a∈ −α
α = (nc-nσ/2)(nσ-1)) defining linearly correlated mutations of the continuous
design variables sc, where n = nd + nc is the total number of design variables.
Let P(t) = {a1,…,aμ} denotes a population of individuals at the t-th generation. The
genetic operators used in the EA method are denoted by the following mappings:
![Page 11: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/11.jpg)
(8)
μ λd c d c
λ λd c d c
k k μμ d c d c
rec : (Ι ,Ι ) (Ι ,Ι ) (recombination)
mut : (Ι ,Ι ) (Ι ,Ι ) (mutation)
sel : (Ι ,Ι ) (Ι ,Ι ) (selection, k {λ, μ+λ})
→
→
→ ∈
A single iteration of the EA, which is a step from the population to the next parent
population is modelled by the mapping
( )P tp
( )P t+1p
11
μ (9) μEA d c t d c t+1opt : (Ι ,Ι ) (Ι ,Ι )→
In Figure 2 a pseudo-code of the EA algorithm is depicted. At the beginning of the
procedure in generation t = 0 the initial parent population , composed by μ design vectors,
is generated randomly (step 3 of the pseudo-code). Steps 5 to 12 correspond to the main part of
the EA algorithm, where in every generation λ offspring vectors are generated by means of
recombination and mutation. D
( )P tp
l is a sub-population with two members selected from the parent
population of the current generation (Step 6) which is used by the recombination operator.
Recombination and mutation operators, described in steps 7 to 10, act on both design variable
vectors s
( )P tp
l and distribution parameter vectors (distribution parameter vectors are denoted as yl in
the pseudo-code). In step 11 the objective and constraint functions are calculated in order to
assess the design vectors in terms of the objective function value and feasibility.
6. SOLVING THE RELIABILITY ANALYSIS PROBLEM
The advancements in structural reliability theory during the last twenty years and the
attainment of more accurate quantification of the uncertainties associated with structural loads
and resistances have stimulated the interest in the probabilistic treatment of structural systems
[23]. The reliability of a structure or its probability of failure is an important factor in the design
procedure since it investigates the probability of the structure to successfully accomplish its
design requirements. Reliability analysis leads to safety measures that a design engineer has to
take into account due to the aforementioned uncertainties. Although from a theoretical point of
view the field has reached a stage where the developed methodologies are becoming
widespread, from a computational point of view serious obstacles have been encountered in
practical implementations. First and Second Order Reliability Methods (FORM and SORM),
that have been developed to perform structural reliability, although they lead to elegant
formulations, they require prior knowledge of the means and variances of the component
![Page 12: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/12.jpg)
random variables and the definition of a differentiable limit-state function. On the other hand
the Monte Carlo Simulation (MCS) method is not restricted by the form and the knowledge of
the limit-state function but is characterized by the high computational cost.
6.1. Monte Carlo simulation
The MCS method is applied in stochastic mechanics when an analytical expression of the
limit-state function is not attainable. This is mainly the case in problems of complex nature with
a large number of random variables, where all other stochastic analysis methods are not
applicable. In structural stochastic analysis problems, the probability of violation of the
behavioural constraints can be written as:
12
d (10) ( ) 0
( )viol xg
p f≥
= ∫x
x x
where fx(x) denotes the joint probability of violation for the random variables, the limit-state
function g(x)<0 defines the safe region and x is the vector of the m random variables.
Considering that MCS is based on the theory of large numbers (N∞) an unbiased estimator of
the probability of violation is given by:
1
1 ( )∞
=∞
= ∑N
viol jj
p IN
x (11)
where xj is the j-th vector of the random structural parameters, and I(xj) is an indicator for
successful and unsuccessful simulations defined as:
1 if ( ) 0
( )0 if ( ) 0
gI
g≥⎧
= ⎨ <⎩
jj
j
xx
x (12)
In order to estimate pviol an adequate number of Nsim independent random samples is produced
using a specific, uniform probability density function of the vector xj. The value of the violation
function is computed for each random sample xj and the Monte Carlo estimation of pviol is given
in terms of sample mean by:
H
sim
NNviolp ≅ (13)
where NH is the number of successful simulations and Nsim the total number of simulations.
The basic MCS is simple to use and has the capability of handling practically every
possible case regardless of its complexity. However, for typical structural reliability problems
the computational effort involved becomes excessive due to the enormous sample size
![Page 13: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/13.jpg)
13
required. To reduce the computational effort more elaborate simulation methods, called
variance reduction techniques, have been developed, their efficiency though is limited for
larger probability values. Moreover, despite the improvements achieved on the efficiency of
computational methods for treating structural analysis problems of large scale structures, they
still require disproportional computational effort for reliability analysis of realistic problems.
This is the reason why very few successful numerical investigations are known in estimating
the probability of failure and are mainly restricted to simple elastic frames and trusses [24].
In a previous work by the authors [17] two methodologies have been proposed where
Neural Networks have been incorporated into the RBO problem in order to reduce the
computational cost of MCS. Both methodologies take advantage of the global approximation
capabilities of the neural networks. The first one exploits the global approximation capabilities
over the design variables space while the second operates on the space of the random variables.
In the first methodology the NN is trained once outside the RBO phase, while in the second one
a new NN is trained for every new candidate design inside the RBO procedure. As it can be
seen from Figure 3a, three steps are required for the first implementation. The first step
corresponds to the construction of the training set, where N design vectors are generated using
the Latin Hypercube Sampling (LHS) technique [25] and their feasibility with respect to both
deterministic and probabilistic constraints is evaluated. The training phase of the NN is
performed in step 2. This phase consists of the selection of a suitable NN architecture and the
training/testing procedures. The trained NN is used to predict the feasibility of new design
vectors both in terms of deterministic and probabilistic constraint checks. The RBO procedure
is performed using Evolutionary Algorithms in the last step of the methodology where the
trained NN is used to assess the feasibility and the quality of every new candidate design
selected by the Evolutionary Algorithm.
In the second methodology, shown schematically in the flowchart of Figure 3b, for every
new candidate design inside the optimization procedure, a new NN is trained in order to replace
the analyses required during the MCS. Figure 3b describes the procedure for assessing the λ
offsprings of an optimization cycle (generation) of the Evolutionary Algorithm. As soon as the
NN is trained over a sample of M random vectors generated using LHS, the trained NN is
applied to predict the structural performance for every new vector of random variables
encountered during the MCS. The second methodology has been proved more efficient [17] in
terms of robustness and computational effort and for this reason it is implemented in the current
work. Special care has been taken in this implementation to alleviate this extrapolation
![Page 14: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/14.jpg)
drawback of NN by generating the training sample using the Latin Hypercube Sampling
method in the range of μi ± 6 σi for, where μi and σi are the mean value and standard deviation of
the i-th random variable.
6.2. First-Order Reliability method
In the general case of a nonlinear limit state function the main objective of the First Order
Reliability Method (FORM) is to calculate the reliability index β. The Hasofer-Lind reliability
index β [26] is calculated by a process of minimization, and the probability of violation is
approximated by:
(14) violp = Φ(-β)
where Φ is the standard normal cumulative distribution function. This equation is exact when
the failure criterion is linear and all random variables have normal distributions. Given a vector
of basic variables x, a failure surface ∂ω on which the failure criterion g(x)=0 is satisfied and a
safe region denoted by g(x)>0, the vector of the reduced variables z is defined as follows:
(15) -1x ( - )≡ ⋅z S x μx
where Sx, is a diagonal matrix of the standard deviations and μx is the vector of mean values.
Then the Hasofer-Lind reliability index β is defined as:
T
ωβ min
∈∂≡
zz z (16)
The point on the failure surface g(x)=0, where its transformation to the z space satisfies
equation (16), is called design or most probable point and will be denoted as zD. The design
point zD is located on the limit-state surface, g(x)=0 and has minimum distance from the origin
in the standard normal space. For applying either first or second-order methods to complex
structural models it is necessary to have an explicit expression or an approximation of either the
entire limit-state function g(x) or of its limit-state surface g(x)=0 in the space of the random
variables x. This is because these methods require not only knowledge of the function but also
of its gradient in the vicinity of its limit-state surface. In the case of unknown expression, the
limit-state function is usually approximated by the Response Surface method.
6.2.1. The Response Surface (RS) method
The RS method was originally proposed by Box [27] as a statistical tool, to find the operating
conditions of a chemical process at which some response was optimized. In order to reduce the
computational effort while maintaining an acceptable accuracy, two important issues should be
14
![Page 15: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/15.jpg)
considered when applying the RS method to the failure probability: (i) The definition of
experimental points for defining the approximation of the limit-state function, (ii) The
analytical expression of the Response Surface function [15]. Usually, a quadratic RS function is
assumed:
m m
2i i i i
i=1 i=1g( ) a + b x + c x= ∑ ∑x (17)
defined in an m-dimensional random variable space where the constants a, bi and ci are
determined by evaluating g(x) at certain specified experimental points, while xi, i=1, …,m are
the random variables. Higher order functions are not generally used for conceptual as well as
for computational reasons.
Bucher and Bourgund [28] proposed an interpolation scheme for the solution of structural
reliability problems, where the quadratic RS function of Eq. (17) is defined with 2m+1
experimental points. It was suggested the experimental points xi, to be taken as xi = μi ± f σi,
where μi and σi are the mean value and standard deviation of the i-th random variable and f is an
arbitrary factor taken equal to 3. Figures 4a and 4b depict the sampling method for defining the
2m+1 experimental points where xi,low = μi - 6 σi and xi,up = μi + 6 σi. In order to avoid the
undesirable case of Figure 4a, where unrealistic values of the random variables are generated
(i.e. a negative value of the modulus of elasticity), a correction of the experimental points is
performed by moving all trial points to an acceptable region as shown in Figure 4b. It was later
shown [15,29] that using a constant value of f = 3 may not always yield good solutions,
particularly when g(x) is highly nonlinear. A better fit of the RS function was obtained by
updating the centre point:
i,D ii,M i
D
(x μ ) (x = μ
( - (− )
+) )
gg g
μμ x
(18)
which helps to locate the new centre point closer to the actual g(x)=0, and by decreasing the
value of the parameter f in subsequent cycles of updating. Then the interpolation scheme of Eq.
(17) is repeated using xM as a new center point. A number of modifications for better estimating
the RS through the generation of the experimental points have been proposed [30-38] in order
to alleviate the shortcomings of the method.
6.2.2. Neural network approximation of the limit-state function
In the previously mentioned implementations of the RS method, it was found that the
performance of the method is significantly influence by the polynomial representation and the
15
![Page 16: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/16.jpg)
16
way the experimental points are generated. In an effort to improve the robustness of the
approximation procedure the polynomial representation of RS is replaced by an approximation
produced by neural networks. In this study local approximation capabilities of the neural
networks in the vicinity of the centre point xM are implemented with the intention to reduce the
number of the training patterns required by the global NN approximator described in Section
6.1.
Hurtado [16] confirmed with statistical learning concepts the sensitivity of the RS
approximation via polynomial expressions compared to the neural network approximations.
The global capabilities of the neural network approximator were examined [39-42] in the
framework of FORM for simple analytic limit-state functions or academic structural reliability
problems. The local approximation capabilities of the neural networks in conjunction with
reliability analysis methods are examined in a simple portal frame by Deng et al. [43].
In this work a novel neural network based methodology, denoted as FORM-NN, is
proposed for the solution of reliability analysis problems in conjunction with structural
optimization considering performance objectives. In the proposed methodology neural
networks are implemented as a local approximator in the vicinity of the centre point xM in a
similar way that the RS based FORM is used. More specifically a neural network
approximation of the limit-state function replaces the polynomial expression of RS. The steps
of the proposed FORM-NN methodology are given in Figure 5. In the first step of the proposed
implementation a number of experimental points are generated through the LHS technique.
Figures 6a and 6b demonstrate schematically the LHS technique with a correction procedure.
This sampling method can generate a variable number of samples well distributed over the
entire range of interest. On the other hand, the number of samples generated with the standard
sampling procedure [28] (Figure 4) is fixed to 2m+1 experimental points. According to LHS
the distribution function of each random variable is divided into a number of stratums of equal
marginal probability. These stratums are randomly selected for each random variable and
randomly shuffled among different variables to define the experimental points. The main
advantage of the LHS technique is that it produces well distributed samples of almost any
sample size of experimental points. This characteristic feature of LHS is demonstrated in the
framework of FORM-NN in the numerical studies where it is shown that a reduced number of
experimental points are required for adequately defining the NN approximation. It has to be
emphasized that FORM-NN does not suffer from the extrapolation drawback since any point
outside the range of interest is corrected as it demonstrated in Figures 6a and 6b.
![Page 17: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/17.jpg)
In the next step the experimental points are used to define the NN approximation of the
limit state function according to the following equation:
k m
H,ij O,ij ki=1 i=1
g( ) = w × w × x⎛ ⎞⎛⎜ ⎜
⎝ ⎠⎝ ⎠∑ ∑x f f ⎞
⎟⎟ (19)
where k is the number of the hidden layer neurons, wH and wO are the weight parameters of the
hidden and output layers, respectively, and f(x) is the sigmoid transfer function:
1(y) =1+ exp(-γy)
f (20)
The coefficient γ defines the slope of the sigmoid function. By the time the NN is trained and
the weight parameters are defined, FORM is performed in order to define the reliability index β
and the corresponding design point xD. Hereafter, the new centre point xM is defined according
to Eq. (18). If there is no convergence with respect to the reliability index β and the
corresponding design point xD, new experimental points are generated with LHS for the new
centre point xM, in order to define the new approximation of limit-state function (Eq. (19)). The
advantages of the proposed NN approximation scheme are demonstrated in the numerical test
section, where it is shown, that contrary to the quadratic polynomial approximation of RS, this
methodology is not affected by the number of the interpolation points and the area from which
the points are generated as defined by the coefficient f.
7. NUMERICAL TESTS
The seven-storey 3D steel building, shown in Figure 7, has been considered in order to perform
both deterministic and reliability-based design optimization considering two types of design
variables: size of member cross-sections and the topology of the vertical structural elements.
All columns are made of X-shapes obtained by welding two HEB profices along the
longitudinal axis. This solution avoids the engagement of the column web in moment resistant
connections as well as the development of high stress connections along the beam flanges
welds which occur in the case of box-columns, due to the transmission of the beam web stresses
through the out of plane bending of the column wall plates [44]. The frame members are
modelled with the force-based fiber beam-column element [45], where four integration points
are considered for each structural element. The initial FE model of the 3D frame consists of 9×6
columns, 1,029 beam-column finite elements with 2,268 degrees of freedom and a slab
thickness of 10 cm. The mean value of the modulus of elasticity is 210 GPa and of the yield
17
![Page 18: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/18.jpg)
stress is fy=235 MPa (S235). The slab distributed loading is G=5.4kPa corresponding to the
permanent action and Q=5 kPa for the live load [46].
The columns and the beams of the structure have double I-shaped cross-sections, and are
grouped into five sets, each corresponding to an independent design variable. The columns on
the corners correspond to the 1st group, the intermediate columns of the perimeter of the plan in
the x-direction belong to the 2nd group, while the intermediate columns in the y-direction of the
plan are grouped in the 3rd set. All remaining columns belong to the 4th group, while the beams
constitute the 5th group. In the case of the combined sizing and topology optimization, two
additional design variables are considered corresponding to the number of the columns along
the two structural dimensions. All optimization runs were performed using the (μ+λ) EA
optimization scheme with μ=10 parents and λ=10 offsprings.
7.1. Random variables
Two types of random variables are considered: randomness of the ground motion excitation
which influences the level of seismic demand, and randomness of the material properties which
affects the structural capacity. The structural stiffness is directly connected to the modulus of
elasticity E, while the strength is influenced by the yield stress fy. For each group of structural
elements three random variables are considered; the modulus of elasticity E, the yield stress fy
and the hardening parameter b of the stress-strain curve. Overall fifteen random variables
related to the structural capacity are considered.
The most common approach to determine seismic actions is the use of a response
spectrum defined by the design code. This approach is general and easy to implement.
However, a more realistic representation of seismic loading can be obtained if a number of
response spectra derived from natural records are used instead. Furthermore, since significant
dispersion on the structural response may arise when different natural records are used, it is
advisable to apply properly scaled seismic records. In this study, the scaling procedure adopted
is based on the peak ground acceleration (PGA) of the seismic records. Three sets of natural
records with both longitudinal and transversal components are used. The records have been
selected from the database of Somerville and Collins [47] and correspond to three hazard
levels, with 50%, 10% and 2% probability of exceedance in 50 years. The records are scaled to
PGA values obtained from hazard curves produced for a specific region. In this work the hazard
curves derived by Papazachos et al. [48] (Table 1) are implemented. Moreover, based on the
assumption that seismic data follow the lognormal distribution [49], the median spectrum x
18
![Page 19: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/19.jpg)
and the standard deviation δ of the spectral acceleration can be calculated using the following
expressions:
n
d,ii 1ln( ( ))
x expR T
n=
⎡ ⎤⎢ ⎥=⎢ ⎥⎣ ⎦
∑ (21)
( )
1 22nd,ii 1
ˆln( ( )) ln(x)1
R Tδ
n=
⎡ ⎤−⎢ ⎥=
−⎢ ⎥⎣ ⎦
∑ (22)
where Rd,i(T) is the spectral acceleration value of the i-th record (i=1,…,n) for a period equal to
T, n is the number of records which depend on the hazard level. In this work, n is taken equal to
22 for the 50% in 50 years hazard level, n = 19 for the 10% in 50 years and n = 6 for the 2% in
50 years [48]. Since three hazard levels are considered in this study, three random variables are
assigned to the problem related to the seismic demand. Figure 8 depicts the two median spectra
for the three hazard levels taken into consideration. The properties, probability density
function, mean value and variance, of each random parameter are given in Table 2.
7.2. Deterministic optimization
In the first part of the numerical studies two deterministic formulations of the optimization
problem are examined, depending on the type of the design variables. The Deterministic Sizing
Optimization (DSO) and the Deterministic combined Sizing-Topology Optimization (DSTO).
The feasibility of every design tested during the optimization procedure is assessed in the
serviceability and ultimate limit states, while three hazard levels are employed for performing
the ultimate limit state checks. According to these checks the maximum inter-storey drift of the
structure for each hazard level should be less than a permissible drift value in agreement with
HAZUS [50]. Both DSO and DSTO problems can be expressed in the following form:
19
![Page 20: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/20.jpg)
s
EC8
50/50
10/50
2/50
min ( )subject to
g ( ) 0θ ( ) (2/3)1.2%θ ( ) (2/3)3.0%θ ( ) (2/3)8.0%
where
INC∈
≤
≤≤
≤
s
ssss
F
dis R , i = 1,..., n ∈
(23)
where CIN(s) is the objective function to be minimized (the volume of the structure), s is the
design variables vector that can take values from a discrete design set Rd, gEC8(s) are the
serviceability constraints, θt/50(s) correspond to the maximum inter-storey drift for the t/50
hazard level, while F is the feasible set. According to HAZUS [50] for a low-rise steel building
the drift limits for the moderate, extensive and complete damage states are defined to 1.2%,
3.0% and 8.0%, respectively. The drift limits for the mid-rise building in question are taken
equal to 2/3 times of the corresponding low-rise ones [50].
The design variables for the DSO formulation are the dimensions of the cross sections of
all structural elements, whereas for the DSTO formulation two additional design variables are
taken into consideration: the number of columns in the horizontal x and y directions of the
structure. The deterministic behavioural constraints, under which the objective function is
minimized, as well as the design variables fulfil the demands of feasibility which are expressed
in terms of displacements. The optimum designs obtained through the two deterministic
formulations are shown in Table 3 where the cross sections and the number of columns are
depicted. One significant observation is the fact that the calculated volume for the DSTO
optimum design is almost half of that obtained with the DSO formulation with fixed topology
of the columns.
7.3. Verification
The proposed NN based methodologies are assessed with a parametric study performed on the
optimum design obtained through the DSTO formulation. Table 4 demonstrates the violation
probabilities with the corresponding reliability indices of the three performance objectives
estimated after 106 Monte Carlo simulations. Table 5 shows the performance of the polynomial
approximation of RS with respect to the value of the parameter f required for the definition of
the 2m+1 experimental points. Moreover, two different methodologies, for defining the
20
![Page 21: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/21.jpg)
21
experimental points, are examined. The standard sampling, where the experimental points are
generated according to the scheme described in Figure 4, and the sampling based on Latin
Hypercube, where the range xi,M ± f σi for each of the m random variables is divided into 2m
non-overlapping segments of equal marginal probability. The results of Table 5 confirm the
observations of other researchers that, the probabilities of violation of the three performance
objectives vary with reference to the value of the coefficient f, irrespective of the procedure
used for generating the sample of experimental points. Worth mentioning is also the fact that
the probability of violation of the performance objective, corresponding to the 50/50 hazard
level, calculated with the FORM (Table 5), is orders of magnitude less compared to the
corresponding one calculated with the MCS (Table 4).
The performance of the two neural network based approximation procedures described in
Figures 3b and 5 is depicted in Tables 6 to 8. The MCS-NN procedure under different number
of training patterns and simulations is shown in Table 6. The neural network employed is the
18×40×3 fully connected network, where the 18 input nodes correspond to the 15 random
variables related to the capacity of the structure plus 3 related to the seismic demand of the
three hazard levels. The three nodes of the output layer correspond to the structural response in
terms of interstorey drift for the three performance objectives of the corresponding hazard
level. Since each simulation run of the NN approximation is computationally inexpensive, two
estimations of the violation probabilities are computed after 106 and 1010 simulations. It can be
observed that the probabilities (and the corresponding reliability indices) estimated after 106
simulations are very close of those given in Table 4 with the conventional MCS which confirms
the global approximation capabilities of NN. The violation probability of 50/50 hazard level
estimated after106 MC simulations is one order of magnitude different compared to the
estimated value after 1010 simulations. As can be seen, 500 training patterns are adequate for an
acceptable approximation of the violation probability by the NN scheme. This size of the
training set is used in the subsequent numerical studies.
Tables 7 and 8 depict the performance of the FORM-NN. Table 7 shows the performance
of FORM-NN with respect to the value of the coefficient f used to define the training set. The
neural network employed is the 16×30×1 fully connected network where the 16 input nodes
correspond to 15 random variables related to the capacity of the structure, plus one node related
to the seismic demand in the corresponding hazard level. The single node of the output layer
corresponds to the performance objective of the structural response in terms of interstorey
drifts. It can be seen that, contrary to the sensitivity observed in the computation of the
![Page 22: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/22.jpg)
22
statistical quantities with respect to the values of f for the conventional FORM, the FORM-NN
demonstrates a relative invariance in the computations of the violation probabilities for f>0.5.
The sensitivity in the computation of the statistical quantities of FORM-NN is examined in
Table 8 with respect to the size of the training set for f=1.0. It is verified that, acceptable
approximations are observed with a sample size greater or equal to m. In the next part of the
numerical studies, m experimental points and f=1.0 are used for the FORM-NN methodology.
For every hazard level, m=16 training samples times the iterations of the FORM-NN procedure
as indicated in Figure 5 are required for training the local approximator. In this case 2 to 3
iterations are adequate for the convergence of the FORM-NN. Thus 16×3(iterations)×3(hazard
levels)=144 training patterns are required compared to the 500 training patterns required for the
MCS-NN methodology.
7.4. Reliability-Based design Optimization
Reliability considerations are recognized as safety constraints in structural engineering and an
optimum design considering these type of constraints should balance cost and safety. In the
case of deterministic optimum design, stress and displacement constraints are taking into
consideration in accordance with the design code safety factors without introducing reliability
as explicit design constraints. In the case of the reliability based design optimization additional
probabilistic constraints related to the strength of the structure are considered. The main goal of
RBO methods is to design for safety and economy with respect to extreme events. The
probabilistic constraints, which are incorporated into the optimization procedure, lead to
unbiased estimates of the structural performance and subsequently, to the determination of
design points that are located within a range of target failure probabilities. The RBO problem
for the examined test case can be formulated, in the framework of PBD, in the following form:
![Page 23: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/23.jpg)
(24)
s
EC8
50/50
10/50
2/50
min ( , )subject to g ( , ) 0 θ ( , ) (2/3)1.2% θ ( , ) (2/3)3.0% θ ( , ) (2/3)8.0%and
IN x
x
x
x
x
C∈
≤
≤
≤
≤
F s μ
s μs μs μs μ
50/50
10/50
2/50
di
P(θ ( , ) > (2/3)1.2%) 0.1% P(θ ( , ) > (2/3)3.0%) 0.05% P(θ ( , ) > (2/3)8.0%) 0.001%where s R , i = 1,..., n
≤
≤
≤
∈
s xs xs x
2x x N( , )∼x μ σ
where CIN(s,μx) is the objective function to be minimised, s is the vector of the design variables,
x is the vector of random variables, P(θt/50(s,x)>θallowable) is the probability of violation of the
constraints which should be less than a specified allowable probability. The probabilistic
constraint enforces the condition that the probability of violating the performance objectives is
less than a certain value. Similarly to the deterministic formulation, two different types of
probabilistic formulation of the optimization problem are examined depending on the type of
the design variables; the Reliability based Sizing Optimization (RSO) and the Reliability based
combined Sizing-Topology Optimization (RSTO).
The optimum designs obtained according to the two probabilistic formulations are shown
in Table 9 where the cross sections and the number of columns are provided. Similarly to the
observation for the two deterministic formulations, the total optimized volume for the RSTO
optimum design is almost half of that obtained with the RSO formulation with fixed topology
of the columns and similar reliability indices. The comparison of the deterministic and
probabilistic optimum designs, with respect to the volume, violation probabilities and
computational cost, is shown in Tables 10 and 11. As can be seen, by examining the violation
probabilities corresponding to the two DBO optima, none of the two DBO formulations (DSO
and DSTO) leads to an optimum that fulfils the reliability constraints. Worth noticing is also the
fact that the RBO designs require 25 to 35 percent more material than the corresponding DBO
and needs three orders of magnitude more computational effort.
In the implementation of the RSO and RSTO formulations both methodologies proposed
for the NN approximation are considered. For the conventional MCS method with 50,000
23
![Page 24: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/24.jpg)
24
samples generated with the LHS method, 1.5 to 2.5 years of computation will be required on a
single processor computing environment. On the other hand, the computational effort is
reduced by two orders of magnitude when NN approximation is considered. This improvement
will be further enhanced when the required number of samples by the conventional MCS is
larger as the allowable violation probabilities become smaller. Moreover, the FORM-NN
methodology, where neural networks are implemented as an approximator of the limit-state
surface in the random variables space, outperforms the MCS-NN where the global
approximation capabilities are required, since it needs almost half the corresponding
computational time.
8. CONCLUSIONS
In most cases the optimum design of structures is based on deterministic formulations that are
focused on the satisfaction of the associated deterministic constraints. In order to find a realistic
optimum design it is required to take into account all necessary random parameters that might
affect the performance of the structural system and incorporate reliability constraints into the
formulation of the optimization problem. Despite the improvements achieved on the methods
dealing with reliability analysis problems the solution of realistic Reliability-Based
Optimization problems in structural mechanics remains an extremely computationally
intensive task. The computational cost is further increased when non-linear structural response
under seismic loading conditions is considered for assessing the behaviour of the tentative
designs encountered during the optimization procedure. In this work the main goal was to
solve, in a computational efficient manner, a performance based design optimization problem
under seismic loading considering probabilistic constraints.
Two distinctive methodologies, which are based on a Neural Network approximator, are
proposed. In the first one the Neural Networks are implemented as a global approximator and
the NN predictions are incorporated into the Monte Carlo simulation method, while in the
second one, the NN are used as a local approximator of the limit-state surface and are
implemented into the First Order Reliability Method framework. Both methods outperform the
conventional implementation of the Monte Carlo simulation method, combined with the Latin
Hypercube sampling, by two or more orders of magnitude in computation time depending on
the required number of MCS samples. Moreover it was found that the implementation of the
Neural Networks as a local approximator seems to be computationally more efficient than its
global counterpart requiring half the corresponding computational time. This is due to the fact
![Page 25: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/25.jpg)
25
that much more training patterns are required for training a Neural Network to perform as a
global approximator over the random variable space, than those required when the Neural
Network is used as a local approximator.
Apart from these two implementations of the Neural Network approximator the quadratic
Response Surface approximator was also examined and was found to be influenced by the
value of the coefficient f used for defining the experimental points of the Response Surface.
This variation of the Response Surface was stabilised with the Neural Network implementation
where the FORM-NN demonstrated a relative invariance in the computation of the violation
probabilities with respect to the coefficient f.
Furthermore, two formulations were examined with reference to the type of the design
variables. It was found that a combination of sizing and topology optimization is considered
has a profound effect on the material volume required compared to the optimum designs
achieved with sizing design variables only.
ACKNOWLEDGEMENTS
The second author acknowledges the financial support of the John Argyris Foundation.
REFERENCES
[1] Park G-J, Lee T-H, Lee KH, Hwang K-H, Robust design: An overview. AIAA Journal, 44 (1): 181-191, (2006).
[2] Beyer H-G, Sendhoff B. Robust optimization - A comprehensive survey. Computer Methods in Applied Mechanics and Engineering, 196(33-34): 3190-3218, (2007).
[3] Frangopol DM, Maute K. Life-cycle reliability-based optimization of civil and aerospace structures. Computers & Structures; 81(7): 397-410, (2003).
[4] Missoum S, Ramu P, Haftka RT. A convex hull approach for the reliability-based design optimization of nonlinear transient dynamic problems. Computer Methods in Applied Mechanics and Engineering; 196(29-30): 2895-2906, (2007).
[5] Tsompanakis Y, Lagaros ND, Papadrakakis M. (Eds.), Structural optimization considering uncertainties, Taylor & Francis (December 2007), ISBN: 9780415452601.
[6] FEMA-350: Recommended Seismic Design Criteria for New Steel Moment-Frame Buildings. Federal Emergency Management Agency, Washington DC, 2000.
[7] FEMA-356: Prestandard and commentary for the seismic rehabilitation of buildings. Federal Emergency Management Agency, Washington DC, SAC Joint Venture, 2000.
[8] Cornell CA, Jalayer F, Hamburger RO, Foutch DA. Probabilistic basis for 2000 SAC Federal Emergency Management Agency steel moment frame guidelines, J. Structural Engineering; 129(4): 526-533, (2002).
[9] Beck JL, Chan E, Irfanoglu A, Papadimitriou C. Multi-criteria optimal structural design under uncertainty. Earthquake Engineering and Structural Dynamics, 28: 741-761, (1999).
[10] Wen YK. Minimum lifecycle cost design under multiple hazards, Reliability Engineering and System Safety; 73(3); 223-231, (2001).
![Page 26: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/26.jpg)
26
[11] Lagaros ND, Fragiadakis M. Robust performance based design optimization of steel moment resisting frames, J. Earthquake Engineering; 11(5): 752-772, (2007).
[12] Foley CM, Pezeshk S, Alimoradi A. Probabilistic performance-based optimal design of steel moment-resisting frames. I: Formulation. Journal of Structural Engineering; 133(6): 757-766, (2007).
[13] Koutsourelakis PS, Pradlwarter HJ, Schueller GI. Reliability of structures in high dimensions, part I: Algorithms and applications. Probabilistic Engineering Mechanics; 19(4): 409-417, (2004).
[14] Au S-K, Beck JL. Estimation of small failure probabilities in high dimensions by subset simulation. Probabilistic Engineering Mechanics; 16(4): 263-277, (2001).
[15] Rajashekhar MR, Ellingwood BR. A new look at the response surface approach for reliability analysis. Structural Safety; 12(3): 205-220, (1993).
[16] Hurtado JE. An examination of methods for approximating implicit limit state functions from the viewpoint of statistical learning theory. Structural Safety; 26(3): 271-293, (2004).
[17] Papadrakakis M, Lagaros ND. Reliability-based structural optimization using neural networks and Monte Carlo simulation. Computer Methods in Applied Mechanics and Engineering; 191(32): 3491-3507, (2002).
[18] Hurtado JE. Neural network in stochastic mechanics, Archives of Computational Methods in Engineering (State of the art reviews); 8(3), 303-342, (2001).
[19] Eurocode 8. Design provisions for earthquake resistance of structures. ENV1998, CEN European Committee for standardization, Brussels, 1996.
[20] Eurocode 3, Design of steel structures. Part1.1: General rules for buildings. CEN- ENV, 1993. [21] Riedmiller M, Braun H. A direct adaptive method for faster back-propagation learning: The
RPROP algorithm, in Ruspini H. (Ed.), Proc. of the IEEE International Conference on Neural Networks (ICNN), San Francisco, USA, 586-591, (1993).
[22] Lagaros ND, Papadrakakis M. Improving the condition of the Jacobian in neural network training, Advances in Engineering Software, 35(1): 9-25, (2004).
[23] Schuëller GI. Developments in stochastic structural mechanics. Archive of Applied Mechanics; 75(10-12): 755-773, (2006).
[24] Jiang C, Han X, Liu GR. Optimization of structures with uncertain constraints based on convex model and satisfaction degree of interval, Computer Methods in Applied Mechanics and Engineering; 196(49-52): 4791-4800 (2007).
[25] Olsson A. Sandberg G, Dahlblom O. On Latin hypercube sampling for structural reliability analysis. Structural Safety; 25(1): 47-68, (2003).
[26] Der Kiureghian A. First- and second-order reliability methods. Chapter 14 in Engineering design reliability handbook, Nikolaidis E, Ghiocel DM, Singhal S. (Eds.), CRC Press, Boca Raton, FL, (2005).
[27] Box GEP. The exploration and exploitation of response surfaces: some general considerations and examples. Biometrics; 10: 16–60, (1954).
[28] Bucher CG, Bourgund U. A fast and efficient response surface approach for structural reliability problems. Structural Safety; 7(1): 57-66, (1990).
[29] Guan XL, Melchers RE. Effect of response surface parameter variation on structural reliability estimates. Structural Safety; 23(4): 429-444, (2001).
[30] Liu YW, Moses F. A sequential response surface method and its application in the reliability analysis of aircraft structural systems. Structural Safety; 16(1): 39-46, (1994).
[31] Kim S-H, Na S-W. Response surface method using vector projected sampling points. Structural Safety; 19(1): 3-19, (1997).
[32] Kaymaz I, McMahon CA. A response surface method based on weighted regression for structural reliability analysis. Probabilistic Engineering Mechanics; 20(1): 11-17, (2005).
[33] Breitung K, Faravelli L. Response surface methods and asymptotic approximations, in: Casciati F, Roberts JB (Eds.), Mathematical Models for Structural Reliability, ch. 5, CRC. Press, Boca Raton, FL, pp. 227-286, (1996).
![Page 27: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/27.jpg)
27
[34] Der Kiureghian A, Dakessian T. Multiple design points in first and second-order reliability. Structural Safety; 20(1): 37-49, (1998).
[35] Rackwitz R, Fiessler B. Structural reliability under combined random load sequences. Computers & Structures; 9: 489-494, (1978).
[36] Mahadevan S, Shi P. Multiple linearization method for nonlinear reliability analysis. J Engineering Mechanics; 127(11): 1165-1173, (2001).
[37] Gupta S, Manohar CS. An improved response surface method for the determination of failure probability and importance measures. Structural Safety; 26(2): 123-139, (2004).
[38] Lu Z-Z, Zhao J, Yue Z-F. Advanced response surface method for mechanical reliability analysis. Applied Mathematics and Mechanics; 28(1): 19-26, (2007).
[39] Gui J, Sun H, Kang H. Structural reliability analysis via global response surface method of BP neural network. Advances in Neural Networks-ISNN 2004: 799-804.
[40] Gomes HM, Awruch AM. Reliability analysis of concrete structures with neural networks and response surfaces. Engineering Computations; 22(1), 110-128, (2005).
[41] Hosni Elhewy A, Mesbahi E, Pu Y. Reliability analysis of structures using neural network method. Probabilistic Engineering Mechanics; 21(1), 44-53, (2006).
[42] Schueremans L, Van Gemert D. Benefit of splines and neural networks in simulation based structural reliability analysis. Structural Safety; 27 (3): 246-261, (2005).
[43] Deng J, Gu D, Li X, Yue ZQ. Structural reliability analysis for implicit performance functions using artificial neural network. Structural Safety; 27(1): 25-48, (2005).
[44] Mazzolani FM, Piluso V, A simple approach for evaluating performance levels of moment-resisting steel frames. In: Fajfar P, Krawinkler H (Eds), Seismic design methodologies for the next generation of codes, 241-252, Balkema, Rotterdam (1997).
[45] Papaioannou I, Fragiadakis M. Papadrakakis M. Inelastic Analysis of framed structures using the fiber approach, Proceedings of the 5th International Congress on Computational Mechanics (GRACM 05), Limassol, Cyprus, 29 June-1 July, (2005).
[46] Eurocode 1: Actions on structures - Part 1-1: General actions – Densities, self-weight, imposed loads for buildings, CEN, prEN 1991-1-1:2001, July 2001.
[47] Somerville P, Collins N. Ground motion time histories for the Humboldt bay bridge, Pasadena, CA, URS Corporation, 2002, www.peertestbeds.net/humboldt.htm.
[48] Papazachos BC, Papaioannou ChA, Theodulidis NP. Regionalization of seismic hazard in Greece based on seismic sources, Natural Hazards, 8(1): 1-18, (1993).
[49] Shome N, Cornell CA. Probabilistic seismic demand analysis of nonlinear structures, Rep. No. RMS-35, Dept. of Civil Engineering, Stanford Univ., Stanford, California, 320, 1999.
[50] FEMA-National Institute of Building Sciences. HAZUS-MH MR1, Multi-hazard Loss Estimation Methodology Earthquake Model, Washington, DC, 2003.
![Page 28: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/28.jpg)
TABLES
Table 1: Seismic hazard levels [48]
Event Recurrence Interval Probability of Exceedance PGA (g)
Frequent 21 years 90% in 50 years 0.06 Occasional 72 years 50% in 50 years 0.11
Rare 475 years 10% in 50 years 0.31 Very Rare 2475 years 2% in 50 years 0.78
Table 2: Properties of the random variables
Random variable Probability density function Mean value Standard
deviation Es (MPa) (5 random variables) N 2.10 E+05 0.10E fy (MPa) (5 random variables) N 235 0.10fy
Hardening (5 random variables) N 1.0% 0.10% Seismic Load (3 random variables) Log-N x (Εq. 21) δ (Εq. 22)
Table 3: Comparison of optimum designs for the Deterministic Sizing Optimization (DSO) and the
combined Sizing-Topology Optimization (DSTO)
Design DSO DSTO
Columns 1 HEB360 HEB500 Columns 2 HEB360 HEB400 Columns 3 HEB400 HEB550 Columns 4 HEB400 HEB400 Beams 5 IPE200 IPE240 No columns x 9 6 No columns y 6 4
Volume (m3) 38.32 22.86
28
![Page 29: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/29.jpg)
29
Table 4: Violation probabilities and reliability indices estimated with MCS after 106 simulations
Hazard level Reliability Index (β) Pviol(%)
50/50 4.76 1.00E-04 10/50 0.47 3.20E+01 2/50 0.10 4.60E+01
Table 5: Violation probabilities and reliability indices β (in parentheses) estimated with FORM-RS. The
quadratic form of RS is implemented with two sampling procedures for generating the experimental
points and different values of f
Hazard level f=0.5 f=1.0 f=2.0 f=3.0
Standard sampling
50/50 1.47E-08 (6.301) 2.08E-08 (6.248) 8.46E-19 (9.280) 1.01E-08 (6.360) 10/50 2.93E+01 (0.546) 2.85E+01 (0.567) 3.04E+01 (0.512) 3.05E+01 (0.511) 2/50 4.51E+01 (0.122) 4.49E+01 (0.128) 4.45E+01 (0.138) 4.45E+01 (0.139)
Latin Hypercube sampling
50/50 4.59E-12 (7.452) 2.79E-08 (6.202) 1.25E-12 (7.622) 6.84E-05 (4.830) 10/50 2.99E+01 (0.528) 2.87E+01 (0.563) 3.13E+01 (0.488) 3.16E+01 (0.479) 2/50 4.42E+01 (0.145) 4.45E+01 (0.137) 4.60E+01 (0.100) 4.52E+01 (0.121)
![Page 30: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/30.jpg)
30
Table 6: Violation probabilities and reliability index β (in parentheses) estimated with MCS-NN for
different number of training patterns and MC simulations
Number of training patterns Hazard level 100 200 500 1000
Estimated for 106 simulations
50/50 4.61E-05 (4.908) 8.43E-05 (4.788) 9.41E-05 (4.766) 8.75E-05 (4.780) 10/50 1.67E+01 (0.965) 2.60E+01 (0.643) 3.07E+01 (0.506) 3.15E+01 (0.480) 2/50 3.63E+01 (0.351) 4.43E+01 (0.144) 4.58E+01 (0.105) 4.54E+01 (0.116)
Estimated for 1010 simulations
50/50 3.44E-06 (5.394) 1.73E-05 (5.096) 9.90E-06 (5.201) 1.09E-05 (5.183) 10/50 1.69E+01 (0.959) 2.88E+01 (0.560) 3.13E+01 (0.478) 3.19E+01 (0.472) 2/50 3.63E+01 (0.347) 4.48E+01 (0.134) 4.59E+01 (0.103) 4.56E+01 (0.107)
Table 7: Violation probabilities and reliability indices β (in parentheses) estimated with FORM-NN, for
2m experimental points and different values of f
Hazard level f=0.5 f=1.0 f=2.0 f=3.0
50/50 1.25E-09 (6.674) 1.21E-05 (5.164) 1.20E-05 (5.165) 1.09E-05 (5.184) 10/50 3.18E+01 (0.474) 3.14E+01 (0.483) 3.11E+01 (0.493) 3.13E+01 (0.488) 2/50 4.56E+01 (0.109) 4.57E+01 (0.108) 4.56E+01 (0.110) 4.55E+01 (0.112)
Table 8: Violation probabilities and reliability indices β (in parentheses) estimated with FORM-NN, for
f=1.0 and different number of experimental points
Number of the experimental points Hazard level m 2m 3m 4m 50/50 4.48E-06 (5.347) 1.21E-05 (5.164) 9.88E-06 (5.202) 1.02E-05 (5.195) 10/50 3.13E+01 (0.487) 3.14E+01 (0.483) 3.13E+01 (0.488) 3.19E+01 (0.472) 2/50 4.50E+01 (0.126) 4.57E+01 (0.108) 4.58E+01 (0.106) 4.58E+01 (0.106)
![Page 31: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/31.jpg)
31
Table 9: Comparison of optimum designs of the Reliability Sizing Optimization (RSO) and the
Reliability combined Sizing-Topology Optimization (RSTO)
Design RSO RSTO
Columns 1 HEB800 HEB450 Columns 2 HEB550 HEB500 Columns 3 HEB400 HEB600 Columns 4 HEB500 HEB700 Beams 5 IPE220 IPE330 No columns x 9 6 No columns y 6 4
β50/50 11.144 11.546 β10/50 3.959 3.413 β2/50 4.415 4.331
Volume (m3) 47.71 31.25
![Page 32: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/32.jpg)
32
Table 10: Performance of the optimum designs
DSO DSTO RSO RSTO
Volume (m3) 38.32 22.86 47.71 31.25
Pviol(50/50)/ Pallowable=1.0E-01(%) 6.45E-05 (%) 1.09E-05 (%) <10-7(%) <10-7(%)
Pviol(10/50)/ Pallowable=5.0E-02(%) 2.51E+01 (%) 3.19E+01 (%) 3.76E-03 (%) 3.21E-02 (%)
Pviol(2/50)/ Pallowable=1.0E-03(%) 4.28E+01 (%) 4.56E+01 (%) 5.05E-04 (%) 7.42E-04 (%)
Table 11: Computational effort of optimization formulations
DSO DSTO RSO MCS
RSO MCS-NN
RSO FORM-NN
RSTO MCS
RSTO MCS-NN
RSTO FORM-NN
Time (hours) 8.00E+00 3.38E+00 2.25E+04* 2.25E+03 6.48E+02 1.44E+04* 1.44E+03 4.14E+02 Time (years) 9.13E-04 3.85E-04 2.57E+00* 2.57E-01 3.70E-02 1.64E+00* 1.64E-01 2.36E-02 * Estimated for 50,000 simulations
![Page 33: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/33.jpg)
FIGURES
Figure 1: Flowchart of the Performance-Based Design (PBD) procedure
33
![Page 34: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/34.jpg)
1. Begin 2. : 0t =3. ( ) ( ) ( ) ( )( )( ){ }( )0 0 0 0initialize : , , , 1,...,p m m mP y s F s m μ= =
4. Repeat 5. For To λ Do Begin : 1l =6. ( )( ): marriage t
l pD P=
7. ( ): s_recombinationl ls D=
8. ( ): y_recombinationl ly D=
9. ( ): s_mutationl ls s=
10. ( ): y_mutationl ly y=11. ( ):l lF F s=
12. End
13. ( ) ( ) ( ) ( )( )( ){ }: , , , 1,...,t t t to l l lP y s F s l λ= =
14. Case selection_type Of
15. ( ) ( ) ( )( )1, : : selection ,t tp oP Pμ λ μ+ =
16. ( ) ( ) ( ) ( )( )1: : selection , ,t t tp oP Pμ λ μ++ = pP
17. End 18. : 1t t= +19. Until termination_criterion 20. End
Figure 2: Pseudo-code of the Evolutionary Algorithm (EA) optimization procedure
34
![Page 35: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/35.jpg)
(a)
(b)
Figure 3: MCS-NN - (a) Methodology 1, (b) Methodology 2
(a) (b) Figure 4: Standard sampling method [28] (a) before and (b) after correction
35
![Page 36: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/36.jpg)
Figure 5: The FORM-NN methodology
(a) (b) Figure 6: LHS technique (a) before and (b) after correction
36
![Page 37: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/37.jpg)
(a)
(b)
Figure 7: 7-storey frame, (a) plan view and (b) side view
37
![Page 38: INNOVATIVE SEISMIC DESIGN OPTIMIZATION …...Two types of random variables are considered: Those which influence the level of seismic demand and those that affect the structural capacity](https://reader033.vdocuments.site/reader033/viewer/2022042203/5ea4d0edb2007f7be840b7d3/html5/thumbnails/38.jpg)
0
5
10
15
20
25
0 1 2 3 4 5 6
Period T (sec)
Acc
eler
atio
n (m
/sec
2)Mean-X (50/50)
Mean-Y (50/50)
Mean-X (10/50)
Mean-Y (10/50)
Mean-X (2/50)
Mean-Y (2/50)
Figure 8: Median response spectra for three hazard levels
38