covid
Buscar en
Geofísica Internacional
Toda la web
Inicio Geofísica Internacional Neural estimation of strong ground motion duration
Información de la revista
Vol. 53. Núm. 3.
Páginas 221-239 (julio - septiembre 2014)
Compartir
Compartir
Descargar PDF
Más opciones de artículo
Visitas
2206
Vol. 53. Núm. 3.
Páginas 221-239 (julio - septiembre 2014)
Open Access
Neural estimation of strong ground motion duration
Visitas
2206
Leonardo Alcántara Nolasco
Autor para correspondencia
leonardo@pumas.ii.unam.mx

Corresponding author:
, Silvia García, Efraín Ovando-Shelley, Marco Antonio Macías Castillo
Instituto de Ingeniería, Universidad Nacional Autónoma de México, Delegación Coyoacán, 04510, México D.F., México
Este artículo ha recibido

Under a Creative Commons license
Información del artículo
Resumen
Texto completo
Bibliografía
Descargar PDF
Estadísticas
Figuras (15)
Mostrar másMostrar menos
Tablas (4)
Table 1. Puebla City Accelerograph network (Alcántara, 1999)
Table 2. RACP Selected records
Table 3. Oaxaca City Accelerograph network (Alcántara et al., 2000)
Table 4. RACO selected records
Mostrar másMostrar menos
Resumen

Este artículo presenta y discute el uso de las redes neuronales para determinar la duración de los movimientos fuertes del terreno. Para tal efecto se desarrolló un modelo neuronal, utilizando datos acelerométricos registrados en las ciudades mexicanas de Puebla y Oaxaca, que predice dicha duración en términos de la magnitud, distancia epicentral, profundidad focal, caracterización del suelo y el azimut. Por lo que, el modelo considera los efectos tanto de la zona sismogénica como del tipo de suelo en la duración del movimiento. El esquema final permite una estimación directa de la duración a partir de variables de fácil obtención y no se basa en hipótesis restrictivas. Los resultados presentados en este artículo indican que la alternativa del cómputo aproximado, particularmente las redes neuronales, es una poderosa aproximación que se basa en los registros sísmicos para explorar y cuantificar los efectos de las condiciones sísmicas y de sitio en la duración del movimiento. Un aspecto esencial y significante de este nuevo modelo es que a pesar de ser extremadamente simple ofrece estimaciones de duración con notable eficiencia. Adicional e importante son los beneficios que arroja esta simplicidad sobre la separación natural de los efectos de la fuente, patrón o directividad y de sitio además de la eficiencia computacional.

Palabras clave:
duración del movimiento de terreno
parámetros de movimientos de terreno
duración significativa
intensidad de Árias
redes neuronales
cómputo aproximado
Abstract

This paper presents and discusses the use of neural networks to determine strong ground motion duration. Accelerometric data recorded in the Mexican cities of Puebla and Oaxaca are used to develop a neural model that predicts this duration in terms of the magnitude, epicenter distance, focal depth, soil characterization and azimuth. According to the above the neural model considers the effect of the seismogenic zone and the contribution of soil type to the duration of strong ground motion. The final scheme permits a direct estimation of the duration since it requires easy-to-obtain variables and does not have restrictive hypothesis. The results presented in this paper indicate that the soft computing alternative, via the neural model, is a reliable recording-based approach to explore and to quantify the effect of seismic and site conditions on duration estimation. An essential and significant aspect of this new model is that, while being extremely simple, it also provides estimates of strong ground motions duration with remarkable accuracy. Additional but important side benefits arising from the model’s simplicity are the natural separation of source, path, and site effects and the accompanying computational efficiency.

Key words:
strong ground motion duration
ground motion parameters
significant duration
Árias Intensity
neural networks
soft computing
Texto completo
Introduction

The principal objective of engineering seismology is to supply quantitative estimations of expected ground-motions for earthquake-resistant design, evaluation of seismic hazards, and seismic risk assessment through the proper characterization of complex time series (accelerograms). Since the first strong-motion accelerograms were recorded a large number of parameters have been defined to characterize movements. The usefulness of strong-motion parameters is dependent primarily upon their intended use. The parameters that can be employed in earthquake-resistant design are few and are directly related to the methods of structural analysis used in current practice. Once a parameter has been selected to characterize the ground motion, it is necessary to develop relationships between this parameter and important seismic features as earthquake source, travel path, and site conditions.

The essence of such predictive relationships for the duration of strong motions depends very heavily on the way duration is defined. In fact many strong-motion duration definitions have been presented; however, all of them attempt to isolate a certain portion of the time series where strongest motion occurs. In general terms, it has been accepted that all of these definitions can be grouped into one of four generic categories (Bommer and Martínez-Pereira, 1996): i) the bracketed duration, the interval between the first and last excursion of particular threshold amplitude, ii) the uniform duration, the sum of all of the time intervals during which the amplitude of the record is above the threshold, iii) the significant duration, which is determined from the Husid plot (Husid, 1969) based on the interval during which a certain portion of the total Árias intensity is accumulated and iv) the structural response duration, determined by applying one of the above three categories to the response of a specific single-degree-of-freedom oscillator.

In this investigation, and considering the definition of significant duration, the connection between data and knowledge is found using a soft computing SC tool: the neural networks NNs. This alternative improves the theory and understanding of the driven parameters (of all kinds including indeterminate ones, possibly expressed in words) of ground-motion duration behavior. SC, NNs particularly, utilize a discovery approach to examine the multidimensional data relationships simultaneously and to identify those that are unique or frequently represented, permitting the acquisition of structured knowledge.

A neuronal empirical model for strong motion duration is proposed here, derived from seismic information registered in Puebla and Oaxaca, México. This model predicts the strong ground motion duration as a function of earthquake magnitude, epicentral distance, focal depth, azimuth (established from epicenters to stations) and soil characterization. The final scheme permits a direct estimation of the duration since it requires easy-to-obtain variables and does not have restrictive hypothesis

Soft Computing

The term Soft Computing SC represents the combination of emerging problem-solving technologies such as Fuzzy Logic FL, Probabilistic Reasoning PR, Neural Networks NN, and Genetic Algorithms GAs. Each of these provides complementary reasoning and searching methods to solve complex, real-world problems. In ideal problem formulations, the systems to be modeled or controlled are described by complete and precise information. In these cases, formal reasoning systems, such as theorem proofs, can be used to attach binary true or false values to statements describing the state or behavior of the physical system.

Soft Computing technologies are flexible computing tools to perform these approximate reasoning and search tasks handling imperfect information. According to Zadeh (Fuzzy Logic pioner): “…in contrast to traditional, hard computing, soft computing is tolerant of imprecision, uncertainty, and partial truth.” The only obvious common point between SC tools (Fuzzy Logic FL, Neural Networks NNs and Genetic Algorithms GAs) is that they have been inspired by the living: the imprecision of human language and its efficiency in conveying and transmitting information for FL, the architecture of the brain for NNs, and the reproduction of living beings for GAs.

Neural Networks

This section will briefly explain the theory of neural networks NN. For a more in depth explanation of these concepts consult Hassoun, (1995); Hertz et al., (1991) and Tettamanzi and Tomassini, (2001).

In the brain, a NN is a network consisting of connected neurons. The nucleus is the center of the neuron and it is connected to other nuclei through the dendrites and the axon. This connection is called a synaptic connection. The neuron can fire electric pulses through its synaptic connections, which are received by the dendrites of other neurons. Figure 1 shows how a simplified neuron looks like. When a neuron receives enough electric pulses through its dendrites, it activates and fires a pulse through its axon, which is then received by other neurons. In this way information can propagate through the NN. The synaptic connections change throughout the lifetime of a neuron and the amount of incoming pulses needed to activate a neuron (the threshold) also change. This process allows the NN to learn (Tettamanzi and Tomassini, 2001).

Figure 1.

Simplified biological neuron.

(0.06MB).

Mimicking the biological process the artificial NN are not “intelligent” but they are capable for recognizing patterns and finding the rules behind complex data-problems. A single artificial neuron can be implemented in many different ways. The general mathematic definition is given by equation 1.

where x is a neuron with n input dendrites (x0,…, xn) and one output axon y(x) and (w0,…,wn) are weights determining how much the inputs should be weighted; g is an activation function that weights how powerful the output (if any) should be from the neuron, based on the sum of the input. If the artificial neuron mimics a real neuron, the activation function g should be a simple threshold function returning 0 or 1. This is not the way artificial neurons are usually implemented; it is better to have a smooth (preferably differentiable) activation function (Bishop, 1996). The output from the activation function varies between 0 and 1, or between -1 and 1, depending on which activation function is used. The inputs and the weights are not restricted in the same way and can in principle be between -∞ and+∞, but they are very often small values centered on zero (Broomhead and Lowe, 1988). Figure 2 provides a schematic view of an artificial neuron.

Figure 2.

An artificial neuron.

(0.05MB).

As mentioned earlier there are many different activation functions, some of the most commonly used are threshold (Eq. 2), sigmoid (Eq.3) and hyperbolic tangent (Eq.4).

where t is the value that pushes the center of the activation function away from zero and s is a steepness parameter. Sigmoid and hyperbolic tangent are both smooth differentiable functions, with very similar graphs. Note that the output range of the hyperbolic tangent goes from -1 to 1 and sigmoid has outputs that range from 0 to 1. A graph of a sigmoid function is given in Figure 3 to illustrate how the activation function looks like. The t parameter in an artificial neuron can be seen as the amount of incoming pulses needed to activate a real neuron. A NN learns because this parameter and the weights are adjusted.

Figure 3.

A graph of a sigmoid function with s=0.5 and t=0

(0.03MB).
NN architecture

The NN used in this investigation is a multilayer feedforward neural network MFNN, which is the most common NN. In a MFNN, the neurons are ordered in layers, starting with an input layer and ending with an output layer. There are a number of hidden layers between these two layers. Connections in these networks only go forward from one layer to the next (Hassoun, 1995). They have two different phases: a training phase (sometimes also referred to as the learning phase) and an execution phase. In the training phase the NN is trained to return a specific output given particular inputs, this is done by continuous training on a set of data or examples. In the execution phase the NN returns outputs on the basis of inputs. In the NN execution an input is presented to the input layer, the input is propagated through all the layers (using equation 1) until it reaches the output layer, where the output is returned. Figure 4 shows a MFNN where all the neurons in each layer are connected to all the neurons in the next layer, what is called a fully connected network.

Figure 4.

A fully connected multilayer feedforward network with one hidden layer and bias neurons

(0.19MB).

Two different kinds of parameters can be adjusted during the training, the weights and the t value in the activation functions. This is impractical and it would be easier if only one of the parameters were to be adjusted. To cope with this problem a bias neuron is introduced. The bias neuron lies in one layer, connected to all the neurons in the next layer, but none in the previous layer and it always emits 1. A modified equation for the neuron, where the weight for the bias neuron is represented as wn+1, is shown in equation 5.

Adding the bias neuron allows the removal of the t value from the activation function, leaving the weights to be adjusted, when the NN is being trained. A modified version of the sigmoid function is shown in equation 6.

The t value cannot be removed without adding a bias neuron, since this would result in a zero output from the sum function if all inputs where zero, regardless of the values of the weights

Training a NN

When training a NN with a set of input and output data, we wish to adjust the weights in the NN to make the NN gives outputs very close to those presented in the training data. The training process can be seen as an optimization problem, where the mean square error between neural and desired outputs must be minimized. This problem can be solved in many different ways, ranging from standard optimization heuristics, like simulated annealing, to more special optimization techniques like genetic algorithms or specialized gradient descent algorithms like backpropagation BP.

The backpropagation algorithm

The BP algorithm works in much the same way as the name suggests: after propagating an input through the network, the error is calculated and the error is propagated back through the network while the weights are adjusted in order to make the error smaller. Although we want to minimize the mean square error for all the training data, the most efficient way of doing this with the BP algorithm, is to train on data sequentially one input at a time, instead of training the combined data.

BP application steps. First the input is propagated through the NN to the output. Then the error ek on a single output neuron k can be calculated as:

where yk is the calculated output and dk is the desired output of neuron k. This error value is used to calculate a δk value, which is again used for adjusting the weights. The δk value is calculated by:

where g’ is the derived activation function. When the δk value is calculated, the δj values can be calculated for preceding layers. The δj values of the previous layer are calculated from the δk values of this layer by the following equation:

where K is the number of neurons in this layer and η is the learning rate parameter, which determines how much the weight should be adjusted. The more advanced gradient descent algorithms does not use a learning rate, but a set of more advanced parameters that makes a more qualified guess to how much the weight should be adjusted. Using these δ values, the Δw values that the weights should be adjusted by, can be calculated:

The Δwjk value is used to adjust the weight wjk by wjk=wjk+Δwjk and the BP algorithm moves on to the next input and adjusts the weights according to the output. This process goes on until a certain stop criteria is reached. The stop criterion is typically determined by measuring the mean square error of the training data while training with the data, when this mean square error reaches a certain limit, the training is stopped.

In this section the mathematics of the BP algorithm have been briefly discussed, but since this report is mainly concerned with the implementation of NN, the details necessary for implementing the algorithm has been left out (for details see Hassoun, 1995 and Hertz et al., 1991).

Duration: predictive relationships

Predictive relationships usually express ground motion parameters as functions of earthquake magnitude, distance, source characteristics, site characteristics, etc. A typical predictive relationship may have the form:

where Y is the ground motion parameter of interest, M the magnitude of the earthquake, R a measure of the distance from the source to the site being considered. C1-C9 are constants to be determined. The σlnY term describes the uncertainty in the value of the ground motion parameter given by the predicative relationship.

Regarding duration parameters many types of predictive relationships have been proposed (Bommer and Martinez-Pereira, 1999), but bracketed duration and significant duration relationships are the most commonly used. The former is defined as the time elapsed between the first and last excursions beyond a specified threshold acceleration. That definition has shown to be sensitive to the threshold acceleration considered and to small events that occur at the final part of a recording. Significant duration is based on the dissipation of energy, within a time interval, and this energy is represented by the integral of the square of the ground motions. In the case of acceleration is related to the Árias intensity IA (Árias, 1970):

here a (t) is the acceleration time history, g is the acceleration of gravity, and T represents the complete duration of recording a (t). Figure 5 present the procedure followed to determine the significant parameters (Husid, 1969). The most common measure of significant duration is a time interval between 5-95% of IA and is denoted by Da5-95.

Figure 5.

Significant duration parameters (acceleration 5-95%).

(0.11MB).

Predictive relationships have also been developed for frequency-dependent duration parameters evaluated from bandpassed accelerograms (e.g., Bolt, 1973, Trifunac and Westermo, 1982; Mohraz and Peng, 1989; and Novikova and Trifunac, 1994). These relationships have several limitations that are basically associated with a deficient representation of magnitude or site effects. Additionally, none of these have been derived from the energy integral. Some other restrictions are related to measured distance (normally the epicentral distance, not the closest site-source distance) and finally there are still others having to do with the regression method used to derive the relationships (Kempton and Stewart, 2006).

In what follows we develop a predictive neuronal model for significant duration that: 1) considers the seismic effects associated to magnitude, focal distance, near-fault rupture directivity and soil conditions and 2) is based on a soft computing procedure that accounts for inter- and intra-event ground-motion variability. Significant duration, from the Árias integral, was selected because of the stability of the method with respect to the definitions of initial and final threshold (Bommer and Martinez-Pereira 1999).

Neural estimation of duration

The ground motion duration model developed here captures the effects of the amount of energy radiated at the source using a neural representation of phenomena implicit in the data, the attenuation of seismic waves along the path due to geometric spreading and energy absorption; it also considers a local modification of the seismic waves as they traverse near-surface materials. The strong-motion duration D is the dependent variable of the NN formulation. The primary predictor variables (independent variables in a typical regression analysis) are M moment magnitude; R epicentral distance; focal depth FD, soil characterization expressed by Ts natural period; and Az azimuth.

NN based on information compiled from PueblaDatabase

The city of Puebla has currently an accelerograph network composed of 11 seismic stations, three of which are located on rock, seven on compressible soil, and one in the basement of a structure. The general characteristics are provided in Table 1 and their locations indicated in Figure 6. Although, the first station (SXPU) was installed in 1972, the number of accelerogram records is relatively low mainly due to the low rate of seismicity in the region and the long process taken to install seismic stations.

Table 1.

Puebla City Accelerograph network (Alcántara, 1999)

Station CodeStation nameSoilGeo-Coordinates
Lat N  Long W 
BHPP  Barranca Honda  Basaltic rock  19.109  -98.227 
CAPP  Central de Abasto  Compressible  19.089  -98.188 
LMPP  Lomas del Mármol  Marble stone  19.001  -98.182 
PBPP  Paseo Nicolás Bravo  Travertine deposit  19.046  -98.208 
PHPU  Parque Habana  Compressible  19.040  -98.167 
PZPU  La Paz, Puebla  Volcanic cone  19.055  -98.227 
RFPP  Río San Francisco  Structure basement  19.050  -98.190 
SAPP  San Alejandro  Lacustrine deposit  19.060  -98.210 
SRPU  San Ramón Castillotla  Compressible  18.965  -98.260 
SXPU  Sismex Puebla  Compressible  19.040  -98.215 
UAPP  Facultad de Ingeniería Civil  Compressible  19.002  -98.202 
Figure 6.

Strong motion network in Puebla (RACP).

(0.2MB).

In the first stage for the integration of our database, records with low signal-to-noise ratios were not taken into account. Hence, only 42 three-component accelerograms associated to three seismic stations (PBPP, SXPU and SRPU) are included in the database. These records were obtained from records of both subduction and normal-faulting earthquakes, originated, respectively, at the contact of the North America and Cocos plates, and by the fracture of the subducted Cocos plate.

The earthquakes in the database have magnitudes ranging from 4.1 to 8.1. Most of the events originated along coast of the Pacific Ocean in the states of Michoacan, Guerrero and Oaxaca. The epicenters of the remaining three events, those of October 24, 1980, April 3, 1997 and June15, 1999 were located in the Puebla-Oaxaca border. Epicentral distances to stations in the city of Puebla range from 300 to 500km and in only one case it reached 800km. That is why accelerations produced by the earthquakes considered in this research did not exceed 10 gal in Puebla.

In a second stage the database was expanded with accelerograms from the Instituto de Ingeniería UNAM Accelerographic Network (Alcántara et al., 2000). The added acceleration histories were recorded in stations on rock located the coastal region of the states of Michoacan, Guerrero and Oaxaca, and forcefully had to be generated by one of the earthquakes we had already catalogued in Table 1. The seismic stations we considered are shown in Figure 7 (filled squares), as well as the locations of the epicenters (inverted triangles). They were 88 three-component accelerograms in the final database.

Figure 7.

Location of epicenters and seismic stations

(0.2MB).

A set of 26 events was used (Table 2) to design the topology of the NNs. These events were selected on the basis of the quality and resolution of the records. Accelerograms with low signal to noise ratios were deleted from the database. Both horizontal components and vertical direction of each seismic event were considered.

Table 2.

RACP Selected records

Event date  Recording station  Event date  Recording station 
7811291  7.8  SXPU  14  9602255  PBPP 
7811292  5.3  SXPU  15  9603271  SRPU 
7811293  4.5  SXPU  16  9607151  6.5  PBPP, SRPU, SXPU 
7903141  SXPU  17  9701111  6.9  PBPP, SRPU, SXPU 
8010241  SXPU  18  9701211  PBPP, SRPU, SXPU 
8110251  7.3  SXPU  19  9704031  4.8  PBPP 
8407021  SXPU  20  9712161  5.9  PBPP 
8502111  5.2  SXPU  21  9801101  6.3  PBPP 
8509191  8.1  SXPU  22  9802031  6.2  PBPP, SRPU, SXPU 
10  8509211  7.5  SXPU  23  9804201  5.5  PBPP, SRPU, SXPU 
11  8904251  6.9  PBPP  24  9906151  6.5  PBPP, SRPU 
12  9509141  7.2  PBPP, SRPU, SXPU  25  9906211  5.8  PBPP, RPU 
13  9602251  6.7  PBPP, SXPU  26  9909301  7.5  PBPP, SXPU 

It is clear that the inputs and output spaces are not completely defined; the phenomena knowledge and monitoring process contain fuzzy stages and noisy sources. Many authors have highlighted the danger of inferring a process law using a model constructed from noisy data (Jones et al., 2007). It is imperative we draw a distinction between the subject of this investigation and that of discovering a process from records. The main characteristic of NN model is unrevealed functional forms. The NN data-driven system is a black-box representation that has been found exceedingly useful in seismic issues but the natural principle that explains the underlying processes remains cryptic. Many efforts have been developed to examine the input/output relationships in a numerical data-set in order to improve the NN modeling capabilities, for example Gamma test (Kemp et al., 2005; Jones et al., 2007; Evans and Jones, 2002), but as far as the authors’ experience, none of these attempts are applicable to the high dimension of the seismic phenomena or the extremely complex neural models for predicting seismic attributes.

Neural approximation

The first step in developing a NN is the representation of the set of input and output cells. There are no clear-cut procedures to define this construction step. While the optimum architecture --hidden nodes and associated weights-- is obtained when the error function is minimized (i.e., the sum of the patterns of the squared differences between the actual and desired outputs is minimum) the numerical or categorical representation of inputs and outputs also depends on the modeler’s experience and knowledge and a trial-and-error procedure must be followed in order to achieve a suitable design.

The RACP database has been modeled using the BP learning algorithm and Feed Forward Multilayer architecture. Time duration in horizontal (mutually orthogonal DH1, N-S, and DH2, E-W) and vertical components (DV) are included as outputs for neural mapping and this attempt was conducted using five inputs (M, R, FD, Ts and AZ). After trying many topologies, we found out that the best model during the training and testing stages has two hidden layers with 200 nodes each. As seen in Figure 8a, the training correlation for DH1, DH2 and DV was quite good, but when the same model is tested (unseen cases are presented to predict the output) considerable differences between measured and estimated duration times are found (Figure 8b). It is important to point out that the results shown in that figure are the best we were able to obtain after trying 25 different topologies. Thus, this can be considered as the model having the best generalization capabilities using the selected learning algorithm, architecture, and nodal hidden structure. In Figure 9 the estimated values obtained for a second set of unseen patterns (validation set) are compared with the numerical predictions obtained using the relationship proposed by Reinoso and Ordaz (2001). The neuronal relationship follows more narrowly the overall trend but fails in some cases, (coefficients of correlation around R2=0.75). It should be stressed that the NN has better interpolation and extrapolation capabilities than the traditional functional approaches. Furthermore, the influence of directivity and fault mechanism on duration can be identified with the NN, based on a multidimensional environment (Figure 10)

Figure 8.

NN results for RACP, training and testing stages

(0.31MB).
Figure 9.

NN results for RACP, validation stage

(0.35MB).
Figure 10.

Input sensitivity analysis for RACP NN

(0.13MB).
NN based on information compiled from OaxacaDatabase

The information used in the study is taken from the Oaxaca accelerographic array (RACO, Red Acelerográfica de la Ciudad de Oaxaca, in Spanish). The first recording station was installed in 1970 and nowadays the network comprises seven stations deployed around the urban area.

The instruments in these stations are located on ground surface. Each station has a digital strong-motion seismograph (i.e., accelerograph) with a wide frequency-band and wide dynamic range. Soil conditions at the stations vary from soft compressible clays to very stiff deposits (see Table 3). Locations of these observation sites are shown in Figure 11. From 1973 to 2004, the network recorded 171 time series from 67 earthquakes with magnitudes varying from 4.1 to 7.8 (Table 4). Events with poorly defined magnitude or focal mechanism, as well as records for which site-source distances are inadequately constrained, or records for which problems were detected with one or more components were removed from the data sets. The final training/testing set contains 147 three-component accelerograms that were recorded in five accelerograph stations OXLC, OXFM, OXAL, OXPM and OXTO. This catalogue represents wide-ranging values of directivity, epicentral distances and soil-type conditions (see Figure 12).

Table 3.

Oaxaca City Accelerograph network (Alcántara et al., 2000)

Station CodeStation nameSoilGeo-Coordinates
Lat N  Log W 
OXFM  Facultad de Medicina  Alluvium  17.084  -96.716 
OXLC  Las Canteras  Rock  17.065  -96.703 
OXPM  Primaria Múgica  Clay  17.061  -96.717 
OXBJ  Primaria Benito Juárez  Clay  17.067  -96.744 
OXAL  Alameda de León  Clay  17.061  -96.725 
OXCU  Ciudad Universitaria  Clay  17.049  -96.713 
OXTO  Instituto Tecnológico  Alluvium  17.078  -96.744 
Figure 11.

Strong motion network in Oaxaca

(0.21MB).
Table 4.

RACO selected records

Event date  Recording station 
197308281  6.8  OXFM 
197811291  7.8  OXFM 
198010241  OXFM 
198206072  OXFM 
198301241  5.3  OXFM 
199407041  5.9  OXFM 
199408271  OXFM 
199408281  5.2  OXFM 
199509141  7.2  OXFM 
10  199604011  OXFM 
11  199802031  6.2  OXFM 
12  199805021  4.4  OXFM 
13  199906151  6.5  OXFM, OXLC 
14  199906152  4.5  OXFM, OXLC 
15  199909301  7.5  OXFM, OXLC 
16  199910071  4.4  OXAL, OXTO 
17  199910251  4.6  OXTO 
18  199911061  4.5  OXFM, OXAL, OXLC, OXPM, OXTO 
19  199911101  4.3  OXAL, OXPM, OXTO 
20  199911291  4.4  OXTO 
21  199912171  4.4  OXPM, OXTO 
22  200001191  4.5  OXLC, OXPM, OXTO 
23  200003011  4.7  OXFM, OXLC, OXPM, OXTO 
24  200003121  6.4  OXLC, OXTO 
25  200005101  4.7  OXFM, OXPM, OXTO 
26  200007211  5.9  OXTO 
27  200008171  4.4  OXTO 
28  200009291  4.7  OXLC, OXPM, OXTO 
29  200010171  5.4  OXTO 
30  200101261  OXTO 
31  200102191  4.8  OXTO 
32  200102201  4.7  OXTO 
33  200109081  4.8  OXAL, OXLC, OXPM, OXTO 
34  200110031  4.4  N/A 
35  200111101  OXPM, OXTO 
36  200111281  OXTO 
37  200201161  6.3  OXLC, OXPM, OXTO 
38  200201301  5.1  OXFM, OXLC, OXPM, OXTO 
39  200202261  OXTO 
40  200205111  4.5  OXPM, OXTO 
41  200206071  4.7  OXFM, OXTO 
42  200206072  5.6  OXFM, OXAL, OXLC, OXPM, OXTO 
43  200206111  4.9  OXTO 
44  200206181  4.5  OXFM 
45  200208051  5.3  OXFM, OXLC, OXPM, OXTO 
46  200208271  4.9  OXFM, OXAL, OXLC, OXPM, OXTO 
47  200211081  5.2  OXTO 
48  200212021  4.7  OXTO 
49  200212291  4.6  OXFM, OXPM, OXTO 
50  200301221  7.6  OXLC 
51  200302011  5.3  OXTO 
52  200306131  4.1  OXTO 
53  200307081  4.6  OXTO 
54  200309251  4.6  OXTO 
55  200312011  4.3  OXAL, OXPM, OXTO 
56  200401131  5.1  OXFM, OXAL, OXLC, OXPM, OXTO 
57  200401132  5.5  OXFM, OXAL, OXLC, OXPM, OXTO 
58  200401141  4.6  OXTO 
59  200401171  4.7  OXFM, OXAL, OXLC, OXPM, OXTO 
60  200402101  4.4  OXFM, OXAL, OXTO 
61  200402181  4.3  OXFM, OXAL, OXLC, OXPM, OXTO 
62  200404201  5.4  OXFM, OXAL, OXPM, OXTO 
63  200405061  4.3  OXTO 
64  200406141  5.8  OXFM, OXAL, OXPM, OXTO 
65  200408011  4.6  OXFM, OXAL, OXTO 
66  200408071  5.3  OXFM, OXAL, OXLC, OXPM, OXTO 
67  200408181  5.7  OXFM, OXAL, OXLC, OXPM, OXTO 
Figure 12.

Epicenters and Seismic stations.

(0.2MB).
Neural modeling

The NN for Oaxaca City was developed using a similar set of independent parameters as those used for Puebla exercise. As the input/output behavior of the previous system is physical meaning the same five descriptors are included as inputs. This action permits to explore both systems’ behaviors and to get wide-ranging conclusions about these variables.

Epicentral distance R was selected as a measure of distance because simple source-site relationships can be derived with it. Focal depth FD, was introduced for identifying data from interface events (FD<50km) and intraslab events (FD>50km). Together with the Azimuth Az, it associates the epicenter with a particular seismogenic zone and directivity pattern (fault mechanism).

To start the neuro training process using the Oaxaca database Ts is disabled and a new soil classification is introduced. Three soil classes were selected: rock, alluvium and clay. The final topology for RACO data contains BP as the learning algorithm and Feed Forward Multilayer as the architecture. Again DH1, DH2, and DV are included as outputs for neural mapping and between the five inputs, four are numerical (M, R, FD, and AZ) and one is a class node (soil type ST). The best model during the training and testing stages has two hidden layers of 150 nodes each and was found through an exhaustive trial and error process.

The results of the RACO NN are summarized in Figure 13. These graphs show the predicting capabilities of the neural system comparing the task-D values with those obtained during the NN training phase. It can be observed that the durations estimated with the NN match quite well calculated values throughout the full distance and magnitude ranges for the seismogenic zones considered in this study. Duration times from events separated to be used as testing patterns are presented and compared with the neuronal blind evaluations in Figure 14. The results are very consistent and remarkably better than those obtained when analyzing RACP database. The linguistic expression of soil type is obviously a superior representation of the soil effect on D prediction.

Figure 13.

NN results for RACO, training stage.

(0.14MB).
Figure 14.

NN results for RACO, testing stage.

(0.18MB).

A sensitivity study for the input variables was conducted for the three neuronal modules. The results are given in Figure 15 and are valid only for the data base utilized. Nevertheless, after conducting several sensitivity analyses changing the database composition, it was found that the RACO trend prevails: ST (soil type) is the most relevant parameter (has the larger relevance), followed by azimuth Az, whereas M, FD and R turned out to be less influential. NNs for the horizontal and vertical components are complex topologies that assign nearly the same weights to the three input variables that describe the event, but an important conclusion is that the material type in the deposit and the seismogenic zone are very relevant to define D. This finding can be explained if we conceptualize the soil deposit as a system with particular stiffness and damping characteristics that determine how will the soil column vibrate and for how long, as seismic waves traverse it and after their passage through the deposit.

Figure 15.

Input sensitivity analysis for RACO NN.

(0.15MB).

Through the {M, R, FD, AZ, ST} → {DH1, DH2, DV} mapping, the neuronal approach we presented offers the flexibility to fit arbitrarily complex trends in magnitude and distance dependence and to recognize and select among the tradeoffs that are present in fitting the observed parameters within the range of variables present in data.

Conclusions

Artificial neural networks were used to estimate strong ground motion duration. These networks were developed using a back propagation algorithm and multi-layer feed-forward architecture in the training stage. In developing the networks it was assumed that the parameters that have the greatest influence on strong motion duration are magnitude, epicentral distance, focal depth, soil characterization and azimuth. These parameters include the effects of seismic source, distance, materials and directivity. The many topologies tested and the input sensitivity developed drive to the conclusion that a broad soil-type classification (in these investigation three soil types) provides a better correlation with seismic phenomena than the more commonly used natural period Ts.

Overall, the results presented here show that artificial neural networks provide good and reasonable estimates of strong ground motion duration in each one of the three orthogonal components of the accelerograms recorded in the cities of Puebla and Oaxaca using easy-to-obtain input parameters: S M, R, FD and Az.

Finally, it is important to highlight that the capabilities of a NN ultimately depend on various factors that require the knowledge of the user about the problem under consideration. This knowledge is essential for establishing the pattern parameters that best represent it. Experience to set and to select the network architecture (including learning rules, transfer functions and hidden nodal structure) and the proper integration of training, test and validation data sets are also very important.

Acknowledgments

The recordings used in this paper were obtained thanks to Instituto de Ingeniería, UNAM and the Benemérita Universidad Autónoma of Puebla (BUAP). We are also grateful to Professor Miguel P Romo for his useful comments.

References
[Adalbjörn et al., 1997]
S. Adalbjörn, N. Končar, A.J. Jones.
A note on the Gamma test.
Neural Computing and Applications, 5 (1997), pp. 131-133
[Alcántara et al., 1999]
L. Alcántara, G. González, D. Almora, A.E. Posada-Sánchez, M. Torres.
Puebla City Accelerograph Network, activities in 1996, RACP-II/BUAP-01.
Internal Report Institute of Engineering, UNAM and Autonomous University of Puebla, México, (1999),
[Alcántara et al., 2000]
L. Alcántara, R. Quaas, C. Pérez, M. Ayala, M.A. Macías, H. Sandoval II-UNAM, C. Javier, E. Mena, E. Andrade, F. González, E. Rodríguez CFE, A. Vil, L. Munguía, M. Luna CICESE, J.M. Espinosa, A. Cuellar, L. Camarillo, S. Ramos, M. Sánchez CIRES, R. Quaas, E. Guevara, J.A. Flores, B. López, R. Ruiz CENAPRED, E. Guevara, J. Pacheco IG-UNAM, M. Ramírez, J. Aguilar, J. Juárez, R. Vera, A. Gama, R. Cruz, F.M. Hurtado, R. del Campo, F. Vera RIIS, L. Alcántara SMIS.
Mexican Strong Motion Data Base.
[Anderson, 2004]
J.G. Anderson.
Quantitative measure of goodness-of-fit of synthetic seismograms, in 13th World Conference on Earthquake Engineering, 243, Vancouver, Canada, (2004), pp. 1-6
[Árias, 1970]
A. Árias.
A measure of earthquake intensity, Seismic Design of Nuclear Power Plants, pp. 438-483
[Bishop, 1996]
C.M. Bishop.
Neural networks for pattern recognition, Oxford University Press, (1996), pp. 482
[Bolt, 1973]
B.A. Bolt.
Duration of strong ground motion.
pp. 25-29
[Bommer and Martinez-Pereira, 1996]
J.J. Bommer, A. Martinez-Pereira.
The prediction of strong-motion duration for engineering design.
pp. 23-28
[Bommer and Martinez-Pereira, 1999]
J.J. Bommer, A. Martinez-Pereira.
The effective duration of earthquake strong motion.
Journal of Earthquake Engineering, 3 (1999), pp. 127-172
[Broomhead and Lowe, 1988]
D.S. Broomhead, D. Lowe.
Multivariable functional interpolation and adaptive networks.
Complex Systems, 2 (1988), pp. 321-355
[Evans and Jones, 2002]
D. Evans, A.J. Jones.
A proof of the Gamma test.
Proc. R. Soc. Lond, 458 (2002), pp. 2759-2799
[Hassoun, 1995]
M.H. Hassoun.
Fundamentals of Artificial Neural Networks, MIT Press, (1995), pp. 511
[Hertz et al., 1991]
J. Hertz, A. Krogh, R.G. Palmer.
Introduction to the Theory of Neural Computing, Addison-Wesley, (1991), pp. 327
[Jones et al., 2007]
A.J. Jones, D. Evans, S.E. Kemp.
A note on the Gamma test analysis of noisy input/ output data and noisy time series.
Physica D, 229 (2007), pp. 1-8
[Tettamanzi and Tomassini, 2001]
A. Tettamanzi, M. Tomassini.
Soft Computing: integrating evolutionary, neural, and fuzzy systems. Springer-Verlag, (2001), pp. 328
[Husid, 1969]
L.R. Husid.
Características de terremotos-análisis general.
Revista del IDIEM, 8 (1969), pp. 21-42
[Kemp et al., 2005]
S.E. Kemp, I.D. Wilson, J.A. Ware.
A tutorial on the Gamma test.
International Journal of Simulation: Systems, Science and Technology, 6 (2005), pp. 1-2
[Kempton and Stewart, 2006]
J. Kempton, J. Stewart.
Prediction equations for significant duration of earthquake ground motions considering site and near-source effects.
Earthquake Spectra, 22 (2006), pp. 985-1013
[Mohraz and Peng, 1989]
B. Mohraz, M.H. Peng.
The use of low-pass filtering in determining the duration of strong ground motion.
Publication PVP-182, Pressure Vessels and Piping Division, ASME, (1989), pp. 197-200
[Novikova and Trifunac, 1994]
E.I. Novikova, M.D. Trifunac.
Duration of strong ground motion in terms of earthquake magnitude, epicentral distance, site conditions and site geometry.
Earthquake Eng. Struct. Dyn, 23 (1994), pp. 1023-1043
[Reinoso and Ordaz, 2001]
E. Reinoso, M. Ordaz.
Duration of strong ground motion during Mexican earthquakes in terms of magnitude, distance to the rupture area and dominant site period.
Earthquake Eng. Struct. Dyn, 30 (2001), pp. 653-673
[Sarma, 1970]
S.K. Sarma.
Energy flux of strong earthquakes.
Tectonophysics, 11 (1970), pp. 159-173
[Trifunac and Westermo, 1982]
M.D. Trifunac, B.D. Westermo.
Duration of strong earthquake shaking.
Soil Dynamics and Earthquake Engineering, 1 (1982), pp. 117-121
Copyright © 2014. Universidad Nacional Autónoma de México
Descargar PDF
Opciones de artículo