Next Article in Journal
Parallel WSAR for Solving Permutation Flow Shop Scheduling Problem
Previous Article in Journal
Maximum Multi-Commodity Flow with Proportional and Flow-Dependent Capacity Sharing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

A Generative Adversarial Network Based Autoencoder for Structural Health Monitoring †

1
Dipartimento di Ingegneria Civile ed Ambientale, Politecnico di Milano, Piazza L. da Vinci 32, 20133 Milano, Italy
2
Laboratoire MSSMat UMR 8579, Université Paris Saclay—CNRS—CentraleSupélec, 3 rue Joliot Curie, 91190 Gif-sur-Yvette, France
3
Dipartimento di Matematica, Politecnico di Milano, Piazza L. da Vinci 32, 20133 Milano, Italy
*
Author to whom correspondence should be addressed.
Presented at the 1st International Electronic Conference on Algorithms, 27 September–10 October 2021; Available online: https://ioca2021.sciforum.net/.
Comput. Sci. Math. Forum 2022, 2(1), 9; https://doi.org/10.3390/IOCA2021-10887
Published: 22 September 2021
(This article belongs to the Proceedings of The 1st International Electronic Conference on Algorithms)

Abstract

:
Civil structures, infrastructures and lifelines are constantly threatened by natural hazards and climate change. Structural Health Monitoring (SHM) has therefore become an active field of research in view of online structural damage detection and long term maintenance planning. In this work, we propose a new SHM approach leveraging a deep Generative Adversarial Network (GAN), trained on synthetic time histories representing the structural responses of both damaged and undamaged multistory building to earthquake ground motion. In the prediction phase, the GAN generates plausible signals for different damage states, based only on undamaged recorded or simulated structural responses, thus without the need to rely upon real recordings linked to damaged conditions.

1. Introduction

Bridges, power generation systems, aircrafts, buildings and rotating machinery are only few instances of structural and mechanical systems which play an essential role in the modern society, even if the majority of them are approaching the end of their original design life [1]. Taking into account that their replacement would be unsustainable from an economic standpoint, alternative strategies for early damage detection have been actively developed so to extend the basis service life of those infrastructures. Furthermore, the advent of novel materials whose long-term behaviour is still not fully understood drives the effort for effective Structural Health Monitoring (SHM), resulting in a saving of human lives and resources [1].
SHM consists of three fundamental steps: (i) measurement, at regular intervals, of the dynamic response of the system; (ii) selection of damage-sensitive features from the acquired data; (iii) statistical analysis of those attributes to assess the current health state of the structure. To characterize the damage state of a system, the method relying on hierarchical phases, originally proposed by [2] represents the currently adopted standard. The latter prescribes several consecutive identification phases (to be tackled in order), namely: check the existence of the damage, the location of the damage, its type, extent and the system’s prognosis. Damaged states are identified by comparison with a reference condition, assumed to be undamaged. The detection of the damage location relies upon a wider awareness of the structural behaviour and the way in which it is influenced by damage. This information, along with the knowledge of how the observed features are altered by different kinds of damage, allows to determine the type of damage. The last two phases require an accurate estimation of the damage mechanisms in order to classify its severity and to estimate the Remaining Useful Life (RUL).
All the steps mentioned above rely on continuous data acquisition and processing to obtain information about the current health condition of a system. In the last few years, the concept of Digital Twin has emerged, combining data assimilation, machine learning and physics-based numerical Simulations [1], the latter being essential to completely understand the physics of the structure and damage mechanisms. A suitable tool able to extract main dominant features from a set of data is represented by neural networks [3], especially generative models such as Generative Adversarial Networks (GANs) [4] and Variational Autoencoders (VAEs) [5].
In this paper, an application of the generative neural network RepGAN, proposed by [6], is presented in the context of SHM. Section 2 provides an overview on existing works. In Section 3, the application of RepGAN to Structural Health Monitoring is presented. In Section 4, extensive numerical results are illustrated, while Section 5 gathers some concluding remarks.

2. Related Work

Generative Adversarial Networks [4] are well known due to their generative capability. Given a multidimensional random variable X ( R d X , E X , P X ) , where ( R d X , E X , P X ) denotes the probabilistic space with σ -algebra E X and probability measure P X , whose samples are collected in the data set S = x ( i ) i = 1 N , with probability density function p X ( X ) , the GAN generator attempts to reproduce synthetic samples x ^ , sampled according to the probability density function p G ( X ) as similar as possible to the original data, i.e., a GAN trains over data samples in order to match p G with p X . G maps a lower dimension manifold ( R d Z , E Z , P Z ) (with d Z < d X in general) into the physics space ( R d X , E X , P X ) . In doing so, G learns to pass the critic test, undergoing the judgement of a discriminator D : R d X 0 , 1 , simultaneously trained to recognize x ^ ( i ) counterfeits. The adversarial training scheme relies on the following two-players Minimax game:
G ; D = arg min G max D V ( D , G ) V ( D , G ) = E X p X [ ln D ( X ) ] + E Z p Z [ ln ( 1 D ( G ( Z ) ) ]
In practice, G is represented by a neural network G θ and D by a neural network D ω , with trainable weights and biases θ and ω , respectively. Moreover, V ( D , G ) is approximated by the Empirical Risk function L S ( ω , θ ) , depending on the data set S , defined as:
θ ; ω = arg min θ max ω L S ( ω , θ ) = = arg min θ max ω 1 n i = 1 n ln D ω ( x ( i ) ) + ln ( 1 D ω ( G θ ( z ( i ) ) )
with z ( i ) sampled from a known latent space probability distribution p Z (for instance the normal distribution N ( 0 , I ) ). The generator G θ induces a sampling probability p G ( X ; θ ) so that, when optimized, passes the critic test, with D being unable to distinguish between x ( i ) and G θ z ( i ) (i.e., D ( x ( i ) ) = 1 2 = D ( G θ z ( i ) ). In other words, x ( i ) and G θ z ( i ) can be associated with the value of a categorical variable C, with two possible values: class “d” (data) and class “g” (generated). x ( i ) and G θ z ( i ) can be therefore sampled with the mixture probability density p M = α χ C = d + ( 1 α ) χ C = g” with χ being the indicator function and α = P ( C = d ) [7]. The optimum solution of the Minimax game in Equation (2) induces a mixture probability distribution 1 2 p C = d + p C = g [4]. The saddle point of V ( D , G ) corresponds to the minimum (with the respect to to D) of the conditional Shannon’s entropy S ( C | X ) (see Appendix A). Moreover, minimizing the conditional Shannon’s entropy S ( C | X ) corresponds to the maximization of the Mutual Information I ( X , C ) = S ( C ) S ( C | X ) (see Appendix B), i.e., it corresponds to extract X samples x ( i ) or x ^ ( i ) that are indistinguishable (belonging to same class), with an uninformative mapping X C .
GANs proved useful in various applications such as generation of artificial data for data-set augmentation, filling gaps in corrupted images and image processing. Especially, deep convolutional generative adversarial networks (DCGANs) [8] proved useful in the field of unsupervised learning. SHM could benefit from GANs as they improve the generalisation performance of models, extracting general features from data, as well as their semantics (damage state, frequency content, etc). However, the adversarial training scheme in Appendix C does not grant a bijective mapping G θ Z : Z X (decoder) and F θ X : X Z (encoder), which is crucial in order to obtain a unique representation of the data into the latent manifold. Autoencoders have been developed for image reconstruction so to learn the identity operator x ^ ( i ) = I ( x ( i ) ) = G θ Z F θ X ( x ( i ) ) . One can leverage the encoder F θ X representation power to sample points z ^ ( i ) = F θ X ( x ( i ) ) belonging to the latent manifold Ω Z and the decoder G θ Z to sample points x ^ ( i ) = G θ Z ( z ( i ) ) belonging to the latent manifold Ω X (see Equation (1)). In order to make the learning process of GANs stable across a range of data-sets and to realize higher resolution and deeper generative models, Convolutional Neural Networks (CNNs) are employed to define F θ X , G θ Z and the discriminators. F θ X and G θ Z induce sampling probability density functions q Z | X = q X Z p X and p X | Z = p X Z p Z , respectively. p X is usually unknown (depending on the data-set at stake), but p Z can be chosen ad hoc (such as, for instance, N ( 0 , I ) ) in order to get a powerful generative tool for realistic data samples x ^ ( i ) . A particular type of Autoencoders, called Variational Autoencoders (VAEs) was introduced by [5], consisting in a probabilistic and generative version of the standard Autoencoder, where the encoder F θ X infers the mean μ Z and variance σ Z 2 of the latent manifold. However, the main contribution provided by VAEs is the straightforward approach that allows to reorganize the gradient computation and reduce variance in the gradients labelled reparametrization trick.
Adversarial Autoencoders (AAEs) [9] employ the adversarial learning framework in Equation (1), replacing G θ Z ( z ( i ) ) by G θ Z F θ X ( x ( i ) ) and adding to the adversarial GAN loss the Mean Square Loss x ( i ) G θ Z F θ X ( x ( i ) ) 2 as an optimization penalty, in order to assure a good reconstruction of the original signal. However, AAEs do not assure a bijective mapping between ( R d X , E X , P X ) and ( R d Z , E Z , P Z ) . In order to achieve the bijection (in a probabilistic sense) between x , z ^ and x ^ , z samples, the distance between the joint probability distributions q X Z ^ = q Z ^ | X p X and p X ^ Z = p X ^ | Z p Z [10], with the posteriors q Z ^ | X and p Z | X ^ must be minimized. A suitable distance operator for probability distributions is the so called Jensen–Shannon distance D J S ( q X Z ^ | | p X ^ Z ) , defined as [10]:
D J S ( q X Z ^ | | p X ^ Z ) = D K L ( q X Z ^ | | p M ) + D K L ( p X ^ Z | | p M ) 2 = S ( p M ) S ( X , Z | M )
with D K L ( p | | q ) = S ( p | | q ) S ( p ) being the Kullback–Leibler divergence (see Appendix B) and p M = q X Z ^ + p X ^ Z 2 being the mixture probability distribution [7], i.e., the probability of extracting X , Z ^ or X ^ , Z from a mixed data set, with α = P ( C = d ) = 1 2 and the entropy of the mixture probability S ( M ) = ln 2 . D J S ( q X Z ^ | | p X ^ Z ) can be rewritten as:
D J S ( q X Z ^ | | p X ^ Z ) = S ( p M ) 1 2 S ( q X Z ^ ) + S ( p X ^ Z ) = = S ( X , Z ) S ( X , Z | M ) = S ( M ) S ( M | X , Z )
The adversarial optimization problem expressed in Equation (1) can be seen as a minimization of the Jensen–Shannon distance for C d , g :
D J S ( q X Z ^ | | p X ^ Z ) + ln 2 = S ( M | X , Z ) = = 1 2 E X , Z ^ q X Z D X , Z ^ + 1 2 E X ^ , Z p X Z 1 D X ^ , Z
that can be combined with the Autoencoder model in order to obtain the following expression [10,11]:
D J S ( q X Z ^ | | p X ^ Z ) + ln 2 = D J S ( q Z ^ | X p X | | p X ^ | Z p Z ) + ln 2 = = 1 2 E X p X D X , F θ X ( X ) + E Z P Z 1 D G θ Z ( Z ) , Z
In this context, F θ X learns to map data into a disentangled latent space, generally following the normal distribution, a good reconstruction is not ensured unless the cross-entropy between X and Z is minimized too [12].
Another crucial aspect of generative models is the semantics of the latent manifold. Most of the standard GAN models trained according to Equation (1) employs a simple factored continuous input latent vector Z and does not enforce any restrictions on the way the generator treats it. The individual dimensions of Z do not correspond to semantic features of the data (uninformative latent manifolds) and Z cannot be effectively used in order to perform meaningful topological operations in the latent manifold (e.g., describing neighborhoods) and to associate meaningful labels to it. An information-theoretic extension to GANs, called InfoGAN [13] is able to learn a meaningful and disentangled representations in a completely unsupervised manner: a Gaussian noise Z is associated with a latent code C to capture the characteristic features of the data distribution (for classification purposes). As a consequence, the generator becomes G θ Z ( Z , C ) and the corresponding probability distribution p G , whose Mutual Information with the respect to to the latent codes C , namely I ( C , G θ Z ( Z , C ) ) . The latter is forced to be high, penalizing the GAN loss in Equation (1) with the variational lower bound L I (G,Q), defined by:
L I ( G , Q ) = E C p C , X p G ln Q ( C | X ) + S ( C ) = E X p G E C p C | X ln q C | X + S ( C )
with q C | X being the probability distribution approximating the real unknown posterior probability distribution p C | X (and represented by the neural network Q Z ). L I ( G , Q ) can be easily approximated via Monte Carlo simulation, and maximized with the respect to to q C | X and p G via reparametrization trick [13].
V InfoGAN ( D , G , Q ) = V ( D , G ) λ L I ( G , Q )

3. Methods

With the purpose of learning a semantically meaningful and disentangled representation of the SHM time-histories, we adopted in this study the architecture called RepGAN, originally proposed in [6]. RepGAN is based on an encoder-decoder structure (both represented by deep CNNs made of stacked 1D convolutional blocks), with a latent space Z = C , S , N . C 0 , 1 d C a categorical variable representing the damage class(es), with C p C which is generally chosen as a categorical distribution over d C classes, i.e., p C = Cat ( d C ) . S R d S is a continuous variable of dimension d S , with S p S , generally p S = N ( 0 , I ) or the uniform distribution p S = U ( 1 , 1 ) . Finally, N R d N is a random noise of d N independent components, with N p N , generally p N N ( 0 , I ) . RepGAN adopts the conceptual frameworks of VAEs and InfoGAN, combining the learning of two representations x z ^ x ^ and z x ^ z ^ , respectively. The x z ^ x ^ scheme must learn to map multiple data instances x ( i ) into their images (via encoder F θ X ) in a latent manifold z ^ ( i ) = F θ X ( x ( i ) ) and back into a distinct instance in data space x ^ ( i ) = G θ Z F θ X ( x ( i ) ) (via decoder G θ Z ), providing satisfactory results in reconstruction. z x ^ z ^ maps multiple data latent instances into the same data representation, in order to guarantee impressive generation and clustering performance. Combining the two surjective mappings, in RepGAN the two learning tasks x z ^ x ^ and z x ^ z ^ are performed together with shared parameters in order to obtain a bijective mapping x z . In practice, the training of z x ^ z ^ is iterated five times more than the x z ^ x ^ . This ability to learn a bidirectional mapping between the input space and the latent space is achieved through a symmetric adversarial process. The Empirical Loss function can be written as:
L S = D J S ( p X ^ | ( C , S , N ) | | p X ) + D J S ( q C ^ | X | | p C ) + D J S ( q S ^ | X | | p S ) + D J S ( q N ^ | X | | p N ) E p C E p X ^ | C ln q C ^ | X E p S E p X ^ | S ln q S ^ | X E p X E q ( C , S , N ) | X ln p X | ( C , S , N )
with the terms:
  • E p C E p X ^ | C ln q C ^ | X minimizing the conditional entropy S ( C | X ) ;
  • E p S E p X ^ | S ln q S ^ | X minimizing the conditional entropy S ( S | X ) .
are introduced in order to constrain a deterministic and injective encoding mapping (see Appendix B). On the other hand, the term
  • E p X [ E q ( C , S , N ) | X ln p X | ( C , S , N ) ] .
penalizes the learning scheme, in order to minimize the conditional entropy S ( X | ( C , S , N ) ) , i.e., in order to grant a good reconstruction.
Following the original RepGAN formulation:
  • E p X E q ( C , S , N ) | X ln p X | ( C , S , N ) is enforced penalizing the L 1 -norm | X G θ Z F θ X ( X ) | ;
  • E p S E p X ^ | S ln q S ^ | X corresponds to the InfoGAN L I penalty, and it is maximized via the reparametrization trick (structuring the S branch of the encoder-decoder structure as a VAE, see [5]).
Finally, E p C E p X ^ | C ln q C ^ | X is maximized in a supervised way, considering the actual class of labeled signals x ( i ) : x d ( i ) corresponding to a damaged structure and x u ( i ) to an undamaged one, respectively. RepGAN provides an informative and disentangled latent space associated with the damage class C . The most significant aspect of the approach is the efficiency in generating reasonable signals for different damage states only on the basis of undamaged recorded or simulated structural responses. Both generators F θ X , G θ Z and discriminators D ω X , D ω C , D ω S and D ω N are parametrized via 1D CNN (and strided 1D CNN), following [8]. Our RepGAN model has been designed using the Keras API, and trained employing a Nvidia Tesla K40 GPU (on the supercomputer Ruche, the cluster of the Mésocentre Moulon of Paris Saclay University).

4. Results and Discussion

In the following, a case study is considered in order to prove the ability of the new architecture to achieve the three fundamental tasks of semantic generation, clustering and reconstruction. The reference example is a shear building subject to an earthquake ground motion whose signals are taken from the STEAD seismic database [14]. STEAD [14] is a high-quality, large-scale, and global data set of local earthquake and non-earthquake signals recorded by seismic instruments. In this work, local earthquake wave forms (recorded at local distances within 350 km of earthquakes) have been considered. Seismic data are constituted by three wave forms of 60 s duration, recorded in east–west, north–south, and up-dip directions, respectively. The structure is composed of 39 storeys. The mass and the stiffness of each floor, in undamaged conditions, are, respectively, m = 625 × 10 3 kg and k = 1 × 10 9   k N m . Damage is simulated through the degradation of stiffness. In the present case, the stiffness reduction has been set equal to 50% of the above mentioned value. The structural response of the system is evaluated considering one degree-of-freedom (dof) per floor. To take into account damping effects, a Rayleigh damping model has been considered.
The following results have been obtained considering 100 signals in both undamaged and damaged conditions for a total of 200 samples, with separated training and validation data sets. Each signal is composed of 2048 time steps with dt = 0.04 s. The training process has been performed over 2000 epochs. The reconstruction capability of the proposed network has been evaluated through the Goodness-of-Fit (GoF) criteria [15] where both the fit in Envelope (EG) and the fit in Phase (FG) are measured. An example is shown in Figure 1. The values 9.17 and 9.69, respectively, related to EG and PG testify the excellent reconstruction quality.
The capability of reproducing signals for different damage scenarios can be appreciated from Figure 2 which presents the original structural response (black) and the corresponding generated one (orange) in both undamaged (left panel in Figure 2) and damaged (right panel in Figure 2) conditions. Regarding the classification capability, the classification report and the confusion matrix in Figure 3 highlight the fact that the model is able to correctly assign the damage class to the considered time histories.

5. Conclusions

In this paper, we introduce a SHM method based on a deep Generative Adversarial Network. Trained on synthetic time histories that represent the structural response of a multistory building in both damaged and undamaged conditions, the new model achieves high classification accuracy (Figure 3) and satisfactory reconstruction quality (Figure 1 and Figure 2), resulting in a good bidirectional mapping between the input space and the latent space. However, the major innovation of the proposed method is the ability to generate reasonable signals for different damage states, based only on undamaged recorded or simulated structural responses. As a consequence, real recordings linked to damaged conditions are not requested. In our future work, we would like to extend our approach to real-time data. We will further consider a dataset constituted by a far larger number of time histories.

Author Contributions

Conceptualization, G.C., L.R., F.G., S.M. and A.C.; methodology, G.C. and F.G.; software, G.C., L.R. and F.G.; validation, G.C. and F.G.; formal analysis, G.C. and F.G.; investigation, G.C. and F.G.; resources, F.G.; data curation, G.C., L.R. and F.G.; writing–original draft preparation, G.C., L.R., M.T., F.G., S.M., A.M. and A.C.; writing–review and editing, G.C. and F.G.; visualization, G.C. and F.G.; supervision, F.G., S.M. and A.C.; project administration, F.G. and A.C.; Funding acquisition, G.C., F.G. and A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated during the study are available from the corresponding author upon reasonable request.

Acknowledgments

The training and testing of the neural network has been performed exploiting the supercomputer resources of the Mésocentre Moulon (http://mesocentre.centralesupelec.fr, last accessed 14 February 2022), the cluster of CentraleSupélec and ENS Paris-Saclay, hosted within the Paris-Saclay University and funded by the Contrat Plan État Région (CPER). This work has been developed thanks to the scholarship “Tesi all’estero—a.y. 2020/2021—second call” funded by Politecnico di Milano.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Shannon’s Entropy

  • Shannon’s entropy for a probability density function p X :
    S ( X ) = S p X = E X p X ln 1 p X = E X p X ln p X 0
  • Conditional Shannon’s entropy for X and Z:
    S X | Z = E Z P Z S p X | Z = E ( X , Z ) p X Z ln 1 p X | Z
    S ( X , Z ) = S ( Z | X ) + S ( X ) = S ( X | Z ) + S ( Z )
  • Cross-entropy:
    S ( p X Z | | q X Z ) = E ( X , Z ) p X Z ln 1 q X Z = E X p X E Z p Z | X ln 1 q X Z
  • Given a data set of identically independent distributed (i.i.d.) samples S = x ( i ) i = 1 N , the true yet unknown probability p X of extracting an instance x ( i ) can be approximated by the likelihood p θ X x ( i ) i = 1 N , whose entropy is
    S ( p θ X ) = ln p θ X x ( i ) i = 1 N = i N ln p θ X x ( i )

Appendix B. Kullback–Leibler Divergence

  • Kullback-Liebler divergence (non-symmetric):
    D K L ( p X Z | | q X Z ) = E ( X , Z ) p X Z ln p X Z q X Z = S ( p X Z ) + S ( p X Z | | q X Z ) S ( p X Z | | q X Z )
    D K L ( p X Z ] | | q X Z ) + S ( X ) = S ( X | Z ) E X p X S ( p Z | X ) + S ( p X Z | | q X Z ) S ( p X Z | | q X Z )
    D K L ( p X Z | | q X Z ) + S ( X ) D K L ( p X Z | | q X Z ) ]
  • Mutual Information between X and X | Z :
    I ( X , Z ) = S ( X ) S ( X | Z ) 0
    If p X | Z = p X ( ( X , Z ) are independent) then I ( X , Z ) = 0 . If p X | Z = δ ( Z f ( X ) ) with f deterministic, then I ( X , Z ) = max X , Z I ( X , Z ) = S ( X ) .
  • S ( Z | X ) = E Z p Z E X p X | Z ln p Z | X = = E X p X E Z p Z | X ln p Z | X q Z | X E Z p Z E X p X | Z ln q X | Z = = E X p X D K L p Z | X | | q Z | X E Z p Z E X p X | Z ln q X | Z E Z p Z E X p X | Z ln q X | Z

Appendix C. Generative Adversarial Networks (GAN)

  • Given X belonging to the probabilistic space ( Ω X , E X , Pr X ) with class C d , g (“d” corresponding to data and “g” to generated, and a discriminator D : Ω X 0 , 1 acting as an expert/critic:
    -
    P ( C = d ) = α ; P ( C = g ) = 1 α
    -
    P ( C = d | x ( i ) ) = D ( x ( i ) )
    -
    P ( C = d | x ( i ) ) = 1 D ( G ( z ( i ) ) )
    S ( C | X ) = E X p X E C p C | X ln p C | X = E C p C E X p X | C ln p C | X
    S ( C | X ) = α E X p X | C = d ln p C = d | X ( 1 α ) E X p X | C = g ln p C = g | X
    S ( C | X ) = α E X p X ln D X ( 1 α ) E Z p Z ln 1 D G ( Z )
    For tuneable conditional probability distributions D ω :
    max I ( X , C ) S ( C ) + max S ( C | X ) = S ( C ) + min S ( C | X )
    max I ( X , C ) S ( C ) + min G max D α E X p X ln D X + ( 1 α ) E Z p Z ln 1 D G ( Z )
    Thus, minimizing S + min G max D S ( C | X ) represents an upper bound for the Mutual Information between C and X , which is maximized by maximizing S ( C | X ) . For an optimum training, D must not be able to discriminate between x ( i ) and x ^ ( i ) , therefore α = 1 2 .

Appendix D. Standard Autoencoder

In the standard Autoencoder formulations [16,17], F and G are trained by maximizing I ( X , Z ) , namely:
F , G = arg max F , G I ( X , Z ) = arg min F , G H ( X | Z ) = arg min F , G E X p X E Z q Z | X ln 1 p X | Z
If the encoder and decoder are parametrized as neural networks, respectively, as F θ X and G θ Z , the AE loss can be approximated by the Empircal Loss:
θ X , θ Z = arg max θ X , θ Z i = 1 N ln p X | Z x ( i ) | Z = F θ X x ( i )
Given the fact that the Gaussian distribution has maximum entropy relative to all probability distributions covering the entire real line, the Empirical Loss in Equation (A2) can be maximized by the Empirical Loss with p X | Z = N G θ Z Z , σ 2 I :
θ X , θ Z = arg max θ X , θ Z i = 1 N 1 2 σ 2 x ( i ) G θ Z F θ X x ( i ) 2 + d Z 2 ln 2 π σ 2

References

  1. Farrar, C.R.; Worden, K. Structural Health Monitoring: A Machine Learning Perspective; Wiley: Oxford, UK, 2013. [Google Scholar] [CrossRef] [Green Version]
  2. Rytter, A. Vibrational Based Inspection of Civil Engineering Structures. Ph.D. Thesis, University of Aalborg, Aalborg, Denmark, 1993. [Google Scholar]
  3. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  4. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 28th Annual Conference on Neural Information Processing Systems (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar] [CrossRef]
  5. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  6. Zhou, Y.; Gu, K.; Huang, T. Unsupervised Representation Adversarial Learning Network: From Reconstruction to Generation. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019. [Google Scholar] [CrossRef] [Green Version]
  7. Lindsay, B.G. Mixture Models: Theory, Geometry and Applications. In NSF-CBMS Regional Conference Series in Probability and Statistics; Institute of Mathematical Statistics and American Statistical Association: Suitland, MD, USA, 1995; Volume 5, p. 163. ISBN 0-940600-32-3. [Google Scholar]
  8. Radford, A.; Luke, L.M.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  9. Makhzani, A.; Shlens, J.; Jaitly, N.; Goodfellow, I. Adversarial Autoencoders. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  10. Dumoulin, V.; Belghazi, I.; Poole, B.; Mastropietro, O.; Lamb, A.; Arjovsky, M.; Courville, A. Adversarially Learned Inference. arXiv 2016, arXiv:1606.00704v3. [Google Scholar]
  11. Donahue, J.; Krähenbühl, P.; Darrell, T. Adversarial Feature Learning. arXiv 2017, arXiv:1605.09782. [Google Scholar]
  12. Li, C.; Liu, H.; Chen, C.; Pu, Y.; Chen, L.; Henao, R.; Carin, L. ALICE: Towards understanding adversarial learning for joint distribution matching. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5501–5509. [Google Scholar]
  13. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  14. Mousavi, S.M.; Sheng, Y.; Zhu, W.; Beroza, G.C. STanford EArthquake Dataset (STEAD): A Global Data Set of Seismic Signals for AI. IEEE Access 2019, 7, 179464–179476. [Google Scholar] [CrossRef]
  15. Kristekova, M.; Kristek, J.; Moczo, P. Time-frequency misfit and goodness-of-fit criteria for quantitative comparison of time signals. Geophys. J. Int. 2009, 178, 813–825. [Google Scholar] [CrossRef]
  16. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.-A. Extracting and Composing Robust Features with Denoising Autoencoders. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1096–1103. [Google Scholar]
  17. Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.-A. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
Figure 1. Time–Frequency Goodness-of-Fit criterion: the black line represents the original time-histories x ( i ) while the red time history depicts the result of the RepGAN reconstructions G Z F X ( x ( i ) ) . GoF is evaluated between 0 and 10: the higher the score, the better is the reconstruction. Frequency Envelope Goodness (FEG), Time–Frequency Envelope Goodness (EG), Time Envelope Goodness (TEG), Frequency Phase Goodness (FPG), Time–Frequency Phase Goodness (PG) and Time Phase Goodness (TPG).
Figure 1. Time–Frequency Goodness-of-Fit criterion: the black line represents the original time-histories x ( i ) while the red time history depicts the result of the RepGAN reconstructions G Z F X ( x ( i ) ) . GoF is evaluated between 0 and 10: the higher the score, the better is the reconstruction. Frequency Envelope Goodness (FEG), Time–Frequency Envelope Goodness (EG), Time Envelope Goodness (TEG), Frequency Phase Goodness (FPG), Time–Frequency Phase Goodness (PG) and Time Phase Goodness (TPG).
Csmf 02 00009 g001
Figure 2. Examples of reconstructed signals for undamaged (left) and damaged (right) time-histories. The black lines represent the original time-histories x u ( i ) and x d ( i ) , respectively. The orange time histories represent the result of the RepGAN reconstructions G Z F X ( x u ( i ) ) and G Z F X ( x d ( i ) ) , respectively. The proposed examples represent the normalized displacement of the 1st floor of the building in object.
Figure 2. Examples of reconstructed signals for undamaged (left) and damaged (right) time-histories. The black lines represent the original time-histories x u ( i ) and x d ( i ) , respectively. The orange time histories represent the result of the RepGAN reconstructions G Z F X ( x u ( i ) ) and G Z F X ( x d ( i ) ) , respectively. The proposed examples represent the normalized displacement of the 1st floor of the building in object.
Csmf 02 00009 g002
Figure 3. Evaluation of the classification ability of the model. On the left panel, precision, recall, f1-score and accuracy values are reported. A precision score of 1.0 for a class C means that every item labelled as belonging to class C does indeed belong to class C, whereas a recall of 1.0 means that every item from class C was labelled as belonging to class C. F1-score is the harmonic mean of the precision and recall. Accuracy represents the proportion of correct predictions among the total number of cases examined. On the right panel, the confusion matrix allows to visualize the performance of the model: each row of the matrix represents the instances in the actual class, while each column depicts the instances in the predicted class.
Figure 3. Evaluation of the classification ability of the model. On the left panel, precision, recall, f1-score and accuracy values are reported. A precision score of 1.0 for a class C means that every item labelled as belonging to class C does indeed belong to class C, whereas a recall of 1.0 means that every item from class C was labelled as belonging to class C. F1-score is the harmonic mean of the precision and recall. Accuracy represents the proportion of correct predictions among the total number of cases examined. On the right panel, the confusion matrix allows to visualize the performance of the model: each row of the matrix represents the instances in the actual class, while each column depicts the instances in the predicted class.
Csmf 02 00009 g003
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Colombera, G.; Rosafalco, L.; Torzoni, M.; Gatti, F.; Mariani, S.; Manzoni, A.; Corigliano, A. A Generative Adversarial Network Based Autoencoder for Structural Health Monitoring. Comput. Sci. Math. Forum 2022, 2, 9. https://doi.org/10.3390/IOCA2021-10887

AMA Style

Colombera G, Rosafalco L, Torzoni M, Gatti F, Mariani S, Manzoni A, Corigliano A. A Generative Adversarial Network Based Autoencoder for Structural Health Monitoring. Computer Sciences & Mathematics Forum. 2022; 2(1):9. https://doi.org/10.3390/IOCA2021-10887

Chicago/Turabian Style

Colombera, Giorgia, Luca Rosafalco, Matteo Torzoni, Filippo Gatti, Stefano Mariani, Andrea Manzoni, and Alberto Corigliano. 2022. "A Generative Adversarial Network Based Autoencoder for Structural Health Monitoring" Computer Sciences & Mathematics Forum 2, no. 1: 9. https://doi.org/10.3390/IOCA2021-10887

APA Style

Colombera, G., Rosafalco, L., Torzoni, M., Gatti, F., Mariani, S., Manzoni, A., & Corigliano, A. (2022). A Generative Adversarial Network Based Autoencoder for Structural Health Monitoring. Computer Sciences & Mathematics Forum, 2(1), 9. https://doi.org/10.3390/IOCA2021-10887

Article Metrics

Back to TopTop