Abstract
Structural monitoring for complex built environments often suffers from mismatch between design, laboratory testing, and actual built parameters. Additionally, real-world structural identification problems encounter many challenges. For example, the lack of accurate baseline models, high dimensionality, and complex multivariate partial differential equations (PDEs) pose significant difficulties in training and learning conventional data-driven algorithms. This paper explores a new framework, dubbed NeuralSI, for structural identification by augmenting PDEs that govern structural dynamics with neural networks. Our approach seeks to estimate nonlinear parameters from governing equations. We consider the vibration of nonlinear beams with two unknown parameters, one that represents geometric and material variations, and another that captures energy losses in the system mainly through damping. The data for parameter estimation is obtained from a limited set of measurements, which is conducive to applications in structural health monitoring where the exact state of an existing structure is typically unknown and only a limited amount of data samples can be collected in the field. The trained model can also be extrapolated under both standard and extreme conditions using the identified structural parameters. We compare with pure data-driven Neural Networks and other classical Physics-Informed Neural Networks (PINNs). Our approach reduces both interpolation and extrapolation errors in displacement distribution by two to five orders of magnitude over the baselines. Code is available at https://github.com/human-analysis/neural-structural-identification.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
- Neural differential equations
- Structural system identification
- Physics-informed machine learning
- Structural health monitoring
1 Introduction
Structural-system identification (SI) [3, 15, 22, 35, 39, 41] refers to methods for inverse calculation of structural systems using data to calibrate a mathematical or digital model. The calibrated models are then used to either estimate or predict the future performance of structural systems and, eventually, their remaining useful life. Non-linear structural systems with spatial and temporal variations present a particular challenge for most inverse identification methods [4, 14, 21]. In dynamic analysis of civil structural systems, prior research efforts primarily focused on matching experimental data with either mechanistic models (i.e., known mechanical models) [31, 38] or with black box models with only input/output information (i.e., purely data-driven approaches), [10, 13, 34]. Examples of these approaches include eigensystem identification algorithms [37], frequency domain decomposition [7], stochastic optimization techniques [30], and sparse identification [8]. A majority of these approaches, however, fail to capture highly non-linear behaviors.
In this paper, we consider the class of non-linear structural problems with unknown spatially distributed parameters (see Fig. 1 for an overview). The parameters correspond to geometric and material variations and energy dissipation mechanisms, which could be due to damping or other system imperfections that are not typically captured in designs. As an instance of this problem class, we consider forced vibration responses in beams with spatially varying parameters. The primary challenges in such problems arise from the spatially variable nature of the properties and the distributed energy dissipation. This is typical for built civil structures, where energy dissipation and other hard-to-model phenomena physically drive the dynamic response behavior. In addition, it is very common to have structural systems with unknown strength distributions, which can be driven by geometric non-linearities or indiscernible/hidden material weaknesses. Finally, a typical challenge in structural systems is the rarity of measured data, especially for extreme loading cases.
We propose a framework, dubbed NeuralSI, for nonlinear dynamic system identification that allows us to discover the unknown parameters of partial differential equations from measured sensing data. The developed model performance is compared to conventional PINN methods and direct regression models. Upon estimating the unknown system parameters, we apply them to the differential model and efficiently prognosticate the time evolution of the structural response. We also investigate the performance of NeuralSI under a limited training data regime across different input beam loading conditions. This replicates the expected challenges in monitoring real structures with limited sensors and sampling capabilities.
NeuralSI contributes to the fields of NeuralPDEs, structural identification, and health monitoring:
-
1.
NeuralSI allows us to learn unknown parameters of fundamental governing dynamics of structural systems expressed in the form of PDEs.
-
2.
We demonstrate the utility of NeuralSI by modeling the vibrations of nonlinear beams with unknown parameters. Experimental results demonstrate that NeuralSI achieves two-to-three orders of magnitude lower error in predicting displacement distributions in comparison to PINN-based baselines.
-
3.
We also demonstrate the utility of NeuralSI in temporally extrapolating displacement distribution predictions well beyond the training data measurements. Experimental results demonstrate that NeuralSI achieves four-to-five orders of magnitude lower error compared to PINN-based baselines.
2 Related Work
Significant efforts have been directed toward physics-driven discovery or approximation of governing equations [15, 21, 26]. Such studies have further been amplified by the rapid development of advanced sensing techniques and machine learning methods [16, 17, 19, 32]. Most of the work to date has mainly focused on ordinary differential equation systems [21, 40]. Neural ODEs [9] have been widely adopted due to their capacity to learn and capture the governing dynamic behavior from directly collected measurements [2, 28, 40]. They represent a significant step above the direct fitting of a relation between input and output variables. In structural engineering applications, Neural ODEs generally approximate the time derivative of the main physical attribute through a neural network.
More recently, data-driven discovery algorithms for the estimation of parameters in differential equations are introduced. These methods typically referred to as physics-informed neural networks (PINNs) include differential equations, constitutive equations, and initial and boundary conditions in the loss function of the neural network and adopt automatic differentiation to compute derivatives of the network parameters [20, 29]. Variational Autoencoders were also learned to build baseline behavioral models, which were then used to detect and localize anomalies [23]. Many other applications have employed Neural ODE for dynamic structure parameter identification in both linear and nonlinear cases [2, 21, 40]. On the other hand, few studies have explored Neural PDEs in other fields such as message passing [6], weather and ocean wave data [11], and fluid dynamics [5]. In [18], a Graph Neural Network was used to solve flow phenomena, and a NeuralPDE solver package was developed in Julia [42] based on PINN.
3 Structural Problem – PDE Derivation
3.1 Problem Description
Many physical processes in engineering can be described as fourth-order time-dependent partial differential problems. Examples include the Cahn-Hilliard type equations in Chemical Engineering, the Boussinesq equation in geotechnical engineering, the biharmonic systems in continuum mechanics, the Kuramoto-Sivashinsky equation in diffusion systems [27] and the Euler-Bernoulli equation considered as an example case study in this paper. The Euler-Bernoulli beam equation is widely used in civil engineering to estimate the strength and deflection of beam structures. The dynamic beam response is defined by:
where u(x, t) is the displacement as a function of space and time. P(x) and \(E_0\) are the modulus coefficient and the reference modulus value of the beam, \(I, \rho \), and A are refereed to the beam geometry and density. F is the distributed force applied to the beam. C(x) represents damping, which is related to energy dissipation in the structure. In this paper, we restrict ourselves only to spatial variation of the beam’s properties and leave the most generalized case with variations in space and time of all variables for a future study.
The fourth-order derivative of the spatial variable and the second-order derivative of time describes the relation between the beam deflection and the load on the beam [1]. Figure 2 shows an illustration of the beam problem considered here, with the deflection u(x, t) as the physical response of interest. The problem can also be formulated as a function of moments, stresses, or strains. The deflection formulation presents the highest order differentiation in the PDE. This was selected to allow for flexibility of the solution to be extended to other applications beyond structural engineering.
To accurately represent the behavior of a structural component, its properties need to be identified. Though the beam geometry is straightforward to measure, the material property and damping coefficient are hard to estimate. The beam reference modulus \(E_0\) is expected to have an estimated range based on the choice of material (e.g., steel, aluminum, composites, etc.) but unforeseen weaknesses in the build conditions can introduce unexpected nonlinear behavior. One of the objectives of this work is to capture this indiscernible randomness from response measurements. In addition, as discussed above, the damping is unpredictable at the design stage and is usually calculated by experiments. For the simply supported beam problem, the boundary conditions are defined as:
where L is the length of the beam. Initially, the beam is static and stable, so the initial conditions of the beam are:
4 NeuralSI
4.1 Discretization of Space
To tackle this high-order PDE efficiently, a numerical approach based on the method of lines is employed to discretize the spatial dimensions of the PDE. Then the system is solved as a system of ordinary differential equations (ODEs). The implemented discretization for the spatial derivatives of different orders are expressed as:
where in the fourth order discretization, \(A_4^*\) is a \(N \times N\) modified band matrix (based on the boundary conditions), and the size depends on the number of elements used for the space discretization, and \(\varDelta x\) is the distance between the adjacent elements discretized in the spatial domain. A similar principle is applied for other order derivatives.
4.2 The Proposed NeuralSI Schematic
A pictorial schematic of NeuralSI is shown in Fig. 1. The Julia differential equation package [28] allows for very efficient computation of the gradient from the ODE solver. This makes it feasible to be used for neural network backpropagation. Thus, the ODE solver can be considered as a neural network layer after defining the ODE problem with the required fields of initial conditions, time span, and any extra parameters. Inputs to this layer can either be output from the previous network layers or directly from the training data.
The network in NeuralSI for the beam problem takes as input the location of the deformation sensors installed on the structure for continuous monitoring of its response. A series of dense layers are implemented to produce the output, which are the parameters that represent the structural characteristics. The parameters are re-inserted into the pre-defined ODE to obtain the final output, i.e., the structure’s dynamic response. The loss is determined by the difference between the dynamic responses predicted by NeuralSI and those measured by the sensors (ground truth).
4.3 Training Data Generation
For experimental considerations in future lab testing, we simulate in this case a beam with length, width, and thickness respectively of 40 cm, 5 cm, and 0.5 cm. The density \(\rho \) is \(2700\,\text {kg}/\text {m}^3\) (aluminum as base material). The force F(t) is defined as a nonlinear temporal function. Considering the possible cases of polynomial or harmonic material properties variations as an example [33], we integrate the beam with a nonlinear modulus E(x) as a sinusoidal function. We use a range for the modulus from 70 GPa to 140 GPa (again using aluminum as a base reference). The damping coefficient C(x) is modeled as a ramp function. The PDE can be rewritten and expressed as:
where the estimated modulus reference \(E_0\) is 70 GPa, and P(x) and C(x) are modulus coefficient and damping that can vary spatially with x. The pre-defined parameters \(P_0 (x)\) and \(C_0 (x)\) are shown in Fig. 3.
The PDE presented in (5) is solved via the differential equation package in Julia. The RK4 solver method is selected for this high-order PDE. The time span was set to 0.045 s to have 3 complete oscillations of the bending response. The number of spatial elements and time steps are chosen as 16 and 160 respectively for balancing the training time cost and response resolution (capture the peak deflections). The deflections u(x, t) are presented as a displacement distribution of size \(16\times 160\), from which ground truth data is obtained for training.
4.4 Network Architecture and Training
The network architecture is presented as a combination of multiple dense layers and an PDE-solver layer. The input to the network is the spatial coordinates x for the measurements, and the network output is the prediction of the dynamic response u(x, t). It is worth mentioning that the structural parameters P and C are produced from the multiple dense layers in separate networks, and the PDE layer takes those parameters to generate a response displacement distribution of size \(16\times 160\). The activation function for predicting the parameter P is a linear scale of the sigmoid function so that the output can be in a reasonable range. For the prediction of parameter C, the network of the same architecture is used, but the last layer does not take any activation function since the range of the damping value is unknown.
The modulus coefficient might be very high during training and lead to erroneous predictions with very high-frequency oscillations. So, we used minibatch training to escape local minima with a batch size of 16. The loss function is defined as the mean absolute error (MAE) between samples from the predicted and ground truth displacement distribution:
where n is the number of samples for training, u and \(\hat{u}\) are the values from true and prediction dynamic responses at different training points in the same minibatch.
Furthermore, inspired by the effectiveness of positional embeddings for representing spatial coordinates in transformers [36], we adopt the same as well as for the spatial input to the network. It is worth noting that the temporal information in the measurements is only used as an aid for mapping and matching the predictions with the ground truth. We use ADAMW [24] as our optimizer, with a learning rate of 0.01 (Fig. 4).
5 Results and Performance
The evaluation of NeuralSI is divided into two parts. In the first part, we evaluate predictions of the parameters P and C from the trained neural network. We assume that each structure has a unique response. To determine how well the model is predicting the parameters, Fréchet distance [12] is employed to estimate the similarity between the ground truth and predicted functions. In this case, the predicted P and C are compared to the original \(P_0\) and, \(C_0\) respectively.
The second part of our evaluation is the prediction of the dynamic responses, which is achieved by solving the PDE using the predicted parameters. The metric to determine the performance of the prediction is the mean average error (MAE) between the predicted and ground truth displacement distribution. The prediction can be extrapolated by solving the PDE for a longer time span and compared with the extrapolated ground truth. The MAE is also calculated from the extrapolated data to examine the extrapolation ability of NeuralSI. Moreover, the dynamic response can be visualized on different elements separately (i.e., separate spatial locations x) for a more fine-grained comparison of the extrapolation results.
5.1 Results
We first trained and evaluated NeuralSI with different combinations of number and size of dense layers, percentage of data used for training, and minibatch size. The best results were achieved by taking a minibatch size of 16, training for a total of 20 epochs, and a learning rate of 0.001 (the first 10 epochs has a learning rate of 0.01).
Figure 5 shows the output of modulus coefficient P and damping C from NeuralSI. For the most part, the predictions match well with the target modulus and damping, respectively. Compared to the modulus coefficient P, the predicted damping C has a larger error since it is less sensitive to the response. A small difference in damping magnitude will not affect the dynamic response as much as a change in the modulus parameter. However, the non-linearity of the modulus and damping are predicted accurately, and it is easy to identify whether the system is under-damped or over-damped based on the predicted damping parameters.
Figure 6 visualizes the ground truth and predicted dynamic displacement response, along with the error between the two. We observe that the maximum peak-peak value in the displacement error is only 0.3% of the ground truth. We also consider the ability of NeuralSI to extrapolate and display the dynamic response by doubling the prediction time span. It is worth mentioning that the peak error in temporal extrapolation does not increase much compared to the peak error in temporal interpolation. The extrapolation results are also examined at different elements from different locations. Figure 7 presents the response at the beam midspan and at quarter length. There are no observed discrepancies between the ground truth and the predicted response.
5.2 Hyperparameter Investigation
Based on the parameters chosen above, we tested the effect of number of dense layers, training sample ratio and minibatch size on the parameter identification and prediction of dynamic responses.
Number of Layers: The number of layers is varied by consecutively adding an extra layer with 32 hidden units right after the input. From Fig. 8, the performance of the network is affected if the number of layers is below 4. This is explained by the fact that the network does not have sufficient capacity to precisely estimate the unknown structural parameters. It is noted that the size of the input and output are determined by the minibatch size and the number of elements used for discretization. A higher input or output size will automatically require a bigger network to improve prediction accuracy. Additionally, the Fréchet distance decreases as the size of the neural network increases, which demonstrates that the prediction of beam parameters is more accurate.
Sample Ratio: The number of training samples plays an important role in the model and in real in-field deployment scenarios. The number and the efficiency of sensor arrangements will be directly related to the number of samples required for accurately estimating the unknown parameters. It is expected that a reduced amount of data is sufficient to train the model given the strong domain knowledge (in the form of PDE) leveraged by NeuralSI. From Fig. 8, when 20% of the ground truth displacement samples are used for training, the loss drops noticeably. With an increased amount of training data, the network performance can still be improved. Furthermore, observe that there is a slight effect of data overfitting when using the full amount of data for training. The Fréchet distance of damping is not stable since our loss function optimizes for accurately predicting the dynamic deflection response, instead of directly predicting the parameters. As such, the same error could be obtained through different combinations of those parameters.
Minibatch Size: The minibatch size plays an important role in the efficiency of the training process and the performance of the estimated parameters. It is worth mentioning that a smaller minibatch size helps escape local minima and reduces errors. However, this induces a higher number of iterations for a single epoch, which is computationally expensive. From Fig. 8 we observe that both the MAE error and the Fréchet distance are relatively low when the minibatch size is smaller than 32.
6 Comparison of NeuralSI with a Direct Response Mapping Deep Neural Network and a PINN
The NeuralSI framework is compared with traditional deep neural networks (DNN) and PINN methods. The tested DNN has 5 dense layers and a Tanh activation. The inputs are the spatial and temporal coordinates x and t, respectively, of the displacement response, and the output is the beam deflection u(x, t) at that spatio-temporal position. The optimizer is LBFGS and the learning rate is 1.0. With a random choice of 20% samples, the loss stabilizes after 500 epochs.
The PINN method is defined with a similar strategy to existing solutions [20, 29]. The Neural network consists of 5 dense layers with Tanh activation function. The loss is defined as a weighted aggregate of the boundary condition loss (second derivative of input x at the boundaries), governing equation loss (fourth-order derivative of x and second-order derivative of the t), and loss between the prediction and ground truth displacement response. We used LBFGS as the optimizer with a learning rate of 1.0. The training was executed for 3700 epochs.
The prediction of the dynamic deformation responses for the two baseline methods and NeuralSI and the corresponding displacement distribution errors are shown in Fig. 9. In NeuralSI, we used ImplicitEulerExtrapolation solver for a 4\(\times \) faster inference. We further optimized the PDE function with ModelingToolkit [25], which provides another 10\(\times \) speedup, for a total of 40\(\times \) speedup over the RK4 solver used for training. Due to a limited amount of data for training, the DNN fails to predict the response. With extra information from the boundary conditions and equation, the PINN method results in an MAE loss of 0.344, and the prediction fits the true displacement distribution well. Most of the values in the displacement distribution error are small, except for some corners. But both methods fail to extrapolate the structural behavior temporally. The extrapolation of DNN predictions produces large discrepancies compared to the ground truth. Similarly, the PINN method fails to match the NeuralSI performance, while fairing much better than the predictions from the DNN, as expected due to the added domain knowledge. The MAE errors were computed and compared with the proposed method trained with 20% data as shown in Fig. 10.
7 Conclusion
In this paper, we proposed NeuralSI, a framework that can be employed for structural parameter identification in nonlinear dynamic systems. Our solution models the unknown parameters via a learnable neural network and embeds it within a partial differential equation. The network is trained by minimizing the errors between predicted dynamic responses and ground truth measurement data. A major advantage of the method is its versatility and flexibility; thus, it can be successfully extended to any PDEs with high-order derivatives and nonlinear characteristics. The trained model can be used to either explore structural behavior under different initial conditions and loading scenarios, which is vital for structural modeling or to determine high-accuracy extrapolation, also essential in systems’ response prognosis. An example beam vibration study case was analyzed to demonstrate the capabilities of the framework. The estimated structural parameters and the dynamic response variations match well with the ground truth (MAE of \(10^{-4}\)). The performance of NeuralSI is also shown to outperform direct regression significantly through deep neural networks and PINN methods by three to five orders of magnitude.
References
Akinpelu, F.O.: The response of viscously damped Euler-Bernoulli beam to uniform partially distributed moving loads. Appl. Math. 3(3), 199–204 (2012)
Aliee, H., Theis, F.J., Kilbertus, N.: Beyond predictions in neural odes: identification and interventions. arXiv preprint arXiv:2106.12430 (2021)
Bagheri, A., Ozbulut, O.E., Harris, D.K.: Structural system identification based on variational mode decomposition. J. Sound Vib. 417, 182–197 (2018)
Banerjee, B., Roy, D., Vasu, R.: Self-regularized pseudo time-marching schemes for structural system identification with static measurements. Int. J. Numer. Meth. Eng. 82(7), 896–916 (2010)
Brandstetter, J., Welling, M., Worrall, D.E.: Lie point symmetry data augmentation for neural PDE solvers. arXiv preprint arXiv:2202.07643 (2022)
Brandstetter, J., Worrall, D., Welling, M.: Message passing neural PDE solvers. arXiv preprint arXiv:2202.03376 (2022)
Brincker, R., Zhang, L., Andersen, P.: Modal identification of output-only systems using frequency domain decomposition. Smart Mater. Struct. 10(3), 441 (2001)
Brunton, S.L., Proctor, J.L., Kutz, J.N.: Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proc. Natl. Acad. Sci. 113(15), 3932–3937 (2016)
Chen, R.T., Rubanova, Y., Bettencourt, J., Duvenaud, D.K.: Neural ordinary differential equations. In: Advances in Neural Information Processing Systems, vol. 31 (2018)
Diez, A., Khoa, N.L.D., Makki Alamdari, M., Wang, Y., Chen, F., Runcie, P.: A clustering approach for structural health monitoring on bridges. J. Civil Struct. Health Monit. 6(3), 429–445 (2016)
Dulny, A., Hotho, A., Krause, A.: NeuralPDE: modelling dynamical systems from data. arXiv preprint arXiv:2111.07671 (2021)
Eiter, T., Mannila, H.: Computing discrete fréchet distance (1994)
Entezami, A., Sarmadi, H., Behkamal, B., Mariani, S.: Big data analytics and structural health monitoring: a statistical pattern recognition-based approach. Sensors 20(8), 2328 (2020)
Entezami, A., Shariatmadar, H., Sarmadi, H.: Structural damage detection by a new iterative regularization method and an improved sensitivity function. J. Sound Vib. 399, 285–307 (2017)
Ghorbani, E., Buyukozturk, O., Cha, Y.J.: Hybrid output-only structural system identification using random decrement and Kalman filter. Mech. Syst. Signal Process. 144, 106977 (2020)
Hasni, H., et al.: A new approach for damage detection in asphalt concrete pavements using battery-free wireless sensors with non-constant injection rates. Measurement 110, 217–229 (2017)
Hasni, H., Alavi, A.H., Lajnef, N., Abdelbarr, M., Masri, S.F., Chakrabartty, S.: Self-powered piezo-floating-gate sensors for health monitoring of steel plates. Eng. Struct. 148, 584–601 (2017)
Horie, M., Mitsume, N.: Physics-embedded neural networks: E(n)-equivariant graph neural PDE solvers. arXiv preprint arXiv:2205.11912 (2022)
Konkanov, M., Salem, T., Jiao, P., Niyazbekova, R., Lajnef, N.: Environment-friendly, self-sensing concrete blended with byproduct wastes. Sensors 20(7), 1925 (2020)
Krishnapriyan, A., Gholami, A., Zhe, S., Kirby, R., Mahoney, M.W.: Characterizing possible failure modes in physics-informed neural networks. In: Advances in Neural Information Processing Systems, vol. 34, pp. 26548–26560 (2021)
Lai, Z., Mylonas, C., Nagarajaiah, S., Chatzi, E.: Structural identification with physics-informed neural ordinary differential equations. J. Sound Vib. 508, 116196 (2021)
Lai, Z., Nagarajaiah, S.: Sparse structural system identification method for nonlinear dynamic systems with hysteresis/inelastic behavior. Mech. Syst. Signal Process. 117, 813–842 (2019)
Li, X., Salem, T., Bolandi, H., Boddeti, V., Lajnef, N.: Methods for the rapid detection of boundary condition variations in structural systems. American Society of Mechanical Engineers (2022)
Loshchilov, I., Hutter, F.: Fixing weight decay regularization in adam (2018)
Ma, Y., Gowda, S., Anantharaman, R., Laughman, C., Shah, V., Rackauckas, C.: ModelingToolkit: a composable graph transformation system for equation-based modeling. arXiv preprint arXiv:2103.05244 (2021)
Maurya, D., Chinta, S., Sivaram, A., Rengaswamy, R.: Incorporating prior knowledge about structural constraints in model identification. arXiv preprint arXiv:2007.04030 (2020)
Modebei, M., Adeniyi, R., Jator, S.: Numerical approximations of fourth-order PDEs using block unification method. J. Nigerian Math. Soc. 39(1), 47–68 (2020)
Rackauckas, C., et al.: Universal differential equations for scientific machine learning. arXiv preprint arXiv:2001.04385 (2020)
Raissi, M., Perdikaris, P., Karniadakis, G.E.: Physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 378, 686–707 (2019)
Reynders, E., De Roeck, G.: Reference-based combined deterministic-stochastic subspace identification for experimental and operational modal analysis. Mech. Syst. Signal Process. 22(3), 617–637 (2008)
Rezaiee-Pajand, M., Entezami, A., Sarmadi, H.: A sensitivity-based finite element model updating based on unconstrained optimization problem and regularized solution methods. Struct. Control. Health Monit. 27(5), e2481 (2020)
Salehi, H., Burgueño, R., Chakrabartty, S., Lajnef, N., Alavi, A.H.: A comprehensive review of self-powered sensors in civil infrastructure: state-of-the-art and future research trends. Eng. Struct. 234, 111963 (2021)
Salem, T., Jiao, P., Zaabar, I., Li, X., Zhu, R., Lajnef, N.: Functionally graded materials beams subjected to bilateral constraints: structural instability and material topology. Int. J. Mech. Sci. 194, 106218 (2021)
Sarmadi, H., Entezami, A., Daneshvar Khorram, M.: Energy-based damage localization under ambient vibration and non-stationary signals by ensemble empirical mode decomposition and Mahalanobis-squared distance. J. Vib. Control 26(11–12), 1012–1027 (2020)
Tuhta, S., Günday, F.: Multi input multi output system identification of concrete pavement using N4SID. Int. J. Interdisc. Innov. Res. Dev. 4(1), 41–47 (2019)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)
Yang, X.M., Yi, T.H., Qu, C.X., Li, H.N., Liu, H.: Automated eigensystem realization algorithm for operational modal identification of bridge structures. J. Aerosp. Eng. 32(2), 04018148 (2019)
Yin, T., Jiang, Q.H., Yuen, K.V.: Vibration-based damage detection for structural connections using incomplete modal data by Bayesian approach and model reduction technique. Eng. Struct. 132, 260–277 (2017)
Yuen, K.V., Au, S.K., Beck, J.L.: Two-stage structural health monitoring approach for phase I benchmark studies. J. Eng. Mech. 130(1), 16–33 (2004)
Zhang, T., et al.: ANODEV2: a coupled neural ODE framework. In: Advances in Neural Information Processing Systems, vol. 32 (2019)
Zhou, X., He, W., Zeng, Y., Zhang, Y.: A semi-analytical method for moving force identification of bridge structures based on the discrete cosine transform and FEM. Mech. Syst. Signal Process. 180, 109444 (2022)
Zubov, K., et al.: NeuralPDE: automating physics-informed neural networks (PINNs) with error approximations. arXiv preprint arXiv:2107.09443 (2021)
Acknowledgements
This research was funded in part by the National Science Foundation grant CNS 1645783.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Li, X., Bolandi, H., Salem, T., Lajnef, N., Boddeti, V.N. (2023). NeuralSI: Structural Parameter Identification in Nonlinear Dynamical Systems. In: Karlinsky, L., Michaeli, T., Nishino, K. (eds) Computer Vision – ECCV 2022 Workshops. ECCV 2022. Lecture Notes in Computer Science, vol 13807. Springer, Cham. https://doi.org/10.1007/978-3-031-25082-8_22
Download citation
DOI: https://doi.org/10.1007/978-3-031-25082-8_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-25081-1
Online ISBN: 978-3-031-25082-8
eBook Packages: Computer ScienceComputer Science (R0)