Informatica logo


Login Register

  1. Home
  2. To appear
  3. Novel Loss Function Construction for Neu ...

Novel Loss Function Construction for Neural Network-Based Prediction of Complex High-Dimensional Nonlinear Dynamical Systems
Kun An   Ying Sun ORCID icon link to view author Ying Sun details   Minghui Yao   Junhua Zhang  

Authors

 
Placeholder
https://doi.org/10.15388/25-INFOR607
Pub. online: 28 October 2025      Type: Research Article      Open accessOpen Access

Received
1 May 2025
Accepted
1 October 2025
Published
28 October 2025

Abstract

Traditional loss functions such as mean squared error (MSE) are widely employed, but they often struggle to capture the dynamic characteristics of high-dimensional nonlinear systems. To address this issue, we propose an improved loss function that integrates linear multistep methods, system-consistency constraints, and prediction-phase error control. This construction simultaneously improves training accuracy and long-term stability. Furthermore, the introduction of recursive loss and interpolation strategies brings the model closer to practical prediction scenarios, broadening its applicability. Numerical simulations demonstrate that this construction significantly outperforms both mean square error and existing custom loss functions in terms of performance.

1 Introduction

Chaotic dynamics are ubiquitous in natural and engineered systems, ranging from weather and climate processes to electronic circuits and structural mechanics. Unlike regular periodic motions, chaotic time series are characterized by sensitive dependence on initial conditions, nonlinearity, and high dimensionality, which make their long-term prediction extremely challenging. Traditional approaches based on analytical modelling or numerical simulation often become intractable when the governing equations are highly nonlinear or the system complexity increases (Zhang et al., 2019). In this context, data-driven methods, particularly neural networks (Sestanovic and Kalinic Milicevic, 2025; Calvo-Rolle et al., 2014), have emerged as effective alternatives for capturing the underlying dynamics and predicting the evolution of chaotic systems.
The success of neural networks in chaotic time series prediction (Fernandes et al., 2020), however, is not determined solely by network architecture; the choice of loss function plays an equally crucial role. A well-designed loss function not only guides the optimization process but also influences the model’s ability to capture long-term dynamics rather than merely short-term patterns. Over time, the development of loss functions has become a central factor in improving both predictive accuracy and generalization performance across different applications. Early loss functions mainly include Mean Square Error (MSE), Mean Absolute Error (MAE), and cross-entropy, the latter rooted Information Theory of Shannon (1948). To address robustness issues, Huber proposed a hybrid of MSE and MAE (Huber, 1964). Chopra et al. (2005) proposed the contrastive loss to minimize distances between similar samples and maximize those between dissimilar ones, on which the triplet loss (Hadsell et al., 2006) introduced additional constraints through triplet comparison. Customized objectives have also been developed: BLEU-based loss for machine translation (Papineni et al., 2002; Ranzato et al., 2015), BPR loss for recommender systems (Rendle et al., 2009), and focal loss for class-imbalance in object detection (Lin et al., 2017). A principled deep learning approach was proposed for multi-task learning, which weighs multiple loss functions by considering homoskedastic uncertainty (Kendall et al., 2018). In physics-related domains, auxiliary loss functions have been integrated into neural networks to incorporate physical dynamics. Examples include PIDynNet (Liu and Meidani, 2023), a physics-informed neural network for structural systems, and a two-coefficient loss function based on linear multistep methods (Zhang et al., 2023). In addition, Ghazvini et al. (2024) combine MSE and LogCosh to mitigate vanishing gradients, while a weight-constrained RNN optimization (Wu et al., 2020) incorporates parameter constraints and regularization to ensure the input-output relationship.
The combination of numerical algorithms and neural networks has advanced deep learning for differential equations. The Runge-Kutta Convolutional Neural Network (RKCNN) builds network models using higher-order Runge-Kutta methods (Zhu et al., 2022). PINNs have been integrated with Runge-Kutta schemes for parameter estimation and dynamic modelling (Zhai et al., 2023; He et al., 2023). Connections between the implicit Euler method and neural networks led to the Adaptive Implicit Network (AIM-NET), allowing flexible convergence for improved parameter estimation (Yuan et al., 2020). Implicit Adams Predictor-Corrector Blocks (IABs) combined with Non-local Sparse Attention (NSA) and Attention Feature Fusion (AFF) form the Implicit Adams Predictor-Corrector Module (IAM) (Yin et al., 2023), while discrete ZNN (DZNN) models of Adams-Bashforth type enhance accuracy for time-varying problems (Yang et al., 2020). A linear multi-step framework based on the implicit Adams-Moulton scheme approximates full-order models in low-dimensional space (Xie et al., 2019). Finally, MultiPINN, a multi-head neural network enriched with PINNs, incorporates RBF interpolation and embeds physics-based priors from differential equations and boundary conditions (Li, 2024). These methods improve efficiency and performance by integrating numerical schemes with physics-informed loss functions.
Building on the integration of numerical algorithms and neural networks, interpolation techniques have become important in deep learning for data augmentation, feature extraction, and regularization, improving accuracy and robustness. Conditional Encoder-Decoder GANs (CEDGANs) treat interpolation as a conditional generation task, capturing spatial relationships (Zhu et al., 2019). Interpolated Adversarial Training enhances adversarial robustness while preserving generalization (Lamb et al., 2019). D’Ambrosio et al. (2021) combine PINNs with a functional interpolation technique, the Theory of Functional Connections (TFC), to learn optimal controls. Interpolation Consistency Training (ICT) ensures predictions for interpolated points align with interpolated predictions, reducing overfitting (Verma et al., 2022). RAKI performs nonlinear interpolation of missing k-space lines, improving noise resilience in MRI reconstruction (Akçakaya et al., 2019), while a CNN-based interpolation method addresses missing projection data in sparse-view CT (Lee et al., 2017).
Previous studies on loss functions remain inadequate for modelling and predicting complex, high-dimensional, nonlinear dynamical systems. Traditional loss functions, such as MSE, are widely adopted for their simplicity, but they often overlook critical dynamic features in long-term predictions. Although several custom loss functions have been proposed, they generally suffer from limited accuracy and poor stability, which restrict their applicability to a broader class of dynamic systems. To address these challenges, this paper introduces an error term in the prediction phase, thereby proposing an improved custom loss function. The improved version further integrates interpolation and recursive strategies to better capture temporal dependencies and enhance long-term prediction accuracy. This paper selects the dynamic modelling of a ring truss antenna as a numerical case study, representing a typical chaotic system within aerospace structural dynamics. Numerical simulations demonstrate that the improved loss function consistently outperforms both the traditional MSE and the original custom loss function in terms of accuracy and stability. This study presents a systematic approach to designing advanced loss functions and demonstrates its effectiveness in predicting complex, high-dimensional nonlinear dynamical systems.

2 Constructing Loss Function

2.1 Mean Square Error

MSE is a common metric for assessing the predictive performance of a model (Zhou et al., 2021; Rumelhart et al., 1986). It measures the average squared difference between the predicted and actual values. The smaller the Mean Squared Error, the closer the model’s predictions are to the actual values and the better the model performance. The formula for the mean square error is as follows:
(1)
\[ MSE=\frac{1}{n}{\sum \limits_{i=1}^{n}}{({y_{i}}-{\tilde{y}_{i}})^{2}},\]
where n is the number of samples, ${y_{i}}$ is the actual value of the i-th sample, ${\hat{y}_{i}}$ is the predicted value of the i-th sample.
When using MSE as a loss function for training a neural network, the aim is to predict the state vector of the next moment using the state vector of the previous moment. Therefore, MSE is mainly used to calculate the difference between the predicted state vector of the next moment and the numerically solved state vector of the next moment. The calculation formula for the MSE used as the loss function in neural networks is as follows:
(2)
\[ {L_{1}}=\frac{1}{N-k}{\sum \limits_{i=0}^{N-k}}{\big|X({t_{i+k}})-Y({t_{i}})\big|^{2}},\]
where $X({t_{i}})$ is the approximate solution at time ${t_{i}}$, $Y({t_{i}})=\tilde{f}(X({t_{i}}))$ is the output of the neural network at time ${t_{i}}$ ($\tilde{f}$ represents the trained neural network), k is the prediction step (i.e. use the i-th point to predict the $(i+k)$-th point). Considering the prediction accuracy, k is taken as 1 in this paper.

2.2 Custom Loss Function

From previous work (Sun et al., 2024), we already know that the linear multistep method is applicable to nonlinear dynamical systems. The general form of the linear multistep method can be expressed as:
(3)
\[ X({t_{i+1}})={\sum \limits_{p=0}^{k-1}}{\alpha _{p}}X({t_{i-p}})+h{\sum \limits_{q=-1}^{k-1}}{\beta _{q}}\dot{X}({t_{i-q}}),\]
here, $X({t_{i}})$ is the approximate solution at time ${t_{i}}$, h is the step size, α and β are constants, and k-1 is the number of steps. This formula indicates that the current point $X({t_{i}})$ is computed as a linear combination of the previous k-1 points and a linear combination of the derivative values at these points.
During the training process of a neural network, all time step information is available because the entire batch of training data is known. Therefore, we use the Adams-Moulton method in the implicit linear multistep method. The Adams-Moulton method requires information about the current time step, but is more numerically stable than the Adams-Bashforth method in the display linear multistep method. The first part of the loss function can then be written as ${\varepsilon _{1i}}={\textstyle\sum _{p=0}^{k-1}}{\alpha _{p}}X({t_{i-p}})+h{\textstyle\sum _{q=-1}^{k-1}}{\beta _{q}}[Y({t_{i-q}})+\bar{f}({t_{i-q}})]-X({t_{i+1}})$. Previous work has utilized both the one-step form (improved Euler method) and the two-step form (Simpson’s method) of the linear multistep method to construct the loss function for neural networks. For the sake of convenience in computing the loss function, this paper adopts the Second-order linear multistep method.
A typical dynamical system we study can be written as:
(4)
\[ \frac{dX(t)}{dt}=f\big(X(t)\big)+\bar{f}(t),\]
where $X(t)={[{\boldsymbol{x}^{T}}(t)\hspace{2.5pt}{\boldsymbol{v}^{T}}(t)]^{T}}\in {\mathbb{R}^{2n}}$ is the state vector of the system, $f(X(t))$ includes the generalized restoring force, and $\bar{f}(t)$ contains the external excitation force. Based on the displacement and velocity, it is divided into the first n dimensions and the last n dimensions. The output of the neural network is denoted as $Y=\tilde{f}({X_{t}})$, which is also divided into ${Y_{1}}$ and ${Y_{2}}$. Since $X(t)={[{\boldsymbol{x}^{T}}(t)\hspace{2.5pt}{\boldsymbol{v}^{T}}(t)]^{T}}\in {\mathbb{R}^{2n}}$ is divided into ${X_{1}}(t)$ and ${X_{2}}(t)$, we can define the second term in the loss function ${L_{2}}$, denoted as ${\varepsilon _{2j}}={Y_{1}}({t_{j}})-{X_{2}}({t_{j}})$. The second term ${\varepsilon _{2j}}$ enforces the inherent consistency between displacement and velocity (i.e. the derivative of displacement must equal the velocity component). More generally, this can be interpreted as a system-specific consistency constraint, ensuring that the neural network respects the intrinsic structure of the dynamical system.The loss function is given by Eq. (5):
(5)
\[ {L_{2}}=\frac{{\gamma _{1}}}{N-k+1}{\sum \limits_{i=k-1}^{N-1}}|{\varepsilon _{1i}}{|^{2}}+\frac{{\gamma _{2}}}{N+1}{\sum \limits_{j=0}^{N}}|{\varepsilon _{2j}}{|^{2}},\]
where ${\gamma _{1}}$ and ${\gamma _{2}}$ are weight coefficients.
Although ${L_{2}}$ successfully embeds system dynamics into the loss function through the implicit linear multistep method term and system-consistency term, it does not explicitly constrain the errors that arise during the iterative prediction phase. As a result, long-term predicting may still suffer from error accumulation. This limitation motivates the improved framework introduced in Section 2.3.

2.3 Improved Custom Loss Function

In Section 2.2, the loss function was designed to constrain only the training residuals. Although this approach improved short-term accuracy, it did not address the error during iterative prediction. To overcome this limitation, we propose an improved framework that simultaneously penalizes both training residuals and prediction-phase discrepancies, thereby enhancing long-term stability.
We next turn our attention to the iterative prediction process, which is fundamentally different from training. During prediction, the true state at the current time step is unknown, so the network must rely on its own previous outputs to generate future states. This inherently necessitates the use of an explicit linear multistep method, as the current state cannot be incorporated implicitly. We denote the abstract form of the explicit linear multistep formula in the prediction process as:
(6)
\[ X({t_{m+1}})={\sum \limits_{p=0}^{k-1}}{\alpha _{p}}X({t_{m-p}})+h{\sum \limits_{q=0}^{k-1}}{\beta _{q}}\big[Y({t_{m-q}})+\bar{f}({t_{m-q}})\big],\]
here, $X({t_{m+1}})$ is the approximate solution at time ${t_{m+1}}$, h is the step size, α and β are constants, and k-1 is the number of steps.
Eq. (6) presents the iterative formula employed during the prediction phase, from which we derive a novel loss term: ${\varepsilon _{3m}}={\textstyle\sum _{p=0}^{k-1}}{\alpha _{p}}X({t_{m-p}})+h{\textstyle\sum _{q=0}^{k-1}}{\beta _{q}}[Y({t_{m-q}})+\bar{f}({t_{m-q}})]-X({t_{m+1}})$. This term primarily constrains errors generated during iterative prediction. Incorporating it into the loss function serves a dual purpose: firstly, it explicitly penalises deviations arising when the model must rely on its own outputs rather than true states during rolling predictions, thereby mitigating long-term error accumulation. Secondly, as the prediction phase relies on explicit numerical methods (which inherently introduce truncation errors), the new term embeds these numerical consistency requirements into the training objective. Unlike the model in Sun et al. (2024), which focused solely on training residuals, the proposed framework introduces a dual-constraint mechanism that integrates both training accuracy and prediction-phase error control, thereby mitigating error accumulation in long-term predictions. Following the previous notation, the complete form of the improved loss function can be expressed as:
(7)
\[ {L_{3}}=\frac{{\gamma _{1}}}{N-k+1}{\sum \limits_{i=k-1}^{N-1}}|{\varepsilon _{1i}}{|^{2}}+\frac{{\gamma _{2}}}{N+1}{\sum \limits_{j=0}^{N}}|{\varepsilon _{2j}}{|^{2}}+\frac{{\gamma _{3}}}{N-k+1}{\sum \limits_{m=k-1}^{N-1}}|{\varepsilon _{3m}}{|^{2}},\]
where
(8)
\[ \left\{\begin{array}{l}{\varepsilon _{1i}}={\textstyle\sum \limits_{p=0}^{k-1}}{\alpha _{p}}X({t_{i-p}})+h{\textstyle\sum \limits_{q=-1}^{k-1}}{\beta _{q}}[Y({t_{i-q}})+\bar{f}({t_{i-q}})]-X({t_{i+1}}),\\ {} {\varepsilon _{2j}}={Y_{1}}({t_{j}})-{X_{2}}({t_{j}}),\\ {} {\varepsilon _{3m}}={\textstyle\sum \limits_{p=0}^{k-1}}{\alpha _{p}}X({t_{m-p}})+h{\textstyle\sum \limits_{q=0}^{k-1}}{\beta _{q}}[Y({t_{m-q}})+\bar{f}({t_{m-q}})]-X({t_{m+1}}).\end{array}.\right.\]
${\gamma _{1}}$, ${\gamma _{2}}$ and ${\gamma _{3}}$ are weight coefficients. In the numerical simulations conducted in this paper, the specific formula employed for Eq. (8) is:
(9)
\[ \left\{\begin{array}{l}{\varepsilon _{1i}}=X({t_{i}})+\frac{h}{2}[Y({t_{i+1}})+\bar{f}({t_{i+1}})+Y({t_{i}})+\bar{f}({t_{i}})]-X({t_{i+1}}),\\ {} {\varepsilon _{2j}}={Y_{1}}({t_{j}})-{X_{2}}({t_{j}}),\\ {} {\varepsilon _{3m}}=X({t_{m}})+h[Y({t_{m}})+\bar{f}({t_{m}})]-X({t_{m+1}}).\end{array}.\right.\]
It is worth highlighting that the improved custom loss function is constructed independently of specific iterative methods, which ensures its generality and applicability across various prediction methods. This construction strategy incorporates the prediction-phase error into the loss function. Consequently, it can be applied to a wide range of data-driven models that rely on numerical methods during the prediction phase, without being restricted to a specific iterative method.
The workflow of the neural network and the composition of the loss function ${L_{3}}$ are illustrated in Fig. 1. Figure 1a illustrates the composition of the loss function, and Fig. 1b shows the underlying model of the two neural networks used in this paper. Firstly, the data X is input into the neural network f, and its parameters θ are optimized to minimize the loss function. Upon completing the training, the trained neural network $\tilde{f}$ is obtained. Subsequently, the last data point from the training set is fed into $\tilde{f}$ as the initial value, and the network outputs the corresponding prediction. The next moment’s data is then iteratively obtained using formula I. Through continuous iterations, the final prediction is produced. In this process, the loss function ${L_{2}}$ consists of ${\varepsilon _{1i}}$ and ${\varepsilon _{2j}}$. Additionally, the error ${\varepsilon _{3m}}$ from the iterative process during the prediction phase is introduced as a new term to ${L_{2}}$, forming the final loss function ${L_{3}}$. Figure 1b illustrates the neural network architectures employed for time-series prediction, including a Transformer-based encoder-decoder model (top) and an LSTM-based encoder-decoder model (bottom). Both models take sequential input data X, representing the system’s dynamic evolution, and aim to predict future states Y. The Transformer architecture incorporates input embedding, positional encoding, multi-head attention mechanisms, and feedforward layers to capture long-range dependencies. In contrast, the LSTM architecture relies on stacked LSTM layers in both the encoder and decoder to model temporal correlations. The output results are compared with numerical solutions to assess prediction accuracy.
infor607_g001.jpg
Fig. 1
Loss functions ${L_{3}}$ and neural networks.

3 Neural Networks Setting

3.1 Neural Network

In this study, we consider three representative neural network models for time series prediction of nonlinear dynamical systems: the LSTM$\text{\_}$ED model proposed in Sun et al. (2024), the S-Transformer, and the Conv-AT. The S-Transformer (Simplified Transformer) is a lightweight variant of the standard Transformer, designed for efficient time series prediction. Unlike the original Transformer, it employs a simplified decoder composed of linear layers, since iterative prediction in this study does not require autoregressive sequence generation. Despite its streamlined architecture, the S-Transformer retains the advantages of multi-head attention mechanisms, such as capturing long-range dependencies and supporting parallel computation. The Conv-AT (Convolutional Attention Transformer) is another variant of the Transformer architecture, tailored to balance local and global feature extraction. Before entering the attention module, the input sequence is processed by a convolutional layer to capture local dependencies and enrich the query representation. The original sequence, after linear transformation, serves as the key and value, allowing the attention mechanism to model relationships between convolution-enhanced local features and the original global representations. This design enables Conv-AT to simultaneously capture fine-grained local structures and long-range dependencies. In this paper, these three neural network models are used solely as validation platforms to test the effectiveness of the proposed improved loss function. To ensure a fair comparison, all numerical simulations are conducted under identical hyperparameter settings.

3.2 Network Setting

The numerical example used in this paper is a beam-ring structure with two degrees of freedom. Based on the framework design, we built three neural networks. The first model is used to predict the linear terms, and the next two are used to predict the nonlinear terms. The parameters of the neural network models using the custom loss function and the improved custom loss function are the same. For different beam-ring structures, different strategies are adopted in setting parameters such as learning rate and number of training times, and we have taken the optimal combination of parameters in the simulation after several numerical simulations. However, to ensure a rigorous comparison of loss functions within the same beam-ring structure, we strictly maintained the same parameter settings throughout.

4 Numerical Simulation

4.1 Beam-Ring Construction

In this section, we consider a nonlinear dynamic system that originates from the beam-ring equivalent model of a circular truss antenna subjected to periodic thermal excitation (Zhang et al., 2019). This system can be proposed to model the truss antenna after it is deployed, which is characterized by light weight, high flexibility, and low stiffness. Under periodic thermal loads, such structures can exhibit complex responses including large deformations, resonances, and even chaotic vibrations, making the model an ideal benchmark for testing prediction frameworks. By applying the improved loss function to this system, we are able to evaluate its effectiveness in handling error accumulation and long-term prediction in nonlinear and chaotic dynamics. It should be emphasised that this antenna model serves merely as a representative case study, and the improved loss function framework is generalisable to a broad class of nonlinear dynamical systems. The following nonlinear dynamics system for the beam-ring equivalent model of a loop truss antenna is considered:
(10a)
\[\begin{aligned}{}& {\ddot{q}_{1}}+\varepsilon {\mu _{1}}{\dot{q}_{1}}+{\omega _{1}^{2}}{q_{1}}+{\varepsilon ^{2}}({f_{11}}{q_{1}}+{f_{12}}{q_{2}})\cos \Omega t+\varepsilon {\alpha _{1}}{q_{1}^{2}}+\varepsilon {\alpha _{2}}{q_{2}^{2}}+\varepsilon {\alpha _{3}}{q_{1}}{q_{2}}\\ {} & \hspace{1em}+\varepsilon {\alpha _{4}}{\dot{q}_{1}}{\dot{q}_{2}}+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}}){\ddot{q}_{1}}+\varepsilon ({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}}){\ddot{q}_{2}}=\varepsilon {f_{1}}\cos \Omega t,\end{aligned}\]
(10b)
\[\begin{aligned}{}& {\ddot{q}_{2}}+\varepsilon {\mu _{2}}{\dot{q}_{2}}+{\omega _{2}^{2}}{q_{2}}+{\varepsilon ^{2}}({f_{21}}{q_{1}}+{f_{22}}{q_{2}})\cos \Omega t+\varepsilon {\beta _{1}}{q_{1}^{2}}+\varepsilon {\beta _{2}}{q_{2}^{2}}+\varepsilon {\beta _{3}}{q_{1}}{q_{2}}\\ {} & \hspace{1em}+\varepsilon {\beta _{4}}{\dot{q}_{1}}{\dot{q}_{2}}+\varepsilon ({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}}){\ddot{q}_{1}}+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}}){\ddot{q}_{2}}=\varepsilon {f_{2}}\cos \Omega t,\end{aligned}\]
where ${q_{1}}$ and ${q_{2}}$ represent dimensionless displacements, ${f_{1}}$ and ${f_{2}}$ are parameters of the external excitation, ${f_{11}}$, ${f_{12}}$, ${f_{21}}$ and ${f_{22}}$ are parameters of the parametric excitation.
Transform Eqs. (10a) and (10b) into the following state equations:
(11a)
\[\begin{aligned}{}& {\dot{q}_{1}}={p_{1}},\end{aligned}\]
(11b)
\[\begin{aligned}{}& {\dot{q}_{2}}={p_{2}},\end{aligned}\]
(11c)
\[\begin{aligned}{}& {\dot{p}_{1}}=\frac{\varepsilon {f_{1}}\cos \Omega t-{E_{1}}-\varepsilon ({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}})\frac{\varepsilon {f_{2}}\cos \Omega t-{E_{2}}}{1+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}})}}{1+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}})-\frac{{\varepsilon ^{2}}({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}})({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}})}{1+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}})}},\end{aligned}\]
(11d)
\[\begin{aligned}{}& {\dot{p}_{2}}=\frac{\varepsilon {f_{2}}\cos \Omega t-{E_{2}}-\varepsilon ({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}})\frac{\varepsilon {f_{1}}\cos \Omega t-{E_{1}}}{1+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}})}}{1+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}})-\frac{{\varepsilon ^{2}}({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}})({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}})}{1+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}})}}.\end{aligned}\]
where
(12a)
\[\begin{aligned}{}{E_{1}}& =\varepsilon {\mu _{1}}{p_{1}}+{\omega _{1}^{2}}{q_{1}}+{\varepsilon ^{2}}({f_{11}}{q_{1}}+{f_{12}}{q_{2}})\cos \Omega t+\varepsilon {\alpha _{1}}{q_{1}^{2}}+\varepsilon {\alpha _{2}}{q_{2}^{2}}\\ {} & \hspace{1em}+\varepsilon {\alpha _{3}}{q_{1}}{q_{2}}+\varepsilon {\alpha _{4}}{p_{1}}{p_{2}},\end{aligned}\]
(12b)
\[\begin{aligned}{}{E_{2}}& =\varepsilon {\mu _{2}}{p_{2}}+{\omega _{2}^{2}}{q_{2}}+{\varepsilon ^{2}}({f_{21}}{q_{1}}+{f_{22}}{q_{2}})\cos \Omega t+\varepsilon {\beta _{1}}{q_{1}^{2}}+\varepsilon {\beta _{2}}{q_{2}^{2}}\\ {} & \hspace{1em}+\varepsilon {\beta _{3}}{q_{1}}{q_{2}}+\varepsilon {\beta _{4}}{p_{1}}{p_{2}}.\end{aligned}\]
Extract an external excitation term from Eq. (11c) and (11d) to generate a generalized form of Eq. (4), which is suitable for our neural network approach:
(13a)
\[\begin{aligned}{}& {\dot{q}_{1}}={p_{1}},\end{aligned}\]
(13b)
\[\begin{aligned}{}& {\dot{q}_{2}}={p_{2}},\end{aligned}\]
(13c)
\[\begin{aligned}{}& {\dot{p}_{1}}=\bigg[\frac{\varepsilon {f_{1}}\cos \Omega t-{E_{1}}-\varepsilon ({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}})\frac{\varepsilon {f_{2}}\cos \Omega t-{E_{2}}}{1+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}})}}{1+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}})-\frac{{\varepsilon ^{2}}({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}})({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}})}{1+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}})}}-\varepsilon {f_{1}}\cos \Omega t\bigg]\\ {} & \phantom{{\dot{p}_{1}}=}+\varepsilon {f_{1}}\cos \Omega t,\end{aligned}\]
(13d)
\[\begin{aligned}{}& {\dot{p}_{2}}=\bigg[\frac{\varepsilon {f_{2}}\cos \Omega t-{E_{2}}-\varepsilon ({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}})\frac{\varepsilon {f_{1}}\cos \Omega t-{E_{1}}}{1+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}})}}{1+\varepsilon ({\beta _{7}}{q_{1}}+{\beta _{8}}{q_{2}})-\frac{{\varepsilon ^{2}}({\alpha _{7}}{q_{1}}+{\alpha _{8}}{q_{2}})({\beta _{5}}{q_{1}}+{\beta _{6}}{q_{2}})}{1+\varepsilon ({\alpha _{5}}{q_{1}}+{\alpha _{6}}{q_{2}})}}-\varepsilon {f_{2}}\cos \Omega t\bigg]\\ {} & \phantom{{\dot{p}_{2}}=}+\varepsilon {f_{2}}\cos \Omega t.\end{aligned}\]
The chaotic time series can be obtained by setting the initial values of the state vectors and the parameters in Eq. (13) as follows:
\[\begin{array}{l@{\hskip4.0pt}l@{\hskip4.0pt}l@{\hskip4.0pt}l}{q_{10}}=0.0414,\hspace{1em}& {q_{20}}=-0.0699,\hspace{1em}& {p_{10}}=0.0288,\hspace{1em}& {p_{20}}=-0.0304,\\ {} {\mu _{1}}=0.0812,\hspace{1em}& {\mu _{2}}=0.0789,\hspace{1em}& {\alpha _{1}}=-0.466,\hspace{1em}& {\alpha _{2}}=-0.495,\\ {} {\alpha _{3}}=0.469,\hspace{1em}& {\alpha _{4}}=-0.4,\hspace{1em}& {\alpha _{5}}=-0.759,\hspace{1em}& {\alpha _{6}}=-0.124,\\ {} {\alpha _{7}}=0.12,\hspace{1em}& {\alpha _{8}}=0.742,\hspace{1em}& {\beta _{1}}=-0.0875,\hspace{1em}& {\beta _{2}}=0.0466,\\ {} {\beta _{3}}=-0.0634,\hspace{1em}& {\beta _{4}}=-0.0769,\hspace{1em}& {\beta _{5}}=-0.058,\hspace{1em}& {\beta _{6}}=0.01,\\ {} {\beta _{7}}=0.0217,\hspace{1em}& {\beta _{8}}=-0.0302,\hspace{1em}& {f_{11}}=10,\hspace{1em}& {f_{12}}=10,\\ {} {f_{21}}=10,\hspace{1em}& {f_{22}}=10,\hspace{1em}& {f_{1}}=83,\hspace{1em}& {f_{2}}=15,\\ {} \varepsilon =0.001,\hspace{1em}& {\omega _{1}}=0.2,\hspace{1em}& {\omega _{2}}=2{\omega _{1}}+0.319\varepsilon ,\hspace{1em}& \Omega ={\omega _{1}}-0.001\varepsilon .\end{array}\]
The dataset is a time series generated by solving Eq. (13) with the ordinary differential equation (ODE) solver. The sampling time is 0–1500s, the time step is 0.0125, so the total number of training data is $4.8\times {10^{5}}$. The predicted time range is 1500–3000s, the time step is 0.0125 and the total number of predicted data is $4.8\times {10^{5}}$. This is due to the fact that we add the error generated by the iterative prediction method to the loss function as well, and therefore need to ensure the consistency of the time step to get a good prediction.

4.2 Results and Discussion

After identifying the equations, neural networks and loss functions, the focus of this section shifts to a comparative analysis of the predictive performance of the three selected loss functions. In this study, we employed three different neural networks to perform numerical simulations for the beam-ring structure, and evaluated the predictive performance using three different loss functions. Due to space limitations, we only display the prediction results on plane $(t,{p_{1}})$ only. While the other images are not shown here, they consistently support the conclusions drawn in this paper. Figures 2–4 illustrate the results of three neural networks’ predictions for the four planes of the beam-ring structure. We plot the significance of each subplot of the image as follows: (a) Prediction results using the loss function ${L_{1}}$, (b) Prediction results using the loss function ${L_{2}}$, (c) Prediction results using the loss function ${L_{3}}$, and (d) Comparison of point-wise errors for different loss functions.
infor607_g002.jpg
Fig. 2
Prediction performance of LSTM$\text{\_}$ED for the plane $(t,{p_{1}})$ using three different loss functions, and error comparison.
infor607_g003.jpg
Fig. 3
Prediction performance of S-Transformer for the plane $(t,{p_{1}})$ using three different loss functions, and error comparison.
infor607_g004.jpg
Fig. 4
Prediction performance of Conv-AT for the plane $(t,{p_{1}})$ using three different loss functions, and error comparison.
When we use ${L_{1}}$ as the loss function for training, the aim is to predict the state vector of the next moment using the state vector of the previous moment. As can be seen in Fig. 2, the prediction result of LSTM$\text{\_}$ED is almost a straight line, which means that constant iterations of the prediction results will eventually converge to the same value. This is due to the model’s inability to capture the complex relationship between the moments before and after the data and the improper choice of the loss function, which prevents the model from being optimized correctly. Similarly, when other models are trained using ${L_{1}}$ as the loss function, the displacement time series curves of S-Transformer and Conv-AT exhibit regular oscillations at high frequencies, which do not match the numerical solutions solved by the ode solver. In summary, none of the three neural networks can effectively predict Eq. (13) when using the MSE loss function.
For the loss functions ${L_{2}}$ and ${L_{3}}$ based on the linear multistep method, the three neural networks can roughly predict the displacement trend. Specifically, on the plane $(t,{p_{1}})$, the neural network using the original loss function ${L_{2}}$ can predict the approximate trend of the displacement time series data, but there is still a considerable error in the details. As can be seen from the figure, it starts to be noticeable around 2200 seconds and the error starts to gradually increase after 2500 seconds. The neural network with the improved loss function ${L_{3}}$ can not only predict the overall trend of the displacement time series, but also control the detail error within a small range. This shows that the improved loss function ${L_{3}}$ has a great advantage in prediction accuracy. The error plots in Figs. 3d–4d also show that the error of the loss function ${L_{3}}$ is relatively small and does not increase further. To summarize, it can be seen from the Figs. 2–4 that the improved loss function ${L_{3}}$ not only has a smaller error, but also predicts the trend more accurately. Detailed results of all the above numerical simulations (including mean square error and improvements for all numerical simulations) are given in Table 1. Rows ${L_{2}}$ and ${L_{3}}$ report the mean squared error between the predicted and numerical solutions obtained using the two respective loss functions. The “%” row indicates the percentage improvement achieved by loss function ${L_{3}}$ over ${L_{2}}$.
Through the numerical simulations, it is evident that the improved loss function proposed herein more effectively controls the accumulation of prediction errors when predicting the response of this chaotic system. Compared to traditional MSE or the original custom loss function, this approach not only maintains high accuracy in short-term predictions but also significantly reduces the error amplification effect during long-term iterative prediction. This advantage stems from the explicit constraint on prediction-phase errors within the loss function. Consequently, the framework demonstrates enhanced applicability and reliability in practical chaotic vibration scenarios.
Table 1
Comparison of errors for loss functions ${L_{2}}$ and ${L_{3}}$.
Neural networks Comparison Planes
$(t,{q_{1}})$ $(t,{q_{2}})$ $(t,{p_{1}})$ $(t,{p_{2}})$
LSTM$\text{\_}$ED ${L_{2}}$ 161.9232 4.7384 5.7993 0.7185
${L_{3}}$ 46.1518 2.9058 1.5152 0.4531
% 71.50% 38.68% 73.87% 36.94%
S-Transformer ${L_{2}}$ 81.5512 3.4842 2.5207 0.5361
${L_{3}}$ 22.9896 3.1256 0.6811 0.4922
% 71.81% 10.29% 72.98% 8.18%
Conv-AT ${L_{2}}$ 102.8211 4.4109 3.1926 0.6818
${L_{3}}$ 36.2087 3.5679 1.0740 0.5601
% 64.78% 19.11% 66.36% 17.85%

5 Advancing the Design of the Loss Function

5.1 Loss Function Design for Recursive Prediction

As can be seen from the previous presentation or Fig. 1, the approach employed in the prediction phase is referred to as Recursive Prediction with Numerical Integration. This method iteratively advances the system state by first predicting the temporal derivative using a trained neural network and then updating the state through a numerical integration scheme. Specifically, starting from an initial state at ${t_{0}}$, the model predicts the corresponding derivative, which is subsequently used to compute the next time step via a numerical integration method (in this study, the Forward Euler method is adopted). The newly obtained time step is then fed back into the network to generate the next prediction, and this process continues recursively until the desired prediction time horizon is reached.
The loss function is computed over an entire batch by evaluating the discrepancy between the numerically integrated predictions and the corresponding ground truth values. Specifically, the computation at each time step relies on the ground truth values rather than the numerically integrated values from the previous step. Unlike the prediction phase, which recursively updates the system state step by step, the loss computation directly assesses the deviations across the entire batch in a non-recursive manner. The following provides an example using the ${\varepsilon _{3m}}$ part of the loss function ${L_{3}}$, which is also the aspect of this section that has been improved. The ${\varepsilon _{3m}}$ part is computed using the Forward Euler method, which can be expressed as:
(14)
\[ X({t_{i}})+h\tilde{f}\big(X({t_{i}})\big)=\tilde{X}({t_{i+1}}).\]
Here, $\tilde{f}(X({t_{i}}))$ represents the prediction of the neural network at time ${t_{i}}$, while $\tilde{X}({t_{i+1}})$ denotes the computed result at time ${t_{i+1}}$ (For clarity in presenting the improvements in this section, we use ~ to denote predicted values rather than ground truth values.). Thus the ${\varepsilon _{3m}}$ part of the loss function ${L_{3}}$ can be written more clearly as: $\tilde{X}({t_{i+1}})-X({t_{i+1}})={\varepsilon _{3m}}$, which is equivalent to $X({t_{i}})+h\tilde{f}(X({t_{i}}))-X({t_{i+1}})={\varepsilon _{3m}}$. Based on the preceding derivation, the ${\varepsilon _{3m}}$ part of the loss function ${L_{3}}$ for the entire batch (with a batch size of 20) can be expressed as Eq. (15):
(15)
\[ \left[\begin{array}{c}X({t_{i}})\\ {} X({t_{i+1}})\\ {} \dots \\ {} X({t_{i+19}})\end{array}\right]+\left[\begin{array}{c}h\tilde{f}(X({t_{i}}))\\ {} h\tilde{f}(X({t_{i+1}}))\\ {} \dots \\ {} h\tilde{f}(X({t_{i+19}}))\end{array}\right]-\left[\begin{array}{c}X({t_{i+1}})\\ {} X({t_{i+2}})\\ {} \dots \\ {} X({t_{i+20}})\end{array}\right]={\varepsilon _{3m}}.\]
Since the loss function ${L_{3}}$ directly evaluates the deviation over the entire batch in a non-recursive manner, it is computed in parallel. Therefore, we can express it in a vectorized form.

5.1.1 RecursiveLoss

As discussed earlier, the loss function is computed in a non-recursive manner, whereas the prediction phase employs a recursive approach. Therefore, to ensure consistency between the loss computation and the prediction process, we reformulate the loss function into a recursive form in this section, allowing it to better capture the error propagation characteristics of the model during prediction. Still considering the entire batch as an example, in Eq. (15), the predictions at subsequent time steps are based on ground truth values. Therefore, we only need to replace these ground truth values with the computed results from the previous time step. After this modification, Eq. (15) takes the following form:
(16)
\[ \left[\begin{array}{c}X({t_{i}})\\ {} \tilde{X}({t_{i+1}})\\ {} \dots \\ {} \tilde{X}({t_{i+19}})\end{array}\right]+\left[\begin{array}{c}h\tilde{f}(X({t_{i}}))\\ {} h\tilde{f}(X({t_{i+1}}))\\ {} \dots \\ {} h\tilde{f}(X({t_{i+19}}))\end{array}\right]-\left[\begin{array}{c}X({t_{i+1}})\\ {} X({t_{i+2}})\\ {} \dots \\ {} X({t_{i+20}})\end{array}\right]={\varepsilon _{3m}}.\]
For the sake of clarity, this formulation is presented in a vectorized form. However, Eq. (16) is inherently computed in a recursive manner rather than in parallel. We write the loss function modified in this way as ${L_{3a}}$. The recursive loss is schematized in Fig. 5.
infor607_g005.jpg
Fig. 5
Loss function ${L_{3a}}$.
infor607_g006.jpg
Fig. 6
Comparison of displacement time series curves and errors predicted by Conv-AT using two loss functions.

5.1.2 Results and Discussion

We still train the model using data from 0 to 1500 seconds and predict time series data from 1500 to 3000 seconds. We plot the displacement time series curve predicted by the Conv-AT and S-Transformer using the new loss function ${L_{3a}}$, as shown in Figs. 6–7.
infor607_g007.jpg
Fig. 7
Comparison of displacement time series curves and errors predicted by S-Transformer using two loss functions.
For both neural networks, the displacement time series curves predicted using RecursiveLoss are much better compared to loss function ${L_{3}}$, and are in better agreement with the numerical solution. The better results of RecursiveLoss can also be seen from the cumulative mean square error.
infor607_g008.jpg
Fig. 8
Three-dimensional phase trajectories of numerical and predicted solutions in the $({q_{1}},{p_{1}},{q_{2}})$ space.
infor607_g009.jpg
Fig. 9
Three-dimensional phase trajectories of numerical and predicted solutions in the $({q_{1}},{p_{1}},{p_{2}})$ space.
infor607_g010.jpg
Fig. 10
Three-dimensional phase trajectories of numerical and predicted solutions in the $({q_{2}},{p_{2}},{q_{1}})$ space.
infor607_g011.jpg
Fig. 11
Three-dimensional phase trajectories of numerical and predicted solutions in the $({q_{2}},{p_{2}},{p_{1}})$ space.
We also draw numerical and predictive solutions for the three-dimensional phase diagram of the beam-ring structure. Figures 8–11 compare the three-dimensional phase portraits of the numerical solution and the predicted solution. The phase portrait illustrates the system’s trajectory in phase space, capturing the interplay between key state variables. The strong agreement between the predicted and numerical solutions demonstrates that the neural network effectively captures the system’s underlying dynamics. The predicted phase portrait closely aligns with the numerical solution, preserving both the trajectory and essential geometric structures. This consistency highlights the neural network’s ability to learn the system’s governing dynamics.
Detailed results of all the above numerical simulations (including mean square error and improvements for all numerical simulations) are given in Table 2.
Table 2
Comparison of errors for loss functions ${L_{3}}$ and ${L_{3a}}$.
Neural networks Comparison Planes
$(t,{q_{1}})$ $(t,{q_{2}})$ $(t,{p_{1}})$ $(t,{p_{2}})$
LSTM$\text{\_}$ED ${L_{3}}$ 46.1518 2.9058 1.5152 0.4531
${L_{3a}}$ 10.8199 3.0217 0.3825 0.4776
% 76.56% −3.99% 74.75% −5.41%
S-Transformer ${L_{3}}$ 22.9896 3.1256 0.6811 0.4922
${L_{3a}}$ 1.3374 0.9517 0.0448 0.1517
% 94.18% 69.55% 93.43% 69.18%
Conv-AT ${L_{3}}$ 36.2087 3.5679 1.0740 0.5601
${L_{3a}}$ 9.1624 2.6030 0.2589 0.4136
% 74.70% 27.04% 75.90% 26.16%

5.2 Variable Step Loss Function

As can be seen from the previous sections, the prediction of the loss functions ${L_{3}}$ and ${L_{3a}}$ is significantly improved, but there is still some error. In addition, the loss functions ${L_{3}}$ and ${L_{3a}}$ have the limitation that they require the step size of the prediction to be the same as the step size of the training data. These two drawbacks motivate us to improve it further.
In numerical methods for solving initial value problems of ordinary differential equations, smaller step sizes provide key advantages. They enhance solution accuracy by reducing local truncation errors, particularly in regions with rapid changes. Additionally, smaller step sizes improve the stability of the numerical solution by minimizing the risk of error accumulation.
Therefore, our next goal is to continue to improve the loss function ${L_{3a}}$ so that the loss function can adapt to different prediction step sizes. The main goal is to adapt the loss function to smaller prediction steps with a fixed step size of the training data, thus taking advantage of higher accuracy and stability in the numerical computation.

5.2.1 Interpolation

The ${\varepsilon _{3m}}$ in the loss function ${L_{3a}}$ is mainly to control the error of the iterative formula in the prediction process, so the step size of the prediction process must be consistent with the step size of the ${\varepsilon _{3m}}$. This limits the step size at prediction time, i.e. we cannot use smaller step sizes in the prediction phase to improve numerical stability. In fact, in some cases small step sizes are very important for the accuracy of the prediction results. Therefore, if the training data is given, we introduce the interpolation method to use a smaller step size in the prediction phase.
From the perspective of computational time and memory usage, this study employs simple linear interpolation for numerical simulations. However, this choice does not preclude the effectiveness of other interpolation methods, and readers are encouraged to explore alternative approaches.
Linear interpolation is a simple and widely used numerical interpolation method for estimating unknown values between known data points. It is based on two known points and interpolation is performed by drawing a straight line between these two points. Given two known points $({x_{0}},{y_{0}})$ and $({x_{1}},{y_{1}})$, linear interpolation is computed arbitrarily by the following equation x values (satisfying ${x_{0}}\leqslant x\leqslant {x_{1}}$) corresponds to the y values:
(17)
\[ y={y_{0}}+\frac{{y_{1}}-{y_{0}}}{{x_{1}}-{x_{0}}}\cdot (x-{x_{0}}).\]
The number of insertion points is determined by the desired step size, and the ${x_{k}}$ values are then selected to be evenly spaced accordingly. These ${x_{k}}$ values can be generated using linear intervals: ${x_{k}}={x_{0}}+k\cdot \frac{{x_{1}}-{x_{0}}}{n+1}$, $k=1,2,\dots ,n$.
Since the interpolated data actually has some error, we only use the interpolated data to train for part ${\varepsilon _{3m}}$. Parts ${\varepsilon _{1i}}$ and ${\varepsilon _{2j}}$ are still trained with the original data. We denote the improved loss function as ${L_{3b}}$, and the schematic is shown in Fig. 12. As can be seen from the Fig. 12, the step size of the ${\varepsilon _{1i}}$ and ${\varepsilon _{2j}}$ parts is notated as h, which is consistent with the sampling step size of the training data. The ${\varepsilon _{3m}}$ part is closely related to the prediction phase from the previous introduction. Therefore, when the step size of the formula of the iterative formula of the prediction phase is set to $h/n$, we interpolate the data input to the ${\varepsilon _{3m}}$ part so that it has a step size of $h/n$. This realizes the use of training data with larger step sizes, but smaller step sizes during the prediction process to improve the accuracy of the prediction.
infor607_g012.jpg
Fig. 12
Loss function ${L_{3b}}$.
Taking the Conv-AT neural network as an example, we refine the step size in the loss function from $1500/120000$ to $1500/480000$ through interpolation. In each batch, since the number of interpolated data points is nearly four times that of the original data, directly applying recursion to the entire interpolated dataset would result in excessive errors, making it difficult for the model to effectively perform backpropagation and update parameters. To address this, we divide the interpolated data into four groups and apply recursive computation independently within each group. This way the error is not too large thus ensuring that the neural network can back propagate and update the parameters effectively.

5.2.2 Results and Discussion

The training and prediction ranges are kept consistent with those in the previous sections. The displacement time series predicted by Conv-AT using the new loss function ${L_{3b}}$ is shown in Fig. 13.
infor607_g013.jpg
Fig. 13
Comparison of displacement time series curves and errors predicted by Conv-AT using two loss functions.
As can be seen in Fig. 13, the predicted values of the neural network using the loss function ${L_{3b}}$ are closer to the numerical solution than ${L_{3a}}$, and the cumulative mean-square error plot also shows that ${L_{3b}}$ accumulates less error. It can be seen that our two improved strategies still give better results when combined.
In fact, we also made the same improvement for LSTM$\text{\_}$ED, S-Transformer, but their prediction errors were instead improved. Therefore, we conjecture that the interpolation method is more effective for neural network models that require high numerical stability. Further investigations are needed to quantify its impact on different types of neural networks.

6 Conclusions

In this paper, based on the original linear multistep loss function, a new loss function strategy is proposed under the condition that the time step is guaranteed to be consistent. In order to test the effect of the traditional loss function (MSE), custom loss function and the improved custom loss function on the network prediction, different neural networks are used to predict the time series of the beam ring system. Based on the analysis, the key conclusions are summarized as follows:
  • 1. For neural networks of the same size, the use of traditional MSE loss functions is very ineffective and cannot accurately capture the complex dynamic behaviour of the system. However, a custom loss function based on three types of errors can more effectively capture the complex dynamics of the system, thereby significantly enhancing prediction accuracy.
  • 2. Numerical simulations show that the improved loss function has higher prediction accuracy with the same neural network size. In addition, the time increase of using the improved loss function is negligible, and the strategy of improving the loss function is generalisable.
  • 3. To further optimise the loss function, we have introduced RecursiveLoss. By calculating the loss function in a recursive form, the loss function can simulate the error that occurs in the prediction phase, which effectively improves the prediction accuracy. Interpolation methods are then introduced on this basis to reduce the step size of the prediction phase under the same input conditions. The addition of interpolation to the prediction error term of the loss function aims to overcome the dependence of the adapted loss function on the step size and broaden its scope of application. Finally, the applicability of the interpolation method is discussed.
It is worth mentioning that our improvement of the loss function does not involve the addition of specific formulas, but rather the incorporation of iterative prediction methods into the loss function in order to control the errors generated during the iterative process. This strategy is applicable to any iterative method and can therefore provide some assistance in interdisciplinary research in applied mechanics and machine learning.

References

 
Akçakaya, M., Moeller, S., Weingärtner, S., Uğurbil, K. (2019). Scan-specific robust artificial-neural-networks for k-space interpolation (RAKI) reconstruction: database-free deep learning for fast imaging. Magnetic Resonance in Medicine, 81, 439–453.
 
Calvo-Rolle, J.L., Fontenla-Romero, O., Pérez-Sánchez, B., Guijarro-Berdiñas, B. (2014). Adaptive inverse control using an online learning algorithm for neural networks. Informatica, 25(3), 401–414.
 
Chopra, S., Hadsell, R., LeCun, Y. (2005). Learning a similarity metric discriminatively, with application to face verification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vol. 1, pp. 539–546.
 
D’Ambrosio, A., Schiassi, E., Curti, F., Furfaro, R. (2021). Pontryagin neural networks with functional interpolation for optimal intercept problems. Mathematics, 9(9), 996.
 
Fernandes, B., Silva, F., Alaiz-Moreton, H., Novais, P., Neves, J., Analide, C. (2020). Long short-term memory networks for traffic flow forecasting: exploring input variables, time frames and multi-step approaches. Informatica, 31(4), 723–749.
 
Ghazvini, A., Sharef, N.M., Sidi, F.B. (2024). Prediction of course grades in computer science higher education program via a combination of loss functions in LSTM model. IEEE Access, 12, 30220–30241.
 
Hadsell, R., Chopra, S., LeCun, Y. (2006). Dimensionality reduction by learning an invariant mapping. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vol. 2, pp. 1735–1742.
 
He, J., Li, X.X., Zhu, H.Q. (2023). An adaptive discrete physics-informed neural network method for solving the Cahn-Hilliard equation. Engineering Analysis with Boundary Elements, 155, 826–838.
 
Huber, P.J. (1964). Robust estimation of a location parameter. Annals of Mathematical Statistics, 35(1), 73–101.
 
Kendall, A., Gal, Y., Cipolla, R. (2018). Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7482–7491.
 
Lamb, A., Verma, V., Kannala, J., Bengio, Y. (2019). Interpolated adversarial training: achieving robust neural networks without sacrificing too much accuracy. In: Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security (AISec’19), pp. 95–103.
 
Lee, H., Lee, J., Cho, S. (2017). View-interpolation of sparsely sampled sinogram using convolutional neural network. In: Medical Imaging 2017: Image Processing, Vol. 10133, p. 1013328.
 
Li, K.J. (2024). MultiPINN: multi-head enriched physics-informed neural networks for differential equations solving. Neural Computing and Applications, 36, 11371–11395.
 
Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollar, P. (2017). Focal loss for dense object detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42, 318–327.
 
Liu, T., Meidani, H. (2023). Physics-informed neural networks for system identification of structural systems with a multiphysics damping model. Journal of Engineering Mechanics, 149(10), 04023079.
 
Papineni, K., Roukos, S., Ward, T., Zhu, W.J. (2002). BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, pp. 311–318.
 
Ranzato, M., Chopra, S., Auli, M., Zaremba, W. (2015). Sequence Level Training with Recurrent Neural Networks. arXiv preprint: arXiv:1511.06732.
 
Rendle, S., Freudenthaler, C., Gantner, Z., Schmidt-Thieme, L. (2009). BPR: Bayesian personalized ranking from implicit feedback. In: Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence, pp. 452–461.
 
Rumelhart, D., Hinton, G., Williams, R. (1986). Learning representations by back-propagating errors. Nature, 323, 533–536.
 
Sestanovic, T., Kalinic Milicevic, T. (2025). Identification of the optimal neural network architecture for prediction of Bitcoin return. Informatica, 36(1), 175–196.
 
Shannon, C.E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27(3), 379–423.
 
Sun, Y., Zhang, L.Y., Yao, M.H., Zhang, J.H. (2024). Neural network models and Shapley additive explanations for a beam-ring structure. Chaos, Solitons & Fractals, 185, 115114.
 
Verma, V., Kawaguchi, K., Lamb, A., Kannala, K., Solin, A., Bengio, Y., Lopez-Paz, D. (2022). Interpolation consistency training for semi-supervised learning. Neural Networks, 145, 90–106.
 
Wu, Z., Rincon, D., Christofides, P.D. (2020). Incorporating structural process knowledge in recurrent neural network modeling of nonlinear processes. In: Proceedings of the American Control Conference (ACC), pp. 2413–2418.
 
Xie, X.P., Zhang, G.N., Webster, C.G. (2019). Non-intrusive inference reduced order model for fluids using deep multistep neural network. Mathematics, 7(8), 757.
 
Yang, M., Zhang, Y.N., Hu, H.F. (2020). Discrete ZNN models of Adams-Bashforth (AB) type solving various future problems with motion control of mobile manipulator. Neurocomputing, 384, 84–93.
 
Yin, S.B., Hu, S.H., Wang, Y.B., Yang, Y.H. (2023). High-order Adams Network (HIAN) for image dehazing. Applied Soft Computing, 139, 110204.
 
Yuan, Q.W., He, J.W., Yu, L., Zheng, G. (2020). Aim-Net: bring implicit Euler to network design. In: 2020 IEEE International Conference on Image Processing (ICIP), pp. 1926–1930.
 
Zhai, W.D., Tao, D.W., Bao, Y.Q. (2023). Parameter estimation and modeling of nonlinear dynamical systems based on Runge-Kutta physics-informed neural network. Nonlinear Dynamics, 111, 21117–21130.
 
Zhang, L.Y., Sun, Y., Wang, A.W., Zhang, J.H. (2023). Neural network modeling and dynamic behavior prediction of nonlinear dynamic systems. Nonlinear Dynamics, 111, 11335–11356.
 
Zhang, W., Wu, R.Q., Behdinan, K. (2019). Nonlinear dynamic analysis near resonance of a beam-ring structure for modeling circular truss antenna under time-dependent thermal excitation. Aerospace Science and Technology, 86, 296–311.
 
Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., Zhang, W. (2021). Informer: beyond efficient transformer for long sequence time-series forecasting. In: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35, pp. 11106–11115.
 
Zhu, D., Cheng, X., Zhang, F., Yao, X., Gao, Y., Liu, Y. (2019). Spatial interpolation using conditional generative adversarial neural networks. International Journal of Geographical Information Science, 34(4), 735–758.
 
Zhu, M., Chang, B., Fu, C. (2022). Convolutional neural networks combined with Runge-Kutta methods. Neural Computing and Applications, 35(2), 1629–1643.

Biographies

An Kun
ankun1906@163.com

K. An is a postgraduate student in the School of Applied Science at Beijing Information Science and Technology University, Beijing, China. He completed a bachelor of science degree in Applied Statistics at Xuzhou University of Technology. His research program encompasses data-driven modelling and deep learning.

Sun Ying
https://orcid.org/0000-0001-5611-2800
sunying0000@126.com

Y. Sun was awarded a PhD by Beijing University of Technology. She currently holds the position of associate professor at the School of Applied Science, Beijing Information Science and Technology University. Her research program encompasses data-driven modelling and the analysis of nonlinear dynamical systems, including chaos and bifurcation theory.

Yao Minghui
merry_mingming@163.com

M. Yao is currently a professor at School of Aeronautics and Astronautics, Tiangong University, China. Her research interests mainly focus on nonlinear vibrations of mechanical structures, energy harvesting, self-power sensors and systems and artificial intelligence.

Zhang Junhua
hua@bistu.edu.cn

J. Zhang received the PhD degree in engineering mechanics from Beijing University of Technology in 2009, and the MSc degree in applied mathematics from Hebei University of Technology in 2005. She is currently a professor with the College of Mechanical and Electrical Engineering, Beijing Information Science and Technology University of China. Her research interests include dynamics and control of nonlinear mechanical systems, design of lightweight metamaterial structures.


Exit Reading PDF XML


Table of contents
  • 1 Introduction
  • 2 Constructing Loss Function
  • 3 Neural Networks Setting
  • 4 Numerical Simulation
  • 5 Advancing the Design of the Loss Function
  • 6 Conclusions
  • References
  • Biographies

Export citation

Copy and paste formatted citation
Placeholder

Download citation in file


Share


RSS

INFORMATICA

  • Online ISSN: 1822-8844
  • Print ISSN: 0868-4952
  • Copyright © 2023 Vilnius University

About

  • About journal

For contributors

  • OA Policy
  • Submit your article
  • Instructions for Referees
    •  

    •  

Contact us

  • Institute of Data Science and Digital Technologies
  • Vilnius University

    Akademijos St. 4

    08412 Vilnius, Lithuania

    Phone: (+370 5) 2109 338

    E-mail: informatica@mii.vu.lt

    https://informatica.vu.lt/journal/INFORMATICA
Powered by PubliMill  •  Privacy policy