We envisage AI architectures to analyze complex time series in an end-to-end fashion, meaning that the quantitative metrics of the time series are learned directly from data, without the use of specific human-thought algorithms. That is, we challenge AI to learn those specific algorithms. We present a first step in this direction, a benchmark test on linear dynamical processes. We tackle the archetypical task of learning the eigenvalues of the state-transition matrix of a linear (discrete-time, stable) dynamical system, from output data. We train a scalable LSTM neural network with artificially generated data from random matrices of dimension 2-to-5. With noise-free data, the performance of the trained network is very good (average R2=0.955), especially in estimating the dominant eigenvalues, whereas there is space for improvements on non-dominant real eigenvalues and on the dimension of the generating matrix. Remarkably, the performance is robust to measurement noise and the network outperforms the mean-square identification of the corresponding AR process (the latter giving exact eigenvalues on noise-free data) at noise standard deviation starting from 10−6.

End-to-end Artificial Intelligence to analyze dynamical processes: A linear benchmark test

Valle, João;Sangiorgio, Matteo;Dercole, Fabio
2025-01-01

Abstract

We envisage AI architectures to analyze complex time series in an end-to-end fashion, meaning that the quantitative metrics of the time series are learned directly from data, without the use of specific human-thought algorithms. That is, we challenge AI to learn those specific algorithms. We present a first step in this direction, a benchmark test on linear dynamical processes. We tackle the archetypical task of learning the eigenvalues of the state-transition matrix of a linear (discrete-time, stable) dynamical system, from output data. We train a scalable LSTM neural network with artificially generated data from random matrices of dimension 2-to-5. With noise-free data, the performance of the trained network is very good (average R2=0.955), especially in estimating the dominant eigenvalues, whereas there is space for improvements on non-dominant real eigenvalues and on the dimension of the generating matrix. Remarkably, the performance is robust to measurement noise and the network outperforms the mean-square identification of the corresponding AR process (the latter giving exact eigenvalues on noise-free data) at noise standard deviation starting from 10−6.
2025
Eigenvalues
Linear systems
Machine learning
Recurrent neural networks
File in questo prodotto:
File Dimensione Formato  
1-s2.0-S0167278925003574-main.pdf

accesso aperto

Descrizione: Paper PDF
: Publisher’s version
Dimensione 1.65 MB
Formato Adobe PDF
1.65 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1296285
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact