Unveiling learning trends in convolutional neural networks with dynamic mode decomposition

dc.contributor.author
Okkath Krishnanunni, Sikha
dc.contributor.author
González Ballester, Miguel Ángel, 1973-
dc.contributor.author
Benitez, Raúl
dc.date.accessioned
2026-03-18T00:32:01Z
dc.date.available
2026-03-18T00:32:01Z
dc.date.issued
2026-03-17T13:34:41Z
dc.date.issued
2026-03-17T13:34:41Z
dc.date.issued
2025
dc.date.issued
2026-03-17T13:34:41Z
dc.identifier
Okkath Krishnanunni S, González Ballester MÁ, Benitez R. Unveiling learning trends in convolutional neural networks with dynamic mode decomposition. Eng Appl Artif Intell. 2025 Mar 15;144:110146. DOI: 10.1016/j.engappai.2025.110146
dc.identifier
0952-1976
dc.identifier
https://hdl.handle.net/10230/72829
dc.identifier
http://dx.doi.org/10.1016/j.engappai.2025.110146
dc.identifier.uri
https://hdl.handle.net/10230/72829
dc.description.abstract
The widespread adoption of pre-trained models in artificial intelligence poses several evaluation challenges, primarily due to limited access to comprehensive information including training data, test data, loss function, and hyperparameter values. Traditional evaluation metrics, like accuracy, precision, and recall, demand the availability of train and test data, making it difficult to determine the expected performance or quality of the deep learning models. This paper addresses this challenge by introducing a comprehensive empirical analysis of Convolutional Neural Networks (CNNs) using Dynamic Mode Decomposition (DMD) theory. Feature maps from the convolutional layers are used to model the trained CNN as a linear dynamical system. By constructing snapshots for DMD analysis, we decompose the system's dynamics into individual modes and eigenvalues, providing a linear approximation of the intricate CNN behavior. We also introduce DMD-based metrics, including entropy of DMD-eigenvalues (EDMD), DMD eigen gap/spectral gap, and zero amplitude DMD modes, to quantify the effectiveness and quality of the learned CNN (well-trained vs poorly-trained) without the need to access the test data. Results indicate that the entropy of DMD-eigenvalues (EDMD) correlates well with the classical evaluation metrics and also distinguishes between well-trained and poorly-trained models. Additionally, we explored the potential of interpreting CNNs through stable DMD modes and their corresponding eigenvalues. Our approach was thoroughly evaluated on both shallow and deep models across various datasets for classification and segmentation tasks, demonstrating its effectiveness in quantifying CNN learning evolution and task performance. We also conducted a comprehensive analysis to examine the correlation between the proposed metrics and classical evaluation metrics on pre-trained models, specifically using ImageNet for classification (20 models) and Pascal VOC for segmentation (3 models) tasks. The code is available at: https://github.com/sikha2552/DMD_XAI.
dc.format
application/pdf
dc.format
application/pdf
dc.language
eng
dc.publisher
Elsevier
dc.relation
Engineering Applications of Artificial Intelligence. 2025;144:110146
dc.rights
© 2025 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
dc.rights
http://creativecommons.org/licenses/by/4.0/
dc.rights
info:eu-repo/semantics/openAccess
dc.subject
Explainable artificial intelligence
dc.subject
Interpreting convolutional neural networks
dc.subject
Empirical analysis
dc.subject
Dynamic mode decomposition
dc.title
Unveiling learning trends in convolutional neural networks with dynamic mode decomposition
dc.type
info:eu-repo/semantics/article
dc.type
info:eu-repo/semantics/publishedVersion


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)