2026-03-18T12:24:46Z
2026-02-17
2026-03-18T12:24:46Z
Biological neurons are emerging as attractive candidates for artificial intelligence and machine learning applications given their natural energy efficiency and self-repair capacity. However, they differ from their idealized artificial counterparts. Biological neurons have highly variable and noisy dynamics and display intrinsic spontaneous activity instead of purely input-driven dynamics. Moreover, biological neuronal networks have physically constrained and highly plastic connections, leading to a complex and ever evolving connectivity structure. Here, we investigate (numerically and with preliminary experimental data) the stability of the input responses of neuronal cultures using a reservoir computing framework. Utilizing a numerical model for the growth and activity of neuronal cultures, previously used to model experimental data, we investigate the effect of large-scale network topology, specifically homogeneous vs modular architectures, on fading memory, reservoir performance under increasingly noisy dynamics, and robustness to network rewiring. We find that modular networks exhibit longer fading memory time, sustain higher performance under noisy conditions, and are more robust to connectivity rewiring than homogeneous networks. Finally, we observe no relationship between some characteristics of the network adjacency matrix (specifically its spectral properties) and reservoir computing performance.
Article
Published version
English
Xarxes neuronals (Neurobiologia); Neurotecnologia; Xarxes neuronals (Informàtica); Neural networks (Neurobiology); Neurotechnology; Neural networks (Computer science)
American Institute of Physics (AIP)
Reproducció del document publicat a: https://doi.org/10.1063/5.0278517
Chaos, 2026, vol. 36, p. 1-14
https://doi.org/10.1063/5.0278517
cc-by (c) Mats Houben, Akke, et al, 2026
https://creativecommons.org/licenses/by/4.0/