Reservoir computing in simulated neuronal cultures: Effectof network structure

Publication date

2026-03-18T12:24:46Z

2026-02-17

2026-03-18T12:24:46Z



Abstract

Biological neurons are emerging as attractive candidates for artificial intelligence and machine learning applications given their natural energy efficiency and self-repair capacity. However, they differ from their idealized artificial counterparts. Biological neurons have highly variable and noisy dynamics and display intrinsic spontaneous activity instead of purely input-driven dynamics. Moreover, biological neuronal networks have physically constrained and highly plastic connections, leading to a complex and ever evolving connectivity structure. Here, we investigate (numerically and with preliminary experimental data) the stability of the input responses of neuronal cultures using a reservoir computing framework. Utilizing a numerical model for the growth and activity of neuronal cultures, previously used to model experimental data, we investigate the effect of large-scale network topology, specifically homogeneous vs modular architectures, on fading memory, reservoir performance under increasingly noisy dynamics, and robustness to network rewiring. We find that modular networks exhibit longer fading memory time, sustain higher performance under noisy conditions, and are more robust to connectivity rewiring than homogeneous networks. Finally, we observe no relationship between some characteristics of the network adjacency matrix (specifically its spectral properties) and reservoir computing performance.

Document Type

Article


Published version

Language

English

Publisher

American Institute of Physics (AIP)

Related items

Reproducció del document publicat a: https://doi.org/10.1063/5.0278517

Chaos, 2026, vol. 36, p. 1-14

https://doi.org/10.1063/5.0278517

Recommended citation

This citation was generated automatically.

Rights

cc-by (c) Mats Houben, Akke, et al, 2026

https://creativecommons.org/licenses/by/4.0/