How does the structure of a neural circuit shape its function? @neuralreckoning.bsky.social & I explore this in our new preprint: doi.org/10.1101/2025... 🤖🧠🧪 🧵1/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.716Z
We start from an artificial neural network with 3 sets of units and 9 possible weight matrices (or pathways). By keeping the two feedforward pathways (W_ih, W_ho) and adding the other 7 in any combination, we can generate 2^7 distinct architectures. All 128 are shown in the post above. 🧵2/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.717Z
This allows us to interpolate between: Feedforward - with no additional pathways. Fully recurrent - with all nine pathways. We term the 126 architectures between these two extremes *partially recurrent neural networks* (pRNNs), as signal propagation can be bidirectional, yet sparse. 🧵3/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.718Z
To compare pRNN function, we introduce a set of multisensory navigation tasks we call *multimodal mazes*. In these tasks, we simulate networks as agents with noisy sensors, which provide local clues about the shortest path through each maze. We add complexity by removing cues or walls. 🧵4/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.719Z
We trained over 25,000 pRNNs on these tasks. And measured their: 📈 Fitness (task performance) 💹 Learning speed 📉 Robustness to various perturbations (e.g. increasing sensor noise) From these data, we reach three main conclusions. 🧵5/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.720Z
First, across tasks and functional metrics, many pRNN architectures perform as well as the fully recurrent architecture. Despite having less pathways and as few as ¼ the number of parameters. This shows that pRNNs are efficient, yet performant. 🧵6/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.721Z
Second, to isolate how each pathway changes network function, we compare pairs of circuits which differ by one pathway. Across pairs, we find that pathways have context dependent effects. E.g. here hidden-hidden connections decrease learning speed in one task but accelerate it in another. 🧵7/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.722Z
Third, to explore why different circuits function differently, we measured 3 traits from every network. We find that different architectures learn distinct sensitivities and memory dynamics which shape their function. E.g. we can predict a network’s robustness to noise from its memory. 🧵8/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.723Z
We’re excited about this work as it: ⭐ Explores a fundamental question: how does structure sculpt function in artificial and biological networks? ⭐ Provides new models (pRNNs), tasks (Multimodal mazes) and tools, in a pip-installable package: github.com/ghoshm/Multi... 🧵9/9
— Marcus Ghosh (@marcusghosh.bsky.social) 2025-08-01T08:26:57.724Z
Partial recurrence enables robust and efficient computation
Abstract
Links
Categories