Investigating the intrinsic top-down dynamics of deep generative models

Abstract Hierarchical generative models can produce data samples based on the statistical structure of their training distribution. This capability can be linked to current theories in computational neuroscience, which propose that spontaneous brain activity at rest is the manifestation of top-down...

Full description

Saved in:
Bibliographic Details
Main Authors: Lorenzo Tausani, Alberto Testolin, Marco Zorzi
Format: Article
Language:English
Published: Nature Portfolio 2025-01-01
Series:Scientific Reports
Online Access:https://doi.org/10.1038/s41598-024-85055-y
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832585772360794112
author Lorenzo Tausani
Alberto Testolin
Marco Zorzi
author_facet Lorenzo Tausani
Alberto Testolin
Marco Zorzi
author_sort Lorenzo Tausani
collection DOAJ
description Abstract Hierarchical generative models can produce data samples based on the statistical structure of their training distribution. This capability can be linked to current theories in computational neuroscience, which propose that spontaneous brain activity at rest is the manifestation of top-down dynamics of generative models detached from action-perception cycles. A popular class of hierarchical generative models is that of Deep Belief Networks (DBNs), which are energy-based deep learning architectures that can learn multiple levels of representations in a completely unsupervised way exploiting Hebbian-like learning mechanisms. In this work, we study the generative dynamics of a recent extension of the DBN, the iterative DBN (iDBN), which more faithfully simulates neurocognitive development by jointly tuning the connection weights across all layers of the hierarchy. We characterize the number of states visited during top-down sampling and investigate whether the heterogeneity of visited attractors could be increased by initiating the generation process from biased hidden states. To this end, we train iDBN models on well-known datasets containing handwritten digits and pictures of human faces, and show that the ability to generate diverse data prototypes can be enhanced by initializing top-down sampling from “chimera states”, which represent high-level features combining multiple abstract representations of the sensory data. Although the models are not always able to transition between all potential target states within a single-generation trajectory, the iDBN shows richer top-down dynamics in comparison to a shallow generative model (a single-layer Restricted Bolzamann Machine). We further show that the generated samples can be used to support continual learning through generative replay mechanisms. Our findings suggest that the top-down dynamics of hierarchical generative models is significantly influenced by the shape of the energy function, which depends both on the depth of the processing architecture and on the statistical structure of the sensory data.
format Article
id doaj-art-52017c8e5d39467ea3747d7de31850e8
institution Kabale University
issn 2045-2322
language English
publishDate 2025-01-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj-art-52017c8e5d39467ea3747d7de31850e82025-01-26T12:30:52ZengNature PortfolioScientific Reports2045-23222025-01-0115111310.1038/s41598-024-85055-yInvestigating the intrinsic top-down dynamics of deep generative modelsLorenzo Tausani0Alberto Testolin1Marco Zorzi2Department of General Psychology and Padova Neuroscience Center, University of PadovaDepartment of General Psychology and Padova Neuroscience Center, University of PadovaDepartment of General Psychology and Padova Neuroscience Center, University of PadovaAbstract Hierarchical generative models can produce data samples based on the statistical structure of their training distribution. This capability can be linked to current theories in computational neuroscience, which propose that spontaneous brain activity at rest is the manifestation of top-down dynamics of generative models detached from action-perception cycles. A popular class of hierarchical generative models is that of Deep Belief Networks (DBNs), which are energy-based deep learning architectures that can learn multiple levels of representations in a completely unsupervised way exploiting Hebbian-like learning mechanisms. In this work, we study the generative dynamics of a recent extension of the DBN, the iterative DBN (iDBN), which more faithfully simulates neurocognitive development by jointly tuning the connection weights across all layers of the hierarchy. We characterize the number of states visited during top-down sampling and investigate whether the heterogeneity of visited attractors could be increased by initiating the generation process from biased hidden states. To this end, we train iDBN models on well-known datasets containing handwritten digits and pictures of human faces, and show that the ability to generate diverse data prototypes can be enhanced by initializing top-down sampling from “chimera states”, which represent high-level features combining multiple abstract representations of the sensory data. Although the models are not always able to transition between all potential target states within a single-generation trajectory, the iDBN shows richer top-down dynamics in comparison to a shallow generative model (a single-layer Restricted Bolzamann Machine). We further show that the generated samples can be used to support continual learning through generative replay mechanisms. Our findings suggest that the top-down dynamics of hierarchical generative models is significantly influenced by the shape of the energy function, which depends both on the depth of the processing architecture and on the statistical structure of the sensory data.https://doi.org/10.1038/s41598-024-85055-y
spellingShingle Lorenzo Tausani
Alberto Testolin
Marco Zorzi
Investigating the intrinsic top-down dynamics of deep generative models
Scientific Reports
title Investigating the intrinsic top-down dynamics of deep generative models
title_full Investigating the intrinsic top-down dynamics of deep generative models
title_fullStr Investigating the intrinsic top-down dynamics of deep generative models
title_full_unstemmed Investigating the intrinsic top-down dynamics of deep generative models
title_short Investigating the intrinsic top-down dynamics of deep generative models
title_sort investigating the intrinsic top down dynamics of deep generative models
url https://doi.org/10.1038/s41598-024-85055-y
work_keys_str_mv AT lorenzotausani investigatingtheintrinsictopdowndynamicsofdeepgenerativemodels
AT albertotestolin investigatingtheintrinsictopdowndynamicsofdeepgenerativemodels
AT marcozorzi investigatingtheintrinsictopdowndynamicsofdeepgenerativemodels