The discovery of resting-state networks shifted the focus from the role of local regions in cognitive tasks to the ongoing spontaneous dynamics in global networks. Recently, efforts have been invested to reduce the complexity of brain activity recordings through the application of nonlinear dimensionality reduction algorithms. Here, we investigate how the interaction between these networks emerges as an organising principle in human cognition. We combine deep variational autoencoders with computational modelling to construct a dynamical model of brain networks fitted to the whole-brain dynamics measured with functional magnetic resonance imaging (fMRI). Crucially, this allows us to infer the interaction between these networks in resting state and seven different cognitive tasks by determining the effective functional connectivity between networks. We found a high flexible reconfiguration of task-driven network interaction patterns and we demonstrate that this reconfiguration can be used to classify different cognitive tasks. Importantly, compared with using all the nodes in a parcellation, we obtain better results by modelling the dynamics of interacting networks in both model and classification performance. These findings show the key causal role of manifolds as a fundamental organising principle of brain function, providing evidence that interacting networks are the computational engines’ brain during cognitive tasks.

The discovery of resting-state networks has greatly influenced the investigation of brain functioning, shifting the focus from local regions involved in cognitive tasks to the ongoing spontaneous dynamics in global networks. This research goes beyond that shift and proposes investigating how human cognition is shaped by the interactions between whole-brain networks embedded in a low-dimensional manifold space. To achieve this, a combination of deep variational autoencoders with computational modelling is used to construct a dynamic model of brain networks, fitted to whole-brain dynamics measured with functional magnetic resonance imaging (fMRI). The results show that during cognitive tasks, highly flexible reconfigurations of task-driven network interaction patterns occur, and these patterns, in turn, can be used to accurately classify different cognitive tasks. Importantly, using this low-dimensional whole-brain network model provides significantly better results than working in the conventional brain space.

The high dimensionality of brain activity recordings is one of the major obstacles hindering experimental and theoretical efforts in neuroscience. At the level of single cells, the human brain is constituted of billions of neurons and synapses where these neurons interact (Sporns et al., 2005); however, coarser scales can also be understood as complex interconnected systems where nodal dynamics represent the aggregate of a macroscopic population of neurons (Deco et al., 2008, 2011). Even if exhaustive recordings to these systems could be obtained, it would remain difficult to reach encompassing theoretical principles due to the sheer complexity of the data. In particular, addressing the behaviour of individual cells does not suffice for this endeavour, since the high interconnectivity of the brain facilitates the emergence of coordinated activity (Biswal et al., 1995; Deco et al., 2008) and, thus, of distributed representations where information is encoded in the global dynamics of large populations of neurons.

The discovery of resting-state networks (RSNs) was a first step towards lowering the dimensionality of brain signals (Beckmann et al., 2005; Damoiseaux et al., 2006), which radically shifted the focus from the role of local regions in cognitive tasks to the ongoing spontaneous dynamics in global networks. In recent years, considerable efforts have been invested to tackle distributed representations by means of nonlinear methods capable of finding low-dimensional representations of brain activity at multiple spatiotemporal scales, from cortical microcircuits (Chaudhuri et al., 2019; Mitchell-Heggs et al., 2023) to whole-brain dynamics measured with functional magnetic resonance imaging (fMRI) (Gao et al., 2021; Glomb et al., 2021; Luppi et al., 2023; Perl et al., 2023; Rué-Queralt et al., 2021; Vidaurre et al., 2018; Vohryzek et al., 2023), among other neuroimaging methods. The success of this approach does rely not only on methodological considerations but also on the fundamental characteristics of brain activity. Despite the very large number of degrees of freedom of brain dynamics, coordinated cognition and behaviour cannot exist without the integration of this activity (Shine et al., 2016; Tononi et al., 1998, 2016). Neural information processing also exhibits redundancy (Hennig et al., 2018; Shine et al., 2019; Tononi et al., 1999), where the activity of several cells is constrained to develop within an abstract geometrical space of lower dimensionality, generally known as manifold. The organisation of brain activity into manifolds with a reduced number of dimensions, named networks or modes, is a frequently replicated phenomenon and has been proposed as a fundamental aspect of brain dynamics, fulfilling different scale-dependent computational roles (Pang et al., 2016). The manifold spans a subspace of modes, in which the brain activity evolves in time following rules that determine the behaviour of the system on that manifold defining a flow (Pillai & Jirsa, 2017). In other words, brain activity can be represented as a dynamical system in a low-dimensional space capturing the time evolution of the system in that space collapsing high-dimensional information to the manifold.

The so-called structured flows on manifolds emerge from basic multiscale processes such as symmetry breaking and near-criticality (Jirsa & Sheheitli, 2022). Indeed, a direct link has been proposed between the behaviour and the flow of the low-dimensional manifold underlying the brain dynamics (Fousek et al., 2024; Jirsa, 2020). In particular, it has been shown that manifold coordinates are relevant in describing cognitive brain processing and behaviour by parametrizing motor control (Gallego et al., 2017), perception (Chandak & Raman, 2023; Stringer et al., 2019), cognition and attention (Song et al., 2023), navigation (Derdikman & Moser, 2010), and sleep (Chaudhuri et al., 2019), among others examples. Overall, several investigations had mapped macroscale brain activities to low-dimensional manifold representations, linking measures of neural activity to cognition; however, it is not clear how these networks interact and how these interactions are related with the cognitive processing.

Therefore, we directly investigated how the interaction between these low-dimensional networks emerge as an organising principle supporting cognitive function, and we asked whether the reconfiguration of these interactions adequately tracks brain activity as participants perform different tasks. To address this question, we first assume that brain activity is most adequately described in terms of a reduced number of abstract variables, here termed as networks, each encoding the simultaneous dynamical behaviour of multiple macroscopic brain regions. These modes are neurobiologically meaningful as they emerge due to the large-scale anatomical and functional organisation of the brain. Formally, modes are the coordinates spanning the subspace, in which the low-dimensional manifold is embedded. The intrinsic dimensionality of brain dynamics corresponds to the minimum number of such modes required for its description. Based on a previous works, we used a deep learning architecture known as variational autoencoders (VAEs) to obtain low-dimensional representations of whole-brain activity from a large-scale fMRI dataset of more than 1,000 participants (Human Connectome Project, HCP). We modelled the dynamics of each coordinate on the manifold by a nonlinear oscillator, representing the mode and operating close to the bifurcation point. The dynamical behaviour changes from fixed point dynamics towards self-sustained oscillations, as successfully applied in several prior works modelling the dynamics of the high-dimensional state space (Deco et al., 2017; Jobst et al., 2017; Perl et al., 2022). To optimise the models, we inferred the effective connectivity between the modes considering the level of nonequilibrium dynamics of brain activity quantified by the nonreversibility of the signals (termed generative connectivity of the arrow of time (GCAT); following the work of Kringelbach et al. (2023)). We explored different manifold dimensions, and we demonstrated that by modelling brain dynamics on the manifold we achieve superior models compared with the traditional ones constructed in the source state space. Crucially, by investigating the reconfiguration of the interaction between networks across different tasks, we were able to track brain activity more precisely during seven cognitive tasks included in the HCP dataset. These results establish that the inferred connectivity in the manifold space provided more informative results compared with those obtained solely in the source space of empirical recordings.

Methodological Overview

Our methodological approach encompasses two complementary stages addressing the main questions we posed. The first analysis involves assessing the optimal dimension of the reduced space that yield the best computational dynamical models of interacting networks, and the second is focused on building these models to investigate how this interaction is modified during cognitive tasks. A summary of the proposed framework is displayed in Figure 1.

Figure 1.

Overview of the pipeline. (A) We trained a variational autoencoder (VAE) using the parcellation with N = 62 brain regions, which consists of the same three structures—encoder, latent space, and decoder—including a regularization loss function that gives VAEs generative properties. To model the dynamics of each network in the latent space, we utilized a nonlinear Stuart-Landau oscillator near the bifurcation point, that is, near the critical regime. In order to determine the connections between the latent variables, we employed the GCAT framework developed by Kringelbach and colleagues. We explored the performance of that models at different numbers of latent dimensions from 5 to 12. The result of this procedure is an optimised effective connectivity between the latent variables, Latent Generative Effective Connectivity (LGEC), which captures the departure from detailed balance and generates the nonequilibrium dynamics observed in brain signals. We also created whole-brain models in the source space to fit the same empirical data by repeating the same optimisation procedure for the latent space mode, that is, the inference of the GEC by implementing the GCAT framework. We compared the performance of both models, and we found that models in the latent space more faithfully reproduce the empirical data quantified by the similarity between the empirical and modelled functional connectivity (FC). (B) We trained the VAE with nine dimensions in the latent space using the spatial patterns over time obtained from the combined fMRI data from seven cognitive tasks from the HCP dataset. We investigated which brain regions contributed to each of the nine latent modes previously found in terms of brain functional networks. To do so, we created a set of surrogate signals in the latent space by introducing standard Gaussian noise into one latent dimension while keeping the other eight dimensions devoid of any signal. We repeated by changing the noise signal for each of the nine modes. We then decoded the surrogate latent signals in each case, obtaining the spatial pattern for each mode in the source space. We then associate these patterns with the activation or deactivation of the seven resting-state networks (RSNs) from Yeo. Finally, we assessed whether the interaction of these networks is driven by task activity by building computational models, obtaining a network interaction matrix for each task. We found that these interactions are flexible, showing high variability across tasks, allowing us to train a high-performance classifier based on that information, surpassing the classification capability of models trained in the source space.

Figure 1.

Overview of the pipeline. (A) We trained a variational autoencoder (VAE) using the parcellation with N = 62 brain regions, which consists of the same three structures—encoder, latent space, and decoder—including a regularization loss function that gives VAEs generative properties. To model the dynamics of each network in the latent space, we utilized a nonlinear Stuart-Landau oscillator near the bifurcation point, that is, near the critical regime. In order to determine the connections between the latent variables, we employed the GCAT framework developed by Kringelbach and colleagues. We explored the performance of that models at different numbers of latent dimensions from 5 to 12. The result of this procedure is an optimised effective connectivity between the latent variables, Latent Generative Effective Connectivity (LGEC), which captures the departure from detailed balance and generates the nonequilibrium dynamics observed in brain signals. We also created whole-brain models in the source space to fit the same empirical data by repeating the same optimisation procedure for the latent space mode, that is, the inference of the GEC by implementing the GCAT framework. We compared the performance of both models, and we found that models in the latent space more faithfully reproduce the empirical data quantified by the similarity between the empirical and modelled functional connectivity (FC). (B) We trained the VAE with nine dimensions in the latent space using the spatial patterns over time obtained from the combined fMRI data from seven cognitive tasks from the HCP dataset. We investigated which brain regions contributed to each of the nine latent modes previously found in terms of brain functional networks. To do so, we created a set of surrogate signals in the latent space by introducing standard Gaussian noise into one latent dimension while keeping the other eight dimensions devoid of any signal. We repeated by changing the noise signal for each of the nine modes. We then decoded the surrogate latent signals in each case, obtaining the spatial pattern for each mode in the source space. We then associate these patterns with the activation or deactivation of the seven resting-state networks (RSNs) from Yeo. Finally, we assessed whether the interaction of these networks is driven by task activity by building computational models, obtaining a network interaction matrix for each task. We found that these interactions are flexible, showing high variability across tasks, allowing us to train a high-performance classifier based on that information, surpassing the classification capability of models trained in the source space.

Close modal

In the first stage (Figure 1A), we created dynamical models of the low-dimensional manifold obtained from the fMRI data of resting-state participants in the HCP in a cortical brain of the DK62 cortical brain parcellation (Deco et al., 2021), which was constructed using Mindboggle-modified Desikan–Killiany parcellation (Desikan et al., 2006), with a total of 62 cortical regions. To do so, we used as input to the VAE the time-by-time spatial patterns from each region empirical time series (Figure 1A), that is, the values of the N brain regions at each time, and we created a training set to optimise the parameters of the VAE. Importantly, a VAE consists of three key components: the encoder network, the middle layer (referred to as the bottleneck or latent space), and the decoder network. In particular, a VAE presents generative properties by forcing a regularization of the latent space through incorporation of a regularization term during the training process; the VAE ensures that the decoding step produces outputs that are both relevant and meaningful (Kingma & Welling, 2013) (see the Methods section). Notice that since our aim was to construct generative models in the latent space with the intention of investigating the decoded modelled signals, regularization of the latent space provided by VAEs was essential.

Subsequently, we constructed computational models in the latent space by describing the dynamics of each latent network using a nonlinear Stuart-Landau oscillator close to the bifurcation point, that is, close to the critical regime (Figure 1A). To infer the connectivity between the latent variables, we implemented the GCAT framework developed by Kringelbach and colleagues, which involves a gradient descent procedure where the effective connectivity is updated at each step, considering the disparities between the empirical and simulated functional connectivity (FC), as well as the forward-shifted connectivity, to generate a good fit of both observables (i.e., zero- and one-lag FC) at the same time. The outcome of this procedure is an optimised effective connectivity between the latent variables (Latent Generative Effective Connectivity [LGEC]), which captures the breakdown of the detailed balance, reproducing the nonequilibrium dynamical behaviour of brain signals (Kringelbach et al., 2023). We then explored the fitting performance of the computational models as a function of the latent dimension. Finally, we were able to decode the modelled signals to further generate observables that we compared with those computed in the original state space to quantify the performance of modelling the networks that constitute the low-dimensional manifold. We also created whole-brain models fitted to the same the empirical data in the source space, that is, we replicated the GCAT framework in the original data consisting of the time series of each brain region (Figure 1A). We then compared the performance of both models in terms of the similarity between the empirical and modelled FC. We found that the optimal dimension of the latent model that more faithfully reproduces the empirical data is nine and that these models are better compared with the source space model.

In the second stage (Figure 1B), we trained the VAE using the spatial patterns obtained over time from the combined fMRI data from seven cognitive tasks from the HCP dataset. Based on the previous analysis, we determined the dimensionality of the latent space in nine networks. To assess the relation within these networks and brain regions, we created a set of surrogate signals in the reduced space by introducing a standard Gaussian noise into one network while keeping the other eight dimensions devoid of any signal. We repeated by changing the noise signal for each of the nine modes. We then decoded the surrogate signals in each case to obtain the corresponding spatial pattern in source space. We then associate these patterns with the activation (mode: positive) or deactivation (mode: negative) of the seven RSNs from Yeo (Yeo et al., 2011) (Figure 1B, upper row). Finally, to investigate how the interaction of these networks is tasks-driven activity reconfigured, we built computational models, obtaining a network interaction matrix for each task. We found that these interactions are flexible, showing high variability across tasks, allowing us to train a high-performance classifier based on that information, surpassing the classification capability of models trained in the source space (Figure 1B, lower row).

Models of Interacting Manifold Networks Outperform High-Dimensional Models in the Original Space

We employed a VAE to obtain a low-dimensional manifold representation of fMRI data obtained from resting-state healthy participants from the HCP dataset. First, we determined the dimensionality of the low-dimensional manifold, that is, the number of networks, by assessing the reconstruction error of the autoencoder using three different grained brain parcellations. We found that, independent of the input dimensionality, the reconstruction error shows a significant decrease in the initial dimensions until it reaches an elbow point at approximately a latent dimension of 10, beyond which the error remains relatively constant (see the Supporting Information Figure S1). We also show in Supporting Information Figure S1 the derivative of the reconstruction error with respect to the latent dimension, demonstrating that for all three cases, around and above a dimension of 10, the value is close to 0. This confirms that no major changes in the reconstruction error occur at higher dimensions. It is worth mentioning that different parcellations show different absolute values of reconstruction error, which, like many other quantities, depend on the parcellation graining (e.g., graph theoretical measures (Domhof et al., 2021); structure function relationship (Messé, 2020); entropy (Kobeleva et al., 2021); in a recent work, the reconstruction of an autoencoder (Jamison et al., 2024); and a systematic evaluation in fMRI (Luppi et al., 2024)) However, here, we focused on the number of dimensions that stabilize the reconstruction error, independent of the absolute value of the error. We then explored in the following sections the optimal low-dimensional space dimension around 10 using a coarse-grained parcellation scheme comprising 62 regions (see the Methods section and Supporting Information).

Subsequently, we modelled the dynamics of each network (i.e., mode) using nonlinear Stuart-Landau oscillators, which exhibit dynamic characteristics determined by a Hopf bifurcation (see the Methods section for the mathematical expression of the model).

Typically, these models are fitted to reproduce empirical observables, such as the FC (Ipiña et al., 2020; Perl et al., 2021). In our case, we built a manifold model to fit the empirical FC between latent space modes (FCLEmp; the subindex L stands for FC computed in the latent space and the super index emp means that uses encoded empirical data), which is computed as the pairwise Pearson’s correlation between the latent modes, and the empirical latent forward-shifted connectivity (FCfLEmp), which represents the pairwise Pearson’s correlation between the signals in latent space and the same signals shifted forward in time (see the Methods section). To improve the model fitting capacities, we averaged the FCLEmpand FCfLEmp by subgroups of 10 participants and created 100 models that fit these average matrices.

To construct the model of the low-dimensional manifold, we fixed the bifurcation parameter for each mode to aj  =  −0.02 and we built the whole manifold model by coupling the oscillators representing each latent neural mode. The connectivity between the modes was inferred using an iteratively pseudo-gradient algorithm, which was proposed by Kringelbach and colleagues (2023) for models in the source space. Briefly, this procedure consists of an iterative optimisation where, in each step, the connections are updated considering the differences between the empirical and the modelled FC in the latent space (shifted and nonshifted) weighted by a learning factor ς (see the Methods section). This iterative process enabled us to accurately capture the underlying relationships between latent modes. By fitting the FCLEmpat the same time with theFCfLEmp, we could effectively capture the nonequilibrium dynamic patterns in latent signals (Kringelbach et al., 2023). The result of this process is an enhanced effective connectivity for each one of the 100 models between the latent modes (LGEC), which also captures the disruption of the detailed balance, leading to the generation of nonequilibrium dynamic patterns in latent signals. We show that this disrupted detailed balance is also captured by the decoded signals.

We proceeded to assess the overall performance of the complete framework by examining the similarity between the modelled and decoded FC (FCLSMod computed as the Pearson’s correlation of the signals modelled in the low-dimensional manifold and then decoded to the source space) and the empirical FC in the source high-dimensional space (FCSEmp) quantified by the correlation between both FC. We also computed the correlation between the forward matrices FCfSEmp and FCfLSMod (Figure 2A). Both matrices exhibited a similar behaviour, displaying an optimal point at Latent Dimension 9, with decreasing performance observed for higher and lower dimensions. To validate the reliability of our approach, we constructed the model directly in the source state space using the GCAT framework, following the exact same methodology as employed in the latent space. However, we observed significantly lower levels of fitting for both observables compared with the results obtained from the low-dimensional manifold modelling (blue boxplot in Figure 2A). This discrepancy was observed not only for the optimal dimension but also across all dimensions under evaluation, demonstrating that even outside of the optimal manifold representation, the low-dimension representation is still more accurate than the high-dimensional original space. Importantly, by modelling the low-dimensional space, we are able to improve the quality of the fitting, avoiding the need to force the model to reproduce redundant and noisy information present in the high-dimensional representation. Consequently, our low-dimensional models accurately capture the interaction between these networks as a signature of cognitive states, allowing for increased performance of the support vector machine (SVM) classifier as we demonstrate later.

Figure 2.

Modelling the low-dimensional brain manifold reconstructs empirical data, outperforming models in the original state space. (A) We generated with a VAE a low-dimensional manifold representation of fMRI data from resting-state healthy participants in the Human Connectome Project (HCP) using a coarse-grained parcellation scheme consisting of 62 regions. To model the dynamics of each latent dimension within the manifold, we employed nonlinear Stuart-Landau oscillators. These oscillators exhibit dynamic characteristics that are governed by a Hopf bifurcation. We constructed a manifold mode model that accurately fit the FCLEmp and FCfLEmp,which represent the pairwise Pearson’s correlation between the signals in the latent space and the same signals with the corresponding shifted forward in time, respectively. The results of the performance of the complete framework in terms of reconstruction of FCSEmp and FCfSEmp are quantified as the correlation between these matrices and the ones obtained through decoding the manifold models for each dimension. The maximum correlations are reaching for Latent Dimension 9 and decreasing for both low and high dimension. Importantly, we demonstrated that these models overcome the performance of a model developed in the high-dimensional source space (blue boxplot), not only for the optimal dimensions but also for all latent dimension explored. (B) The hybrid matrices display in the upper diagonal triangle—the FCSEmp—and in the lower diagonal triangle—the FCLSmod—to observe the similarity between both matrices when exploring different latent dimensions ranging from 5 to 12.

Figure 2.

Modelling the low-dimensional brain manifold reconstructs empirical data, outperforming models in the original state space. (A) We generated with a VAE a low-dimensional manifold representation of fMRI data from resting-state healthy participants in the Human Connectome Project (HCP) using a coarse-grained parcellation scheme consisting of 62 regions. To model the dynamics of each latent dimension within the manifold, we employed nonlinear Stuart-Landau oscillators. These oscillators exhibit dynamic characteristics that are governed by a Hopf bifurcation. We constructed a manifold mode model that accurately fit the FCLEmp and FCfLEmp,which represent the pairwise Pearson’s correlation between the signals in the latent space and the same signals with the corresponding shifted forward in time, respectively. The results of the performance of the complete framework in terms of reconstruction of FCSEmp and FCfSEmp are quantified as the correlation between these matrices and the ones obtained through decoding the manifold models for each dimension. The maximum correlations are reaching for Latent Dimension 9 and decreasing for both low and high dimension. Importantly, we demonstrated that these models overcome the performance of a model developed in the high-dimensional source space (blue boxplot), not only for the optimal dimensions but also for all latent dimension explored. (B) The hybrid matrices display in the upper diagonal triangle—the FCSEmp—and in the lower diagonal triangle—the FCLSmod—to observe the similarity between both matrices when exploring different latent dimensions ranging from 5 to 12.

Close modal

Figure 2B depicts hybrid matrices as a function of the latent dimension that compare the empirical FC FCSEmp (upper triangle) and one obtained as FCLSMod (lower triangle).

Source Network Identification of Tasks-Activity Latent Networks

We investigated how cognitive processing is related to dynamics in the low-dimensional networks by analysing fMRI data from seven cognitive tasks from the HCP dataset (Social [SOC], Language [LAN], Working Memory [WM], Motor [MOT], Gambling [GAM], Emotion [EMO], Relational [REL]). Similar to the previous section, we trained the VAE using the spatial patterns over time obtained from the combined data of the rest condition and the seven tasks (see the Methods section). We fixed the latent dimension in nine based on the previous results and we investigated which brain regions contributed to each of the nine latent networks in terms of brain functional networks. To assess this, we obtained one spatial pattern on the empirical original space associated to each latent mode (see the Methods section). We evaluated these patterns in terms of the reference functional brain networks estimated by Yeo and colleagues: Visual (VIS), Somatomotor (SM), Dorsal Attention (DA), Ventral Attention (VA), Limbic (Lim), Frontoparietal (FP), and Default mode (DMN), known as the seven RSNs from Yeo et al. (2011). We computed the percentage of participation of each brain region within the DK62 parcellation to each of the seven RSNs. We then computed the Pearson’s correlation between each of the nine patterns corresponding to each neural mode and the percentage of participation of each brain region to each RSN. In Figure 3A, we show, as an example, that brain renders corresponding to the spatial pattern of the Network 9 is divided in positive and negative, standing for brain regions that present high correlation and anticorrelation to certain RSNs (DMN and Vis, respectively). Figure 3B shows the level of correlation between each latent mode (N“X,” with “X” varying from 1 to 9), and RSNs with an indicating star are the ones that are statistically significant after false discovery rate (FDR) correction. We found that some modes could be associated with the activation of the DMN at the same time with the deactivation of the VIS network (Latent Networks 6 and 9), while other can be associated with the activation of the Lim network (N1 and N7). Other latent networks could be described as the activation of VIS areas at the same time of the deactivation of the FP and DMN (N8). Finally, Latent Networks 4 and 5 could be characterised by the activation of primary/sensory motor function in high correlation with VA and DA networks. Interestingly, our association of latent dimensions with the activation and deactivation patterns in the brain shows high similarity with previous results based on nonsupervised clustering on brain activity considering the well-known coactivation patterns (CAPs) as is presented by Huang and colleagues (2020) and with functional gradients analysis (Margulies et al., 2016). Importantly, we replicated this analysis using seven dimensions for the latent space and we found that the networks show similar correspondence with the seven Yeo RSNs than the results for N = 9 (Supporting Information Figure S2). We also included in the Supporting Information the association between DK62 regions and the Yeo7 functional networks.

Figure 3.

RSNs are formed from the latent networks revealed by VAE. (A) We determined the brain regions associated with each mode. We show as an example the brain renders corresponding to the spatial pattern of the Mode 3 divided in Mode 3 positive and Mode 3 negative, standing for brain regions that present high correlation and anticorrelation to the DMN and Vis network, respectively. (B) We identified each latent mode with a pattern in the source space of 62 brain regions. We associated each spatial patterns with the Yeo7 RSNs by computing the correlation of each pattern with the percentage of belonging of each region to each RSN (* indicates the correlation that are significant after false discovery rate correction). The reference functional brain networks estimated by Yeo and colleagues are named as follows: Visual (VIS), Somatomotor (SM), Dorsal Attention (DA), Ventral Attention (VA), Limbic (Lim), Frontoparietal (FP), and Default Mode (DMN).

Figure 3.

RSNs are formed from the latent networks revealed by VAE. (A) We determined the brain regions associated with each mode. We show as an example the brain renders corresponding to the spatial pattern of the Mode 3 divided in Mode 3 positive and Mode 3 negative, standing for brain regions that present high correlation and anticorrelation to the DMN and Vis network, respectively. (B) We identified each latent mode with a pattern in the source space of 62 brain regions. We associated each spatial patterns with the Yeo7 RSNs by computing the correlation of each pattern with the percentage of belonging of each region to each RSN (* indicates the correlation that are significant after false discovery rate correction). The reference functional brain networks estimated by Yeo and colleagues are named as follows: Visual (VIS), Somatomotor (SM), Dorsal Attention (DA), Ventral Attention (VA), Limbic (Lim), Frontoparietal (FP), and Default Mode (DMN).

Close modal

Reconfiguration of Low-Dimensional Network Interaction During Cognitive Tasks

We then built a model for each task by inferring the effective connectivity between networks in the manifold to obtain the best performance to reproduce FCLEmp and FCfLEmp (see the Methods section). Thus, we built 100 models for each task and we assessed each model performance by measuring the correlation between the empirical FC (FCSEmp) and the latent model-decoded FC (FCLSmod) for each task (Figure 3A). For each task, we obtained the LGEC for each one of the 100 models, and in Figure 3B, we display the average LGEC across the model, representing the interaction between networks. Note that the interaction matrices are capturing the time asymmetry interaction between the networks and are consequently asymmetric. We show how interactions are reconfigured, comparing resting-state and social tasks (Figure 3C and D). To quantify how the interaction between low-dimensional networks is shaped by cognitive states, we computed the total level of connectivity of each network as the sum of all outcome and income interactions. We found significant differences (Wilcoxon rank-sum test, FDR corrected) in the level of TC for almost all networks (except for N7) in the comparison between resting state and social task (Figure 3C). The comparison between rest and other tasks is displayed in Supporting Information Figure S3, also showing a significant network interaction reconfiguration. We represented with directed graphs the interactions between networks above a threshold (LGECij  >  0.1) divided in the outcomes of each network, that is, how each network impacts others (Figure 3D, first column), and the incomes of each network, that is, how each network is driven by the others (Figure 3D, second column). Interestingly, the most important interactions are between networks involving the DMN (N6, N8, and N9), which is known to play an important role in the cortical organisation during cognition (Margulies et al., 2016). In particular, the interaction between N2 and N6 is observed during social tasks and not during rest, showing a reconfiguration of that interaction during social.

Finally, we trained a SVM using the 100 LGEC matrices generated from modelling the manifold for each task presented in Figure 4B. The objective was to determine if the information captured in each matrix could serve as a unique fingerprint for each task. We also trained the SVM classifier using the elements of the empirical FC (FCSEmp), the FC modelled with the GCAT framework (FCSMod), and the LGEC matrices with randomised labels (Figure 5A). Our results showed that the best classification accuracy (0.89 ± 0.01) was achieved using the LGEC matrices, outperforming the LGEC with randomised classes (0.13 ± 0.03), the FCSMod (0.78 ± 0.01), and the FCSEmp (0.76 ± 0.02) (Figure 5B). In Figure 5C, we also displayed in the rightmost column the confusion matrix obtained for the classification of the eight labels, the seven tasks, and resting state.

Figure 4.

Modelling the low-dimensional manifold network shows a flexible reconfiguration of network interaction during cognitive tasks. (A) We constructed models for each task and evaluated the performance of each model by quantifying the correlation between the FCSEmp and FCLSmod in each case. (B) The output of each model was the inferred connectivity between the latent variables, called the LGEC, for each task. We show the average LGEC across 100 models for each case. (C) We computed the total level of connectivity (TC) of each network as the sum of all outcome and income interactions. We found significant differences (Wilcoxon rank-sum test, false discovery rate corrected) in the level of TC for almost all networks (except for N7) in the comparison between resting state and social task. (*** means p < 0.001;* means 0.01 < p < 0.05). (D) The highest interactions between networks (above 0.1) represented as a graph. In the left column is the outcome’s connections, representing how each network drives the others, and in the right column is the income’s connections standing for the impact that the rest of the networks have on each network.

Figure 4.

Modelling the low-dimensional manifold network shows a flexible reconfiguration of network interaction during cognitive tasks. (A) We constructed models for each task and evaluated the performance of each model by quantifying the correlation between the FCSEmp and FCLSmod in each case. (B) The output of each model was the inferred connectivity between the latent variables, called the LGEC, for each task. We show the average LGEC across 100 models for each case. (C) We computed the total level of connectivity (TC) of each network as the sum of all outcome and income interactions. We found significant differences (Wilcoxon rank-sum test, false discovery rate corrected) in the level of TC for almost all networks (except for N7) in the comparison between resting state and social task. (*** means p < 0.001;* means 0.01 < p < 0.05). (D) The highest interactions between networks (above 0.1) represented as a graph. In the left column is the outcome’s connections, representing how each network drives the others, and in the right column is the income’s connections standing for the impact that the rest of the networks have on each network.

Close modal
Figure 5.

Interaction between low-dimensional manifold networks that better distinguishes between cognitive tasks. (A) We trained a support vector machine (SVM) using the 100 LGEC generated by modelling the manifold in each task to evaluate whether the information condensed in each matrix is a fingerprint for each task. We also trained the same SVM classifier by using as an input the elements of the FCSEmp, elements of the FCSMod, and also the LGEC with scrambled labels. (B) We found that the best classification is reached using the LGEC (0.89 ± 0.01) compared with the FCSEmp (0.76 ± 0.02), the FCSMod (0.78 ± 0.01), and the scrambled LGEC (0.13 ± 0.03). (C) The confusion matrix obtained for the SVM classifier in the task that distinguishes between eight classes (resting and the seven cognitive states).

Figure 5.

Interaction between low-dimensional manifold networks that better distinguishes between cognitive tasks. (A) We trained a support vector machine (SVM) using the 100 LGEC generated by modelling the manifold in each task to evaluate whether the information condensed in each matrix is a fingerprint for each task. We also trained the same SVM classifier by using as an input the elements of the FCSEmp, elements of the FCSMod, and also the LGEC with scrambled labels. (B) We found that the best classification is reached using the LGEC (0.89 ± 0.01) compared with the FCSEmp (0.76 ± 0.02), the FCSMod (0.78 ± 0.01), and the scrambled LGEC (0.13 ± 0.03). (C) The confusion matrix obtained for the SVM classifier in the task that distinguishes between eight classes (resting and the seven cognitive states).

Close modal

We demonstrated that interactions of low-dimensional manifold networks underlying whole-brain dynamics reveal how large-scale brain organisation is flexibly coordinated during cognitive processing. In this sense, our results expand traditional whole-brain models and low-dimensional brain representation analysis (e.g., RSNs (Damoiseaux et al., 2006) or gradients (Margulies et al., 2016)) by modelling the interaction between these low-dimensional networks. To do so, we combined deep learning VAEs and computational modelling to quantitatively infer the interaction between these data-driven obtained networks. Importantly, we showed that generating dynamical models within the subspace in which the low-dimensional manifold is embedded could yield more accurate models in terms of reproducing empirical data compared with modelling the original high-dimensional whole-brain space. Both results are to be expected under the assumption that large-scale dynamics are orchestrated into a reduced number of independent degrees of freedom, which emerge as part of a broader neurobiological principle underlying brain activity. In the following section, we discuss the implications of these results, in terms of both their contribution to systems neuroscience and their practical applications to neuroimaging.

Whole-Brain fMRI Dynamics are Intrinsically Low-Dimensional

Several past studies investigated the compressibility of resting-state fMRI data using linear methods such as principal (Carbonell et al., 2011; Margulies et al., 2016) and/or independent component analysis (Damoiseaux et al., 2006), revealing a set of canonical RSN overlapping with distinct neuroanatomical regions subserving specific cognitive functions (Smith et al., 2009). Other methods, such as the clustering of CAPs (Liu & Duyn, 2013), result in similar information. These networks also show alterations in certain neuropsychiatric disorders (Greicius, 2008) and contain information that can be used to predict and decode the contents of ongoing consciousness and cognition (Cole et al., 2016); moreover, the neurobiological underpinnings of the RSN are highlighted by multimodal imaging studies, revealing that their activity correlates with different frequency bands measured using simultaneous EEG (Mantini et al., 2007). As shown in Figure 3, the networks representing cognitive processing, revealed using autoencoders, are related to the RSN, although this relationship is not a simple one-to-one mapping. Instead, these networks are a combination of different RSNs, indicating that whole-brain dynamics can be spanned by different sets of functional networks. We also observed the same behaviour when considering the seven latent dimensions, showing that despite the number of networks is the same, there is not a one-to-one mapping between both sets of networks. Despite the mapping between RSN and networks, the latter have distinct properties due to the nature of the generative autoencoder algorithm. Besides its capacity to capture nonlinear relationships between the networks in the manifold and the high-dimensional data, autoencoders are generative, which is of key importance to translate the dynamics in the latent space towards that of resting-state high-dimensional fMRI data. Crucially, this generative capacity allows us to create dynamical models of interactive networks fitted to the whole-brain dynamics, revealing that changes in these interactions are reorganised during cognitive processing.

Large-Scale Manifolds and Cognition

The use of dimensionality reduction techniques is a common strategy in statistics and machine learning, applied with the purpose of simplifying the description of the data by performing a reasonable grouping of the variables (e.g., clustering) or expressing them as a linear or nonlinear combination of a reduced set of components (e.g., Independent Component Analysis, Principal Component Analysis) (Pang et al., 2016). The application of these methods to neural activity recordings is not only useful in terms of data processing, as expected, but also reveals important features of how the brain represents information relevant to cognition and behaviour (Chung & Abbott, 2021; Vidaurre et al., 2018). Research conducted using a variety of methods and experimental organisms shows converging evidence of the manifold organisation of neural activity, where manifold dimensions can be parametrized by variables with correspondence to behavioural observations. Examples include the representing of movement parameters in the motor cortex (Gallego et al., 2017), the mapping of spatiotemporal coordinates to the hippocampal cortex during navigation (Derdikman & Moser, 2010), and the encoding of sensory (e.g., visual (Stringer et al., 2019), olfactory (Chandak & Raman, 2023)) information. While these manifolds can change in response to learning (Nieh et al., 2021), they are generally stable (Gallego et al., 2020), and it is believed that they relate to the network of synaptic connections and their strengths; however, the relationship between manifolds and structural features remains difficult to demonstrate (Langdon et al., 2023). Our results suggest that large-scale brain activity also benefits from the organisation into low-dimensional manifolds. Indeed, the corresponding dimensions (i.e., networks) may parametrize the engagement of different cognitive functions in spontaneous mentation. Thus, the organisation of brain activity in the form of manifolds appears to occur across scales, where coarser scales index the participation of large assemblies of cells, whose dynamics may also be understood in terms of a low-dimensional space.

Interaction of Networks in the Manifold and Cognition

As we discussed above, low-dimensional representation of brain activity has been widely used to understand large-scale brain organisation during cognition. Here, we focused on modelling the low-dimensional space that not only allows us to enhance the fitting quality, eliminating the need to make the model replicate redundant or noisy data from the high-dimensional space, but also allows us to infer the interaction between the networks. Consequently, we went beyond the pure low-dimensional representation and focused how the interactions of these networks embedded in that low-dimensional manifold are reconfigured with task activity. We found that these interactions are reorganised during cognition and the classification of tasks from the HCP is optimal when based on the low-dimensional network representation, suggesting that these manifold dimensions are not arbitrary but encode organisational principles of brain function of cognitive processing instead (Figure 5). We also found the most important interactions are between networks involving the DMN (N6, N8, and N9), which is known to play an important role in the cortical organisation during cognition and primary/sensory motor function such as the Latent Network 4 (Margulies et al., 2016) (Figure 4). Overall, our results not only broadcast the relevance of the low-dimensional brain representation but also the interaction between networks in that space provides crucial information for understanding cognitive processing.

Possible Disruptions in Disease and Unconsciousness

As in the case of RSN, we expect that networks and its interactions are sensitive to different neurological and psychiatric conditions. If these networks reflect the representation of ongoing cognition and consciousness in large-scale patterns of brain activity, we could expect the most salient alterations in disorders leading to severe impairments in these domains. An example is the case of patients with brain injury who show a state of stable and persisting unresponsiveness, which can be interpreted as a lack of conscious thought (Schiff et al., 2014). Diminished or absent conscious content can also be present in certain transient states, either spontaneously occurring (such as deep sleep) or induced by pharmacological means (Brown et al., 2010). Previous research shows that a low-dimensional representation of brain dynamics captures the differences between the aforementioned states and conscious wakefulness, the former showing a reduced repertoire of states visited during the recording, and less structured and complex transition between (Rué-Queralt et al., 2021) these states (Varley et al., 2021). Future studies should adopt the framework developed here to investigate different pathological and physiological brain states, with the purpose of obtaining complementary information on the landmark features of brain dynamics during health and disease.

Modelling the Coupled Modes Dynamics

Our work provides novel methodological perspectives for the analysis of large-scale neuroimaging data. The majority of whole-brain modelling studies begin by adopting a certain parcellation of the human brain into regions of interest (ROI). Assuming that the anatomical connectivity between ROI is known, it is then possible to couple the local equations and simulate whole-brain dynamics. Our work suggests that this methodology may be suboptimal since brain dynamics are better represented by modes that span multiple ROIs; instead, models should strive to capture the relationship between networks. While the anatomical connectivity between the distributed sets of regions spanning the neural nodes is ill-defined, our result shows that equivalent results can be obtained using the effective connectivity. The generative capacity of autoencoders represents another attractive characteristic of our work, as it allows us to create new data following similar statistics as empirical recordings, with potential applications to machine learning and data augmentation. Importantly, our approach can be applied to different neuroimaging modalities such EEG/MEG (Magnetoencephalography), allowing to create models of the low-dimensional networks.

Limitations and Future Work

Our work presents some limitations that should be addressed in future studies. Perhaps the most salient one stems from our use of fMRI data, with known limitations to its biological interpretability. This should be assisted by other sources of information capable of complementing the recordings and shedding light on the origin of the neural models. Also, the use of improved MRI machines with higher fields could transcend the spatial organisation of the cortex into maps, allowing researchers to focus on layer-specific contributions. Finally, the set of tasks to be distinguished was relatively small. A larger repertoire of conditions should be included to tackle questions concerning the specificity of the underlying modes to classify pathological and physiological brain states.

Conclusions

We demonstrated that interactions of low-dimensional manifold networks underlying whole-brain dynamics reveal how large-scale brain organisation is flexibly coordinated during cognitive processing. Importantly, our results change the focus not only from brain regions to low-dimensional networks but also from network description towards network interaction. We quantified the interactions between these networks, combining deep learning autoencoders and dynamical modelling, leading to better results than the ones obtained in the original high-dimensionality space. We conclude that standard parcellations of brain activity are prone to overlook the underlying manifold organisation of fMRI and that future studies should attempt to characterise and model brain states, adopting the perspective of interactive networks.

Neuroimaging Ethics

The Washington University–University of Minnesota (WU-Minn HCP) Consortium obtained full informed consent from all participants, and research procedures and ethical guidelines were followed in accordance with Washington University Institutional Review Board approval.

Neuroimaging Participants

The neuroimaging dataset used in this investigation was obtained from the HCP March 2017 public data release. From the total of 1,003 participants available in the release, a sample of 1,000 was chosen for the first part when only considered rest and a sample of 990 participants was chosen for the second part because not all participants performed all tasks.

The HCP task battery comprises seven tasks, namely, working memory, motor, gambling, language, social, emotional, and relational tasks. Detailed descriptions of these tasks can be found on the HCP website (Barch et al., 2013). The tasks were intentionally designed by HCP to cover seven major cognitive domains, which aim to capture the diversity of neural systems: (a) visual, motion, somatosensory, and motor systems; (b) working memory, decision-making, and cognitive control systems; (c) category-specific representations; (d) language processing; (e) relational processing; (f) social cognition; and (g) emotion processing.

Apart from the resting-state scans, all 1,003 participants completed all tasks in two separate sessions. The first session included working memory, gambling, and motor tasks, while the second session involved language processing, social cognition, relational processing, and emotion processing tasks.

Neuroimaging Extraction of Functional Time Series

The processing and extracting of data from both resting-state and task-based fMRI datasets are detailed in the HCP. The preprocessing steps, described in detail on the HCP website, utilised standardised methods from FSL (FMRIB Software Library), FreeSurfer, and the Connectome Workbench software. The process included the correction of spatial and gradient distortions, addressing head motion, intensity normalization, bias field removal, registration to the T1-weighted structural image, and transformation to the 2-mm Montreal Neurological Institute (MNI) space. The FIX artefact removal procedure (Schröder et al., 2015; Smith et al., 2013) and ICA + FIX processing (Arslan et al., 2018; Griffanti et al., 2017) were employed to remove structured artefacts and head motion parameters, respectively.

The preprocessed time series data of all grayordinates were in HCP CIFTI grayordinates standard space and available in surface-based Connectivity Informatics Technology Initiative (CIFTI) files for each participant during resting state and each of the seven tasks. A custom-made MATLAB script using the ft_read_cifti function from the FieldTrip toolbox was used to extract the average time series for all the grayordinates in each region of the each parcellation used, as defined in the HCP CIFTI grayordinates standard space. The BOLD time series were filtered using a second-order Butterworth filter with a frequency range of 0.008–0.08 Hz and extracted the first and last 5% volumes to avoid the border effect of the filtering.

Parcellations

The neuroimaging data underwent processing using the three standard parcellations with different amount of brain cortical regions: (a) the Mindboggle-modified Desikan–Killiany parcellation, called DK62 (Desikan et al., 2006); (b) the Schaefer considering 500 regions; and (c) the Schaefer parcellation considering 1,000 regions. This DK62 parcellation consists of a total of 62 cortical regions, with 31 regions in each hemisphere, as defined by Klein and Tourville in 2012. Consequently, the DK80 parcellation contains a total of 62 regions, precisely defined in the common HCP CIFTI grayordinates standard space. Then, the Schaefer 500 and 1,000 parcellations were also used with 250 and 500 cortical regions per hemisphere, respectively, and well defined in the HCP CIFTI (Schaefer et al., 2018).

VAE Architecture and Training

The VAE consists of the same three structures as the autoencoder, but it incorporates certain differences that give VAEs generative properties. To guarantee this feature, during training, errors were propagated through gradient descent to minimize a loss function consisting of two terms. The first term is a standard reconstruction error, computed from the units in the output layer of the decoder. The second term is a regularization term, calculated as the Kullback-Leibler divergence between the distribution in the latent space and a standard Gaussian distribution. This regularization term ensures continuity and completeness in the latent space, meaning that similar values are decoded into similar outputs and that these outputs represent meaningful combinations of the encoded inputs. Our encoder network is composed of a deep neural network with rectified linear units as activation functions and two dense layers. This network funnels down to a low-dimensional variational layer, spanning the latent space explored from Dimensions 5 to 12. The encoder network applies a nonlinear transformation to map the values of each time pattern to Gaussian probability distributions within the latent space. Conversely, the decoder network replicates the architecture of the encoder, generating reconstructed temporal patterns by sampling from these distributions. We trained the VAE with time-by-time spatial patterns for the training set, creating from the 90%/10% (training/testing) resting-state HCP participants’ full dataset. The network training procedure consisted of batches with 128 samples and 50 training epochs using an Adam optimizer, and the training consists of error backpropagation via gradient descent to minimize a loss function that composed the two terms.

In the last section of this work, we extended the framework by including to the resting-state (REST) data the recordings from the seven cognitive tasks from HCP (SOC, LAN, WM, MOT, GAM, EMO, REL) to find the low-dimensional manifold that represents the dynamics that support cognitive processing. We repeated the procedure as explained above, but in this case, we train the VAE using the time-by-time spatial patterns from the full data by concatenating the rest condition and the seven tasks. We fixed the latent dimensions to nine and we used the DK62 parcellation based on our previous results to build models in the latent space.

Modes Coupled Model and LGEC Inference

We modelled the local dynamics of each latent neural mode as a Stuart-Landau oscillator (a dynamical system that can be expressed as the normal form of a supercritical Hopf bifurcation); this model presents the capability to shift from noisy to oscillatory dynamics. Importantly, this kind of applied to the whole-brain dynamics that have been able to replicate key aspects of brain dynamics observed in electrophysiology, magnetoencephalography (Cabral et al., 2014), and fMRI (Deco et al., 2017; Ipiña et al., 2020; Perl et al., 2022). Specifically, given a parcellation that number M of modes, the whole-brain dynamics of the coupled modes can be expressed by the local dynamics of M Stuart-Landau oscillators connected via the connectivity matrix C, which is defined by:
(1)
where the complex variable zj denotes the state (zj  =  xj + iyj) of mode j, ηj is the additive uncorrelated Gaussian noise with variance σ2 (for all j), and ωj is the intrinsic mode frequency. The intrinsic frequencies ωj (which lie in the 0.008–0.08 Hz band) were estimated from the empirical data as the averaged peak frequencies of the narrowband signals of the different modes. Finally, the mode’s bifurcation parameter is aj, which defines the dynamics of the mode with respect to the Hopf bifurcation. This bifurcation is characterised by bifurcation parameter (a) that induces qualitative changes in the dynamical behaviour when the system crosses the bifurcation point (a  =  0). In the supercritical regime (a  >  0), the oscillator generates self-sustained oscillations, whereas in the subcritical regime (a  <  0), the dynamics converge to a fixed point. Importantly, the addition of noise to the model causes the dynamics near the bifurcation point (a ~ 0) to stochastically transition between both regimes, resulting in oscillations with complex amplitude modulations. Previous studies have demonstrated that this regime represents the optimal point for fitting empirical data in whole-brain models (Deco et al., 2017; Ipiña et al., 2020; Perl et al., 2022). We fix the bifurcation parameter of each mode in that regime (a  =  −0.02).
We fit C such that the model optimally reproduces the empirical correlation matrix in the latent space FCLEmp (i.e., the pairwise correlation between the latent modes) and the empirical forward time-shifted correlation between the latent modes FCfLEmp(τ). We selected the parameter τ  =  3, which led to a decrease in the averaged autocorrelation also based on previous works (Deco et al., 2022). We note that fitting the time-shifted correlations can lead to asymmetries in the connectivity C, which, in turn, can produce nonequilibrium dynamics. Using a heuristic pseudo-gradient algorithm, we proceeded to update the C until the fit is fully optimised. More specifically, the updating uses the following form:
(2)
The model was run repeatedly with the updated C until the fit converges towards a stable value. We initialised C as a zero matrix of dimension MxM (the number of latent modes), and we update all the connections from this matrix (in either hemisphere). For the Stuart-Landau model, we used ς  =  0.00001 and continue until the algorithm converges. We generate 100 models by repeating this procedure to fit the average empirical matrices, which was obtained by averaging a set of 10 (9) participants for the resting state (resting state plus tasks). In this way, we manage to improve the fitting of the models. In each model, for each iteration, we average over 10 (9) simulations as we made for the empirical data. Overall, we use the term LGEC for the optimised C, which reflects the connectivity between the modes.

Obtaining Source Networks by Decoding the Latent Modes

We created a set of surrogate signals by introducing standard Gaussian noise into one latent dimension while keeping the other eight dimensions devoid of any signal. We repeated this process, varying the dimension with noise and the dimensions without any signal, and we called N “X,” with X from 1 to 9. Subsequently, we decoded each of these nine sets of surrogate latent signals, mapping them from the latent dimensions to the 62 brain regions of the original empirical data for each case. Next, we examined the regions in the source space that were activated in response to the presence of noise in each latent dimension. We quantified the variance of the decoded time signals to evaluate the statistical response of the activity associated with each latent dimension. This analysis allowed us to obtain one spatial pattern in the empirical original space associated to each latent dimension.

SVM Classifier

We use a SVM with Gaussian kernels as implemented in the MATLAB function fitcecoc. The function returns a full, trained, multiclass, error-correcting output codes (ECOC) model. This is achieved using the predictors in the input with class labels. The function uses K(K − 1)/2 binary SVM models using the one-versus-one coding design, where we used K = 8 as the number of unique class labels. In other words, the SVM had as inputs the elements of the LGEC obtained in each of the 100 models for each one of the seven tasks and the resting-state condition. The output was eight classes corresponding to the conditions (rest and seven tasks). We subdivided the output into 90% training and 10% validation, repeated and shuffled 100 times. For the classification using the FC in the original source space, we average the FC across nine participants, as we did with the LGEC computation, and use as an input the upper diagonal elements of the matrices.

Statistical Analyses

We applied the Wilcoxon rank-sum method to test the significance, and we applied the FDR at the 0.05 level of significance to correct for multiple comparisons (Hochberg & Benjamini, 1990).

Supporting information for this article is available at https://doi.org/10.1162/netn_a_00434.

Yonatan Sanz Perl: Conceptualization; Formal analysis; Methodology; Software; Validation; Visualization; Writing – original draft; Writing – review & editing. Sebastian Geli: Methodology; Software; Writing – review & editing. Eider Pérez-Ordoyo: Software; Writing – review & editing. Lou Zonca: Methodology; Writing – review & editing. Sebastian Idesis: Writing – review & editing. Jakub Vohryzek: Data curation; Methodology; Writing – review & editing. Viktor K. Jirsa: Writing – original draft; Writing – review & editing. Morten L. Kringelbach: Conceptualization; Data curation; Methodology; Writing – original draft; Writing – review & editing. Enzo Tagliazucchi: Conceptualization; Formal analysis; Methodology; Supervision; Validation; Writing – original draft; Writing – review & editing. Gustavo Deco: Conceptualization; Formal analysis; Investigation; Methodology; Supervision; Validation; Writing – original draft; Writing – review & editing.

Yonatan Sanz Perl, EU ERC Synergy Horizon Europe, Award ID: NEMESIS project (ref. 101071900). Gustavo Deco, EU ERC Synergy Horizon Europe, Award ID: NEMESIS project (ref. 101071900). Morten L. Kringelbach, Center for Music in the Brain, funded by the Danish National Research Foundation, Award ID: DNRF117. Morten L. Kringelbach, Centre for Eudaimonia and Human Flourishing at Linacre College funded by the Pettit and Carlsberg Foundations. Gustavo Deco, Department of Research and Universities of the Generalitat of Catalunya, Award ID: AGAUR Research Support Grant (ref. 2021 SGR 00917). Gustavo Deco, MCIN /AEI /10.13039/501100011033 / FEDER, UE, the Ministry of Science and Innovation, the State Research Agency and the European Regional Development Fund, Award ID: PID2022-136216NB-I00. Enzo Tagliazucchi, FONCYT-PICT, Award ID: 2019-02294. Enzo Tagliazucchi, CONICET-PIP, Award ID: 11220210100800CO. Enzo Tagliazucchi, ANID/FONDECYT, Award ID: 1220995.

The dataset utilised in this study originated from an independent publicly accessible collection of fMRI data. In this case, we opted for a sample comprising 1,003 participants, drawn from the HCP’s March 2017 public data release (https://www.humanconnectome.org/study/hcpyoung-adult).

The code used in this paper is deposited in https://github.com/yonisanzperl/ModellingManifoldModes. The software dependencies are MATLAB (2018b), Python (3.6), and Keras. From time to time, the code might be updated.

Manifold:

Refers to a low-dimensional space that reveals an underlying structure encapsulating the constraints and degrees of freedom of the system, thereby capturing the main dynamical properties of the entire system.

Criticality:

At the brink of a bifurcation, the system exhibits certain characteristic dynamic features, most of which are associated with enhanced fluctuations.

Autoencoder:

Refers to a deep learning network architecture composed of three structures: an encoder, a bottleneck (or latent space), and a decoder. The encoder reduces the dimensionality from N to M (MN), and the decoder restores the dimensionality from M to N.

Nonequilibrium dynamics:

The branch of thermodynamics that deals with systems not in thermodynamic equilibrium. It studies systems where macroscopic properties (e.g., temperature, pressure) vary spatially and temporally. A fundamental feature of nonequilibrium systems is their temporal irreversibility, that is, the emergence of an arrow of time.

Stuart-Landau oscillator:

A nonlinear dynamical system defined by the normal form of a Hopf bifurcation, which exhibits qualitatively different regimes, ranging from a fixed point to self-sustained oscillations, depending on the value of the bifurcation parameter.

Latent Generative Effective Connectivity (LGEC):

Refers to the connectivity between the latent dimensions. This connectivity is inferred using whole-brain modelling, which incorporates the nonequilibrium information of the empirical time series.

Latent dimension:

Refers to each dimension in the low-dimensional representation of whole-brain activity. Typically, the bottleneck layer of an autoencoder is known as the latent space, which is composed of latent dimensions.

Whole-brain model:

A tool for modelling brain dynamics using whole-brain neuroimaging techniques. Whole-brain models are constructed based on the anatomical connectivity of a reduced set of anatomically defined brain regions, incorporating a model of the local dynamics.

Generative mechanisms:

Refers to the fundamental rules that drive the temporal evolution of a system. These mechanisms can be derived by constructing a model of the system and investigating the causal influence of manipulating its elements.

Arslan
,
S.
,
Ktena
,
S. I.
,
Makropoulos
,
A.
,
Robinson
,
E. C.
,
Rueckert
,
D.
, &
Parisot
,
S.
(
2018
).
Human brain mapping: A systematic comparison of parcellation methods for the human cerebral cortex
.
NeuroImage
,
170
,
5
30
. ,
[PubMed]
Barch
,
D. M.
,
Burgess
,
G. C.
,
Harms
,
M. P.
,
Petersen
,
S. E.
,
Schlaggar
,
B. L.
,
Corbetta
,
M.
, …
WU-Minn HCP Consortium
.
(
2013
).
Function in the human connectome: Task-fMRI and individual differences in behavior
.
NeuroImage
,
80
,
169
189
. ,
[PubMed]
Beckmann
,
C. F.
,
DeLuca
,
M.
,
Devlin
,
J. T.
, &
Smith
,
S. M.
(
2005
).
Investigations into resting-state connectivity using independent component analysis
.
Philosophical Transactions of the Royal Society B: Biological Sciences
,
360
(
1457
),
1001
1013
. ,
[PubMed]
Biswal
,
B.
,
Yetkin
,
F. Z.
,
Haughton
,
V. M.
, &
Hyde
,
J. S.
(
1995
).
Functional connectivity in the motor cortex of resting human brain using echo-planar MRI
.
Magnetic Resonance in Medicine
,
34
(
4
),
537
541
. ,
[PubMed]
Brown
,
E. N.
,
Lydic
,
R.
, &
Schiff
,
N. D.
(
2010
).
General anesthesia, sleep, and coma
.
New England Journal of Medicine
,
363
(
27
),
2638
2650
. ,
[PubMed]
Cabral
,
J.
,
Kringelbach
,
M. L.
, &
Deco
,
G.
(
2014
).
Exploring the network dynamics underlying brain activity during rest
.
Progress in Neurobiology
,
114
,
102
131
. ,
[PubMed]
Carbonell
,
F.
,
Bellec
,
P.
, &
Shmuel
,
A.
(
2011
).
Global and system-specific resting-state fMRI fluctuations are uncorrelated: Principal component analysis reveals anti-correlated networks
.
Brain Connectivity
,
1
(
6
),
496
510
. ,
[PubMed]
Chandak
,
R.
, &
Raman
,
B.
(
2023
).
Neural manifolds for odor-driven innate and acquired appetitive preferences
.
Nature Communications
,
14
(
1
),
4719
. ,
[PubMed]
Chaudhuri
,
R.
,
Gerçek
,
B.
,
Pandey
,
B.
,
Peyrache
,
A.
, &
Fiete
,
I.
(
2019
).
The intrinsic attractor manifold and population dynamics of a canonical cognitive circuit across waking and sleep
.
Nature Neuroscience
,
22
(
9
),
1512
1520
. ,
[PubMed]
Chung
,
S.
, &
Abbott
,
L. F.
(
2021
).
Neural population geometry: An approach for understanding biological and artificial neural networks
.
Current Opinion in Neurobiology
,
70
,
137
144
. ,
[PubMed]
Cole
,
M. W.
,
Ito
,
T.
,
Bassett
,
D. S.
, &
Schultz
,
D. H.
(
2016
).
Activity flow over resting-state networks shapes cognitive task activations
.
Nature Neuroscience
,
19
(
12
),
1718
1726
. ,
[PubMed]
Damoiseaux
,
J. S.
,
Rombouts
,
S. A. R. B.
,
Barkhof
,
F.
,
Scheltens
,
P.
,
Stam
,
C. J.
,
Smith
,
S. M.
, &
Beckmann
,
C. F.
(
2006
).
Consistent resting-state networks across healthy subjects
.
Proceedings of the National Academy of Sciences of the United States of America
,
103
(
37
),
13848
13853
. ,
[PubMed]
Deco
,
G.
,
Jirsa
,
V. K.
, &
McIntosh
,
A. R.
(
2011
).
Emerging concepts for the dynamical organization of resting-state activity in the brain
.
Nature Reviews Neuroscience
,
12
(
1
),
43
56
. ,
[PubMed]
Deco
,
G.
,
Jirsa
,
V. K.
,
Robinson
,
P. A.
,
Breakspear
,
M.
, &
Friston
,
K.
(
2008
).
The dynamic brain: From spiking neurons to neural masses and cortical fields
.
PLOS Computational Biology
,
4
(
8
),
e1000092
. ,
[PubMed]
Deco
,
G.
,
Kringelbach
,
M. L.
,
Arnatkeviciute
,
A.
,
Oldham
,
S.
,
Sabaroedin
,
K.
,
Rogasch
,
N. C.
, …
Fornito
,
A.
(
2021
).
Dynamical consequences of regional heterogeneity in the brain’s transcriptional landscape
.
Science Advances
,
7
(
29
),
eabf4752
. ,
[PubMed]
Deco
,
G.
,
Kringelbach
,
M. L.
,
Jirsa
,
V. K.
, &
Ritter
,
P.
(
2017
).
The dynamics of resting fluctuations in the brain: Metastability and its dynamical cortical core
.
Scientific Reports
,
7
(
1
),
3095
. ,
[PubMed]
Deco
,
G.
,
Sanz Perl
,
Y.
,
Bocaccio
,
H.
,
Tagliazucchi
,
E.
, &
Kringelbach
,
M. L.
(
2022
).
The INSIDEOUT framework provides precise signatures of the balance of intrinsic and extrinsic dynamics in brain states
.
Communications Biology
,
5
(
1
),
572
. ,
[PubMed]
Derdikman
,
D.
, &
Moser
,
E. I.
(
2010
).
A manifold of spatial maps in the brain
.
Trends in Cognitive Sciences
,
14
(
12
),
561
569
. ,
[PubMed]
Desikan
,
R. S.
,
Ségonne
,
F.
,
Fischl
,
B.
,
Quinn
,
B. T.
,
Dickerson
,
B. C.
,
Blacker
,
D.
, …
Killiany
,
R. J.
(
2006
).
An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest
.
NeuroImage
,
31
(
3
),
968
980
. ,
[PubMed]
Domhof
,
J. W. M.
,
Jung
,
K.
,
Eickhoff
,
S. B.
, &
Popovych
,
O. V.
(
2021
).
Parcellation-induced variation of empirical and simulated brain connectomes at group and subject levels
.
Network Neuroscience
,
5
(
3
),
798
830
. ,
[PubMed]
Fousek
,
J.
,
Rabuffo
,
G.
,
Gudibanda
,
K.
,
Sheheitli
,
H.
,
Petkoski
,
S.
, &
Jirsa
,
V.
(
2024
).
Symmetry breaking organizes the brain’s resting state manifold. Scientific Reports
,
14
(
1
),
31970
. ,
[PubMed]
Gallego
,
J. A.
,
Perich
,
M. G.
,
Chowdhury
,
R. H.
,
Solla
,
S. A.
, &
Miller
,
L. E.
(
2020
).
Long-term stability of cortical population dynamics underlying consistent behavior
.
Nature Neuroscience
,
23
(
2
),
260
270
. ,
[PubMed]
Gallego
,
J. A.
,
Perich
,
M. G.
,
Miller
,
L. E.
, &
Solla
,
S. A.
(
2017
).
Neural manifolds for the control of movement
.
Neuron
,
94
(
5
),
978
984
. ,
[PubMed]
Gao
,
S.
,
Mishne
,
G.
, &
Scheinost
,
D.
(
2021
).
Nonlinear manifold learning in functional magnetic resonance imaging uncovers a low-dimensional space of brain dynamics
.
Human Brain Mapping
,
42
(
14
),
4510
4524
. ,
[PubMed]
Glomb
,
K.
,
Kringelbach
,
M. L.
,
Deco
,
G.
,
Hagmann
,
P.
,
Pearson
,
J.
, &
Atasoy
,
S.
(
2021
).
Functional harmonics reveal multi-dimensional basis functions underlying cortical organization
.
Cell Reports
,
36
(
8
),
109554
. ,
[PubMed]
Greicius
,
M.
(
2008
).
Resting-state functional connectivity in neuropsychiatric disorders
.
Current Opinion in Neurology
,
21
(
4
),
424
430
. ,
[PubMed]
Griffanti
,
L.
,
Douaud
,
G.
,
Bijsterbosch
,
J.
,
Evangelisti
,
S.
,
Alfaro-Almagro
,
F.
,
Glasser
,
M. F.
, …
Smith
,
S. M.
(
2017
).
Hand classification of fMRI ICA noise components
.
NeuroImage
,
154
,
188
205
. ,
[PubMed]
Hennig
,
J. A.
,
Golub
,
M. D.
,
Lund
,
P. J.
,
Sadtler
,
P. T.
,
Oby
,
E. R.
,
Quick
,
K. M.
, …
Chase
,
S. M.
(
2018
).
Constraints on neural redundancy
.
eLife
,
7
,
e36774
. ,
[PubMed]
Hochberg
,
Y.
, &
Benjamini
,
Y.
(
1990
).
More powerful procedures for multiple significance testing
.
Statistics in Medicine
,
9
(
7
),
811
818
. ,
[PubMed]
Huang
,
Z.
,
Zhang
,
J.
,
Wu
,
J.
,
Mashour
,
G. A.
, &
Hudetz
,
A. G.
(
2020
).
Temporal circuit of macroscale dynamic brain activity supports human consciousness
.
Science Advances
,
6
(
11
),
eaaz0087
. ,
[PubMed]
Ipiña
,
I. P.
,
Kehoe
,
P. D.
,
Kringelbach
,
M.
,
Laufs
,
H.
,
Ibañez
,
A.
,
Deco
,
G.
, …
Tagliazucchi
,
E.
(
2020
).
Modeling regional changes in dynamic stability during sleep and wakefulness
.
NeuroImage
,
215
,
116833
. ,
[PubMed]
Jamison
,
K. W.
,
Gu
,
Z.
,
Wang
,
Q.
,
Tozlu
,
C.
,
Sabuncu
,
M. R.
, &
Kuceyeski
,
A.
(
2024
).
Release the Krakencoder: A unified brain connectome translation and fusion tool
.
bioRxiv
. ,
[PubMed]
Jirsa
,
V.
(
2020
).
Structured Flows on Manifolds as guiding concepts in brain science
.
Selbstorganisation – Ein Paradigma Für Die Humanwissenschaften
,
89
102
.
Jirsa
,
V.
, &
Sheheitli
,
H.
(
2022
).
Entropy, free energy, symmetry and dynamics in the brain
.
Journal of Physics: Complexity
,
3
(
1
),
015007
.
Jobst
,
B. M.
,
Hindriks
,
R.
,
Laufs
,
H.
,
Tagliazucchi
,
E.
,
Hahn
,
G.
,
Ponce-Alvarez
,
A.
, …
Deco
,
G.
(
2017
).
Increased stability and breakdown of brain effective connectivity during slow-wave sleep: Mechanistic insights from whole-brain computational modelling
.
Scientific Reports
,
7
(
1
),
4634
. ,
[PubMed]
Kingma
,
D. P.
, &
Welling
,
M.
(
2013
).
Auto-encoding variational bayes
.
2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings
. https://arxiv.org/abs/1312.6114v10.
Kobeleva
,
X.
,
López-González
,
A.
,
Kringelbach
,
M. L.
, &
Deco
,
G.
(
2021
).
Revealing the relevant spatiotemporal scale underlying whole-brain dynamics
.
Frontiers in Neuroscience
,
15
,
715861
. ,
[PubMed]
Kringelbach
,
M. L.
,
Perl
,
Y. S.
,
Tagliazucchi
,
E.
, &
Deco
,
G.
(
2023
).
Toward naturalistic neuroscience: Mechanisms underlying the flattening of brain hierarchy in movie-watching compared to rest and task
.
Science Advances
,
9
(
2
),
eade6049
. ,
[PubMed]
Langdon
,
C.
,
Genkin
,
M.
, &
Engel
,
T. A.
(
2023
).
A unifying perspective on neural manifolds and circuits for cognition
.
Nature Reviews Neuroscience
,
24
(
6
),
363
377
. ,
[PubMed]
Liu
,
X.
, &
Duyn
,
J. H.
(
2013
).
Time-varying functional network information extracted from brief instances of spontaneous brain activity
.
Proceedings of the National Academy of Sciences
,
110
(
11
),
4392
4397
. ,
[PubMed]
Luppi
,
A. I.
,
Gellersen
,
H. M.
,
Liu
,
Z.-Q.
,
Peattie
,
A. R. D.
,
Manktelow
,
A. E.
,
Adapa
,
R.
, …
Stamatakis
,
E. A.
(
2024
).
Systematic evaluation of fMRI data-processing pipelines for consistent functional connectomics
.
Nature Communications
,
15
(
1
),
4745
. ,
[PubMed]
Luppi
,
A. I.
,
Vohryzek
,
J.
,
Kringelbach
,
M. L.
,
Mediano
,
P. A. M.
,
Craig
,
M. M.
,
Adapa
,
R.
, …
Stamatakis
,
E. A.
(
2023
).
Distributed harmonic patterns of structure-function dependence orchestrate human consciousness
.
Communications Biology
,
6
(
1
),
117
. ,
[PubMed]
Mantini
,
D.
,
Perrucci
,
M. G.
,
Del Gratta
,
C.
,
Romani
,
G. L.
, &
Corbetta
,
M.
(
2007
).
Electrophysiological signatures of resting state networks in the human brain
.
Proceedings of the National Academy of Sciences of the United States of America
,
104
(
32
),
13170
13175
. ,
[PubMed]
Margulies
,
D. S.
,
Ghosh
,
S. S.
,
Goulas
,
A.
,
Falkiewicz
,
M.
,
Huntenburg
,
J. M.
,
Langs
,
G.
, …
Smallwood
,
J.
(
2016
).
Situating the default-mode network along a principal gradient of macroscale cortical organization
.
Proceedings of the National Academy of Sciences of the United States of America
,
113
(
44
),
12574
12579
. ,
[PubMed]
Messé
,
A.
(
2020
).
Parcellation influence on the connectivity-based structure–function relationship in the human brain
.
Human Brain Mapping
,
41
(
5
),
1167
1180
. ,
[PubMed]
Mitchell-Heggs
,
R.
,
Prado
,
S.
,
Gava
,
G. P.
,
Go
,
M. A.
, &
Schultz
,
S. R.
(
2023
).
Neural manifold analysis of brain circuit dynamics in health and disease
.
Journal of Computational Neuroscience
,
51
(
1
),
1
21
. ,
[PubMed]
Nieh
,
E. H.
,
Schottdorf
,
M.
,
Freeman
,
N. W.
,
Low
,
R. J.
,
Lewallen
,
S.
,
Koay
,
S. A.
, …
Tank
,
D. W.
(
2021
).
Geometry of abstract learned knowledge in the hippocampus
.
Nature
,
595
(
7865
),
80
84
. ,
[PubMed]
Pang
,
R.
,
Lansdell
,
B. J.
, &
Fairhall
,
A. L.
(
2016
).
Dimensionality reduction in neuroscience
.
Current Biology
,
26
(
14
),
R656
R660
. ,
[PubMed]
Perl
,
Y. S.
,
Escrichs
,
A.
,
Tagliazucchi
,
E.
,
Kringelbach
,
M. L.
, &
Deco
,
G.
(
2022
).
Strength-dependent perturbation of whole-brain model working in different regimes reveals the role of fluctuations in brain dynamics
.
PLOS Computational Biology
,
18
(
11
),
e1010662
. ,
[PubMed]
Perl
,
Y. S.
,
Pallavicini
,
C.
,
Ipiña
,
I. P.
,
Demertzi
,
A.
,
Bonhomme
,
V.
,
Martial
,
C.
, …
Tagliazucchi
,
E.
(
2021
).
Perturbations in dynamical models of whole-brain activity dissociate between the level and stability of consciousness
.
PLOS Computational Biology
,
17
(
7
),
e1009139
. ,
[PubMed]
Perl
,
Y. S.
,
Pallavicini
,
C.
,
Piccinini
,
J.
,
Demertzi
,
A.
,
Bonhomme
,
V.
,
Martial
,
C.
, …
Tagliazucchi
,
E.
(
2023
).
Low-dimensional organization of global brain states of reduced consciousness
.
Cell Reports
,
42
(
5
),
112491
. ,
[PubMed]
Pillai
,
A. S.
, &
Jirsa
,
V. K.
(
2017
).
Symmetry breaking in space-time hierarchies shapes brain dynamics and behavior
.
Neuron
,
94
(
5
),
1010
1026
. ,
[PubMed]
Rué-Queralt
,
J.
,
Stevner
,
A.
,
Tagliazucchi
,
E.
,
Laufs
,
H.
,
Kringelbach
,
M. L.
,
Deco
,
G.
, &
Atasoy
,
S.
(
2021
).
Decoding brain states on the intrinsic manifold of human brain dynamics across wakefulness and sleep
.
Communications Biology
,
4
(
1
),
854
.
Schaefer
,
A.
,
Kong
,
R.
,
Gordon
,
E. M.
,
Laumann
,
T. O.
,
Zuo
,
X.-N.
,
Holmes
,
A. J.
, …
Yeo
,
B. T. T.
(
2018
).
Local-global parcellation of the human cerebral cortex from intrinsic functional connectivity MRI
.
Cerebral Cortex
,
28
(
9
),
3095
3114
. ,
[PubMed]
Schiff
,
N. D.
,
Nauvel
,
T.
, &
Victor
,
J. D.
(
2014
).
Large-scale brain dynamics in disorders of consciousness
.
Current Opinion in Neurobiology
,
25
,
7
14
. ,
[PubMed]
Schröder
,
T. N.
,
Haak
,
K. V.
,
Jimenez
,
N. I. Z.
,
Beckmann
,
C. F.
, &
Doeller
,
C. F.
(
2015
).
Functional topography of the human entorhinal cortex
.
eLife
,
4
,
e06738
. ,
[PubMed]
Shine
,
J. M.
,
Bissett
,
P. G.
,
Bell
,
P. T.
,
Koyejo
,
O.
,
Balsters
,
J. H.
,
Gorgolewski
,
K. J.
, …
Poldrack
,
R. A.
(
2016
).
The dynamics of functional brain networks: Integrated network states during cognitive task performance
.
Neuron
,
92
(
2
),
544
554
. ,
[PubMed]
Shine
,
J. M.
,
Breakspear
,
M.
,
Bell
,
P. T.
,
Ehgoetz Martens
,
K. A.
,
Shine
,
R.
,
Koyejo
,
O.
, …
Poldrack
,
R. A.
(
2019
).
Human cognition involves the dynamic integration of neural activity and neuromodulatory systems
.
Nature Neuroscience
,
22
(
2
),
289
296
. ,
[PubMed]
Smith
,
S. M.
,
Beckmann
,
C. F.
,
Andersson
,
J.
,
Auerbach
,
E. J.
,
Bijsterbosch
,
J.
,
Douaud
,
G.
, …
WU-Minn HCP Consortium
.
(
2013
).
Resting-state fMRI in the Human Connectome Project
.
NeuroImage
,
80
,
144
168
. ,
[PubMed]
Smith
,
S. M.
,
Fox
,
P. T.
,
Miller
,
K. L.
,
Glahn
,
D. C.
,
Fox
,
P. M.
,
Mackay
,
C. E.
, …
Beckmann
,
C. F.
(
2009
).
Correspondence of the brain’s functional architecture during activation and rest
.
Proceedings of the National Academy of Sciences
,
106
(
31
),
13040
13045
. ,
[PubMed]
Song
,
H.
,
Shim
,
W. M.
, &
Rosenberg
,
M. D.
(
2023
).
Large-scale neural dynamics in a shared low-dimensional state space reflect cognitive and attentional dynamics
.
eLife
,
12
,
e85487
. ,
[PubMed]
Sporns
,
O.
,
Tononi
,
G.
, &
Kötter
,
R.
(
2005
).
The human connectome: A structural description of the human brain
.
PLOS Computational Biology
,
1
(
4
),
e42
. ,
[PubMed]
Stringer
,
C.
,
Pachitariu
,
M.
,
Steinmetz
,
N.
,
Carandini
,
M.
, &
Harris
,
K. D.
(
2019
).
High-dimensional geometry of population responses in visual cortex
.
Nature
,
571
(
7765
),
361
365
. ,
[PubMed]
Tononi
,
G.
,
Boly
,
M.
,
Massimini
,
M.
, &
Koch
,
C.
(
2016
).
Integrated information theory: From consciousness to its physical substrate
.
Nature Reviews Neuroscience
,
17
(
7
),
450
461
. ,
[PubMed]
Tononi
,
G.
,
Edelman
,
G. M.
, &
Sporns
,
O.
(
1998
).
Complexity and coherency: Integrating information in the brain
.
Trends in Cognitive Sciences
,
2
(
12
),
474
484
. ,
[PubMed]
Tononi
,
G.
,
Sporns
,
O.
, &
Edelman
,
G. M.
(
1999
).
Measures of degeneracy and redundancy in biological networks
.
Proceedings of the National Academy of Sciences
,
96
(
6
),
3257
3262
. ,
[PubMed]
Varley
,
T. F.
,
Denny
,
V.
,
Sporns
,
O.
, &
Patania
,
A.
(
2021
).
Topological analysis of differential effects of ketamine and propofol anaesthesia on brain dynamics
.
Royal Society Open Science
,
8
(
6
),
201971
. ,
[PubMed]
Vidaurre
,
D.
,
Abeysuriya
,
R.
,
Becker
,
R.
,
Quinn
,
A. J.
,
Alfaro-Almagro
,
F.
,
Smith
,
S. M.
, &
Woolrich
,
M. W.
(
2018
).
Discovering dynamic brain networks from big data in rest and task
.
NeuroImage
,
180
,
646
656
. ,
[PubMed]
Vohryzek
,
J.
,
Cabral
,
J.
,
Timmermann
,
C.
,
Atasoy
,
S.
,
Roseman
,
L.
,
Nutt
,
D. J.
, …
Kringelbach
,
M. L.
(
2023
).
Harmonic decomposition of spacetime (HADES) framework characterises the spacetime hierarchy of the DMT brain state
.
bioRxiv
.
Yeo
,
B. T. T.
.,
Krienen
,
F. M.
,
Sepulcre
,
J.
,
Sabuncu
,
M. R.
,
Lashkari
,
D.
,
Hollinshead
,
M.
, …
Buckner
,
R. L.
(
2011
).
The organization of the human cerebral cortex estimated by intrinsic functional connectivity
.
Journal of Neurophysiology
,
106
(
3
),
1125
1165
. ,
[PubMed]

Author notes

Competing Interests: The authors have declared that no competing interests exist.

Handling Editor: Lucina Uddin

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/legalcode.

Supplementary data