Abstract
Generative models of brain activity have been instrumental in testing hypothesized mechanisms underlying brain dynamics against experimental datasets. Beyond capturing the key mechanisms underlying spontaneous brain dynamics, these models hold an exciting potential for understanding the mechanisms underlying the dynamics evoked by targeted brain stimulation techniques. This paper delves into this emerging application, using concepts from dynamical systems theory to argue that the stimulus-evoked dynamics in such experiments may be shaped by new types of mechanisms distinct from those that dominate spontaneous dynamics. We review and discuss (a) the targeted experimental techniques across spatial scales that can both perturb the brain to novel states and resolve its relaxation trajectory back to spontaneous dynamics and (b) how we can understand these dynamics in terms of mechanisms using physiological, phenomenological, and data-driven models. A tight integration of targeted stimulation experiments with generative quantitative modeling provides an important opportunity to uncover novel mechanisms of brain dynamics that are difficult to detect in spontaneous settings.
Author Summary
Generative models are important tools for testing hypothesized mechanisms of brain dynamics against experimental data. This review highlights an application of generative models in analyzing a form of brain activity evoked by emerging targeted stimulation techniques. We argue that analyzing targeted stimulation dynamics can uncover mechanisms that are hidden during commonly analyzed spontaneous dynamics and explore how integrating diverse targeted stimulation experiments with existing generative models offer a significant opportunity to uncover these novel mechanisms and thereby expand our mechanistic understanding of brain dynamics.
INTRODUCTION
Statistical analyses of experimental neuroimaging data have long demonstrated that the brain supports a rich repertoire of spatiotemporal dynamics (Biswal, Yetkin, Haughton, & Hyde, 1995; Honey, Kötter, Breakspear, & Sporns, 2007; Lopes da Silva, van Rotterdam, Barts, van Heusden, & Burr, 1976; Vidaurre, Smith, & Woolrich, 2017). To better understand the intrinsic mechanisms that give rise to the brain’s dynamical properties, generative models of brain activity—also known as computational models, dynamic models, or mathematical models—can be used to test hypothesized mechanisms against properties of brain dynamics that they aim to explain or predict. While the term “generative model” can have a range of interpretations across fields, here, we use it as shorthand for “generative model of brain dynamics” to describe a model that can generate (through simulation) a set of time series of brain activity that can be compared with data measured experimentally (Ramezanian-Panahi, Abrevaya, Gagnon-Audet, Voleti, Rish, & Dumas, 2022). Existing generative models allow us to simulate brain activity based on a wide variety of encoded mechanisms across spatial scales (Acharya, Ruf, & Nozari, 2022; Breakspear, 2017; Cabral, Kringelbach, & Deco, 2017; Deco, Jirsa, Robinson, Breakspear, & Friston, 2008; Kim & Bassett, 2020; Ramezanian-Panahi et al., 2022), allowing the hypothesized mechanisms to be directly evaluated against the dynamics of activity accessed in experimental neural recordings.
A crucial advantage that generative models provide is the ability to rigorously test candidate mechanisms against datasets from different experimental settings. Early generative models were used to test mechanisms underlying dynamics of brain activity recorded during cognitive or motor tasks, or in response to sensory stimuli (Brunel & Wang, 2001; David, Kilner, & Friston, 2006; Jansen & Rit, 1995; Jirsa, Friedrich, Haken, & Kelso, 1994; Jirsa & Haken, 1996, 1997; Kerr, Rennie, & Robinson, 2008; Moran et al., 2007; Rennie, Robinson, & Wright, 2002; Zipser, 1991). Rich dynamical properties were later found even in spontaneous brain activity, that is, activity not attributable to any externally applied task or stimulus (Fox & Raichle, 2007). Many of these properties have hypothesized mechanistic explanations that have been tested through generative models (Abdelnour, Dayan, Devinsky, Thesen, & Raj, 2018; Breakspear et al., 2006; Deco, Kringelbach, Jirsa, & Ritter, 2017; Freyer et al., 2011; Honey et al., 2009; Nozari et al., 2024; Robinson, Loxley, O’Connor, & Rennie, 2001; Robinson, Rennie, & Rowe, 2002). As datasets from different experimental settings reveal novel spatiotemporal properties of brain dynamics, generative models present new opportunities to refine and hypothesize new underlying mechanistic accounts that can accurately capture them.
A new opportunity of growing interest to generative modeling is to test our understanding of the dynamics of brain activity evoked by targeted stimulation. Unlike traditionally used sensory stimulation paradigms, targeted stimulation techniques directly target neurons or neural populations using concentrated external inputs of energy. This approach bypasses the constraints imposed by the brain’s natural sensory pathways, offering significantly greater spatial precision in target selection and flexibility in stimulation strength. Commonly used targeted stimulation techniques include optogenetics (Deisseroth, 2015), electrode stimulation (Flesher et al., 2016; Kringelbach, Jenkinson, Owen, & Aziz, 2007), transcranial magnetic stimulation (TMS) (Rogasch & Fitzgerald, 2013), and chemogenetics (Roth, 2016), all of which are applicable in vivo and allow for simultaneous measurement of evoked responses with compatible measurement techniques (Bonnard et al., 2016; Luboeinski & Tchumatchenko, 2020; Markicevic et al., 2020; O’Shea et al., 2022; Sanzeni et al., 2023). This review focuses on the new opportunities that datasets of activity measured from targeted stimulation experiments present in validating and refining our generative models of brain activity and thereby better understanding the important mechanisms that underpin brain dynamics.
Why should one be interested in the mechanisms underlying the brain’s response to targeted stimulation? There are both theoretical and practical motivations. Theoretically, despite the high dimensionality of brain activity recordings, brain dynamics of spontaneous activity is known to be constrained to a lower-dimensional manifold (Chaudhuri, Gerçek, Pandey, Peyrache, & Fiete, 2019; Churchland et al., 2012; Cunningham & Yu, 2014; Humphries, 2021; Shine, Breakspear, et al., 2019; Shine, Hearne, et al., 2019). This implies the existence of a vast space of unexplored states that the brain does not typically access, unless perturbed with a sufficiently strong stimulus. Targeted stimulation techniques enable us to access states that are not only inaccessible during spontaneous activity but also difficult to access during sensory-evoked activity, hence initiating new forms of artificially evoked activity that transcend the brain’s natural activation patterns. Consequently, generative models present an important opportunity to test and calibrate mechanisms on datasets obtained through a unique experimental paradigm, potentially driving the discovery of new mechanisms of brain dynamics distinct from those that sufficiently capture spontaneous dynamics. Practically, there is growing interest in developing clinical neuromodulation technologies using targeted stimulation to address a wide range of neurodegenerative and neurological disorders (Kurtin et al., 2023). While trial-and-error methods for determining effective stimulation parameters have shown promising results for specific patients (Cole et al., 2020; Ramasubbu, Lang, & Kiss, 2018), a wider deployment of neuromodulation necessitates a systematic approach that can more efficiently explore the vast stimulation parameter space and tailor the technology to each patient’s needs (Capogrosso & Lempka, 2020; Wang, Hutchings, & Kaiser, 2015). An enhanced mechanistic understanding of the brain’s response to the targeted stimulation techniques behind these technologies can therefore accelerate the development of more systematic personalized protocols by guiding the interpretation of individual responses to different stimulation parameter combinations.
In this review, we focus on the theoretically motivated opportunities in understanding the brain’s response to targeted stimulation (for a review on the practically motivated opportunities mentioned above, see Acharya et al., 2022; Kurtin et al., 2023; Tang & Bassett, 2018; and Wang et al., 2015 for examples). With reference to a theoretical abstracted picture of brain dynamics, we argue for the existence of hidden nonlinear mechanisms of brain dynamics, which are challenging to detect from spontaneous dynamics. We highlight three available tools that we argue are essential to uncover these mechanisms: (a) targeted stimulation techniques to make precise perturbations that can transition brain activity away from spontaneous patterns, (b) measurement techniques to simultaneously record the brain’s response to the perturbation as it reverts to spontaneous patterns, and (c) generative models to test mechanisms that we hypothesize to capture the dynamics of recorded activity evoked by targeted stimulation. Anchored around this argument, we explore in detail the opportunities that integrating these tools—that is, integrating different targeted stimulation and measurement techniques with different physiological, phenomenological, and data-driven models—provided in expanding our understanding of the various mechanisms underlying brain dynamics.
A DYNAMICAL SYSTEMS FRAMING OF TARGETED BRAIN STIMULATION
To concretely establish our theoretical motivation for understanding the brain’s response to targeted stimulation, we present a geometric depiction of brain dynamics that abstracts the complexities of brain activity measurements across different spatial scales and species, which is shown in Figure 1. With reference to dynamical systems theory, we use this picture to argue that complex brain dynamics are likely underpinned by mechanisms that are not clearly apparent in the dynamics of commonly analyzed spontaneous activity and that targeted stimulation techniques and measurement techniques are necessary tools to generate brain dynamics that manifest these hidden mechanisms. We also argue that the dynamics evoked by targeted stimulation are likely to be nonlinear and that to rigorously test and validate the hidden mechanisms underlying these dynamics necessitates the use of generative modeling and simulation.
An abstracted picture of brain dynamics that conceptualizes the dynamics measured from distinct experimental settings, while simplifying the complexities of brain activity measurements across other experimental conditions. (A) Brain activity dynamics can be represented as the evolution of a set of key state variables, such as (i) the membrane potentials of individual neurons in a network or (ii) activation of distributed modes of whole-brain activity. (B) In this abstracted picture, spontaneous brain dynamics can be viewed as stochastic fluctuations about an attractor (a manifold that attracts nearby trajectories). Common attractor classes include fixed points (Sip et al., 2023; van den Heuvel & Hulshoff Pol, 2010), labeled (i), and limit cycles (Shine, Breakspear, et al., 2019; Shine, Hearne, et al., 2019), labeled (ii). (C) In contrast to spontaneous dynamics, stimulus-evoked brain dynamics consists of a perturbation, labeled (i), to a point that is typically far from the attractor, followed by a dynamic relaxation, labeled (ii), back to the attractor. Stimulus-evoked dynamics can exhibit interesting nonlinearities, where the response (red) qualitatively differs with the strength of the perturbation (blue), as depicted at (iii).
An abstracted picture of brain dynamics that conceptualizes the dynamics measured from distinct experimental settings, while simplifying the complexities of brain activity measurements across other experimental conditions. (A) Brain activity dynamics can be represented as the evolution of a set of key state variables, such as (i) the membrane potentials of individual neurons in a network or (ii) activation of distributed modes of whole-brain activity. (B) In this abstracted picture, spontaneous brain dynamics can be viewed as stochastic fluctuations about an attractor (a manifold that attracts nearby trajectories). Common attractor classes include fixed points (Sip et al., 2023; van den Heuvel & Hulshoff Pol, 2010), labeled (i), and limit cycles (Shine, Breakspear, et al., 2019; Shine, Hearne, et al., 2019), labeled (ii). (C) In contrast to spontaneous dynamics, stimulus-evoked brain dynamics consists of a perturbation, labeled (i), to a point that is typically far from the attractor, followed by a dynamic relaxation, labeled (ii), back to the attractor. Stimulus-evoked dynamics can exhibit interesting nonlinearities, where the response (red) qualitatively differs with the strength of the perturbation (blue), as depicted at (iii).
In this abstracted picture, the brain’s activity is characterized as the evolution of a brain state over time. Brain states can be encapsulated by a set of time-varying state variables, {x1(t), x2(t), x3(t), …, xN(t)}, and brain dynamics thus correspond to trajectories, x(t) = [x1(t), x2(t), …, xN(t)], through the state space. A specific spatial scale of interest can be addressed through appropriate selection of the state variables, from the activities of individual neurons in a population (Buzsáki, 2004) (depicted in Figure 1A(i)) to the spatially distributed functional gradients or modes that describe patterns of population-scale activity over the whole brain (Huntenburg, Bazin, & Margulies, 2018; Pang et al., 2023) (depicted in Figure 1A(ii)).
By abstracting brain dynamics as trajectories through a state space, we can conceptualize the dynamics of brain activity measured from qualitatively distinct experimental settings. Consider the commonly examined setting of spontaneous activity, which is measured when the brain is at rest and absent from any explicit external stimulus. Measurements indicate that spontaneous brain states are concentrated at a subset of the state space with dimension much less than that of the state space and do not deviate significantly from this subset (Chaudhuri et al., 2019; Churchland et al., 2012; Cunningham & Yu, 2014; 2019; Shine, Hearne, et al., 2019). Consistent with these properties, spontaneous dynamics in our abstracted picture can be conceptualized as trajectories concentrated near attractors, which are low-dimensional subsets of the state space that attract neighboring trajectories (Strogatz, 2018). Additional sensory inputs and background biological processes can also cause trajectories to temporarily deviate from and return to the attractor, effectively yielding stochastic fluctuations about the attractor. Attractors, which define spontaneous dynamics across different experimental conditions, can be further categorized into classes of similar geometric features, such as stable fixed points (Sip et al., 2023; van den Heuvel & Hulshoff Pol, 2010) (depicted in Figure 1B(i)), or flows along a more complex low-dimensional manifold such as a limit cycle (Churchland et al., 2012; Shine, Breakspear, et al., 2019; Shine, Hearne, et al., 2019) (depicted in Figure 1B(ii)). Through classes of attractors, we can concisely capture the intrinsic dynamical properties of spontaneous brain activity in this abstracted picture while aggregating the impact of other experimental conditions.
In addition to the low-dimensional attractor where trajectories of spontaneous dynamics are concentrated, there are a myriad of possible trajectories of brain dynamics through states far from the attractor. Unfortunately, as brain states move further from the attractor, they become less likely to be observed in spontaneous settings. However, our abstracted picture can conceptualize an alternative stimulus-evoked setting that employs external intervention to reach these distant states, as shown in Figure 1C. Initially positioned near the attractor (depicted here as a fixed point), the brain is first perturbed with a sufficiently strong stimulus to a distant state (Figure 1C(i)) and subsequently relaxes back to the attractor following stimulus termination (Figure 1C(ii)). According to nonlinear dynamical systems theory, a central characteristic of stimulus-evoked dynamics is the nonlinear nature of the relaxation trajectory (Strogatz, 2018), where its shape is expected to be highly sensitive to the magnitude of the perturbation, as illustrated in Figure 1C(iii). However, while such nonlinearities can produce a diverse range of trajectories far from the attractor, their effects tend to vanish as the trajectory approaches the attractor governing spontaneous dynamics, implying that spontaneous dynamics can be sufficiently approximated by simpler linear models. Generative modeling studies have indeed shown that linear models can capture properties of macroscale spontaneous brain dynamics as effectively as nonlinear models, such as the established resting-state functional connectivity benchmark (Hosaka, Hieda, Hayashi, Jimura, & Matsui, 2024; Messé, Rudrauf, Giron, & Marrelec, 2015; Nozari et al., 2024). These findings, supported by predictions of nonlinear dynamical systems theory, suggest that the nonlinear mechanisms shaping stimulus-evoked brain dynamics in this dynamical systems picture are difficult to be detected in spontaneous settings. Thus, to uncover and understand these nuanced hidden mechanisms of stimulus-evoked dynamics in practice (beyond this abstracted picture), we require an experimental paradigm that allows us to perturb the brain away from its spontaneous attractor (Figure 1C(i)) and resolve the nonlinear relaxation trajectory (Figure 1C(ii)), alongside a model that allows us to question and validate the nonlinear mechanisms that underpin these trajectories.
To this end, we bring attention to a practically feasible strategy that integrates targeted stimulation techniques and measurement techniques with generative models. Targeted stimulation techniques can perturb the brain from the attractor to distant states, with spatial coverage and precision that surpass traditional sensory stimuli (Deisseroth, 2015; Keller et al., 2014; Rogasch & Fitzgerald, 2013; Roth, 2016). High precision selection of states is crucial, as the nonlinear dynamics of stimulus-evoked activity would be highly sensitive to perturbation strengths. Compatible measurement techniques play an indispensable role by simultaneously tracking the brain’s relaxation trajectory back to the attractor at sufficient spatial and temporal resolution. Finally, the simulation approach of generative models is essential for linking hypothesized mechanisms to the dynamics observed in measurements, as a mechanistic understanding of nonlinear properties is challenging to pursue from static statistical measures alone (John et al., 2022).
Targeted stimulation techniques, measurement techniques, and generative models are therefore three important tools which, when integrated, can uncover and provide understanding of new mechanisms of brain dynamics that are concealed in spontaneous settings. By conducting experiments with stimulation and measurement techniques, we can generate datasets of stimulus-evoked dynamics with potentially new forms of nonlinearities. By constraining generative models with datasets from these experiments, we can rigorously test and validate hypotheses about the novel mechanisms which drive their nonlinear dynamics. To explore the opportunities that emerge from this integration, we first delve into the different experiments that can be executed.
TARGETED STIMULATION EXPERIMENTS
Through an abstracted picture of brain dynamics, illustrated in Figure 1, we have argued why targeted stimulation experiments play an important role in revealing novel mechanisms of brain dynamics in measured brain activity. This picture serves another purpose: It describes the fundamental dynamical properties of activity in any targeted stimulation experiment while abstracting away details of the specific techniques used. This interpretation bridges a wide body of stimulation and measurement techniques developed from disjoint fields, each with their own mechanics of action, spatial scale, and species compatibility. Here, we review a diverse range of key stimulation and measurement techniques that can be used in a targeted stimulation experiment, contextualizing their features within our unified picture.
In a targeted stimulation experiment, a stimulation technique perturbs the brain to a state distant from the attractor representing spontaneous dynamics (Figure 1(i)), and a measurement technique captures the brain’s subsequent relaxation trajectory back to the attractor (Figure 1(ii)). Key targeted stimulation and measurement techniques are illustrated in Figure 2. Based on the spatial scale in which brain states are defined, it is crucial to select an appropriate targeted stimulation technique to elicit a precise perturbation in this state space and an appropriate measurement technique to accurately capture the relaxation.
Stimulus-evoked brain activity can be generated through a diverse range of targeted stimulation techniques and measured by a range of measurement techniques at different spatial scales. (A) Targeted stimulation techniques can perturb the brain to novel brain states at different degrees of spatial precision. From neuron-scale techniques, such as optogenetics; to mesoscale techniques, such as ICMS and DBS; and to macroscale techniques, such as chemogenetics and TMS. (B) Different measurement techniques can measure the brain’s response to targeted stimulation at different spatial resolutions. From neuron-resolution modalities, such as Neuropixels electrodes, patch clamps, and calcium imaging; to mesoscale modalities, such as iEEG systems; and to macroscale modalities, such as EEG and fMRI.
Stimulus-evoked brain activity can be generated through a diverse range of targeted stimulation techniques and measured by a range of measurement techniques at different spatial scales. (A) Targeted stimulation techniques can perturb the brain to novel brain states at different degrees of spatial precision. From neuron-scale techniques, such as optogenetics; to mesoscale techniques, such as ICMS and DBS; and to macroscale techniques, such as chemogenetics and TMS. (B) Different measurement techniques can measure the brain’s response to targeted stimulation at different spatial resolutions. From neuron-resolution modalities, such as Neuropixels electrodes, patch clamps, and calcium imaging; to mesoscale modalities, such as iEEG systems; and to macroscale modalities, such as EEG and fMRI.
Targeted Stimulation Techniques
Different targeted stimulation techniques can be selected to elicit perturbations at a spatial precision of choice. Microscale techniques can precisely stimulate individual neurons (Deisseroth, 2015), while meso-to-macroscale techniques can stimulate neural populations ranging from millimeters to centimeters in size (Flesher et al., 2016; Kringelbach et al., 2007; Rogasch & Fitzgerald, 2013). Figure 2A illustrates key targeted stimulation techniques, ordered by degrees of spatial precision. We review these techniques below.
At the microscale, techniques can stimulate a single target neuron or a set of neurons. The most prevalent microscale stimulation technique is optogenetics, in which target neurons are genetically modified with injected photoreceptor proteins called opsins and later exposed to light of a specified wavelength to generate action potentials (Deisseroth, 2015). Two-photon microscopy allows the light beam to be made sufficiently narrow to be directed exclusively to the target neurons (Rickgauer & Tank, 2009; Shemesh et al., 2017; Tong et al., 2023).
At the mesoscale, neural populations can be similarly stimulated using optogenetics, or via electrical methods such as intracortical microstimulation (ICMS) and deep brain stimulation (DBS). A compelling benefit of using optogenetics at this scale is its ability to target specific cell types, cortical layers, and projection targets through viral-vector gene-delivery techniques (Yizhar, Fenno, Davidson, Mogri, & Deisseroth, 2011). The alternative methods, ICMS and DBS, are electrical techniques which stimulate local populations in the vicinity of an implanted electrode using electric fields (Flesher et al., 2016; Kringelbach et al., 2007). While these methods lack the forms of specificity that can be achieved with optogenetics, they can be applied in human brains as it does not require genetic modification of the target neural tissue. Local populations near the cortical surface can be targeted by ICMS through a chronically implanted multielectrode array (Flesher et al., 2016), whereas populations of varying depth down to subcortical structures can also be targeted by DBS through electrodes implanted deep in the brain (Kringelbach et al., 2007).
Finally, macroscale neural populations can be stimulated using electrical methods such as TMS, or chemogenetics, which is a genetic modification technique similar to optogenetics. TMS uses an external electromagnetic coil to generate a magnetic field, which then induces an electric field over the target population (Rogasch & Fitzgerald, 2013; Tremblay et al., 2019). TMS is a noninvasive technique, as the generated magnetic field passes across the scalp and skull; however, the difficulty in downsizing the coil has mainly restricted its use to primate brains (Alekseichuk et al., 2019). In smaller rodent brains, macroscale populations can be stimulated using chemogenetics, in which the target population is injected with an engineered protein called Designer Receptors Exclusively Activated by Designer Drugs (DREADDs) and consequently activated by an orally administered designer drug (Roth, 2016). Similar to optogenetics, chemogenetics benefits from the forms of specificity provided by viral-vector techniques, but it is able to stimulate much larger neural populations, as the designer drug diffuses throughout the brain when orally administered (Sternson & Roth, 2014). Chemogenetic stimulations also have the unique ability to modulate brain activity for sustained periods ranging from hours to weeks (Muir, Lopez, & Bagot, 2019).
Measurement Techniques
Following the perturbation elicited by the selected targeted stimulation technique, we can simultaneously measure the brain’s relaxation trajectory back to the attractor of spontaneous activity. Collections of such measurements form a dataset of stimulus-evoked activity, which we have argued to reveal undiscovered mechanisms of brain dynamics upon further analysis. To capture the relaxation, the selected measurement technique must measure the exhibited brain activity at a matching spatial resolution. Microscale measurement techniques can resolve the response from individual neurons, while mesoscale and macroscale measurement techniques capture aggregated activity of neural populations. Figure 2B illustrates key measurement techniques, ordered by degrees of spatial resolution. We review these techniques below.
At a microscale resolution, single-unit recording methods capture the spiking activity of individual neurons. The classic single-unit recording method is whole-cell patch-clamp electrophysiology, which reads the intracellular potential from an electrode of a micropipette tightly suctioned to the target neuron membrane (Hill & Stephens, 2021). The spatial coverage of patch clamps is however limited to a few neurons. An alternate emerging modality is calcium imaging, which measures changes in levels in calcium concentration as an indirect proxy for spiking activity (Grienberger & Konnerth, 2012). Modern microscopy techniques can measure calcium levels of hundreds of neurons, recently measuring activity across all neurons in C. elegans in response to neuron scale optogenetics (Randi, Sharma, Dvali, & Leifer, 2023). With emerging microelectrode arrays, such as the recently developed Neuropixels probes (Steinmetz et al., 2021), it is now possible to directly measure neuronal spiking with a spatial coverage comparable with that of calcium imaging.
The activity of a mesoscale population is conventionally measured by the local field potential (LFP) of its extracellular medium, which aggregates the spiking of its constituent neurons (Buzsáki, Anastassiou, & Koch, 2012). LFPs can be recorded by intracranial electroencephalographic (iEEG) electrodes penetrating the cortex. Commonly used iEEG electrode designs include subdural strip or grid electrodes, which can monitor LFPs from multiple positions from the cortical surface, or depth electrodes, which can measure at multiple depths at a single surface position (Lachaux, Rudrauf, & Kahane, 2003).
The activity of a macroscale neural population could be measured by simultaneously tracking mesoscale populations with multiple iEEG systems; however, this option can only be implemented in rare surgical scenarios (e.g., the Functional Brain Tractography project (Jedynak et al., 2023)), as many electrodes become excessively invasive. Instead, there exist noninvasive functional neuroimaging techniques that make proxy measurements at this spatial scale. Electroencephalography (EEG) can measure electrical potentials from electrodes attached around the scalp at high temporal resolution, making it suitable for measuring transient responses to stimulation such as TMS (Rogasch & Fitzgerald, 2013). Complementing EEG is functional magnetic resonance imaging (fMRI), which measures the blood oxygen level dependent (BOLD) signals, tracking underlying neural activity. Despite its limited temporal resolution, fMRI with its high spatial coverage can be used to measure responses to a macroscale modulatory stimulation method such as chemogenetics (Markicevic et al., 2020, 2023; Zerbi et al., 2019).
In summary, there are a range of sophisticated targeted stimulation techniques that can perturb precise neurons or broad neural populations and measurement techniques that can simultaneously resolve the subsequent relaxation of neuronal spiking activity or aggregated neural population activity. Thanks to these technological advancements, datasets of stimulus-evoked activity can be generated through targeted stimulation experiments at any spatial scale of our interest: micro-, meso-, or macroscale. As argued through an abstracted picture of brain dynamics (Figure 1), stimulus-evoked brain activity may manifest highly nonlinear dynamics, driven by novel mechanisms. However, integrating these experiments with generative models allows us to rigorously test hypotheses about these nonlinear mechanisms at any spatial scale. In the next section, we detail how different generative models can be utilized to harness this significant opportunity.
GENERATIVE MODELS OF BRAIN ACTIVITY
Generative models provide a mechanistic account of the spatiotemporal dynamics of brain activity observed under any experimental setting. Mechanisms are encoded as a set of dynamical rules, such as difference equations or differential equations, which govern the evolution of brain activity over time. By simulating these governing rules, we can generate a spatially distributed time series of brain activity, which can be compared against activity from experimental data. Mechanisms of generative models can also be designed to generate nonlinear time series, allowing us to capture the nonlinear dynamical properties of stimulus-evoked activity. Aside from the ability to generate neuronal dynamics, the hypothesized mechanisms of generative models can also be assessed and validated by integrating the model with a target experimental dataset. This process may involve model selection and model calibration, where the model’s parameters and possible structural characteristics are optimized to explain the dataset as accurately as possible (Ramezanian-Panahi et al., 2022). One can additionally assess the model’s predictions of the target dynamics through an uncertainty quantification phase, in which statistical methods quantify the reliability of a model’s parameter estimates and predictions (Jha, Hashemi, Vattikonda, Wang, & Jirsa, 2022; Penas, Hashemi, Jirsa, & Banga, 2024), and sensitivity analyses evaluate how robust the output dynamics are to small changes in the inferred parameters (Baldy, Breyton, Woodman, Jirsa, & Hashemi, 2024; Hashemi et al., 2024). Through the generation of nonlinear dynamics and the integration with targeted stimulation datasets, generative models are therefore indispensable tools in the endeavor to infer mechanisms of brain dynamics revealed in the activity evoked by targeted stimulation.
Generative models encode various forms of mechanisms to provide different forms of accounts of brain dynamics. Figure 3 illustrates a spectrum of key generative models, organized by the type of mechanisms they encode. By choosing and employing an appropriate model, we can test and validate the novel mechanisms that we hypothesize to drive the dynamics evoked by targeted stimulation. Physiological models, for example, use physically measurable quantities to quantitatively encapsulate complex physiological processes in the brain (Breakspear, 2017; Deco et al., 2008). In contrast, phenomenological models use mathematical abstractions to capture the core spatiotemporal properties of brain dynamics hypothesized by the mechanism (Cabral et al., 2017; Kim & Bassett, 2020). Finally, we distinguish data-driven models, which learn statistical associations from sample data to make accurate predictions of brain dynamics (Acharya et al., 2022; Ramezanian-Panahi et al., 2022). Despite their varying mechanisms, all generative models in Figure 3 can simulate both spontaneous and stimulus-evoked activity as abstracted by the dynamical systems picture in Figure 1. The relevant output variables construct a state space where brain activity follows a trajectory of brain states (Figure 1A). Introducing stochastic noise as an input variable concentrates the trajectory around an attractor, reflecting the model’s depiction of spontaneous dynamics at the corresponding spatial scale (Figure 1B). Contrastingly, setting the input variable to a deterministic spatiotemporal profile aligned with the experimental stimulus would perturb the brain state away from the attractor before drawing it back via the relaxation trajectory (Figure 1C).
Generative models come in a variety of flavors, and can be positioned along a spectrum between three extremes, depending on the type of mechanistic account of brain dynamics they provide. (A) Physiological models—such as biological neural networks, neural mass models, and neural field models—model physiological processes of brain activity with physically measurable biological quantities (Breakspear, 2017; Deco et al., 2008). (B) Phenomenological models—such as linear network models, coupled oscillator models, and conductance-based models—model abstracted mechanisms that can capture essential dynamical properties of brain activity with minimal physiological detail (Cabral et al., 2017; Kim & Bassett, 2020). (C) Data-driven models—such as models constructed through system identification or artificial neural networks—learn statistical associations of brain dynamics from sample data to make accurate predictions of brain dynamics (Acharya et al., 2022; Ramezanian-Panahi et al., 2022).
Generative models come in a variety of flavors, and can be positioned along a spectrum between three extremes, depending on the type of mechanistic account of brain dynamics they provide. (A) Physiological models—such as biological neural networks, neural mass models, and neural field models—model physiological processes of brain activity with physically measurable biological quantities (Breakspear, 2017; Deco et al., 2008). (B) Phenomenological models—such as linear network models, coupled oscillator models, and conductance-based models—model abstracted mechanisms that can capture essential dynamical properties of brain activity with minimal physiological detail (Cabral et al., 2017; Kim & Bassett, 2020). (C) Data-driven models—such as models constructed through system identification or artificial neural networks—learn statistical associations of brain dynamics from sample data to make accurate predictions of brain dynamics (Acharya et al., 2022; Ramezanian-Panahi et al., 2022).
In the subsections below, we review how physiological, phenomenological, and data-driven models are constructed and how the specific types of mechanisms they encode can contribute to our understanding of the dynamics evoked by targeted stimulation. For each model category, we provide examples of modeling studies which have tested mechanisms underlying brain dynamics through an integration with datasets of stimulus-evoked activity, with a primary focus on emerging datasets of targeted stimulation. We also highlight how findings of some studies support our core argument presented through the abstracted picture in Figure 1: that stimulus-evoked dynamics could be driven by novel mechanisms of brain dynamics which play a more peripheral role in shaping spontaneous dynamics.
Physiological Models
In physiological models, variables and parameters represent physical quantities, such as cell-body potentials or synaptic coupling strengths, and interact in accordance to experimentally supported physiological processes (Breakspear, 2017; Deco et al., 2008; Ma & Tang, 2017). Physically measurable variables and parameters assist in interpreting the functional significance of individual variables and the effects of changes in static parameters on the resultant dynamics. Crucially, the physiological formulation allows predictions of the model to be tested not only against the model’s ability to capture the predicted brain dynamics but also against physical measurements from physiological experiments. Physiological models can study the dynamics of responses to a wide range of targeted stimulation techniques, with input variables quantifying the technique’s specific mechanics of action. For example, a physiological model could be formulated to directly compare responses with optogenetics, which directly depolarizes neuron membranes, with responses to electrode stimulation, which feeds a current into the extracellular medium. We can construct physiological models to test mechanisms of physiological processes at different spatial scales, ranging from the microscale to whole brain scale.
To capture microscale brain dynamics, we can use biological neural networks, which simulate the activity of interconnected neurons (Ma & Tang, 2017). Separate single-neuron models, such as the Hodgkin–Huxley model (Hodgkin & Huxley, 1952), are coupled to capture neuron–neuron interactions, reflecting underlying synaptic processes. By assigning separate models to each neuron, biological neural networks can simulate dynamics with realistic heterogeneity among neurons. Specific parameters can be assigned to heterogeneous classes of neurons, such as excitatory and inhibitory neurons (Arkhipov et al., 2018; Billeh et al., 2020; Gast, Solla, & Kennedy, 2024; Landau, Egger, Dercksen, Oberlaender, & Sompolinsky, 2016; Stefanescu & Jirsa, 2008), or classes of intricate morphological features (Aberra, Peterchev, & Grill, 2018; Aberra, Wang, Grill, & Peterchev, 2020). Ongoing projects to create expansive classification databases such as the Blue Brain Project (Markram, 2006) and the Allen Cell Types Database (Gouwens et al., 2018), alongside technological advances in computing power, suggest that future biological neural networks could simulate brain dynamics at neuronal resolution while spanning large spatial scales.
Zooming out from the microscale, mean-field population models instead track the mean activity of all neurons in a population, without tracking each constituent neuron (Breakspear, 2017; Deco et al., 2008). Crucially, this mean-field approach to model construction requires assumptions on the connectivities between populations in space. For example, there are neural field models (Jirsa & Haken, 1996, 1997), which assume isotropic connectivity between populations to allow for efficient computation of population-scale activity across continuous space (Deco et al., 2008; Robinson, 2005; Robinson et al., 2002; Robinson, Rennie, & Wright, 1997). Another class of mean-field models across populations are neural mass models or brain-network models (Breakspear, 2017; Pathak, Roy, & Banerjee, 2022), which coarse-grain neural populations into discrete spatial regions, and then couple these regions to one another in a network, using measured structural connectome data (Chaudhuri, Knoblauch, Gariel, Kennedy, & Wang, 2015; Deco, Jirsa, McIntosh, Sporns, & Kötter, 2009; Deco et al., 2021). Notwithstanding the limitations in accuracy and physical realism due to mean-field assumptions (Robinson, 2019), the significant reduction in the number of variables allows mean-field population models to be computed much more efficiently than the equivalent hypothetical biological neural network of the same spatial scale.
By designing stimulation inputs with unique spatiotemporal profiles, physiological models integrated with targeted stimulation datasets can help investigate the mechanisms behind varying responses to different stimulation forms. For instance, a biological neural network of the rat somatosensory cortex revealed how morphological features like axon length shape neuron activation under ICMS and TMS under various stimulation parameters (Aberra et al., 2018). Another biological neural network of the mouse and monkey somatosensory cortex investigated recurrent mechanisms underpinning the difference in responses between optogenetic and visual sensory stimulation (Sanzeni et al., 2023). Additionally, physiological models can account for nonlinearities observed in experimental responses to targeted stimulation. For example, a biological neural network of the macaque motor cortex demonstrated potential nonlinear mechanisms underpinning distortive effects of ICMS on the dynamics of ongoing task-related activity (O’Shea et al., 2022). At the macroscale, a neural field model with nonlinear synaptic mechanisms studied the calcium-dependent mechanisms underlying the nonlinear effects of TMS-induced plasticity (Fung & Robinson, 2013). A similar neural field model demonstrated significant differences in local synaptic strength parameters when fitted to spontaneous EEG versus evoked potentials (Kerr et al., 2008). This discrepancy was reconciled with a dynamic synaptic strength governed by a nonlinear gain modulation mechanism (Babaie-Janvier & Robinson, 2020).
Phenomenological Models
While physiological models offer realistic biophysically constrained explanations of how brain activity stems from independently measurable physiological processes, it is also useful to avoid excessive physiological details that complicate the model’s dimensionality and human interpretability (Robinson, 2022). An alternative approach, taken by phenomenological models, is to replace a full physiological description with simpler canonical dynamical systems, which preserve the dynamics hypothesized to be exhibited by the underlying complex physiological processes (Cabral et al., 2017; Kim & Bassett, 2020). This abstraction is supported by the observation that many physiologically realistic models of stable brain dynamics exhibit canonical dynamical structures such as fixed points and limit cycles (Freyer et al., 2011; Ramezanian-Panahi et al., 2022; Siu, Müller, Zerbi, Aquino, & Fulcher, 2022). While the variables and parameters of phenomenological models are not physiologically interpretable (and thus not experimentally measurable), they simplify the model’s complexity and the systems-level interpretation of the resulting simulated dynamics. Phenomenological models are thus valuable for investigating the core spatially distributed properties of the dynamics of networks of neurons or brain regions in response to targeted stimulation, without comprehensive detail of the local physiological mechanisms that underpin each neuron or brain region’s dynamics.
Phenomenological models vary in the sophistication of dynamics exhibited by their canonical structures. A simple example is a linear network model, which uses a linear system of equations to model the dynamics of a network of nodes representing neurons or neural populations (Kim & Bassett, 2020; Tang & Bassett, 2018). Each node’s dynamics are driven by a linear combination of activities from coupled nodes, and any external perturbation such as targeted brain stimulation. Linear models are a means to capture correlations between the activity of coupled neurons or neural populations with the simplest form of node-to-node coupling interactions (Abdelnour, Voss, & Raj, 2014; Raj, Kuceyeski, & Weiner, 2012). They may also allow for closed-form expressions for network responses to inputs, mathematically linking the system dynamics directly to its underlying coupling structure (Gu et al., 2017; Parkes et al., 2022).
Generalizing the linear setting, there are nonlinear network models governed by nonlinear systems of equations, which can capture more sophisticated forms of local dynamics. A simple example is linear threshold models (Mišić et al., 2015), which extend linear models by incorporating an additional threshold to the linear combination input. More sophisticated coupled oscillator models aim to capture oscillatory properties of brain dynamics as observed in recordings of brain activity across spatial scales (Breakspear, Heitmann, & Daffertshofer, 2010; Deco, Kringelbach, et al., 2017). Commonly used oscillators include the Kuramoto oscillator which oscillates with constant amplitude (Cabral, Hugues, Sporns, & Deco, 2011; Gollo, Roberts, & Cocchi, 2017) and the Stuart–Landau (Hopf) oscillator which can also change in amplitude (Deco et al., 2018; Deco, Cabral, et al., 2017; Deco, Kringelbach, et al., 2017). A more physiologically inspired nonlinear model is a conductance-based model, which models population-level activity abstracted as a simpler biological neural network (Breakspear, Terry, & Friston, 2003; Honey et al., 2009; Larter, Speelman, & Worth, 1999). Here, the aggregate activity of each population is tracked by the dynamics of a single neuron model, which assumes a sufficiently strong coherence between constituent neurons of each population (Breakspear, 2017). Despite this strong physiological assumption, conductance-based models are capable of exhibiting population-scale spatiotemporal dynamics with interesting properties that are commonly observed in biological neural networks.
In comparison with physiological models, the abstractions incorporated in the local dynamics of phenomenological models limit their ability to distinguish between the dynamics evoked by different targeted stimulation techniques. However, phenomenological models can be used to more directly investigate how the brain’s underlying structure shapes the spatially distributed properties of its stimulus-evoked dynamics under an arbitrary stimulus input, as they can isolate the effects of complex local dynamics. For example, a coupled oscillator model integrated with mouse optogenetic stimulation data demonstrated how heterogeneous long-range structural connections allow stimulus-evoked activity to reveal a rich repertoire of novel dynamical responsive functional networks, extending the established repertoire of resting-state networks observed in spontaneous dynamics (Spiegler, Abadchi, Mohajerani, & Jirsa, 2016; Spiegler, Hansen, Bernard, McIntosh, & Jirsa, 2020). Phenomenological models can also investigate the extent to which deviations in spatiotemporal dynamics between stimulus-evoked and spontaneous activity are influenced by the stimulus position. For example, a coupled oscillator model demonstrated how TMS-induced changes in functional connectivity varied between stimulating highly connected hub regions in hub regions and stimulating weakly connected peripheral regions (Gollo et al., 2017). Another coupled oscillator model studied how the influence of stimulus position varied between different global brain states such as wakefulness and deep sleep (Deco et al., 2018).
Data-Driven Models
The governing rules of phenomenological and physiological models summarized above reflect qualitatively interpretable mechanisms underlying brain dynamics, such as the internal physiological processes of neurons in biological neural networks or the oscillatory behaviors of population-scale neural recordings in a whole-brain oscillator model. In contrast, data-driven models encode statistical mechanisms that can describe the quantitative structure of the governing rules without a qualitative interpretation (Acharya et al., 2022; Ramezanian-Panahi et al., 2022). Through the use of statistical techniques, data-driven models are capable of exploring a wide range of possible governing rules which reflect encoded statistical mechanisms, enabling highly accurate predictions of brain dynamics that are unconstrained by predefined physiological or phenomenological mechanisms. This flexibility is particularly advantageous when studying responses to targeted stimulation, where underlying qualitative mechanisms may be difficult to hypothesize in advance. At the same time, choosing rule structures requires careful judgment: Models need enough coefficients to predict complex dynamics, but not too many underconstrained coefficients that undermine the model’s generalization ability.
The range of different possible statistical properties of brain activity that a data-driven model can capture varies with the range of governing rules that it can explore. For example, there are data-driven models that use the process of system identification to infer a set of mathematical equations that govern the dynamics of brain activity from the dataset (Acharya et al., 2022). While system identification algorithms are limited to exploring and discovering equation-based governing rules, these equations enable the interpretation of the learned statistical properties of individual time series and pairwise dependencies between time series. Commonly used system identification procedures include linear time series models which learn simple linear equation structures (Ljung, 1998) and procedures which can estimate nonlinear equation structures (Billings, 2013; Brunton, Proctor, & Kutz, 2016; Williams, Kevrekidis, & Rowley, 2015). Comparing different system identification procedures therefore provides an opportunity to detect nonlinearities in stimulus-evoked dynamics by comparing the capacities of linear and nonlinear equations in predicting responses (Acharya, Davis, & Nozari, 2024; Chang et al., 2012; Yang et al., 2021).
There are also more flexible data-driven models such as artificial neural networks (not to be confused with biological neural networks), which can be used to solve a more general problem of learning a predictive mapping between an input and output variable from a dataset of observations (Acharya et al., 2022). The majority of artificial neural network architectures, including the simplest feedforward architecture, learn nonlinear mappings between variables as a sequence of layers of units, including the input layer, output layer, and hidden layers (Ljung, Andersson, Tiels, & Schön, 2020). Each unit receives inputs from units of the previous layer, then feeds the input into an activation rule before sending it to units of the next layer. While less interpretable than the systems of equations discovered by system identification methods, artificial neural networks are more flexible in replicating a wider range of governing rule structures, including those with high orders of nonlinearities that would otherwise be computationally expensive to be captured as a set of equations (Hornik, Stinchcombe, & White, 1989; Vyas, Golub, Sussillo, & Shenoy, 2020). This flexibility potentially enables artificial neural networks to create future personalized models that make accurate predictions of individual responses to targeted stimulation, adapting important individual-specific structural and functional characteristics that are difficult to encode as physiological or abstracted mechanisms (Misra et al., 2021).
A systematic comparison of data-driven models can be used to test whether hypothesized statistical mechanisms are necessary in capturing responses to targeted stimulation. For example, a comparative study of responses to DBS found significant differences in fitted linear model parameters between spontaneous and stimulus-evoked dynamics, thus suggested the use of a switched linear model as a simple nonlinear model that switches between different parameters with and without stimulation (Acharya et al., 2024). Similar predictive powers were also found between switched linear models and less-interpretable feedforward neural networks, demonstrating that switched linear models sufficiently explained responses to DBS without significantly trading off predictive capabilities (Acharya et al., 2024). Another avenue of systematic comparison involves an architecturally distinct family of recurrent neural networks, which use recurrent connections to accumulate memory of an indefinite number of past inputs and observed states (Ljung et al., 2020). While recurrent neural networks have mostly focused on modeling dynamics during task stimuli (Mante, Sussillo, Shenoy, & Newsome, 2013; Vyas et al., 2020), they may also investigate how dynamic working-memory mechanisms can impact variations in time-dependent responses to sequences of targeted or sensory stimulation. For example, a comparison of recurrent neural networks with other memoryless neural network architectures demonstrated the necessity of internal memory mechanisms in predicting responses to visual stimuli sequences of varying images (Güçlü & van Gerven, 2017).
In summary, in this section, we have explored the variety of different generative models, as well as their different aims, advantages, and disadvantages for modeling dynamics evoked by targeted stimulation. Choosing an appropriate model is crucial for investigating different types of questions about this form of dynamics: Testing mechanisms related to the physiology of neurons would benefit from a biological neural network, or exploring the role of the connectome in the spatially distributed response to TMS could benefit from a phenomenological linear network model, or a project to accurately predict the firing rates of single neurons in response to optogenetic stimulation could benefit from system identification models or artificial neural network models. Integrated with datasets of activity from targeted stimulation experiments, generative models complete our toolkit for uncovering the mechanisms underlying the complex, high-dimensional dynamics measured from this unique experimental paradigm.
CONCLUSION
Emerging targeted stimulation techniques such as optogenetics and electrode stimulation, combined with measurement techniques such as calcium imaging and EEG, now allow us to measure the intricate spatiotemporal responses of neurons and neural populations to a range of perturbations with unprecedented precision (Deisseroth, 2015; Keller et al., 2014; Rogasch & Fitzgerald, 2013; Roth, 2016). Using an abstracted picture of brain dynamics, and drawing on results from dynamical systems theory, we presented an argument for why the dynamics from novel states accessed in these targeted stimulated experiments may be driven by new types of mechanisms, distinct to those underpinning the simpler low-dimensional dynamics of spontaneous activity. The mechanisms which govern dynamics evoked by targeted stimulation are also functionally significant, as they reflect how the brain recruits diverse neural populations while processing complex streams of sensory inputs. Fortunately, through an integration of targeted stimulation and measurement techniques with different generative models, we can uncover, analyze, and better understand these mechanisms at a chosen spatial scale. Targeted stimulation techniques allow us to precisely perturb the brain to a range of states distant from the attractor manifold on which spontaneous dynamics unfold. Measurement techniques can simultaneously capture the brain’s trajectory as it relaxes back to the attractor, generating datasets of activity that could reveal the novel (and likely highly nonlinear) mechanisms that govern the relaxation’s dynamics. Finally, by integrating datasets of activity evoked by targeted stimulation with generative models, it is possible to rigorously test and validate mechanisms that we hypothesize to drive the dynamics. Different generative modeling approaches are suited to testing different types of hypothesized mechanisms, from different physiological processes that can be incorporated in physiological models (Breakspear, 2017; Deco et al., 2008) to the abstracted dynamical mechanisms treated in phenomenological models (Cabral et al., 2017; Kim & Bassett, 2020) and the statistical mechanisms empirically learned using data-driven models (Ljung, 1998; Ljung et al., 2020). Our exploration touches on some of the many opportunities to uncover mechanistic insights into complex brain dynamics, ultimately guiding the practical implementation of emerging neuromodulation technologies in both research and clinical settings.
While our dynamical systems framework provides a foundation for interpreting the unique characteristics of brain dynamics evoked by targeted stimulation compared with spontaneous brain activity, there are alternative ways to conceptualize these dynamics. One such approach involves the separation of timescales (Kuehn, 2015). Some mechanisms may act too quickly to significantly influence the trajectories of fluctuations near the spontaneous activity attractor but become crucial in capturing the far-from-attractor dynamics evoked by targeted stimulation. Brain activity may therefore possibly be modeled as a multitimescale system, with mechanisms operating over various timescales in response to targeted stimulation (Babaie-Janvier & Robinson, 2020; Chaudhuri et al., 2015). Targeted stimulation can also trigger new forms of nonequilibrium dynamics if multiple attractors exist in the state space, where specific perturbations can cause the relaxation trajectory to migrate to different attractors. Each attractor may represent unique forms of spontaneous brain activity, such as the states of deep sleep versus wakefulness (Deco et al., 2019) or sequential discrete thoughts during cognition (Spivey & Dale, 2006). Generative models of targeted stimulation can therefore inform the development of neuromodulation treatments to restore healthy brain function, by designing optimal perturbations that can force brain states to transition from “diseased” to “healthy” attractor dynamics (Kringelbach & Deco, 2020; Muldoon et al., 2016; Perl et al., 2021).
As a novel experimental paradigm, datasets generated from targeted stimulation experiments allow us to not only test newly hypothesized mechanisms of brain dynamics but also to refine existing mechanisms that have been validated mostly on datasets of spontaneous activity. Integrating existing models with new datasets can potentially contribute toward resolving standing issues of degeneracy in computational neuroscience, where generative models founded on differing mechanisms can lead to similar dynamics of spontaneous activity (Bernard, 2023). It is possible that the relaxation dynamics measured in targeted stimulation experiments, given their potential nonlinearities, can be used to arbitrate between different generative quantitative models of brain dynamics, beyond the relatively simple statistical properties of spontaneous activity that both forms of models are known to adequately capture, such as their pairwise linear correlation structure (as “functional connectivity”) (Messé et al., 2015; Nozari et al., 2024).
Looking at and beyond the targeted stimulation setting, we suggest that more synergistic interactions between experimentalists and theoreticians can accelerate progress in our mechanistic understanding of brain dynamics. As experimentalists forge more sophisticated paradigms that each generate unique forms of brain dynamics, theoreticians can build more robust generative models that can explain or predict the dynamics of brain activity from a wider range of experimental settings. Conversely, refined generative models motivate the development of further experimental designs tailored to test and validate new theoretical accounts and predictions. This reciprocal relationship between experimental precision and theoretical rigor can uncover novel hypotheses and validate theoretical predictions, fostering a deeper and more integrative comprehension of the intricate mechanisms underlying brain function.
ACKNOWLEDGMENTS
The authors thank Nigel Rogasch for feedback and suggestions for the manuscript. R.M. would like to thank the Australian Government Research Training Program (RTP) Scholarship for financial support. B.D.F. would like to thank the Selby Scientific Foundation for financial support.
AUTHOR CONTRIBUTIONS
Rishikesan Maran: Conceptualization; Visualization; Writing – original draft; Writing – review & editing. Eli J. Müller: Writing – review & editing. Ben D. Fulcher: Conceptualization; Supervision; Writing – review & editing.
TECHNICAL TERMS
- Targeted stimulation experiment:
An experiment that generates stimulus-evoked dynamics through targeted stimulation and measurement techniques.
- Generative model of brain activity:
A quantitative model that generates brain dynamics as a time series, which can be compared with experimental data.
- Spontaneous dynamics:
Brain dynamics in the absence of any external task or stimulus, which can be conceptualized as noisy deviations about an attractor.
- Physiological model:
A generative model of brain activity that models physiological processes in the brain using physically measurable variables and parameters.
- Phenomenological model:
A generative model of brain activity that captures core spatiotemporal properties of brain dynamics with abstracted dynamical mechanisms.
- Data-driven model:
A generative model of brain activity that learns statistical associations of brain dynamics from experimental data.
- Attractor:
A low-dimensional subset of the state space to which nearby trajectories are attracted.
- Stimulus-evoked dynamics:
Brain dynamics resulting from an external stimulus, which can be conceptualized as a perturbation away from an attractor followed by a relaxation back to the attractor.
REFERENCES
Competing Interests
Competing Interests: The authors have declared that no competing interests exist.
Author notes
Handling Editor: Olaf Sporns