Signal interactions in brain network communication have been little studied. We describe how nonlinear collision rules on simulated mammal brain networks can result in sparse activity dynamics characteristic of mammalian neural systems. We tested the effects of collisions in “information spreading” (IS) routing models and in standard random walk (RW) routing models. Simulations employed synchronous agents on tracer-based mesoscale mammal connectomes at a range of signal loads. We find that RW models have high average activity that increases with load. Activity in RW models is also densely distributed over nodes: a substantial fraction is highly active in a given time window, and this fraction increases with load. Surprisingly, while IS models make many more attempts to pass signals, they show lower net activity due to collisions compared to RW, and activity in IS increases little as function of load. Activity in IS also shows greater sparseness than RW, and sparseness decreases slowly with load. Results hold on two networks of the monkey cortex and one of the mouse whole-brain. We also find evidence that activity is lower and more sparse for empirical networks compared to degree-matched randomized networks under IS, suggesting that brain network topology supports IS-like routing strategies.

How do mammal brains control the communication of signals across their entire network? A fundamental goal for any large-scale communication system is managing signal interactions. Yet because brain dynamics are nonlinear and emergent, signal interactions on brain networks have been little studied. Here we investigate two forms of nonlinear signal interaction on the mammal mesoscale connectome: collisions and duplication (redundancy). Using explicit numerical simulations on high-accuracy connectomes of the mouse and monkey, we find evidence that mammal brain networks operate efficiently under a routing strategy that employs destructive collisions and the redundant spread of information. In contrast, standard random walk strategies with the same collision rule are less efficient in terms of global activity and sparseness of activity. Comparisons to randomized networks support these findings.

How does the mammal brain manage the communication of signals across its network? At the level of mesoscale mammal brain network structure, comprising connectivity of tens to hundreds of brain areas over thousands to hundreds of thousands of connections, an increasingly detailed picture of network connectivity is emerging (Markov et al., 2014; Oh et al., 2014; Gămănuţ et al., 2018). Researchers are now aiming to “animate” these networks in order to understand patterns of whole-brain signaling dynamics (a research area termed dynomics; Kopell et al., 2014).

However, a given topology can support a wide range of observable dynamics (see, e.g., Friston, 2011; Knoblauch et al., 2016). Adding to the challenge is that activity on the whole network is poorly understood, with severe limitations in sampling at the single-neuron level, and in spatial and temporal precision for imaging methods. Moreover, though single-neuron spiking dynamics for some neuron classes are well understood, global activity will not necessarily be a linear summation of single-neuron dynamics.

One approach to understanding brain network dynamics is to consider the system of intercommunication among nodes as a bridge between network topology and network dynamics. Communication dynamics can serve as a linkage between structural connectivity and functional connectivity (Avena-Koenigsberger et al., 2018). More generally, nodes in the network can be considered as communicators that attempt to pass signals or messages to other nodes according to routing rules (Graham, 2014).

Communication demands such as flexibility and reliability place important constraints on the system (Poggio, 1984; Hahn et al., 2018). Effective and robust routing strategies are especially important on the mammal connectome because the topology is such that paths of just a few synapses exist between any node and practically any other node. Brain networks also need to dynamically exchange multimodal information among a large number of nodes in real time without appreciably changing network topology. In addition, routing in the brain lacks central control.

To begin to understand what routing strategy is in use in the brain, one can look to design strategies in engineered systems (Graham & Rockmore, 2011; Graham, 2014, 2017; Navlakha et al., 2015, 2017; Fornito, Zalesky & Bullmore, 2016). A fundamental engineering goal for any large-scale communication system is the management of signal interactions. Signal interactions could take many forms such as summation, thresholding, collision, and duplication. As we describe in the Discussion, brain-like networks with signal interactions involving summations/thresholds tend to produce undesirable behavior such as activity die-off or overload (Kaiser et al., 2007; Kaiser & Hilgetag, 2010). Here we perform initial simulations of two interactions that have received little attention: signal collision and signal duplication (i.e., redundancy).

Collisions occur when signals from different sources converge on the same target at the same time. They are the price paid for the ability to route signals selectively to many possible destinations on a fixed network. Collision dynamics are emergent: they are a nonlinear effect of topology, node dynamics, and current traffic. All large-scale engineered communication systems have means for managing collisions, through redress, arbitration, and other strategies (see, e.g., Kleinrock, 1976; Mišić & Mišić, 2014).

Given the a priori likelihood of collisions on small-world-like brain networks (compared to, for example, lattice networks), we argue that it is in the interest of brain networks to establish dynamic interactions that manage collisions successfully across the entire brain. Management of collisions is especially important in mesoscale brain networks since behavior-related information must be exchanged among a large number of subsystems.

However, collisions introduce nonlinearities, which are difficult to study analytically. Nonlinear signal interactions are increasingly investigated with explicit agent-based models in the study of other communication networks (e.g., epidemics: Min et al., 2020), but little work in network neuroscience has used explicit approaches. For example, most RW models (e.g., Noh & Rieger, 2004; Abdelnour, Voss, & Raj, 2014), which route signals to randomly chosen outgoing edges use analytical methods that cannot account for signal interactions such as collisions.

An exception is Mišić et al. (2014a, 2014b). In this pair of studies, buffers (node memory allocations, a form of nonlinearity) were employed to manage collisions in a RW routing model. In this scheme, colliding signals at the inputs of a node are lined up in a queue and stored in node memory until they can be directed to a randomly chosen outgoing edge. Buffers are ubiquitous on the Internet and may be useful in brains. But though several neurobiological mechanisms for buffering have been proposed, they remain hypothetical (see, e.g., Goldman-Rakic, 1996; Graham, 2014; Funahashi, 2015).

An alternative to RW models is shortest path routing, which invokes the logic of evolved optimality (e.g., Bullmore & Sporns, 2009; Rubinov & Sporns, 2010; Goñi et al., 2013, 2014). Shortest path strategies also have parallels in engineered systems like the Internet (e.g., OSPF: open shortest path first routing; see, e.g., Sosnovich et al., 2017). However, shortest path models typically ignore signal interactions such as collisions, again because these models are studied analytically. Yet signal interactions are especially important in a shortest path context. Though it has not been well recognized, successful shortest path routing requires knowledge of current network traffic: a short path is not necessarily short if there is congestion. Even when short path models include small buffers, they show substantial message loss due to collisions (Graham & Hao, 2018). In any case, shortest path routing is considered implausible because it requires global knowledge of network architecture to select shortest paths for all signals (Seguin, Razi, & Zalesky, 2018, but see Mišić et al., 2015).

In the present work, we investigate a strategy of fully destructive collisions. The brain may in fact use a less punitive strategy than this, but neurophysiological evidence increasingly supports the notion that signals in the brain regularly collide and are destroyed. For example, Sardi et al. (2017) demonstrated the failure of coincident signals above threshold to elicit spikes in vitro. They also showed that excitatory signals can cancel each other when they collide “head on” (Sardi et al., 2017; this is in fact predicted by the Hodgkin–Huxley model: Scott, 1977). Gidon et al. (2020) have shown exclusive-or (XOR) activity in single ex vivo human pyramidal cortical neurons, a behavior that effectively destroys some colliding signals. At the circuit level, gating circuits (e.g., Steriade & Paré, 2007; Gollisch & Meister, 2010), when closed, can also be conceived as destructive collisions between signals. If collisions need to be managed at the single-cell level and the level of circuits, the same logic applies at the level of mesoscale brain networks (see Discussion).

Collisions need not incapacitate a communication network. They may be useful in establishing stable and robust dynamics with locally implemented routing rules. In particular, we suggest that collisions could—almost paradoxically—help promote system efficiency.

Destructive collisions may on average help ensure a low level of energy use. Energy efficiency is imperative given that the dynamic connectome of mammals has strong metabolic constraints (Bullmore & Sporns, 2012; Goñi et al., 2013). Destructive collisions may also promote homeostasis, which appears desirable for maintaining relatively constant activity in the face of changing system demands (e.g., Aeschbach et al., 1997; Turrigiano, 1999).

But the brain cannot achieve efficiency simply by using as little energy as possible (Poldrack, 2015). The brain instead operates in sparse fashion, in part because the sensory environment is sparse. To avoid confusion, we note that sparseness (or, equivalently, sparsity) in the present context refers to the distribution of activity across nodes and over time, rather than to the topology of the network (on a scale from “sparse” to “dense” connectivity). Sparse operation of a system is defined by the achievement of low average activity ratios (see, e.g., Földiák, 2002).

Although sparse activity can confer low average energy usage, it does not necessarily imply minimal energy use. Instead, it involves a characteristically non-Gaussian distribution of activity across units, in particular a distribution with a strong peak near zero activity (most units are “off” at a given time) and heavy tails (a small number of units are likely to be highly active).

In the brain, it has been understood for decades that only a small fraction of units (10% or less) in a given part of the network can be highly active at a given time (Levy & Baxter, 1997; Lennie, 2000; Attwell & Laughlin, 2003); this is known as population sparseness. In addition, it is estimated that a unit can only be highly active over a small fraction of its lifetime; this is known as lifetime sparseness (see Willmore & Tolhurst, 2001; Graham & Field, 2006). Sparse activity in the visual system, for example, is thought to be in part a result of sparse physical environments (Field, 1994; Olshausen & Field, 1996; Bell & Sejnowski, 1997), in addition to metabolic constraints and other factors (Graham & Field, 2009). We suggest here that collision dynamics that result in emergent sparse activity are good candidates for how the mammal brain routes information.

If collisions are destructive, it may be necessary for the brain to adopt strategies to generate redundancy in order to promote reliability. Brain network communication models increasingly include not just a pathfinding dimension (spanning a spectrum from random walks to shortest paths: Avena-Koenigsberger et al., 2019), but also a second dimension that can be seen as redundancy (Avena-Koenigsberger et al., 2017; Bettinardi et al., 2017; Tipnis et al., 2018).

A simple way to promote redundancy is to produce redundant copies of messages at each node. Based on this idea, we introduce an information spreading (IS) model of brain network communication. Signal interactions in a system with collisions and information spreading will produce nonlinear, emergent behavior. For example, as more signals and copies of those signals propagate on the network, the influence of collisions is increasingly important. Using numerical simulations, we investigate emergent behavior of the IS model, in comparison to a standard random walk (RW) model. As described more fully in the Discussion section, IS and RW models should be seen as spanning a spectrum of distributed routing strategies, rather than as a test model and a control.

We show that mammal brain networks using IS achieve a balance between collisions and redundancy and they produce sparse activity. In particular, we show that, across a range of signal loads, low activity and high sparseness are emergent properties of network-wide communication under an IS model. Performance of the IS model contrasts with RW models, which exhibits high and increasing energy use with increasing signal loads, and lower and decreasing sparseness across loads. To investigate whether empirical topology in the mammal connectome supports sparse, efficient activity patterns, we also compare dynamics of the anatomical network to randomized networks with matched degree distributions.

Overview

Using a Markovian agent-based model on monkey and mouse tracer-based connectomes, we implemented an information spreading (IS) model wherein nodes pass copies of incoming signals to all outgoing edges. Simulations utilized a nonlinear collision rule whereby all colliding messages are destroyed. We also tested a standard random walk (RW) model following the same collision rule. In RW, each message is passed from one node to another through a randomly chosen outgoing edge. At each time step, a set number of messages, which we term the load, is injected to randomly chosen nodes; this is the primary independent variable in this study.

Connectomes

We utilized three mesoscale mammal connectomes: two of the macaque monkey (Markov et al., 2014, termed monkey1 in the present study as shorthand; and the CoCoMac database, see Bakker et al., 2012, termed monkey2), and one for the mouse (Oh et al., 2014, termed mouse). All edges are directed and have a weight of 1.

The monkey1 connectome comprises 91 cortical nodes and 1,615 edges, primarily in the visual system (ipsilateral). Of these nodes, 29 have in- and out-degree > 0, while the remainder are source nodes (in-degree = 0). This means that 62 nodes in monkey1 are not reachable from the rest of the network, and only pass messages injected at those nodes. This limitation is a result of the retrograde tracing method used by Markov et al. (2014), and is the reason we included the monkey2 network as well. For monkey2, we used only nodes that had an in-degree and an out-degree of at least 1. This set comprised nodes corresponding to 163 cortical regions and 5,093 edges. The mouse connectome spans the entire mouse brain (ipsilateral), comprising 213 nodes and 16,954 edges (all have in- and out-degree > 0). For all connectomes, isolated nodes with no incoming or outgoing edges (present only in monkey2) had been removed. We note that both mouse and monkey2 contain self-loops (i.e., nodes with edges to themselves), but removing these edges did not change the outcome of the experiments so self-loops are included in the presented data. Including nodes with in-degree or out-degree of 0 also did not affect the outcome. Adjacency matrices for the datasets are shown in Figure 1.

Figure 1. 

Adjacency matrices of the connectomes tested. Mouse includes the ipsilateral whole brain of the mouse (Oh et al., 2014), while monkey1 (Markov et al., 2014) and monkey2 (Bakker et al., 2012) include monkey cortex. All edges (blue dots) are directed and have weight 1. Nonzero (nz) entries are shown in numerical and percentage terms below each matrix.

Figure 1. 

Adjacency matrices of the connectomes tested. Mouse includes the ipsilateral whole brain of the mouse (Oh et al., 2014), while monkey1 (Markov et al., 2014) and monkey2 (Bakker et al., 2012) include monkey cortex. All edges (blue dots) are directed and have weight 1. Nonzero (nz) entries are shown in numerical and percentage terms below each matrix.

Close modal

Message Passing Models

In our models, time is discretized and message passing at each node is synchronous. During a time step, a node can be active or inactive (0 or 1). Each time step t, L new messages are injected into the system at randomly chosen nodes, representing load. Variable loads are to be expected in brain networks, but are most likely restricted to a relatively low range given metabolic constraints. Because the three connectomes have different numbers of nodes N, load is expressed as a percentage of the number of nodes in the network. We also report the case of 1 new message injected per time step in each case. We restrict simulations to the load regime with a number of messages less than or equal to 50% of N.

During each time step, a node looks to all incoming messages, including messages that are already in the system and new messages that are scheduled to be injected; if there is more than one message coming in, all are deleted. The full set of operations in the model is given as pseudocode in Supporting Information Box 1. We are also making code to generate these models available on GitHub (Hao & Graham, 2020).

Measures of Dynamic Network Activity

Simulations consist of 500 runs of 1,000 time steps each. Data from the first 500 time steps is excluded to allow the analysis of equilibrium dynamics (see, e.g., Schruben & Margolin, 1978). However, we note that data for the full test run produces the same overall pattern of results.

First, we measured the fraction of nodes active on a given time step. We measured both attempted activity before accounting for collisions, and actual (net) activity after collisions.

We also measured the sparseness over time in five time step windows (bins) using the Treves--Rolls measure of sparseness (Treves & Rolls, 1991; Rolls & Tovee, 1995); recall that sparseness here relates to network activity rather than to network structure. The Treves–Rolls measure is sensitive to higher order statistical regularities in data (see Willmore & Tolhurst, 2001). We calculate population sparseness by constructing a histogram of the frequency of firing xn within a time window per node, then sum over i nodes. The Treves–Rolls measure S (Equation 1) divides the square of the sum of the distribution by the sum of the squares of the distribution:
(1)
This measure ranges between maximal sparseness at N−1 if one unit is active per window and all other units are inactive; minimal sparseness is 1.0 for uniformly distributed activity (sparseness of standard normal distribution by this measure is 0.64). See Figure 2 for an illustration of this measure. Lower values of Treves–Rolls sparseness indicate more sparse activity. We note that tests of 10 time step windows for the sparseness measure produced similar results. We also calculated lifetime sparseness as above, summing over time instead of over nodes.
Figure 2. 

Illustration of the Treves–Rolls measure of sparseness S for hypothetical histograms of activity data. Shown are histograms that are well approximated by common functions, along with corresponding sparseness values as calculated from Equation 1. Top left: (half) Gaussian, S ≈ 0.64; top right: exponential, S ≈ 0.5; bottom left: generalized Pareto (shape parameter = 0.25), S ≈ 0.33; bottom right: maximal sparseness (closest to zero), S = N−1 , is achieved for a histogram where xn are all zero except for one nonzero entry, indicated here by “High.” A uniform histogram (not shown) has minimal sparseness of S = 1.0 .

Figure 2. 

Illustration of the Treves–Rolls measure of sparseness S for hypothetical histograms of activity data. Shown are histograms that are well approximated by common functions, along with corresponding sparseness values as calculated from Equation 1. Top left: (half) Gaussian, S ≈ 0.64; top right: exponential, S ≈ 0.5; bottom left: generalized Pareto (shape parameter = 0.25), S ≈ 0.33; bottom right: maximal sparseness (closest to zero), S = N−1 , is achieved for a histogram where xn are all zero except for one nonzero entry, indicated here by “High.” A uniform histogram (not shown) has minimal sparseness of S = 1.0 .

Close modal

Comparison to Randomized Networks

To investigate the dependency of network dynamics on the specific wiring patterns of the mammal connectome, we compared network activity under our models to the same models implemented on randomized networks. Randomized networks have the same distribution of in- and out-degree as the empirical networks. This amounts to shuffling the in- and outgoing edges of nodes in a given network (Maslov & Sneppen, 2002).

However, the nature of the connectomes in this study complicates the randomization process. When a network is dense (in a topological sense; except for the present discussion of randomization, all references to density and sparseness refer to network activity rather than topology), as in the mouse connectome and the 29-node core of monkey1, network topology changes little after randomization. We therefore thresholded the mouse connectome to make it less dense for randomization. Since edge weights in the mouse are known (corresponding to tracer volumes) we thresholded the network to exclude edges with a weight below 0.0136 (a value close to the modal weight, and one implied as meaningful by Oh et al., 2014), then randomized the resulting network. Thresholding of the mouse network reduced degree in a way that brought edge density to a level comparable to that of monkey2 (but interestingly, the global topology changed relatively little after removing more than half of the edges, suggesting self-similarity). The density problem is compounded in monkey1 because the full 91-node network is low density as compared to the other two networks. Therefore, we have omitted randomization of monkey1; results of randomized network comparison in the monkey are for the monkey2 connectome, which has intermediate connectivity. Note that each trial in the randomized network simulations corresponds to a different randomized network. Other randomization approaches, such as that of Tipnis et al. (2018), have used iterative search among randomized networks (based on the method of Maslov & Sneppen, 2002) to find those that differ most from the empirical network. However, this approach has potential for bias since only a single randomized network is generated, whereas our method compares the empirical network to a large family of randomizations.

Qualitative differences in network dynamics are apparent in visualizations of node activity over time. In Figure 3, we show attempted and actual (net) activity through a representative simulation of the mouse connectome, with nodes shown along the vertical axis and the last 500 time steps of the simulation running from left to right ( L = 10 %).

Figure 3. 

(A) Attempted activity (before collisions) and (B) actual activity (net, following collisions) under IS and RW models in the mouse connectome at 10% load. Vertical axis corresponds to nodes and time runs left to right. Colors indicated in the vertical axis in attempted activity plots (A) represent the number of colliding messages at a given node and time step. Net activity per node per time step (B) is binary (black = off, white = active).

Figure 3. 

(A) Attempted activity (before collisions) and (B) actual activity (net, following collisions) under IS and RW models in the mouse connectome at 10% load. Vertical axis corresponds to nodes and time runs left to right. Colors indicated in the vertical axis in attempted activity plots (A) represent the number of colliding messages at a given node and time step. Net activity per node per time step (B) is binary (black = off, white = active).

Close modal

Figure 3A shows attempted message passing per node according to the color scale shown (i.e., the number of messages colliding). In general, attempted messages are distributed in unimodal, monotonically decreasing fashion in RW, whereas attempted messages are bimodally distributed in IS with a high peak at zero (not shown).

Figure 3B shows actual activity for the IS model (left panel) and the RW model (right panel). From inspection, the net activity is more “bursty” and sparse in IS while activity is more uniformly distributed across nodes and time in RW.

We can quantify these observations by calculating the average net activity and the sparseness of activity across simulations. Figure 4 shows average net activity (expressed as a fraction of the number of nodes active in a given model, averaged over time and trials) as a function of load L (expressed as a percentage of nodes creating new messages on each time step, except for the left-most tick, which represents injection of one new message per time step). IS models are shown with solid lines and RW models are shown with dashed lines. Though RW models have lower net activity with L = 1 message, net activity increases rapidly and monotonically as load increases. The relationship is well fit by a log function (not shown). In contrast, IS models show lower activity that is relatively uniform across load.

Figure 4. 

Mean of net simulated activity on the mammal connectome as a function of load under IS (solid lines) and RW (dashed lines) models for mouse (circles), monkey1 (squares), and monkey2 (triangles) networks. In IS, activity remains mostly constant across load and is lower compared to the corresponding value for RW at all load levels except 1 message per time step. Activity in RW increases by more than two octaves across loads tested. Shaded areas represent 2 standard deviations of variability from the mean across simulations (note that in some cases this dispersion is smaller than the width of the markers/lines). Data points show significant differences (t test, p < 0.01 ) for corresponding IS and RW measurements at all loads.

Figure 4. 

Mean of net simulated activity on the mammal connectome as a function of load under IS (solid lines) and RW (dashed lines) models for mouse (circles), monkey1 (squares), and monkey2 (triangles) networks. In IS, activity remains mostly constant across load and is lower compared to the corresponding value for RW at all load levels except 1 message per time step. Activity in RW increases by more than two octaves across loads tested. Shaded areas represent 2 standard deviations of variability from the mean across simulations (note that in some cases this dispersion is smaller than the width of the markers/lines). Data points show significant differences (t test, p < 0.01 ) for corresponding IS and RW measurements at all loads.

Close modal

Figure 5 shows population sparseness of activity as a function of load. As with net activity across nodes, sparseness of activity in a given network is greater (closer to zero) for IS compared to RW at all loads except the minimum load. As a function of load, we observed the same pattern for sparseness over time within a simulation (lifetime sparseness of activity) as was seen for population sparseness, namely that sparseness over time in a given network is greater (closer to zero) for IS compared to RW except at the lowest loads (see Supporting Information Figure S1). We note that, like attempted activity, net activity over time for IS is bimodally distributed with a high peak at zero, whereas net activity under RW is distributed in roughly Gaussian fashion (see Supporting Information Figure S2).

Figure 5. 

Population sparseness of activity as a function of load under IS (solid lines) and RW (dashed lines) models for mouse (circles), monkey1 (squares), and monkey2 (triangles) networks. As with net activity, IS models show relatively constant sparseness of activity across load, with greater sparseness of activity (closer to zero) than RW models at all loads except 1 message per time step. Sparseness in RW decreases by more than a factor of 2. Shaded areas represent 2 standard deviations of variability from the mean of population sparseness across simulations. Data points show significant differences (t test, p < 0.01 ) for corresponding IS and RW measurements at all loads.

Figure 5. 

Population sparseness of activity as a function of load under IS (solid lines) and RW (dashed lines) models for mouse (circles), monkey1 (squares), and monkey2 (triangles) networks. As with net activity, IS models show relatively constant sparseness of activity across load, with greater sparseness of activity (closer to zero) than RW models at all loads except 1 message per time step. Sparseness in RW decreases by more than a factor of 2. Shaded areas represent 2 standard deviations of variability from the mean of population sparseness across simulations. Data points show significant differences (t test, p < 0.01 ) for corresponding IS and RW measurements at all loads.

Close modal

Results of comparison to randomized networks (Figure 6) in monkey2 and the thresholded mouse show that IS models have lower net activity and greater sparseness in the empirical network at almost all loads tested. RW models on empirical networks generally show very small decreases (monkey2) or slight increases in activity compared to randomized networks (thresholded mouse).

Figure 6. 

(A) Net activity and (B) population sparseness for empirical and randomized networks of the monkey2 and the (thresholded) mouse under IS and RW models. In IS models, both the empirical monkey2 network (triangles) and the (thresholded) mouse network (circles) show lower activity and greater sparseness (showing a difference in percentage terms of between 2–7%) compared to corresponding randomized networks. Randomized RW models show largely the same behavior as the empirical networks (differing by less than 1%) in terms of activity and sparseness. Filled symbols indicate significant differences (t test, p < 0.01) between empirical and corresponding randomized networks, whereas open symbols indicate no significant difference.

Figure 6. 

(A) Net activity and (B) population sparseness for empirical and randomized networks of the monkey2 and the (thresholded) mouse under IS and RW models. In IS models, both the empirical monkey2 network (triangles) and the (thresholded) mouse network (circles) show lower activity and greater sparseness (showing a difference in percentage terms of between 2–7%) compared to corresponding randomized networks. Randomized RW models show largely the same behavior as the empirical networks (differing by less than 1%) in terms of activity and sparseness. Filled symbols indicate significant differences (t test, p < 0.01) between empirical and corresponding randomized networks, whereas open symbols indicate no significant difference.

Close modal

More specifically, empirical monkey2 and thresholded mouse networks under IS differ significantly (t test, p < 0.01 ) at all loads compared to corresponding randomized networks. Under IS, the mean net activity of the empirical networks across loads is lower by 0.013 (monkey2) and 0.0039 (thresholded mouse) activity units compared to corresponding randomized networks. In percentage terms, this equates to a difference in activity between empirical and randomized networks of 7.3% and 2.3%, respectively. The mean sparseness of the empirical network across loads under IS is greater (closer to zero) by 0.026 (monkey2) and 0.0092 (thresholded mouse) units compared to corresponding randomized networks. This equates to a difference in sparseness of 5.6% and 1.7%, respectively. Empirical monkey2 and thresholded mouse networks under RW also differ significantly (t test, p < 0.01 ) at all but 2 load values compared to corresponding randomized networks. However, compared to corresponding randomized networks, the mean net activity of the empirical networks across loads under RW is marginally lower by 0.00074 (monkey2) or greater by 0.0027 (thresholded mouse) activity units. This equates to a difference in net activity of 0.30% and 0.93%, respectively. Likewise, the mean sparseness of the empirical network across loads under RW is marginally greater by 0.0013 (monkey2) or lower (closer to one) by 0.0012 (mouse) units compared to corresponding randomized networks. This equates to a difference in sparseness of 0.25% and 0.027%, respectively.

We note that the same pattern of results also obtains on the unthresholded mouse network. Although the differences between empirical and randomized networks are not large, they are uniformly significant. This evidence is consistent with the notion that empirical network topology in mammals promotes sparse, efficient activity under an information spreading routing strategy.

In the context of collision interactions among signals, we used numerical simulations to show how the mammal brain can exploit mesoscale network structure by spreading information widely, and in particular by sending multiple copies of an incoming signal. Under an information spreading (IS) strategy for routing, the network achieves globally low net activity and high sparseness of activity like what is found in real mammal brains. Net activity and sparseness of activity change relatively little as load increases under the IS model. This result suggests a way that the brain could achieve ongoing sparse activity spread across the entire system using local protocol, but without deviating too far from globally homeostatic operation. In contrast, random walk (RW) models with the same collision rule produce substantially higher summed activity. In RW, substantial numbers of nodes are active in a given time window, resulting in less sparse activity. Net activity and sparseness of activity under RW change substantially as load increases over the same range, with activity increasing more than fourfold, and sparseness decreasing by around a factor of two. Our findings hold in both the monkey cortex and mouse whole brain. In addition, empirical networks under IS models show lower activity and greater sparseness of activity than randomized networks with the same degree distributions.

Together, these results suggest that nonlinear collision protocol and mammal connectome topology could promote sparse, efficient routing of communication traffic across the mammal brain network. This protocol is biologically plausible, and does not require centralized control.

However, one can ask: Are RW and IS models really comparable to each other? It is certainly the case that the actual number of messages circulating on the network is different under RW and IS models. For example, a single message injected as load will—on the next time step, and in the absence of incoming messages—lead to only one outgoing message under RW, while in IS it will lead to k outgoing messages (where k is out-degree) under these circumstances.

Yet such nonlinear interactions are precisely our object of study. Rather than seeing RW models as a control relative to IS models, the two models should be seen as extreme ends of a spectrum of diffusion-like routing strategies with distributed protocol. This spectrum is orthogonal to the spectrum of pathfinding strategies that span from shortest paths to random walks (see Avena-Koenigsberger et al., 2019). At one end of the spectrum of routing strategies, we have the random walk approach that is employed extensively to model diffusion-like processes with a conservation law. At the other end of the spectrum are models that generate spreading dynamics (e.g., those used to model infectious disease).

Random walk models are in any case very common in network neuroscience, so their collision dynamics are of inherent interest to the field. We have shown that RW models have clear limitations in terms of achieving sparse, low, and relatively uniform activity across loads given simple collision rules.

Taking a wider perspective, the management of signals arising from different sources that impinge on the same target is a critical problem in network neuroscience. In the mesoscale mammal connectome, where all nodes are only a few hops from each other (and where signals are excitatory), the system would seem to need a strategy for managing coincident excitations in order to route signals appropriately.

This conception stands in contrast to models of dynamics of the mesoscale connectome (particularly cortex) inspired by integrate-and-fire spiking neuron models, wherein the goal of the system is to have many excitatory signals arrive at the same time in order for the unit to be highly activated. Collisions—in the sense of multiple coincident efferents—are seen as signals to be integrated. However, studies such as Kaiser et al. (2007) and Kaiser and Hilgetag (2010) that approach dynamics from this perspective confirm the necessity of keeping global activity at a low and relatively constant level. Kaiser and colleagues tested a threshold routing model on surrogate hierarchical networks with biologically plausible parameters, and suggested that brain networks must seek regimes of “limited sustained activity.” Using a “spreading” model that is somewhat similar to our IS model, their simulations indicated that specific parameters of activity likelihood promoted limited sustained activity of around 10–20% of nodes. However, many if not most parameter settings led to “dying-out” activity or overload (Kaiser et al., 2007; Kaiser & Hilgetag, 2010). We have shown here that limited sustained activity of around 10–20% of nodes is robustly achieved on mammal connectomes under a nonlinear routing approach employing destructive collisions and information spreading. However, in contrast to Kaiser and Hilgetag (2010), our model shows that this pattern of activity is distributed across the entire network over time, rather than concentrated in a minority of nodes, and this behavior is demonstrated on all simulation runs.

In separate tests, we found that the structure of mammal brain networks is such that collisions that sum or threshold signals lead to very high activity and very low sparseness in both RW and IS models. This was suggested by tests of a “let one pass” collision rule whereby exactly one message in a collision of m messages is allowed to pass on a randomly chosen outgoing edge. Even at low load values, the network quickly overloads (especially under IS) with this collision rule producing data that are not meaningful, and therefore not reported here.

Our results show that strategies with nonlinear, self-organizing mechanisms that align with basic physiology can generate network activity that offsets collisions, requires no central agent, operates sparsely in time and over nodes, and maintains relatively uniform global activity despite large increases in signal load. However, automatic deletion is not the only or necessarily the best solution for managing collisions. There are many other possible strategies and combinations of strategies that could be at play, including buffering (Mišić et al., 2014a), resending (Graham & Rockmore, 2011), and deflection routing (Fornito et al., 2016).

We find some evidence for one alternative strategy, which is to reduce load almost to zero. In this regime, redundancy may not be needed because collisions are rare, and a given path is unlikely to be congested. This is suggested in the finding that RW models achieve lower net activity and higher sparseness of activity than IS models at the lowest load (1 message per time step) in all cases. It remains possible that the brain could adopt this kind of extremely sparse strategy (see, e.g., Ovsepian, 2015). However, such a system, being so close to floor values, would be restricted to low activity ranges and would be vulnerable to large changes in dynamics if load varies more than a small amount. In any case, if collisions are indeed very rare, this is also a major problem for models based on integrate-and-fire strategies and linear approaches.

We note that redundancy in the brain such as correlated firing among neighboring neurons (see, e.g., Meytlis, Nichols & Nirenberg, 2012) is typically thought to be something that the system should minimize (Wainwright, 1999; Barlow, 2001). Following Shannon’s theory of information, parallel communication channels should seek to reduce mutual information to theoretical minima. However, the brain’s network topology is not parallel but rather small world–like (see, e.g., Achard et al., 2006; see also Knoblauch et al., 2016). In this situation, there is no straightforward way to apply Shannon’s theory for arbitrary node-to-node communication (El Gamal & Kim, 2011). Our results suggest that redundancy of signals may allow brain networks to strike a dynamic balance between signal creation and destruction.

However, copying all incoming signals has its drawbacks. More realistic solutions may be found between the extreme ends of the spectrum of routing strategies. For example, it may be more efficient to send redundant signals only on some fraction of outgoing edges. Tests exploring this parameter are underway; we predict a smooth transition in dynamics between the observed extremes of RW and IS conditions. In any case, the purpose of the present investigation is not to find the optimal solution for intrabrain communication but rather to address the question, how could a mammal brain efficiently route signals, given the likelihood of collisions, the need for distributed protocol, and a demand for low, sparse activity levels? Much as vision scientists have gained knowledge from asking what an efficient retinal code would be given the statistical regularities of the environment (e.g., Field, 1994; Graham & Field, 2009), we aim to advance network model building by asking what an efficient whole-brain routing protocol would be given its architecture and operational demands.

We note that collisions cannot be ignored by assuming that signals arriving at a node at the same time do not interact. Of course, at the mesoscale, inputs arising in different nodes that travel to a target node do not necessarily synapse with the same population within the target node. But if there were dedicated links from sending nodes via a target node to destinations one synapse away, this would imply a different network topology (i.e., the target node would not be considered a node). If the brain performs dynamic routing at nodes—and functional demands appear to necessitate this (Olshausen et al., 1993; Graham & Rockmore, 2011)—collisions must be managed.

Nevertheless, the brain probably does not annihilate all colliding signals at all nodes. The rules governing collisions may vary in time and space from those that are fully punitive for coincident signals to those that require greater coincidence. This is a limitation of our study. Collision rules are a dimension worth exploring but as an initial study, we adopted a simplified model that promotes interpretability.

The difficulty of studying nonlinear interactions of brain network communication is due in part to the problem of simulating dynamics explicitly. The challenge lies in tracking every signal’s path and interactions. RW models and shortest path models admit analytical solutions that capture the likelihood of a set of states of the network. But this is generally possible only when collisions and other nonlinear interactions are ignored. As a result, explicit models capable of accounting for collisions have been relatively little studied to date (see Graham & Hao, 2018). With further study of explicit models we may be able to understand fine-grained activity patterns such as regional differences and correlations in activity. However, though we were able to model signal interactions explicitly, the lack of tracking of messages is a limitation of the present study. For example, we would like to know how many messages are delivered, and how quickly. We have identified a way to infer message “age” in our simulations and will present results of this work in a subsequent report.

This study has other limitations. One is the idiosyncrasies of the connectomes studied. For example, the CoCoMac network is known to be incomplete (Chen et al., 2020), and the data Markov et al. (2014) are limited by retrograde tracing procedures. In addition, we have not constrained or tailored activity in terms of functional demands.

Finally, we are limited in terms of the interpretation of our network randomization results. As noted earlier, randomizing networks in which nodes have a degree that is a significant fraction of the number of nodes leads to random networks that share edges with their empirical counterparts. However, in preliminary studies of the monkey2 network, we found that a non-degree-preserving random network (an Erdõs–Rényi network with the same number of nodes and edges, but where no edges are shared between empirical and randomized networks) shows significantly higher activity and lower sparseness compared to empirical networks under both IS and RW. In fact, the difference between empirical and randomized networks is somewhat more pronounced compared to the degree-preserving case, indicating that degree distributions do play a role in generating low, sparse activity, but they do not fully explain this effect.

We conclude that the mammal brain network appears capable of striking a balance between activity and collisions such that sparse activity is robustly maintained across a wide range of loads. At present we cannot say what properties of network structure are necessary for supporting this behavior. Indeed, it remains unknown whether the “creative destruction” we observe here is unique to brains or constitutes a more widespread behavior of complex networks. Nevertheless, we argue that some strategy for managing collisions is necessary, and whatever solution the mammal brain has adopted, it must generate low, stable net activity that is sparsely distributed across the network.

Supporting information for this article is available at https://doi.org/10.1162/netn_a_00165. Matlab code for performing simulations under IS and RW models is posted on GitHub: https://github.com/saltiie/collisions-and-redundancy (Hao & Graham, 2020).

YH and DG conceptualized the study, designed the framework, performed analyses, interpreted the results, and wrote the manuscript.

We thank Barbara Finlay, David Field, Marc-Thorsten Hütt, and Ao Tang for helpful discussions.

Collision:

An event in which signals on two or more of a node’s incoming edges arrive at the same time.

Sparseness (of activity):

A characteristic distribution displaying low average activity ratios across time or across units.

Population sparseness:

A measure of the degree of sparseness of activity across units during a given time window.

Lifetime sparseness:

A measure of the degree of sparseness of activity of a given unit over time.

Information spreading:

A model of network activity where incoming signals at a node are copied to all outgoing nodes.

Random walk:

A model system where activity is passed by “walkers” who move at random from node to node.

Randomized network:

A network whose edges are randomized with respect to an original network, but that preserves the in- and out-degree distributions of the nodes.

Load:

Number of new messages injected in a time step at randomly chosen nodes.

Attempted activity:

Network activity before destructive collisions are accounted for.

Actual (net) activity:

Network activity after destructive collisions are accounted for.

Treves–Rolls sparseness measure:

A measure of the sparseness of a distribution with 0 corresponding to maximal sparseness and 1 corresponding to minimal sparseness.

Routing protocol:

The rules or strategy employed by nodes to direct the interchange of signals across a network.

Abdelnour
,
F.
,
Voss
,
H. U.
, &
Raj
,
A.
(
2014
).
Network diffusion accurately models the relationship between structural and functional brain connectivity networks
.
NeuroImage
,
90
,
335
347
. DOI: https://doi.org/10.1016/j.neuroimage.2013.12.039, PMID: 24384152, PMCID: PMC3951650
Achard
,
S.
,
Salvador
,
R.
,
Whitcher
,
B.
,
Suckling
,
J.
, &
Bullmore
,
E. D.
(
2006
).
A resilient, low-frequency, small-world human brain functional network with highly connected association cortical hubs
.
Journal of Neuroscience
,
26
(
1
),
63
72
. DOI: https://doi.org/10.1523/JNEUROSCI.3874-05.2006, PMID: 16399673, PMCID: PMC6674299
Aeschbach
,
D.
,
Matthews
,
J. R.
,
Postolache
,
T. T.
,
Jackson
,
M. A.
,
Giesen
,
H. A.
, &
Wehr
,
T. A.
(
1997
).
Dynamics of the human EEG during prolonged wakefulness: evidence for frequency- specific circadian and homeostatic influences
.
Neuroscience Letters
,
239
(
2–3
),
121
124
. DOI: https://doi.org/10.1016/S0304-3940(97)00904-X
Attwell
,
D.
, &
Laughlin
,
S. B.
(
2001
).
An energy budget for signaling in the grey matter of the brain
.
Journal of Cerebral Blood Flow and Metabolism
,
21
(
10
),
1133
2114
. DOI: https://doi.org/10.1097/00004647-200110000-00001, PMID: 11598490
Avena-Koenigsberger
,
A.
,
Mišić
,
B.
,
Hawkins
,
R. X.
,
Griffa
,
A.
,
Hagmann
,
P.
,
Goñi
,
J.
, &
Sporns
,
O.
(
2017
).
Path ensembles and a tradeoff between communication efficiency and resilience in the human connectome
.
Brain Structure and Function
,
222
(
1
),
603
618
. DOI: https://doi.org/10.1007/s00429-016-1238-5, PMID: 27334341
Avena-Koenigsberger
,
A.
,
Mišić
,
B.
, &
Sporns
,
O.
(
2018
).
Communication dynamics in complex brain networks
.
Nature Reviews Neuroscience
,
19
(
1
),
17
. DOI: https://doi.org/10.1038/nrn.2017.149, PMID: 29238085
Avena-Koenigsberger
,
A.
,
Yan
,
X.
,
Kolchinsky
,
A.
,
van den Heuvel
,
M.
,
Hagmann
,
P.
, &
Sporns
,
O.
(
2019
).
A spectrum of routing strategies for brain networks
.
PLoS Computational Biology
,
155
(
3
),
e1006833
. DOI: https://doi.org/10.1371/journal.pcbi.1006833, PMID: 30849087, PMCID: PMC6426276
Bakker
,
R.
,
Wachtler
,
T.
, &
Diesmann
,
M.
(
2012
).
CoComac 2.0 and the future of tract-tracing databases.
DOI: https://doi.org/10.3389/fninf.2012.00030, PMID: 23293600, PMCID: PMC3530798
Barlow
,
H.
(
2001
).
Redundancy reduction revisited
.
Network: Computation in Neural Systems
,
12
(
3
),
241
253
. DOI: https://doi.org/10.1080/net.12.3.241.253, PMID: 11563528
Bell
,
A. J.
, &
Sejnowski
,
T. J.
(
1997
).
The ’independent components’ of natural scenes are edge filters
.
Vision Research
,
37
,
3327
3338
. DOI: https://doi.org/10.1016/S0042-6989(97)00121-1
Bettinardi
,
R. G.
,
Deco
,
G.
,
Karlaftis
,
V. M.
,
Van Hartevelt
,
T. J.
,
Fernandes
,
H. M.
,
Kourtzi
,
Z.
, …
Zamora-López
,
G.
(
2017
).
How structure sculpts function: Unveiling the contribution of anatomical connectivity to the brain’s spontaneous correlation structure
.
Chaos: An Interdisciplinary Journal of Nonlinear Science
,
27
(
4
),
047409
. DOI: https://doi.org/10.1063/1.4980099, PMID: 28456160
Bullmore
,
E.
, &
Sporns
,
O.
(
2009
).
Complex brain networks: Graph theoretical analysis of structural and functional systems
.
Nature Reviews Neuroscience
,
10
(
3
),
18 6
. DOI: https://doi.org/10.1038/nrn2575, PMID: 19190637
Bullmore
,
E.
, &
Sporns
,
O.
(
2012
).
The economy of brain network organization
.
Nature Reviews Neuroscience
,
13
(
5
),
336
349
. DOI: https://doi.org/10.1038/nrn3214, PMID: 22498897
Chen
,
Y.
,
Zhang
,
Z. K.
,
He
,
Y.
, &
Zhou
,
C.
(
2020
).
A large-scale high-density weighted structural connectome of the macaque brain acquired by predicting missing links
.
Cerebral Cortex
. DOI: https://doi.org/10.1093/cercor/bhaa060, PMID: 32313935, PMCID:PMC7391281
El Gamal
,
A.
, &
Kim
,
Y. H.
(
2011
).
Network information theory
.
Cambridge, UK
:
Cambridge University Press
. DOI: https://doi.org/10.1017/CBO9781139030687
Field
,
D. J.
(
1994
).
What is the goal of sensory coding?
Neural Computation
,
6
(
4
),
559
601
. DOI: https://doi.org/10.1162/neco.1994.6.4.559
Földiák
,
P.
(
2002
).
Sparse coding in the primate cortex
. In
M. A.
Arbib
(Ed.),
The handbook of brain theory and neural networks
.
Cambridge, MA
:
MIT Press
.
Fornito
,
A.
,
Zalesky
,
A.
, &
Bullmore
,
E.
(
2016
).
Fundamentals of brain network analysis
.
London
:
Academic Press
.
Friston
,
K. J.
(
2011
).
Functional and effective connectivity: A review
.
Brain Connectivity
,
1
(
1
),
13- -36
. DOI: https://doi.org/10.1089/brain.2011.0008, PMID: 22432952
Funahashi
,
S.
(
2015
).
Functions of delay-period activity in the prefrontal cortex and mnemonic scotomas revisited
.
Frontiers in Systems Neuroscience
,
9
,
2
. DOI: https://doi.org/10.3389/fnsys.2015.00002, PMID: 25698942, PMCID: PMC4318271
Gămănuţ
,
R.
,
Kennedy
,
H.
,
Toroczkai
,
Z.
,
Ercsey-Ravasz
,
M.
,
Van Essen
,
D. C.
,
Knoblauch
,
K.
, &
Burkhalter
,
A.
(
2018
).
The mouse cortical connectome, characterized by an ultra-dense cortical graph, maintains specificity by distinct connectivity profiles
.
Neuron
,
97
(
3
),
698
715
. DOI: https://doi.org/10.1016/j.neuron.2017.12.037, PMID: 29420935, PMCID: PMC5958229
Gidon
,
A.
,
Zolnik
,
T. A.
,
Fidzinski
,
P.
,
Bolduan
,
F.
,
Papoutsi
,
A.
,
Poirazi
,
P.
,
Larkum
,
M. E.
… (
2020
).
Dendritic action potentials and computation in human layer 2/3 cortical neurons
.
Science
,
367
(
6473
),
83
87
. DOI: https://doi.org/10.1126/science.aax6239, PMID: 31896716
Goldman-Rakic
,
P. S.
(
1996
).
Regional and cellular fractionation of working memory
.
Proceedings of the National Academy of Sciences
,
93
(
24
),
13473
13480
. DOI: https://doi.org/10.1073/pnas.93.24.13473, PMID: 8942959PMCID: PMC33633
Gollisch
,
T.
, &
Meister
,
M.
(
2010
).
Eye smarter than scientists believed: Neural computations in circuits of the retina
.
Neuron
,
65
(
2
),
150
164
. DOI: https://doi.org/10.1016/j.neuron.2009.12.009, PMID: 20152123, PMCID: PMC3717333
Goñi
,
J.
,
Avena-Koenigsberger
,
A.
,
de Mendizabal
,
N. V.
,
van den Heuvel
,
M. P.
,
Betzel
,
R. F.
,
Sporns
,
O.
(
2013
).
Exploring the morphospace of communication efficiency in complex networks
.
PLoS One
,
8
(
3
). DOI: https://doi.org/10.1371/journal.pone.0058070, PMID: 23505455, PMCID: PMC3591454
Goñi
,
J.
,
van den Heuvel
,
M. P.
,
Avena-Koenigsberger
,
A.
,
de Mendizabal
,
N. V.
,
Betzel
,
R. F.
,
Griffa
,
A.
, &
Sporns
,
O.
(
2014
).
Resting-brain functional connectivity predicted by analytic measures of network communication
.
Proceedings of the National Academy of Sciences
,
111
(
2
),
833
838
. DOI: https://doi.org/10.1073/pnas.1315529111, PMID: 24379387, PMCID: PMC3896172
Graham
,
D. J.
(
2014
).
Routing in the brain
.
Frontiers in Computational Neuroscience
,
8
,
44
. DOI: https://doi.org/10.3389/fncom.2014.00044, PMID: 24782752, PMCID: PMC3989587
Graham
,
D. J.
(
2017
).
Commentary: Building brains that communicate like machines
.
Behavioral and Brain Sciences
,
e266
,
37
38
. DOI: https://doi.org/10.1017/S0140525X17000152, PMID: 29342700
Graham
,
D. J.
, &
Field
,
D. J.
(
2006
).
Sparse coding in the neocortex
. In
J. H.
Kaas
and
L. A.
Krubitzer
(Eds.),
Evolution of nervous systems
(Vol.
III
, pp.
181
187
).
Oxford, UK
:
Elsevier
. DOI: https://doi.org/10.1016/B0-12-370878-8/00064-1
Graham
,
D. J.
, &
Field
,
D. J.
(
2009
).
Natural images: Coding efficiency
. In
L. R.
Squire
(Ed.),
Encyclopedia of neuroscience
(Vol.
VI
, pp.
19
27
).
Oxford, UK
:
Academic Press
. DOI: https://doi.org/10.1016/B978-008045046-9.00212-6
Graham
,
D. J.
, &
Hao
,
Y.
(
2018
).
A selective diffusion model of brain network activity
.
Proceedings of the Conference on Cognitive Computational Neuroscience
,
1195
. https://ccneuro.org/2018/proceedings/1195.pdf, DOI: https://doi.org/10.32470/CCN.2018.1195-0
Graham
,
D. J.
, &
Rockmore
,
D. N.
(
2011
).
The packet switching brain
.
Journal of Cognitive Neuroscience
,
23
(
2
),
26 7–3276
. DOI: https://doi.org/10.1162/jocn.2010.21477, PMID: 20350173
Hahn
,
G.
,
Ponce-Alvarez
,
A.
,
Deco
,
G.
,
Aertsen
,
A.
, &
Kumar
,
A.
(
2019
).
Portraits of communication in neuronal networks
.
Nature Reviews Neuroscience
,
20
(
2
),
117
127
. DOI: https://doi.org/10.1038/s41583-018-0094-0, PMID: 30552403
Hancock
,
J. T.
(
2017
).
Cell signalling
.
Oxford, UK
:
Oxford University Press
.
Hao
,
Y.
, &
Graham
,
D.
(
2020
).
Collisions and redundancy
, GitHub, https://github.com/saltiie/collisions-and-redundancy
Kaiser
,
M.
,
Görner
,
M.
, &
Hilgetag
,
C. C.
(
2007
).
Functional criticality in clustered networks without inhibition
.
New Journal of Physics
,
9
,
110
. DOI: https://doi.org/10.1088/1367-2630/9/5/110
Kaiser
,
M.
, &
Hilgetag
,
C. C.
(
2010
).
Optimal hierarchical modular topologies for producing limited sustained activation of neural networks
.
Frontiers in Neuroinformatics
,
4
,
8
. DOI: https://doi.org/10.3389/fninf.2010.00008, PMID: 20514144, PMCID: PMC2876872
Kleinrock
,
L.
(
1976
).
Queueing systems
In
Computer applications
(
Vol. II
).
New York, NY
:
Wiley
.
Knoblauch
,
K.
,
Ercsey-Ravasz
,
M.
,
Kennedy
,
H.
, &
Toroczkai
,
Z.
(
2016
).
The brain in space
. In
Micro-, meso-and macro-connectomics of the brain
(pp.
45
74
).
Cham, Switzerland
:
Springer
. DOI: https://doi.org/10.1007/978-3-319-27777-6_5, PMID: 28590672
Kopell
,
N. J.
,
Gritton
,
H. J.
,
Whittington
,
M. A.
, &
Kramer
,
M. A.
(
2014
).
Beyond the connectome: The dynome
.
Neuron
,
83
(
6
),
1319
1328
. DOI: https://doi.org/10.1016/j.neuron.2014.08.016, PMID: 25233314, PMCID: PMC4169213
Lennie
,
P.
(
2003
).
The cost of cortical computation
.
Current Biology
,
13
(
6
),
493
497
. DOI: https://doi.org/10.1016/S0960-9822(03)00135-0
Levy
,
W. B.
, &
Baxter
,
R. A.
(
1996
).
Energy efficient neural codes
.
Neural Computation
,
8
(
3
),
531
543
. DOI: https://doi.org/10.1162/neco.1996.8.3.531, PMID: 8868566
Markov
,
N. T.
,
Ercsey-Ravasz
,
M. M.
,
Ribeiro Gomes
,
A. R.
,
Lamy
,
C.
,
Magrou
,
L.
,
Vezoli
,
J.
,
Misery
,
P.
,
Falchier
,
A.
,
Quilodran
,
R.
,
Gariel
,
M. A.
,
Sallet
,
J.
,
Gamanut
,
R.
,
Huissoud
,
C.
,
Clavagnier
,
S.
,
Giroud
,
P.
,
Sappey-Marinier
,
D.
,
Barone
,
P.
,
Dehay
,
C.
,
Toroczkai
,
Z.
,
Kennedy
,
H.
(
2014
).
A Weighted and Directed Interareal Connectivity Matrix for Macaque Cerebral Cortex
.
Cerebral Cortex
,
24
(
1
),
17
36
. DOI: https://doi.org/10.1093/cercor/bhs270
Maslov
,
S.
, &
Sneppen
,
K.
(
2002
).
Specificity and stability in topology of protein networks
.
Science
,
296
,
910
913
. DOI: https://doi.org/10.1126/science.1065103, PMID: 11988575
Meytlis
,
M.
,
Nichols
,
Z.
, &
Nirenberg
,
S.
(
2012
).
Determining the role of correlated firing in large populations of neurons using white noise and natural scene stimuli
.
Vision Research
,
70
,
44
53
. DOI: https://doi.org/10.1016/j.visres.2012.07.007, PMID: 22885035, PMCID: PMC3980944
Min
,
B.
, &
Castellano
,
C.
(
2020
).
Message-passing theory for cooperative epidemics
.
Chaos: An Interdisciplinary Journal of Nonlinear Science
,
30
(
2
),
023131
. DOI: https://doi.org/10.1063/1.5140813, PMID: 32113239
Mišić
,
B.
,
Sporns
,
O.
, &
McIntosh
,
A. R.
(
2014a
).
Communication efficiency and congestion of signal traffic in large-scale brain networks
.
PLoS Computational Biology
,
10
(
1
),
e1003427
. DOI: https://doi.org/10.1371/journal.pcbi.1003427, PMID: 24415931, PMCID: PMC3886893
Mišić
,
B.
,
Goñi
,
J.
,
Betzel
,
R. F.
,
Sporns
,
O.
, &
McIntosh
,
A. R.
(
2014b
).
A network convergence zone in the hippocampus
.
PLoS Computational Biology
,
10
(
12
),
e1003982
. DOI: https://doi.org/10.1371/journal.pcbi.1003982, PMID: 25474350, PMCID: PMC4256084
Mišić
,
B.
,
Betzel
,
R. F.
,
Nematzadeh
,
A.
,
Goi
,
J.
,
Griffa
,
A.
,
Hagmann
,
P.
, &
Sporns
,
O.
(
2015
).
Cooperative and competitive spreading dynamics on the human connectome
.
Neuron
,
86
(
6
),
1518
1529
. DOI: https://doi.org/10.1016/j.neuron.2015.05.035, PMID: 26087168
Mišić
,
V. B.
&
Mišić
,
J.
(
2014
).
Machine-to-machine communications: Architectures, technology, standards, and applications
.
Boca Raton, FL
:
CRC Press
. DOI: https://doi.org/10.1201/b17141
Navlakha
,
S.
,
Barth
,
A. L.
, &
Bar-Joseph
,
Z.
(
2015
).
Decreasing-rate pruning optimizes the construction of efficient and robust distributed networks
.
PLoS Computational Biology
,
11
(
7
),
e1004347
. DOI: https://doi.org/10.1371/journal.pcbi.1004347, PMID: 26217933, PMCID: PMC4517947
Navlakha
,
S.
,
Bar-Joseph
,
Z.
, &
Barth
,
A. L.
(
2017
).
Network design and the brain
.
Trends in Cognitive Sciences
,
22
,
64
78
. DOI: https://doi.org/10.1016/j.tics.2017.09.012, PMID: 29054336
Noh
,
J. D.
, &
Rieger
,
H.
(
2004
).
Random walks on complex networks
.
Physical Review Letters
,
92
(
11
),
118701
. DOI: https://doi.org/10.1103/PhysRevLett.92.118701, PMID: 15089179
Oh
,
S. W.
,
Harris
,
J. A.
,
Ng
,
L.
,
Winslow
,
B.
,
Cain
,
N.
,
Mihalas
,
S.
,
Wang
,
Q.
,
Lau
,
C.
,
Kuan
,
L.
,
Henry
,
A. M.
,
Mortrud
,
M. T.
,
Ouellette
,
B.
,
Nguyen
,
T. N.
,
Sorensen
,
S. A.
,
Slaughterbeck
,
C. R.
,
Wakeman
,
W.
,
Li
,
Y.
,
Feng
,
D.
,
Ho
,
A.
,
Nicholas
,
E.
, …
Zeng
,
H.
(
2014
).
A mesoscale connectome of the mouse brain
.
Nature
,
508
,
207
214
. DOI: https://doi.org/10.1038/nature13186
Olshausen
,
B. A.
,
Anderson
,
C. H.
, &
Van Essen
,
D. C.
(
1993
).
A neurobiological model of visual attention and invariant pattern recognition based on dynamic routing of information
.
Journal of Neuroscience
,
13
(
11
),
4700
4719
. DOI: https://doi.org/10.1523/JNEUROSCI.13-11-04700.1993, PMID: 8229193, PMCID: PMC6576339
Olshausen
,
B. A.
, &
Field
,
D. J.
(
1996
).
Emergence of simple-cell receptive field properties by learning a sparse code for natural images
.
Nature
,
381
(
6583
),
607
. DOI: https://doi.org/10.1038/381607a0, PMID: 8637596
Ovsepian
,
S. V.
(
2019
).
The dark matter of the brain
.
Brain Structure and Function
,
224
(
3
),
973
983
. DOI: https://doi.org/10.1007/s00429-019-01835-7, PMID: 30659350
Poggio
,
T.
(
1984
).
Routing thoughts. Massachusetts Institute of Technology Artificial Intelligence Laboratory Working Paper
,
258
.
Poldrack
,
R. A.
(
2015
).
Is “efficiency” a useful concept in cognitive neuroscience?
Developmental Cognitive Neuroscience
,
11
,
12
17
. DOI: https://doi.org/10.1016/j.dcn.2014.06.001, PMID: 24981045, PMCID: PMC6989750
Rolls
,
E. T.
, &
Tovee
,
M. J.
(
1995
).
Sparseness of the neuronal representation of stimuli in the primate temporal visual cortex
.
Journal of Neurophysiology
,
73
,
713
726
. DOI: https://doi.org/10.1152/jn.1995.73.2.713, PMID: 7760130
Rubinov
,
M.
, &
Sporns
,
O.
(
2010
).
Complex network measures of brain connectivity: Uses and interpretations
.
NeuroImage
,
52
(
3
),
1059
1069
. DOI: https://doi.org/10.1016/j.neuroimage.2009.10.003, PMID: 19819337
Sardi
,
S.
,
Vardi
,
R.
,
Sheinin
,
A.
,
Goldental
,
A.
, &
Kanter
,
I.
(
2017
).
New types of experiments reveal that a neuron functions as multiple independent threshold units
.
Scientific Reports
,
7
(
1
),
1
17
. DOI: https://doi.org/10.1038/s41598-017-18363-1, PMID: 29269849, PMCID: PMC5740076
Schruben
,
L. W.
, &
Margolin
,
B. H.
(
1978
).
Pseudorandom number assignment in statistically designed simulation and distribution sampling experiments
.
Journal of the American Statistical Association
,
73
(
363
),
504
520
. DOI: https://doi.org/10.1080/01621459.1978.10480044, https://doi.org/10.2307/2286595
Scott
,
A.
(
1977
).
Neurophysics
.
New York, NY
:
John Wiley and Sons
.
Seguin
,
C.
,
Razi
,
A.
, &
Zalesky
,
A.
(
2019
).
Inferring neural signalling directionality from undirected structural connectomes
.
Nature Communications
,
10
(
1
),
1
13
. DOI: https://doi.org/10.1038/s41467-019-12201-w, PMID: 31537787, PMCID: PMC6753104
Sosnovich
,
A.
,
Grumberg
,
O.
, &
Nakibly
,
G.
(
2017
).
Formal black-box analysis of routing protocol implementations
. arXiv preprint
arXiv:1709.08096
.
Steriade
,
M.
, &
Paré
,
D.
(
2007
).
Gating in cerebral networks
.
Cambridge, UK
:
Cambridge University Press
. DOI: https://doi.org/10.1017/CBO9780511541735
Tipnis
,
U.
,
Amico
,
E.
,
Ventresca
,
M.
, &
Goni
,
J.
(
2018
).
Modeling communication processes in the human connectome through cooperative learning
.
IEEE Transactions on Network Science and Engineering
,
7
,
476
488
. DOI: https://doi.org/10.1109/TNSE.2018.2878487
Treves
,
A.
, &
Rolls
,
E. T.
(
1991
).
What determines the capacity of autoassociative memories in the brain?
Network: Computation in Neural Systems
,
2
,
371
397
. DOI: https://doi.org/10.1088/0954-898X_2_4_004
Turrigiano
,
G. G.
(
1999
).
Homeostatic plasticity in neuronal networks: The more things change, the more they stay the same
.
Trends in Neurosciences
,
22
(
5
),
221
227
. DOI: https://doi.org/10.1016/S0166-2236(98)01341-1
Wainwright
,
M. J.
(
1999
).
Visual adaptation as optimal information transmission
.
Vision Research
,
39
(
23
),
3960
3974
. DOI: https://doi.org/10.1016/S0042-6989(99)00101-7
Waxman
,
S. G.
(
1972
).
Regional differentiation of the axon: A review with special reference to the concept of the multiplex neuron
.
Brain Research
,
47
(
2
),
269
288
. DOI: https://doi.org/10.1016/0006-8993(72)90639-7
Willmore
,
B.
, &
Tolhurst
,
D. J.
(
2001
).
Characterizing the sparseness of neural codes
.
Network: Computation in Neural Systems
,
12
(
3
),
255
270
. DOI: https://doi.org/10.1080/net.12.3.255.270, PMID: 11563529

Author notes

Competing Interests: The authors have declared that no competing interests exist.

Handling Editor: Olaf Sporns

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/legalcode.

Supplementary data