netn_a_00246
netn_a_00246
netn_a_00246
DOI:
https://doi.org/10.1162/netn_a_00246
AUTHOR SUMMARY
Received: 3 October 2021
Accepted: 11 March 2022 This research presents a novel framework to study functional connectivity on the largest yet
brain connectome of the fruit fly, revealing that synaptic excitation-inhibition (E/I) balance
Competing Interests: The authors have
declared that no competing interests characterizes the formation of dynamic brain connectivity as a percolation process. Various
exist.
remarkable properties of brain functions, such as information transmission efficiency, robust
Corresponding Authors: flexibility, and economy, emerge as byproducts of percolation transition. These advantages
Yang Tian
tiany20@mails.tsinghua.edu.cn can be simultaneously ensured at an optimal E/I balance 1:1, consistent with previous in vitro
Pei Sun
peisun@tsinghua.edu.cn
experimental predictions. Our work demonstrates percolation as a potential way to understand
the emergence of brain function characteristics through connectivity.
Handling Editor:
Alex Fornito
Copyright: © 2022
INTRODUCTION
Massachusetts Institute of Technology
Published under a Creative Commons To survive through the evolution, our brain should be efficient enough to process external
Attribution 4.0 International
(CC BY 4.0) license information, robust to accidental damages (e.g., lesions), and economic in energy using. Over
the last decades, this evolutionary inference has been corroborated by numerous neuroscience
studies. The brain has been discovered to support highly efficient information transmission
between neurons, circuits, and cortices, making it possible to promptly gather and distribute
external information (Amico et al., 2021; Avena-Koenigsberger, Misic, & Sporns, 2018;
Graham, Avena-Koenigsberger, & Misic, 2020; Mišić, Sporns, & McIntosh, 2014). Such infor-
mation transmission efficiency, manifested as the low time cost of communications between
neurons or high broadcasting capacity of information, is demonstrated to vary across different
topological attributes of brain connectivity (Avena-Koenigsberger et al., 2018). Meanwhile, the
brain is revealed to feature robust flexibility, a kind of capacity to tolerate the large-scale
destruction of neurons or synaptic connections (e.g., by lesions) (Aerts, Fias, Caeyenberghs,
& Marinazzo, 2016; Joyce, Hayasaka, & Laurienti, 2013; Kaiser, Martin, Andras, & Young,
2007) while maintaining robust brain functions (Achard, Salvador, Whitcher, Suckling, &
Till now, it remains unclear where these remarkable properties of brain originate from. The
close relationships between these properties and the brain network naturally lead to an emer-
gentism hypothesis that argues these properties may originate from specific characteristics of
brain connectivity. In the recent decades, abundant corollaries of this hypothesis have been
verified from different perspectives. For instance, shortest paths in brain connectivity are
inferred as a principal communication substrate in the brain according to the properties of
information transmission efficiency (Avena-Koenigsberger et al., 2018). Although having a
short average path is costly, real brain connectivity still possesses near-minimal path length
(Betzel et al., 2016; Bullmore & Sporns, 2012; Kaiser & Hilgetag, 2006; Rubinov, Ypma, Watson,
& Bullmore, 2015) in functional interactions (Goñi et al., 2014; Hermundstad et al., 2013) to
support efficient information transmission. Moreover, brain connectivity is inferred as scale-free
according to the robustness of scale-free networks (Albert, Jeong, & Barabási, 2000) and implied
as small-world by the near-minimal path length (Bullmore & Sporns, 2012). While mammalian
brains with scale-free connectivity are robust against random lesions, they are significantly vul-
nerable to hub-targeted attacks (Kaiser & Hilgetag, 2004; Kaiser et al., 2007). However, once the
connectivity topology approaches a small-world network while maintaining scale-free property
(e.g., the macroscopic human brain) (Achard et al., 2006; Alstott et al., 2009; Crossley et al.,
2014; Joyce et al., 2013), the brain becomes more resilient to hub-targeted attacks than a
comparable scale-free network and keeps equally robust to random attacks. Besides these
mentioned corollaries, many other verified corollaries could be found, demonstrating that
brain connectivity pattern critically shapes brain functions.
However, these corollaries alone are not sufficient for a complete demonstration of the
emergentism hypothesis. Key supporting evidence remains absent because it is technically
infeasible to capture and control the emergence of these properties in vitro or vivo, at least
in the near future. One challenge arises from the scarcity of technology to record multi-mode
Functional connectivity: (including both static and functional connectivity), fine-grained, and high-throughput connec-
The connectivity formed by dynamic tome data (Sporns, Tononi, & Kötter, 2005). Another challenge is lacking experimental
interactions among neurons. methods to modify the connectivity to control the emergence of these properties. Although
the theoretical study may be an alternative choice, current classic models in neuroscience are
either experiment driven and proposed for ex post facto analysis or simulation driven and
designed for imitating phenomena rather than explaining mechanisms. These models are inap-
plicable for an assumption-free analysis when there is no experiment for reference. The
absence of direct verification due to these challenges makes the validity of the hypothesis
questionable.
Here we discuss the possibility of a feasible and physically fundamental demonstration of
the emergentism hypothesis. These advantageous properties of brain functions are all relevant
to the benefits or costs of forming specific functional connectivity (interactions among neurons)
on the static connectivity (anatomical structure). Therefore, the emergence of these properties
will be detectable if we can formalize the evolution of functional connectivity patterns on static
Topology properties of static connectivity. Let us begin with the static or anatomical connectivity of
the fruit fly brain. The data is acquired from the open source brain connectome lately released by
FlyEM project (Xu et al., 2020). In the present study, neurons and synapses are considered only
when the cell bodies are positioned precisely (assigned with a 10-nm spatial resolution coordi-
nate). The selected data set, including 23,008 neurons, 4,967,364 synaptic connections (synaptic
clefts), and 635,761 pairs of directionally adjacent relations (two neurons are directionally adja-
cent if one synaptic connection comes out of one neuron and leads to another), supports us to
analyze static connectivity on the brain region or macroscopic scale (Figure 1A–B) and the cell
or microscopic scale (Figure 1C–D). Please see the Materials and Methods for data acquisition.
The macroscopic static connectivity is analyzed in terms of the potential input projections
or potential output projections that a brain region can receive from or cast to another brain
region. These projections are treated as potential because static connectivity may not be
equivalent to the functional one. In the analysis, we count these two kinds of projections
between brain regions, based on which the total potential input projections (TPIP, the
macroscopic in-degree) received by each region and the total potential output projections
(TPOP, the macroscopic out-degree) coming from each region can be measured (Figure 1A).
Please see the Materials and Methods and Figure 1B for variable descriptions.
In Figure 1B and Table 1, we analyze the power law distributions of TPIP and TPOP with a
maximum likelihood estimation approach (R. Virkar & Clauset, 2014; Y. Virkar & Clauset,
2014), suggesting that the macroscopic static connectivity is plausibly scale-free (power law
exponent α 2 (2, 3) is estimated with ideal goodness). More details of power law analysis can
be seen in Materials and Methods. Meanwhile, we verify the symmetry (e.g., with a balance
between input and output projections) of macroscopic static connectivity using the Pearson
correlations and the average change fractions (see Table 2). The connectivity is suggested as
symmetric since (1) there are significant positive correlations between TPIP and TPOP (e.g.,
larger than 0.9); (2) The average change fraction of TPIP compared with TPOP is sufficiently
small (e.g., smaller than 1). We also present other corroborative evidence derived from related
Note. Being scale-free requires P ∝ n−α, where α 2 (2, 3). Goodness of estimation is expected as to be less than 0.05.
Note. Strong symmetry requires a strong positive correlation (e.g., correlation > 0.9 and p < 10−3). Strong symmetry implies a small average change fraction
(e.g., fraction < 1). The term “strictly strong” means that the strictest criterion of strong symmetry is completely satisfied. The term “less strong” means that the
strictest criterion of strong symmetry is partly satisfied.
variables to support these findings (please see Tables 5–6 and Figure 5 in Materials and
Methods for additional information).
Variable Meaning
PIP Potential input projections that a brain region can receive from another region
POP Potential input projections that a brain region can cast to another region
TPIP Total potential input projections that a brain region can receive from all other regions (macroscopic in-degree)
TPOP Total potential input projections that a brain region can cast to all other regions (macroscopic out-degree)
downstream relations can be found among brain regions during information processing, these
relations may merely exist in functional connectivity (static connectivity is principally symmet-
ric). This finding reminds us that functional connectivity can not be simply reflected by static
connectivity. Moreover, we speculate that the uncertain scale-free property is relevant with an
existing controversy of whether static brain connectivity is scale-free or not (see pieces of sup-
porting (Kaiser & Hilgetag, 2004; Kaiser et al., 2007)) and opposing evidence (Breskin, Soriano,
Moses, & Tlusty, 2006; Humphries, Gurney, & Prescott, 2006). Scale-free property of the brain
may critically rely on the granularity and the variables used in connectivity characterization.
To this point, the static connectivity of the fruit fly brain has been characterized, below we
turn to formalize the formation of functional connectivity based on the static connectivity.
decline with λ because specific clusters become larger gradually (the first moment of cluster size
maintains relatively constant while the second moment increases significantly). In sum, the com-
putational costly coactivation analysis (∼1.2 × 109 times of LIF model computation) enables us
to study the E/ I balance λ as a control parameter of functional connectivity formation. As
expected, a higher E/I balance creates stronger functional connectivity because coactivation
occurs more often. More neurons are included in the same cluster rather than maintain isolated,
making it possible for large-scale communication between neurons to emerge.
However, the concordant increase of dynamic connection degree with the increase of E/I
balance λ discussed above is insufficient to give a whole picture of all important information.
A piece of missing information lies in that the observed formation of functional connectivity is
a sigmoid-like process rather than a uniform growth process. While functional connectivity
forms promptly when λ is relatively small, the formation speed becomes stable at a large λ.
The nonuniform speed is not a trivial consequence of λ nor of an artificially manipulation.
Therefore, we conjecture it as an emergence phenomenon triggered by λ and restricted by
specific mechanisms.
Functional connectivity characterized by percolation. Let us step back from the above analysis
and rethink the nature of functional connectivity. Functional connectivity is rooted in the coac-
tivation probability between neurons and is affected by both static connectivity and the E/I
balance λ. One can interpret functional connectivity as a communication pattern between
neurons, where static connectivity serves as a network of information channels, and λ modifies
the information transmission probability in these channels. Although this idea has been studied
previously in neuroscience computationally (Amico et al., 2021; Graham et al., 2020; Shew,
Yang, Yu, Roy, & Plenz, 2011), we discuss it from a more physically fundamental perspective—
percolation. Percolation is a universal characterization of critical phenomena and phase tran-
sitions in a probabilistic form (Agliari, Cioli, & Guadagnini, 2011; Amini & Fountoulakis, 2014;
Balogh & Pittel, 2007; Baxter, Dorogovtsev, Goltsev, & Mendes, 2010; Ben-Naim & Krapivsky,
Equations 1 and 2. Here self-consistency means that P in and P out can be represented as the
functions of themselves (Dorogovtsev et al., 2001; Li et al., 2021).
1 ∂
P in ¼ 1− Gðx; y Þx¼1;y¼1−P in ; (1)
hk i ∂x
1 ∂
P out ¼ 1− Gðx; y Þx¼1−P out ;y¼1 : (2)
hk i ∂y
The normalization term hki in Equations 1 and 2 denotes the average in-degree (or identically,
the average out-degree) in Equation 3.
X X
hk i ¼ P ðu; v Þu ¼ P ðu; v Þv: (3)
Equations 1 and 2 are derived based on the probability generating function (Equation 4), a
standard and practical approach to study random graphs, especially in real data sets (Newman,
Strogatz, & Watts, 2001).
X
Gðx; y Þ ¼ P ðu; v Þx u y v : (4)
u;v
Merely requiring the knowledge of P(u, v) in Equation 4, Equations 1 and 2 have been pow-
erful enough in studying the formation of the GSCC. In other words, they can predict when the
brain connectivity becomes percolate. Specifically, the sufficient and necessary condition for
the GSCC to emerge is that Equations 1 and 2 have nontrivial solutions in (0, 1] (note that the
trivial solution is P in = P out = 0). In practice, it is unnecessary to analytically study the non-
trivial solutions of Equations 1 and 2. Instead, potential solutions can be numerically explored
in a comprehensible way. Specifically, we only need to rewrite Equations 1 and 2 as functions
1 ∂
W in ¼ 1 − P in − Gðx; y Þx¼1;y¼1−P in ; (5)
hk i ∂x
1 ∂
W out ¼ 1 − P out − Gðx; y Þx¼1−P out ;y¼1 ; (6)
hk i ∂y
and explore when W in and W out go through the lines W in = 0 and W out = 0 (see Figure 3C).
^ in, P
When there exist non-trivial solutions (P ^ out), the probability ϕ that a neuron belongs to the
GSCC can be calculated in Equation 7. Otherwise, the probability maintains closing to 0.
ϕ ¼ 1−G 1−P ^ in þ G 1 − P
^ out ; 1 − G 1; 1 − P ^ in :
^ out ; 1 − P (7)
In Figure 3D, we compare between ϕ and the real probability that a neuron belongs to the
GSCC in the experiment to quantify the error of our theoretical predictions. High consistency
can be found between the predictions and experiment.
As can be seen in Figure 3B and Figure 3D, a phenomenon referred to as percolation transition
happens at λ = 0.3, where the GSCC emerges suddenly (ϕ = 0 when λ ≤ 0.3 and ϕ ≥ 0.06 when λ >
0.3). This is a transition of the brain from being fragmentized to percolate. In other words, neurons
become extensively coupled with each other to form system-level functional connectivity after λ
reaches over 0.3. Therefore, λ = 0.3 serves as the percolation threshold. Moreover, probability ϕ in
Equation 5, as well as the corresponding experimental observations, will eventually approximate
to an upper bound (the growing speed approaches to 0 after λ reaches over 0.6). This phenomenon
hints that the GSCC, reflecting functional connectivity, has a size intrinsically bounded by static
connectivity. Functional connectivity forms through synaptic connections and, therefore, must
be a subgraph of static connectivity. By recalculating Equations 1–5 in static connectivity, the
upper bound of ϕ is obtained in Figure 3D and it is reached by ϕ when λ ≥ 0.6. Note intrinsic
consistency between the observations in Figure 2A, Figure 2D, and Figure 2E and that in
Figure 3D, even quantitative details are different as they concern about different parameters.
Furthermore, the above-observed percolation transition can also be confirmed from the per-
spective of diluted percolation. Under dilution condition, functional connectivity formation is
not only constrained by the coactivation probability between neurons but also by the activa-
tion probability of each neuron itself. In other words, the dilution condition represents a more
realistic situation where neurons are conditionally activated (e.g., by external stimulus) and
functional connectivity may form only between activable neurons. The nondilution percola-
tion analyzed above serves as a special case of the diluted one (see Figure 3E). Considering the
dilution of neurons (e.g., each neuron is activated following a probability ρ), another version of
In Figure 3F, we can see that a meaningful ρc 2 [0, 1] (the solution of a probability is mean-
ingful if it is in the interval of [0, 1]) emerges only when λ ≥ 0.3. It decreases with λ until
reaching to its lower bound when λ ≥ 0.6. The existence of a meaningful percolation thresh-
old serves as a necessary condition for percolation transition to happen (e.g., when λ = 0.5,
percolation transition may happen if ρ ≥ ρc = 0.017; when λ = 0.2, percolation transition never
happens since ρc = 6.321 is meaningless). Therefore, the GSCC can form only after λ reaches
over 0.3. These findings are consistent with the above nondilution percolation analysis. In
Figure 3F, we directly show the potential solution of percolation threshold ρc as a function
of λ. Later we will show the benefits of such an illustration.
In summary, our analysis presented above demonstrates that percolation, a universal for-
malism of criticality and phase transitions, can characterize the formation of brain functional
connectivity without other top-down modeling or assumptions. All analytic calculations only
require the knowledge of degree distributions in the brain connectome, which is accessible in
practice. Below, we suggest that the percolation analysis alone is sufficient to explain the
emergence of three key properties of brain functions.
its characteristic path length (average shortest path length between all neurons) (Albert &
Barabási, 2002) and global efficiency (average inverse shortest path length between all neu-
rons) (Latora & Marchiori, 2001) (Figure 4). In general, the characteristic path length reflects
the average cost of optimal information transmission (transmission always follows the min-
imal path length principle). The global efficiency is the average optimal information trans-
mission efficiency. In Figure 4A, once λ reaches over the nondilution percolation threshold
0.3, the characteristic path length of functional connectivity drops sharply while the global
efficiency increases significantly. Once λ reaches over 0.6, the variation speeds of these two
metrics approximate 0. High consistency between these variation trends and the percolation
process in Figure 3D can be observed. Meanwhile, we measure the number ratio and aver-
age size ratio between SCCs and WCCs in functional connectivity, respectively. We also
calculate the size ratio between the GSCC and the GWCC (giant weakly connected cluster).
These ratios are shown as functions of λ in Figure 4B. They principally reflect the proportion
of the close-knit neural community, where information broadcasting capacity is high, within
all communicable neurons (a neuron is communicable if it may communicate with at least
one other neuron). In Figure 4B, the size ratio between the GSCC and the GWCC has a
similar rising trend with the global efficiency, suggesting that the giant close-knit neural com-
munity will occupy more communicable neurons after percolation transition. Although the
other two ratios (the number ratio and average size ratio between SCCs and WCCs) can not
reflect the improvement of information broadcasting capacity by percolation transition, they
fluctuate significantly near the nondilution percolation threshold and may function as
observable markers of percolation transition.
In summary, information transmission efficiency improvement is demonstrated as a bypro-
1998) and the transitivity (Newman, 2003). In Figure 4G–H, these two metrics are shown as
functions of λ, respectively. While they increase sharply after percolation transition, their
increase rates drop once λ reaches over 0.5. In other words, modular structures experience
massive mergers after percolation transition, becoming larger and more close-knit. However,
the ever-rising λ > 0.5 can not improve functional segregation without limitation. These mod-
ular structures in functional connectivity are ultimately restricted by static connectivity. As for
functional integration, it refers to the ability to collect and combine specialized information
from distributed neural clusters (Rubinov & Sporns, 2010b). The characteristic path length and
the global efficiency shown in Figure 4A are practical metrics of this property, which have
been demonstrated as explicable by percolation.
Combine the information in Figure 4E–H, we can see that both network wiring economy
DISCUSSION
The Optimal Synaptic E/I Balance for Brain Functions
Let us move forward our analysis by incorporating all above presented findings to solve a crit-
ical question concerned in neuroscience: what is the optimal synaptic E/I balance?
There are several vital values of the E/I balance λ according to our previous analysis: the
nondilution percolation threshold λ = 0.3 (E/I balance is 3:7) where percolation transition hap-
pens; the approximate value λ = 0.5 (E/I balance is 1:1) which reconciles the actual size of the
GSCC with its increasing rates; the approximate value λ = 0.6 (E/I balance is 3:2), after which
percolation approximates to its bound. The advantageous characteristics of brain functions,
including efficiency (Avena-Koenigsberger et al., 2018), robustness (Aerts et al., 2016; Joyce
et al., 2013; Kaiser et al., 2007), and economy (Bullmore & Sporns, 2012), principally expe-
rience sharp increases after percolation transition at λ = 0.3 and become changeless after λ =
0.6 because they are intrinsically bounded by the characteristics of static or anatomical con-
nectivity. Information transmission efficiency, robust flexibility, and network wiring economy
have relatively large actual quantities and high increasing speeds near λ = 0.5, after which
their increasing speeds gradually approximate 0 and are not sufficient to act as payoffs of
the rising of λ. Therefore, the actual value of network running economy is very likely opti-
mized near λ = 0.5. Above this value a sharp drop is observed. Moreover, a significantly high
E/I balance may damage information encoding efficiency (Barta & Kostal, 2019; Sprekeler,
2017) and membrane potential stabilizing (Sadeh & Clopath, 2021; Sprekeler, 2017). Taking
all these pieces of evidence together, it is suggested that an optimal E/I balance for the brain to
guarantee advantageous properties simultaneously may be λ ∼ 0.5, consistent with the previ-
ous in vitro experimental finding (Shew et al., 2011). Furthermore, this inferred optimal E/I
Percolation theory: balance by percolation theory corroborates the findings of the sufficient condition for neural
A physics theory that characterizes dynamics in the brain to be critical (Poil, Hardstone, Mansvelder, & Linkenkaer-Hansen,
critical phenomena and phase 2012). In other words, it provides explanations of the origin of cortical criticality, a widespread
transitions from a probabilistic and phenomenon in multiple species’ brains (Beggs & Timme, 2012; Chialvo, 2004; Fontenele
geometric perspective.
et al., 2019; Fosque, Williams-García, Beggs, & Ortiz, 2021; Gautam, Hoang, McClanahan,
Grady, & Shew, 2015; Millman, Mihalas, Kirkwood, & Niebur, 2010; Williams-García, Moore,
Beggs, & Ortiz, 2014), from a new perspective. In our research, the E/I balance λ is defined as
the fraction of excitatory synapses in all synapses. Because the excitation/inhibition strength
in Equation 10 is uniformly generated, the predicted optimal E/I balance λ ∼ 0.5 implies a
balance between excitation WE and inhibition WI strengths on the whole brain. This result is
consistent with a well-known finding that neural activities with high and robust entropy
occur when WE and WI are balanced (Agrawal et al., 2018). Meanwhile, we can further
transform the optimal E/I balance λ ∼ 0.5 to parameter ψ, the fraction of excitatory neurons
in all neurons, since WE, WI, and ψ are mathematically related (Agrawal et al., 2018). To guar-
antee high and robust entropy, a large ψ can be derived according to Agrawal et al. (2018). These
results corroborate the abundance of excitatory neurons in mammalian cortices (Hendry,
Schwark, Jones, & Yan, 1987; Meinecke & Peters, 1987; Sahara, Yanagawa, O’Leary, & Stevens,
2012).
In the present study, we have suggested percolation as a window to understand how brain
function properties emerge during functional connectivity formation. The congruent relation-
ship between brain connectivity and percolation is natural and comprehensible. Just like the
porous stone immersed in a pool of water, the brain is “porous” in terms of neurons and
immersed in a “pool” of information. Through simple analytical calculations, percolation anal-
ysis has shown strong consistency both in theoretical and experimental observations. Perhaps
because of these advantages, percolation theory has attracted emerging interest in neurosci-
ence. From early explorations that combine limited neural morphology data with computa-
tional simulations to analyze percolation (Costa, 2005; da Fontoura Costa & Coelho, 2005;
da Fontoura Costa & Manoel, 2003; Stepanyants & Chklovskii, 2005) to more recent works
that study percolation directly on the relatively small-scale and coarse-grained brain connec-
tome and electrically stimulated neural dynamics data captured from living neural networks
(e.g., primary neural cultures in rat hippocampus) (Amini, 2010; Breskin et al., 2006; Cohen
et al., 2010; Eckmann et al., 2007), the efforts from physics have inspired numerous follow-up
explorations in neuroscience (Bordier et al., 2017; Carvalho et al., 2020; Del Ferraro et al.,
2018; Kozma & Puljic, 2015; Lucini et al., 2019; Zhou et al., 2015). These studies capture
an elegant and enlightening view about optimal neural circuitry, neural collective dynamics,
criticality, and the relation between brain connectivity and brain functions. Built on these
works, we present a more systematic and biologically justified framework to formalize func-
tional connectivity formation as percolation on random directed graphs. The merit of our
framework is fourfold. First, different from the bootstrap (e.g., see Eckmann et al., 2007) or
quorum (e.g., see Cohen et al., 2010) percolation analysis in previous studies, we leave the
criterion of occupation as a work of neural dynamics computation rather than preset the occu-
pation profile of neurons. In the future, this criterion can be optimized by synchronous neural
dynamics recording when the required technology matures. Second, percolation on directed
graphs can capture the directionality of neural dynamics diffusion. Third, the percolation
defined on random graphs is a natural reflection of the inequivalence between static and func-
tional connectivity, consistent with the fact that static connectivity serves as a platform of the
dynamic one. Finally, we distinguish between dilution and nondilution conditions, enabling us
to characterize functional connectivity formation in stimulus-driven or attacked situations.
These properties enhance the capacity of our framework to accommodate the real character-
istics of brain connectivity.
Built on the equivalence between functional connectivity formation and the percolation
controlled by synaptic E/I balance, we reveal that brain function properties emerge as bypro-
ducts of percolation transition. Although the present percolation framework is universal, we
Rooted in emergentism, the present study attempts to explore the origin mechanisms of brain
function characteristics from a physics perspective and avoid assumption-based modeling.
The underlying mechanisms are suggested to be described by percolation, a universal char-
acterization of critical phenomena and phase transitions (Dorogovtsev et al., 2001; Li et al.,
2021). Our experiments in the largest yet brain connectome of the fruit fly, Drosophila mel-
anogaster (Pipkin, 2020; Schlegel et al., 2021; Xu et al., 2020), have demonstrated the
capacity of percolation to explain the formation of functional connectivity and the emer-
gence of brain function properties without additional assumptions. The strong explanatory
power of percolation theory toward these neuroscience concepts is experimentally observed,
even though they are initially studied in different scientific fields. Such an intriguing connection
reemphasizes the nature of the brain as a physical system. Recently, this physics perspec-
tive has seen substantial progress in neuroscience. For instance, the free-energy principle is
demonstrated as a unified foundation of perception, action, and learning (Friston, 2009,
2010, 2012; Friston & Kiebel, 2009; Guevara, 2021). Cortical criticality is discovered to
account for the efficient transformation between cortical states (Beggs & Timme, 2012;
Chialvo, 2004; Fontenele et al., 2019; Fosque et al., 2021; Gautam et al., 2015; Millman
et al., 2010; Williams-García et al., 2014). Moreover, information thermodynamics is
revealed as a bridge between the physical and the informational brain (Capolupo, Freeman,
& Vitiello, 2013; Collell & Fauquet, 2015; Sartori, Granger, Lee, & Horowitz, 2014; Sengupta,
Stemmler, & Friston, 2013; Street, 2020). In the future, this rising direction may continue con-
tributing to neuroscience as a window to understand the intrinsic bases of brain function
characteristics.
To conclude, the parallels discovered between functional connectivity formation and the
percolation on random directed graphs controlled by synaptic E/I balance serve as a new path-
way toward understanding the emergence of brain function characteristics. The percolation,
an elegant formalism of critical phenomena and phase transitions, may elucidate the emer-
gence of brain function characters from brain connectivity.
The brain connectome data is released as a part of the FlyEM project (Xu et al., 2020). The
connectome imaging is implemented applying the focused ion beam scanning electron
microscopy (FIB-SEM) technology. Then images are cleaned and enhanced by a series of pipe-
lines (e.g., automated flood-filling network segmentation), generating the most fine-grained
and high-throughput connectome of the fruit fly central brain (∼2.5 × 105 nm in each dimen-
sion) to date (Xu et al., 2020). There are 21,662 traced (all main branches within the volume
are reconstructed) and uncropped (main arbors are contained in the volume) neurons as well
as 4,495 traced, cropped, and large (≥1,000 synaptic connections) neurons in the connec-
tome. The complex wiring diagram between neurons consists of ∼6 × 106 traced and
Variable Meaning
SC All synaptic connections featured by one neuron
ISC All synaptic connections received by one neuron from other neurons
OSC All synaptic connections coming from one neuron to other neurons
DA All directionally adjacent relations between one neuron and other neurons
IDA All directionally adjacent relations between one neuron and its presynaptic neurons
ODA All directionally adjacent relations between one neuron and its postsynaptic neurons
ISC/IDA Synaptic connections between one neuron and one of its presynaptic neuron (the number of ISCs per IDA)
OSC/ODA Synaptic connections between one neuron and one of its postsynaptic neuron (the number of OSCs per ODA)
and we can similarly distinguish between IDAs and ODAs. The fine-grained information is
necessary in neural dynamics computation while the coarse-grained information in useful in
analyzing microscopic functional connectivity. Therefore, we also calculate SC/DA, ISC/IDA,
and OSC/ODA to reflect how a neuron allocates synaptic connections to each of its neighbors.
These variables offer mappings between fine-grained and coarse-grained information. In our
research, variables relevant with DA are used in our main analyses (please see Tables 1 and 2
and Figure 1 in the main text) while other variables are used to provide auxiliary information
(please see Tables 5 and 6 and Figure 5).
Note. Being scale-free requires P ∝ n−α, where α 2 (2, 3). Goodness of estimation is expected as to be less than 0.05.
Figure 5. Supplementary information of static connectivity characterization. (A) Supplementary variables of macroscopic static connectivity
and microscopic static connectivity are graphically illustrated in an instance. (B) The power law models of synaptic connection number (SC),
directionally adjacent relation number (DA), and the number of synaptic connections per directionally adjacent relation (SC/DA). (C) The
power law models of input synaptic connection number (ISC) and output synaptic connection number (OSC). (D) The power law models
of the number of synaptic connections between one neuron and one of its presynaptic neuron (ISC/IDA) and the number of synaptic connec-
tions between one neuron and one of its postsynaptic neuron (OSC/ODA).
Note. Strong symmetry requires a strong positive correlation (e.g., correlation > 0.9 and p < 10−3). Strong symmetry implies a small average change fraction
(e.g., fraction < 1). The term “strictly strong” means that the strictest criterion of strong symmetry is completely satisfied. The term “less strong” means that the
strictest criterion of strong symmetry is partly satisfied. The term “relatively weak” means that the strictest criterion of strong symmetry is not satisfied.
suggest that three levels of symmetry can be specified. The symmetry is suggested as strictly
Here we introduce the method of functional connectivity generation. Specifically, every direc-
tionally adjacent relation (DA) Ni → Nj (here Ni and Nj are neurons) has a static connection
strength Sij defined by Equation 10. In the definition, random variable Xi 2 {0, 1} determines
whether Ni is an excitatory or inhibitory neuron. Random variables Yij ∼ U(0,1] and Zij 2∼
U [−1,0] are generated uniformly to measure excitation/inhibition strength (e.g., conductance).
Notion αij denotes the number of synaptic connections (SC) from Ni to Nj (SC/DA, the precise
α
number of SCs per DA) and h·i denotes the expectation. By calculating α ij in Equation 10, we
h ij ii;j
can quantify the effects of SCs on Sij.
αij
Sij ¼ Xi Yij þ ð1 − Xi ÞZij : (10)
αij i;j
Under each condition of E/I balance λ, a set of variable {X1, …, Xn} (here n denotes the number of
neurons) is randomly initialized. Then, we update {X1, …, Xn} by the following algorithm: (1)
measure ^λ, the fraction of excitatory SCs in all SCs based on the current configuration of {X1,
…, Xn}, and calculate the difference |λ ^ − λ|; (2) randomly select a neuron Ni and change the
corresponding Xi (e.g., change Xi from 1 to 0 or from 0 to 1). (the change will not be kept unless
^ − λ|); (3) repeat steps (1 and 2) until ^λ ∼ λ. Based on this algorithm, the fraction of
it reduces |λ
excitatory synapses in all synapses is principally controlled by λ with reasonable errors in
practice.
Then, we measure coactivation probability P ij in Equation 11 to define the dynamic con-
nection strength, where δ(·) denotes the Dirac delta function. Probability P ij is quantified in
term of that an activated Ni can activate Nj in m times of experiments (m = 100).
1X m
P ij ¼ δ I k − Ijk : (11)
m k¼1 i
Each time all neurons are initialized at the resting potential, −70 mV. Then we continuously
activate Ni (make Iik = 1) and several other randomly selected presynaptic neurons of Nj during
an interval of [0, t] (we define t = 50 ms). We ensure that at least 20% of presynaptic neurons of
Nj are activated. Such activation intensity is no longer accidental and should be strong enough
to affect Nj in the real brain. Any activation will be marked by Equation 12, where Ijk = 1 if Nj is
activated at least one time during [0, t], otherwise Ijk = 0. The criterion of activation in Equation
12 is to verify if the actual membrane potential Vj(τ) reaches the spiking threshold V ^ = −50 mV
based on the unit step function ν(·).
" #
X
t
Ijk ¼ν ^ :
ν Vτ − V (12)
We calculate the actual membrane potential Vj(τ) applying the standard leaky integrate-and-
fire (LIF) model (Gerstner et al., 2014) in Equation 13, where τm is the membrane time constant
(τm = 10 ms), notion Nh searches through all presynaptic neurons of Nj, and τ fh represents the
arrival time of f-th spike of neuron Nh.
( )
∂Vj ðτ Þ 1 X X f f
¼− ^ −
Vj ðτ Þ − V Shj δ τ − τ h Vj τ h − Vh τ h
f
: (13)
∂τ τm N f
h
To make such a computationally costly experiment accessible for readers, we release our
results in Tian and Sun (2021). Readers can freely download the dataset to obtain the co-
activation probability P ij under each λ condition.
In our research, the measurements of the characteristic path length (Albert & Barabási, 2002),
the global efficiency (Latora & Marchiori, 2001), the assortativity coefficient (Newman, 2002),
the rich-club coefficient (Ball et al., 2014; Colizza et al., 2006; Van Den Heuvel & Sporns,
2011), the physical Rentian exponent (Bassett et al., 2010; Chen, 1999; Ozaktas, 1992), the
clustering coefficient (Watts & Strogatz, 1998), and the transitivity (Newman, 2003) are imple-
mented based on the toolbox released by Rubinov and Sporns (2010b). This widely used tool-
box and its instructions can be found in Rubinov and Sporns (2010a). Readers can use the
adjacent matrix of functional connectivity as an input to calculate these properties. One thing
to note is that the original version of the toolbox defines the graph distance between two
unconnected neurons (there is no path between them) as infinity. Although this definition is
mathematically justified, it may bring inconvenience in computational implementations
because inf usually leads to inf or nan in computer programs. Common practices to avoid
the influence of potential infinite values on target results are either excluding these infinite
values during calculations or replacing them with n + 1 (here n denotes the total number of
neurons). To maintain the sample size in analyses, our research chooses the second approach.
ACKNOWLEDGMENTS
Authors are grateful for discussions and assistance of Drs. Ziyang Zhang and Yaoyuan Wang
from the Laboratory of Advanced Computing and Storage, Central Research Institute, 2012
Laboratories, Huawei Technologies Co. Ltd., Beijing, 100084, China.
AUTHOR CONTRIBUTIONS
FUNDING INFORMATION
Pei Sun, Artificial and General Intelligence Research Program of Guo Qiang Research Institute
at Tsinghua University, Award ID: 2020GQG1017.
Achard, S., Salvador, R., Whitcher, B., Suckling, J., & Bullmore, E. between communication efficiency and resilience in the human
(2006). A resilient, low-frequency, small-world human brain connectome. Brain Structure and Function, 222(1), 603–618.
functional network with highly connected association cortical https://doi.org/10.1007/s00429-016-1238-5, PubMed:
hubs. Journal of Neuroscience, 26(1), 63–72. https://doi.org/10 27334341
.1523/JNEUROSCI.3874-05.2006, PubMed: 16399673 Avena-Koenigsberger, A., Misic, B., & Sporns, O. (2018). Commu-
Aerts, H., Fias, W., Caeyenberghs, K., & Marinazzo, D. (2016). nication dynamics in complex brain networks. Nature Reviews
Brain networks under attack: Robustness properties and the Neuroscience, 19(1), 17–33. https://doi.org/10.1038/nrn.2017
impact of lesions. Brain, 139(12), 3063–3083. https://doi.org/10 .149, PubMed: 29238085
.1093/brain/aww194, PubMed: 27497487 Ball, G., Aljabar, P., Zebari, S., Tusor, N., Arichi, T., Merchant, N.,
Agliari, E., Cioli, C., & Guadagnini, E. (2011). Percolation on cor- … Counsell, S. J. (2014). Rich-club organization of the newborn
related random networks. Physical Review E, 84(3), 031120. human brain. Proceedings of the National Academy of Sciences,
https://doi.org/10.1103/ PhysRevE.84.031120, PubMed: 111(20), 7456–7461. https://doi.org/10.1073/pnas.1324118111,
22060341 PubMed: 24799693
Agrawal, V., Cowley, A. B., Alfaori, Q., Larremore, D. B., Restrepo, Balogh, J., & Pittel, B. G. (2007). Bootstrap percolation on the ran-
J. G., & Shew, W. L. (2018). Robust entropy requires strong and dom regular graph. Random Structures & Algorithms, 30(1–2),
balanced excitatory and inhibitory synapses. Chaos: An Interdis- 257–286. https://doi.org/10.1002/rsa.20158
ciplinary Journal of Nonlinear Science, 28(10), 103115. https:// Barta, T., & Kostal, L. (2019). The effect of inhibition on rate code
doi.org/10.1063/1.5043429, PubMed: 30384653 efficiency indicators. PLoS Computational Biology, 15(12),
Albert, R., & Barabási, A.-L. (2002). Statistical mechanics of com- e1007545. https://doi.org/10.1371/journal.pcbi.1007545,
plex networks. Reviews of Modern Physics, 74(1), 47. https://doi PubMed: 31790384
.org/10.1103/RevModPhys.74.47 Bassett, D. S., Greenfield, D. L., Meyer-Lindenberg, A., Weinberger,
Albert, R., Jeong, H., & Barabási, A.-L. (2000). Error and attack tol- D. R., Moore, S. W., & Bullmore, E. T. (2010). Efficient physical
erance of complex networks. Nature, 406(6794), 378–382. embedding of topologically complex information processing net-
https://doi.org/10.1038/35019019, PubMed: 10935628 works in brains and computer circuits. PLoS Computational Biol-
Alstott, J., Breakspear, M., Hagmann, P., Cammoun, L., & Sporns, ogy, 6(4), e1000748. https://doi.org/10.1371/journal.pcbi
O. (2009). Modeling the impact of lesions in the human brain. .1000748, PubMed: 20421990
PLoS Computational Biology, 5(6), e1000408. https://doi.org/10 Baxter, G. J., Dorogovtsev, S. N., Goltsev, A. V., & Mendes, J. F.
.1371/journal.pcbi.1000408, PubMed: 19521503 (2010). Bootstrap percolation on complex networks. Physical
Amico, E., Abbas, K., Duong-Tran, D. A., Tipnis, U., Rajapandian, Review E, 82(1), 011103. https://doi.org/10.1103/PhysRevE.82
M., Chumin, E., … Goñi, J. (2021). Towards an information the- .011103, PubMed: 20866561
oretical description of communication in brain networks. Net- Beggs, J. M., & Timme, N. (2012). Being critical of criticality in the
work Neuroscience, 5(3), 646–665. https://doi.org/10.1162/netn brain. Frontiers in Physiology, 3, 163. https://doi.org/10.3389
_a_00185, PubMed: 34746621 /fphys.2012.00163, PubMed: 22701101
Amini, H. (2010). Bootstrap percolation in living neural networks. Ben-Naim, E., & Krapivsky, P. (2011). Dynamics of random graphs
Journal of Statistical Physics, 141(3), 459–475. https://doi.org/10 with bounded degrees. Journal of Statistical Mechanics: Theory
.1007/s10955-010-0056-z and Experiment, 2011(11), P11008. https://doi.org/10.1088
Amini, H., & Fountoulakis, N. (2014). Bootstrap percolation in /1742-5468/2011/11/P11008
power-law random graphs. Journal of Statistical Physics, 155(1), Betzel, R. F., Avena-Koenigsberger, A., Goñi, J., He, Y., De Reus,
72–92. https://doi.org/10.1007/s10955-014-0946-6 M. A., Griffa, A., … Sporns, O. (2016). Generative models of the
Avena-Koenigsberger, A., Mišić, B., Hawkins, R. X., Griffa, A., Hagmann, human connectome. NeuroImage, 124, 1054–1064. https://doi
P., Goñi, J., & Sporns, O. (2017). Path ensembles and a tradeoff .org/10.1016/j.neuroimage.2015.09.041, PubMed: 26427642
Bollobás, B. (2013). Modern graph theory (Vol. 184). New York, NY: disorders. Brain, 137(8), 2382–2395. https://doi.org/10.1093
Springer Science & Business Media. /brain/awu132, PubMed: 25057133
Bordier, C., Nicolini, C., & Bifone, A. (2017). Graph analysis and da Fontoura Costa, L., & Coelho, R. C. (2005). Growth-driven per-
modularity of brain functional connectivity networks: Searching colations: The dynamics of connectivity in neuronal systems. The
for the optimal threshold. Frontiers in Neuroscience, 11, 441. European Physical Journal B-Condensed Matter and Complex
https://doi.org/10.3389/fnins.2017.00441, PubMed: 28824364 Systems, 47(4), 571–581. https://doi.org/10.1140/epjb/e2005
Breskin, I., Soriano, J., Moses, E., & Tlusty, T. (2006). Percolation in -00354-5
living neural networks. Physical Review Letters, 97(18), 188102. da Fontoura Costa, L., & Manoel, E. T. M. (2003). A percolation
https://doi.org/10.1103/ PhysRevLett.97.188102, PubMed: approach to neural morphometry and connectivity. Neuroinfor-
17155581 matics, 1(1), 65–80. https://doi.org/10.1385/ NI:1:1:065,
Bridi, M. C., Zong, F.-J., Min, X., Luo, N., Tran, T., Qiu, J., … PubMed: 15055394
Kirkwood, A. (2020). Daily oscillation of the excitation- Del Ferraro, G., Moreno, A., Min, B., Morone, F., Pérez-Ramírez,
inhibition balance in visual cortical circuits. Neuron, 105(4), Ú., Pérez-Cervera, L., … Makse, H. A. (2018). Finding influential
Goltsev, A. V., Dorogovtsev, S. N., & Mendes, J. F. (2008). Percola- Karbowski, J. (2007). Global and regional brain metabolic scaling
tion on correlated networks. Physical Review E, 78(5), 051105. and its functional consequences. BMC Biology, 5(1), 1–11.
https://doi.org/10.1103/PhysRevE.78.051105, PubMed: 19113093 https://doi.org/10.1186/1741-7007-5-18, PubMed: 17488526
Goñi, J., Van Den Heuvel, M. P., Avena-Koenigsberger, A., De Kiebel, S. J., & Friston, K. J. (2011). Free energy and dendritic
Mendizabal, N. V., Betzel, R. F., Griffa, A., … Sporns, O. self-organization. Frontiers in Systems Neuroscience, 5, 80.
(2014). Resting-brain functional connectivity predicted by ana- https://doi.org/10.3389/fnsys.2011.00080, PubMed: 22013413
lytic measures of network communication. Proceedings of the Kozma, R., & Puljic, M. (2015). Random graph theory and neuro-
National Academy of Sciences, 111(2), 833–838. https://doi.org percolation for modeling brain oscillations at criticality. Current
/10.1073/pnas.1315529111, PubMed: 24379387 Opinion in Neurobiology, 31, 181–188. https://doi.org/10.1016/j
Graham, D., Avena-Koenigsberger, A., & Misic, B. (2020). Network com- .conb.2014.11.005, PubMed: 25460075
munication in the brain. Network Neuroscience, 4(4), 976–979. Latora, V., & Marchiori, M. (2001). Efficient behavior of small-world
https://doi.org/10.1162/netn_e_00167, PubMed: 33195944 networks. Physical Review Letters, 87(19), 198701. https://doi
Grent, T., Gross, J., Goense, J., Wibral, M., Gajwani, R., Gumley, .org/10.1103/PhysRevLett.87.198701, PubMed: 11690461
A. I., … Uhlhaas, P. J. (2018). Resting-state gamma-band power
jointly emerge from balanced excitation/inhibition in neuronal Sporns, O. (2011). The non-random brain: Efficiency, economy, and
networks. Journal of Neuroscience, 32(29), 9817–9823. https://doi complex dynamics. Frontiers in Computational Neuroscience, 5,
.org/10.1523/JNEUROSCI.5990-11.2012, PubMed: 22815496 5. https://doi.org/10.3389/fncom.2011.00005, PubMed:
Radicchi, F., & Fortunato, S. (2009). Explosive percolation in scale- 21369354
free networks. Physical Review Letters, 103(16), 168701. https:// Sporns, O., Tononi, G., & Kötter, R. (2005). The human connec-
doi.org/10.1103/PhysRevLett.103.168701, PubMed: 19905730 tome: A structural description of the human brain. PLoS Compu-
Ramírez-Toraño, F., Bruña, R., de Frutos-Lucas, J., Rodŕıguez-Rojo, tational Biology, 1(4), e42. https://doi.org/10.1371/journal.pcbi
I., Marcos de Pedro, S., Delgado-Losada, M., … Maestú, F. .0010042, PubMed: 16201007
(2021). Functional connectivity hypersynchronization in rela- Sprekeler, H. (2017). Functional consequences of inhibitory plastic-
tives of alzheimer’s disease patients: An early E/I balance dys- ity: Homeostasis, the excitation-inhibition balance and beyond.
function? Cerebral Cortex, 31(2), 1201–1210. https://doi.org/10 Current Opinion in Neurobiology, 43, 198–203. https://doi.org
.1093/cercor/bhaa286, PubMed: 33108468 /10.1016/j.conb.2017.03.014, PubMed: 28500933
Rubinov, M., & Sporns, O. (2010a). Brain connectivity toolbox. Stepanyants, A., & Chklovskii, D. B. (2005). Neurogeometry and