Brain computation by assemblies of neurons
Contributed by Christos H. Papadimitriou, March 27, 2020 (sent for review February 6, 2020; reviewed by Angela D. Friederici and Tomaso A. Poggio)
Significance
Our expanding understanding of the brain at the level of neurons and synapses, and the level of cognitive phenomena such as language, leaves a formidable gap between these two scales. Here we introduce a computational system which promises to bridge this gap: the Assembly Calculus. It encompasses operations on assemblies of neurons, such as project, associate, and merge, which appear to be implicated in cognitive phenomena, and can be shown, analytically as well as through simulations, to be plausibly realizable at the level of neurons and synapses. We demonstrate the reach of this system by proposing a brain architecture for syntactic processing in the production of language, compatible with recent experimental results.
Abstract
Assemblies are large populations of neurons believed to imprint memories, concepts, words, and other cognitive information. We identify a repertoire of operations on assemblies. These operations correspond to properties of assemblies observed in experiments, and can be shown, analytically and through simulations, to be realizable by generic, randomly connected populations of neurons with Hebbian plasticity and inhibition. Assemblies and their operations constitute a computational model of the brain which we call the Assembly Calculus, occupying a level of detail intermediate between the level of spiking neurons and synapses and that of the whole brain. The resulting computational system can be shown, under assumptions, to be, in principle, capable of carrying out arbitrary computations. We hypothesize that something like it may underlie higher human cognitive functions such as reasoning, planning, and language. In particular, we propose a plausible brain architecture based on assemblies for implementing the syntactic processing of language in cortex, which is consistent with recent experimental results.
Sign up for PNAS alerts.
Get alerts for new articles, or get an alert when an article is cited.
How does the brain beget the mind? How do molecules, cells, and synapses effect cognition, behavior, intelligence, reasoning, and language? The remarkable and accelerating progress in neuroscience, both experimental and theoretical–computational, does not seem to bring us closer to an answer: The gap is formidable, and seems to necessitate the development of new conceptual frameworks. As Richard Axel (1) recently put it, “we do not have a logic for the transformation of neural activity into thought and action. I view discerning [this] logic as the most important future direction of neuroscience.” What kind of formal system, embodying and abstracting the realities of neural activity, would qualify as the sought “logic”?
We propose a formal computational model of the brain based on assemblies of neurons; we call this system the Assembly Calculus. In terms of detail and granularity, the Assembly Calculus occupies a position intermediate between the level of individual neurons and synapses and the level of the whole-brain models useful in cognitive science (e.g., refs. 2 and 3).
The basic elementary object of our system is the assembly of excitatory neurons. The idea of assemblies is, of course, not new. They were first hypothesized seven decades ago by Donald O. Hebb (4) to be densely interconnected sets of neurons whose loosely synchronized firing in a pattern is coterminous with the subject thinking of a particular concept or idea. Assembly-like formations have been sought by researchers during the decades following Hebb’s prediction (see, e.g., ref. 5), until they were clearly identified more than a decade ago through calcium imaging (6, 7). More recently, assemblies (sometimes called ensembles) and their dynamic behavior have been studied extensively in the animal brain (see, e.g., ref. 8), while assemblies are the central theme of György Buzsáki’s recent book (9).
Our calculus outfits assemblies with certain operations that create new assemblies and/or modify existing ones: project, reciprocal project, associate, merge, and a few others. These operations reflect known properties of assemblies observed in experiments, and they can be shown, either analytically or through simulations (often both), to result from the activity of neurons and synapses. In other words, the high-level operations of this system can be “compiled down” to the world of neurons and synapses—a fact reminiscent of the way in which high-level programming languages are translated into machine code.
Model
Our mathematical results, as well as most of our simulation results, employ a simplified and analytically tractable, yet plausible, model of neurons and synapses. We assume a finite number of brain areas denoted , etc., intended to stand for an anatomically and functionally meaningful partition of the cortex (Fig. 1). Each area contains a population of excitatory neurons with random recurrent connections.* By this, we mean that each ordered pair of neurons in an area has the same small probability of being connected by a synapse, independently of what happens to other pairs—this is a well-studied framework usually referred to as Erdõs–Renyi graph or (10). In addition, for certain ordered pairs of areas, say , there are random afferent synaptic interconnections from to ; that is, for every neuron in and every neuron in , there is a chance that there is a synaptic connection from the former to the latter.†
Fig. 1.

We use this model to fathom, quantitatively, the creation and manipulation of assemblies of neurons. Since the model is probabilistic (by virtue of the random synaptic connectivity), our analytical results postulating the effectiveness of the various operations must contain the clause “with high probability,” where the event space is implicitly the underlying random graph. We assume that all cells in an assembly belong to the same brain area, denoted .
Our model also encompasses simplified forms of plasticity and inhibition. We assume multiplicative Hebbian plasticity: If at a time step neuron fires and, at the next time step, neuron fires, and there is a synapse from to , the weight of this synapse is multiplied by , where is the final parameter of our model (along with , and ). Larger values of the plasticity coefficient result in the operations converging faster, and render many of our proofs simpler. Finally, we model inhibition and excitatory–inhibitory balance by postulating that neurons fire in discrete time steps, and, at any time, only a fixed number of the excitatory neurons in any area fire—in particular, those neurons which receive the highest excitatory inputs.
The four basic parameters of our model are these: (the number of excitatory neurons in an area, and the basic parameter of our model), (the probability of recurrent and afferent synaptic connectivity), (the maximum number of firing neurons in any area), and the plasticity coefficient . Typical values of these parameters in our simulations are , and .
Our model, as described so far, would result, through plasticity, in gigantic synaptic weights after a long time of operation. We further assume that synaptic weights are renormalized, at a slower time scale, so that the sum of presynaptic weights at each neuron stays relatively stable. This process of homeostasis through renormalization is orthogonal to the phenomena we describe here, and it interferes minimally with our arguments and simulations.
We emphasize that our model is generic in the sense that it is not assumed that neural circuits specific to various tasks are already in place, beyond random connectivity between excitatory neurons within an area and between selected pairs of areas. Its functionality—the needed apparatus for each task, such as implementing an assembly operation—emerges from the randomness of the network and the selection of the neurons with highest synaptic input as an almost certain consequence of certain simple events—such as the repeated firing of an assembly.
Assembly Projection.
How do assemblies in the association cortex come about? It has been hypothesized (see, e.g., ref. 13) that an assembly imprinting, for example, a familiar face in a subject’s medial temporal lobe (MTL) is created by the projection of a neuronal population, perhaps in the inferotemporal cortex, encoding this face as a whole object. By projection of an assembly , we mean the creation of an assembly in a downstream area that can be thought of as a “copy” of , and such that will henceforth fire every time fires.
How is the new assembly formed in a downstream area by the repeated firing of in area ? The process was vividly predicted in the discussion section of ref. 14, for the case in which is the olfactory bulb and is the piriform cortex. Once has fired once, synaptic connectivity from area to area excites many neurons in area . Inhibition will soon limit the firing in area to a smaller set of neurons, let us call it , consisting, in our framework, of neurons (Fig. 2A). Next, the simultaneous firing of and creates a stronger excitation in area (one extra reason for this is plasticity, which has already strengthened the connections from to ), and, as a result, a new set of neurons from area will be selected to fire, call it . One would expect to overlap substantially with —this overlap can be calculated in our mathematical model to be roughly for a broad range of parameters. If continues firing, a sequence of sets of neurons of size in area will be created. For a large range of parameters and for high enough plasticity, this process can be proved to converge exponentially fast, with high probability, to create an assembly , the result of the projection.
Fig. 2.

We denote the projection process described above as . Assembly projection has been demonstrated both analytically (16, 17) and through detailed simulations (18, 19); simulation results in our model, as well as improved analysis, are presented in Fig. 2B4. Once the operation has taken place, we write and .
Dense Intraconnection of Assemblies.
Hebb (4) hypothesized that assemblies are densely intraconnected—that is, the chance that two neurons have a synaptic connection is significantly larger when they belong to the same assembly than when they do not—and our analysis and simulations verify this hypothesis (Fig. 2C). From the point of view of computation, this is rather surprising, because the problem of finding a dense subgraph of a certain size in a sparse graph is a known difficult problem in computer science (20), and thus the very existence of an assembly may seem surprising. How can the brain deal with this difficult computational problem? The creation of an assembly through projection as outlined in the previous paragraphs provides an explanation: Since the elements of assembly were selected to have strong synaptic input from the union of and , one intuitively expects the synaptic recurrent connectivity of to be higher than random. In addition, the weights of these synapses should be higher than average because of plasticity.
The Random Projection and Cap Primitive.
It can be said that assembly operations, as described here, are powered exclusively by two forces known to be crucial for life more generally: randomness and selection (in addition to plasticity). No special purpose neural circuits are required to be in place; all that is needed is random synaptic connectivity between, and recurrently within, areas, and selection, through inhibition in each area, of the out of cells currently receiving highest synaptic input. All assembly computation described here consists of applications of this operator, which we call random projection and cap (RP&C). We believe that RP&C is an important primitive of neural computation, and computational learning more generally, and can be shown to have a number of potentially useful properties. For example, we establish, analytically and through simulations, that RP&C preserves overlap of assemblies remarkably well (as first noticed empirically in ref. 21). Finally, we used RP&C in experiments as the nonlinearity in each layer of a deep neural network, in the place of the sigmoid or the rectified linear unit typically used in machine leaning, and we found that it seems to perform at a level competitive with these.
Association and Pattern Completion.
In a recent experiment (15), electrocorticography recordings of human subjects revealed that a neuron in a subject’s MTL consistently responding to the image of a particular familiar place—such as the Pyramids—starts to also respond to the image of a particular familiar person—say, the subject’s sibling—once a combined image has been shown of this person in that place. A compelling parsimonious explanation of this and many similar results is that two assemblies imprinting two different entities adapt to the cooccurrence, or other observed affinity, of the entities they imprint by increasing their overlap, with cells from each migrating to the other while other cells leave the assemblies to maintain its size to ; we say that the two assemblies are associating with one another. The association of two assemblies and in the same brain area is denoted associate, with the common area implicit. We can show, analytically and through simulations, that the simultaneous sustained firing of the two parents of and does effect such increase in overlap, while similar results have been obtained by simulations of networks of spiking neurons through spiked timing-dependent plasticity (STDP) (18).
Assemblies are large and, in many ways, random sets of neurons, and, as a result, any two of them, if in the same area, may overlap a little by chance. If the assembly size is about less than the square root of , as we often assume in our simulations, this random overlap, if any, should be very few cells. In contrast, overlap resulting from the operation is quite substantial: The results of ref. 15 suggest an overlap between associated assemblies in the MTL of about 8 to 10% of the size of an assembly. The association between assemblies evokes a conception of a brain area as the arena of complex association patterns between the area’s assemblies; for a discussion of certain mathematical aspects of this view, see ref. 22.
One important and well-studied phenomenon involving assemblies is pattern completion: the firing of the whole assembly in response to the firing of a small subset of its cells (8); presumably, such completion happens with a certain a priori probability depending on the particular subset firing. In our experiments, pattern completion happens in a rather striking way, with small parts of the assembly being able to complete, very accurately, the whole assembly (Fig. 2 F1 and F2).
We believe that association and pattern completion open up fascinating possibilities for a genre of probabilistic computation through assemblies, a research direction which should be further pursued.
Merge.
The most sophisticated and complex operation in the repertoire of the Assembly Calculus is merge. Denoted , this operation starts with two assemblies and , in different brain areas, and creates a new assembly , in a third area , such that there is ample synaptic connectivity from and to , and also vice versa, from to both and .
Linguists had long predicted that the human brain is capable of combining, in a particularly strong sense, two separate entities to create a new entity representing this specific combination (23, 24), and that this ability is recursive in that the combined entity can, in turn, be combined with others. This is a crucial step in the creation of the hierarchies (trees of entities) that seem necessary for the syntactic processing of language, but also for hierarchical thinking more generally (deduction, discourse, planning, storytelling, etc.). Recent fMRI experiments (25) have indicated that, indeed, the completion of phrases and sentences (the completion of auditory stimuli such as “hill top” and “ship sunk”) activates parts of Broca’s area—in particular, the pars opercularis BA 44 for phrases, and the pars triangularis BA 45 for sentences. In contrast, unstructured word sequences such as “hill ship” do not seem to activate Broca’s area. Recall that Broca’s area has long been believed to be implicated in the syntactic processing of language.
A parsimonious explanation of these findings is that phrases and sentences are represented by assemblies in Broca’s area that are the results of the merge of assemblies representing their constituents (that is, assemblies for words such as “ship” and “sunk”); presumably, the word assemblies reside in Wernicke’s area implicated in word selection in language. As these hierarchies need to be traversed both in the upward and in the downward direction (e.g., in the processes of language parsing and language generation, respectively), it is natural to assume that merge must have two-way connections between the new assembly and the constituent assemblies.
Our algorithm for implementing merge, explained in Results and Methods, is by far the most complex in this work, as it involves the coordination of five different brain areas with ample reciprocal connectivity between them, and requires stronger plasticity than other operations; our simulation results are reported in Section 2; see also Fig. 2 G1 and G2.
Finally, a simpler operation with similar yet milder complexity is : It is an extension of , with the additional functionality that the resulting has strong backward synaptic connectivity to . Reciprocal projection has been hypothesized to be instrumental for implementing variable binding in the brain—such as designating “cats” as the subject of the sentence “cats chase mice” (see ref. 19). The plausibility of has been experimentally verified through detailed simulations of networks of spiking neurons with STDP (19), as well as in our simplified model.
Readout and Control Operations.
The purpose of the Assembly Calculus is to provide a formal system within which high-level cognitive behavior can be expressed. Ultimately, we want to be able to write meaningful programs—in fact, parallel programs—in this system, for example, containing segments such as
.
With this goal in mind, we next introduce certain additional low-level control operations, sensing and affecting the state of the system. First, a simple read operation. In an influential review paper on assemblies, Buzsáki (7) proposes that, for assemblies to be functionally useful, readout mechanisms must exist that sense the current state of the assembly system and trigger appropriate further action. Accordingly, the Assembly Calculus contains an operation that identifies the assembly which has just fired in area , and returns null otherwise.
Finally, the Assembly Calculus contains two simple control operations. We assume that an assembly in an area can be explicitly caused to fire by the operation . That is to say, at the time an assembly is created, a mechanism is set in place for activating it; in view of the phenomenon of pattern completion discussed above, in which firing a tiny fragment of an assembly leads to the whole assembly firing, this does not seem implausible. We also assume that, by default, the excitatory cells in an area are inhibited, unless explicitly disinhibited by the operation for a limited period, whose end is marked by the operation ; the plausibility of the disinhibition–inhibition primitives is argued in ref. 26 in terms of vasoactive intestinal peptide cells (27).
Example.
For a simple example of a program in the Assembly Calculus, the command where , is equivalent to the following:
.
Simulations show that, with typical parameters, a stable assembly is formed after about steps. An alternative version of this program, relying on the function , is the following:
.
For another simple example, the command , whose effect is to increase the overlap of two assemblies in the same area , is tantamount to this:
A far more elaborate program in the Assembly Calculus, for a proposed implementation of the creation of syntax trees in the generation of natural language, is described in Discussion.
Computational Power.
We can show (SI Appendix) that the Assembly Calculus as defined above is capable of implementing, under appropriate assumptions, arbitrary computations on bits of memory—and, under much weaker assumptions, bits. This suggests—in view of the well-established equivalence between parallel computation and space-bounded computation (28)—that any parallel computation with a few hundred parallel steps, and using a few hundred registers, can, in principle, be carried out by the Assembly Calculus.
The Assembly Hypothesis.
The Assembly Calculus is a formal system with a repertoire of rather sophisticated operations, where each of these operations can be ultimately reduced to the firing of randomly connected populations of excitatory neurons with inhibition and Hebbian plasticity. The ensuing computational power of the Assembly Calculus may embolden one to hypothesize that such a computational system—or rather a neural system far less structured and precise, which, however, can be usefully abstracted this way—underlies advanced cognitive functions, especially in the human brain, such as reasoning, planning, and language.
Related Work.
Assemblies of excitatory neurons are, of course, not new: They have been hypothesized (4, 5), identified in vivo (6, 7), studied experimentally (8), and discussed extensively (29) over the past decades—even, occasionally, in connection to computation (see ref. 30). However, we are not aware of previous work in which assemblies, with a suite of operations, are proposed as the basis of a computational system intended to explain cognitive phenomena.
Assemblies and their operations as treated here bear a certain resemblance to the research tradition of hyperdimensional computing (see refs. 31–33), systems of high-dimensional sparse vectors equipped with algebraic operations, typically component-wise addition, component-wise multiplication, and permutation of dimensions. Indeed, an assembly can be thought of as a high-dimensional vector, namely, the characteristic vector of its support; but this is where the similarity ends. While assembly operations as introduced here are meant to model and predict cognitive function in the brain, hyperdimensional computing is a machine learning framework—inspired, of course, by the brain, like many other such frameworks—and used successfully, for example, for learning semantics in natural language (33). In sparse vector systems, there is no underlying synaptic connectivity between the dimensions, or partition of the dimensions into brain areas. Finally, the operations of the Assembly Calculus (project, associate, and merge) are very different in nature, style, detail, and intent from the operations in sparse vector systems (add, multiply, and permute).
Assembly computation is closer in spirit to Valiant’s model of neuroids and items (34), which was an important source of inspiration for this work. One difference is that, whereas, in Valiant’s model, the neuron (called neuroid there) has considerable local computational power—for example, to set its presynaptic weights to arbitrary values—in our formalism, computational power comes from a minimalistic model of inhibition and plasticity; both models assume random connectivity. Another important difference is that, in contrast to an item in Valiant’s theory, an assembly is densely intraconnected, while the mechanism for its creation is described explicitly.
Results and Methods
Projection.
The operation entails activating, repeatedly, assembly while is disinhibited. Such repeated activation creates, in the disinhibited area , a sequence of sets of cells, let us call them . The mathematical details are quite involved, but the intuition is the following: Cells in can be thought of as competing for synaptic input. At the first step, only provides synaptic input, and thus consists of the cells in which happen to have the highest sum of synaptic weights originating in —note that these weights are subsequently increased by a factor of due to plasticity. At the second step, neurons in both and spike, and, as a result, a new set of “winners” from among cells of is selected; for typical parameters, overlaps heavily with . This continues as long as keeps firing, with certain cells in replaced by either “new winners”—cells that never participated in a with —or by “old winners”—cells that did participate in some with . We say that the process has converged when there are no new winners. Upon further firing of , may evolve further slowly, or cycle periodically, with past winners coming in and out of ; in fact, this mode of assembly firing (cells of the assembly alternating in firing) is very much in accordance with how assemblies have been observed to fire in Ca+ imaging experiments in mice; see, for example, ref. 35. It is theoretically possible that a new winner cell may come up after convergence; but it can be proved that this is a highly unlikely event, and we have never noticed it in our simulations. The number of steps required for convergence depends on the parameters, but, most crucially, on the plasticity coefficient ; this dependence is fathomed analytically and through simulations in Fig. 2B4.
Density.
It was postulated by Hebb (4) that assemblies are densely interconnected—presumably such density was thought to cause their synchronized firing. Since assemblies are created by projection, increased density is intuitively expected: Cells in are selected to be highly connected to , which is increasingly closer to as increases. It is also observed in our simulations (Fig. 2C) and predicted analytically in our model.
Association and Pattern Completion.
Our simulations (Fig. 2E), as well as analytical results, show that the overlap of two assemblies in the same area increases substantially in response to simultaneous firing of the two parent assemblies and (assumed to be in two different areas). The amount of postassociation overlap observed in our simulations is compatible with the estimates in the literature (36, 37), and increases with the extent of cooccurrence (the number of consecutive simultaneous activations of the two parents).
Since association between assemblies is thought to capture affinity of various sorts, the question arises: If two associated assemblies and are both projected in another area, will the size of their overlap be preserved? Our results, both analytical and through simulations, strongly suggest that assembly overlap is indeed conserved reasonably well under projection (see Section 2 and Fig. 2). This is important, because it means that the signal of affinity between two assemblies is not lost when the two assemblies are further projected in a new brain area.
Pattern completion involves the firing, with some probability, of an assembly following the firing of a few of its neurons (8, 35). Simulations in our model (Fig. 2F2) show that, indeed, if the creation of an assembly is completed with a sufficient number of activations of its parent, then firing fewer than half of the neurons of the assembly will result in most, or essentially all, of the assembly firing.
Reciprocal Projection and Merge.
Reciprocal projection, denoted , is a more elaborate version of projection. It involves firing, which causes in area to fire, which, in turn, causes, as with ordinary projection, a set in to fire. The difference is that now there is synaptic connectivity from area to area (in addition to connectivity from to ), which causes, in the next step, to move slightly to a new assembly , while has become . This continuing interaction between the and the eventually converges, albeit slower than with ordinary projection, and under conditions of ampler synaptic connectivity and plasticity. The resulting assembly has strong synaptic connectivity both to and from (instead of only from to , as is the case with ordinary projection). That reciprocal projection works as described above has been shown both analytically and through simulations in our model, as well as in simulations in a more realistic neural model with explicit inhibitory neurons and STDP in ref. 26.
The operation is essentially a double reciprocal projection. It involves the simultaneous repeated firing, in different areas, of the parents of both and , which causes the simultaneous repeated firing, also in two different areas and , of and . In the presence of enabled afferent two-way connectivity between and , and also between and , this initiates a process whereby a new assembly is eventually formed in area , which, through its firing, modifies the original assemblies and . In the resulting assemblies, there is strong two-way synaptic connectivity between and , as well as between and . Analytical and simulation results are similar to those for reciprocal projection (Fig. 2G1).
Simulations.
We gain insights into the workings of our model, and validate our analytical results, through simulations. In a typical simulation task, we need to simulate a number of discrete time steps in two or three areas, in a random graph with nodes (where the notation means ”a small constant multiple of ). Creating this graph requires computation. Next, simulating each firing step entails selecting, in each area, the out of cells that receive that largest synaptic input. This takes computations per step. Since the number of steps needed for convergence is typically much smaller than the ratio , the computation for the creation of the graph dominates the computational requirements of the whole simulation (recall that ).
In our simulator, we employ a maneuver which reduces this requirement to , enabling simulations of the required scale on a laptop. The trick is this: We do not generate all edges of the random graph a priori, but generate them “on demand” as the simulation proceeds. Once we know which cells fired at the previous step, we generate the cells in the area of interest that receive the most synaptic input, as well as their incoming synapses from the firing cells, by sampling from the tail of the appropriate binomial distribution; we then compare with previously generated cells in the same area to select the cells that will fire next.
Our simulation system is available online, making it possible to reproduce and extend our simulation results: http://github.com/dmitropolsky/assemblies.
Discussion
We have defined a formal system intended to model the computations underlying cognitive functions. Its elementary object is an assembly, a set of excitatory neurons all belonging to the same brain area, and capable of near-simultaneous firing. The operations of this system enable an assembly to create a “copy” of itself in a new brain area through projection, two assemblies to increase their overlap in order to reflect observed cooccurrence in the world or other similarity, and, furthermore, two assemblies to create a new assembly with ample connectivity to and from the two original ones. These operations do not rely on preexisting neural circuits; instead, their apparatus is generated on-the-fly, with high probability, through plasticity and randomness of synaptic connectivity. The resulting formal system, equipped with certain simple control operations, is fraught with considerable computational power. Central to our work is the speculation that something akin to this formal system may underlie cognitive functions, especially in the human brain.
What is an assembly, exactly? The precise definition is a conundrum. In our narrative so far, assembly is simply a set of excitatory neurons in a brain area capable of synchronous firing. According to Hebb’s prediction (4) as well as current observations (8) and neurorealistic simulations (18), the neurons of an assembly are not all activated simultaneously, but instead fire in a “pattern” with different neurons firing at different times.‡ In our formalism, and in our analytical proofs and our experiments, the assumption of discrete time steps suppresses, to a large extent, this sequential behavior. But, even in our model, once the projection has stabilized (no new neurons in area fire that had never fired before), a small number of neurons often keep coming in and out of the spiking set . One possible principled definition of an assembly is this: An assembly in an area is a sparse distribution on the neurons of area , whose support (the set of neurons with nonzero probability) is not much larger than .
One aspect of our assembly operations is the crucial role of plasticity. While plasticity is well studied as an important aspect of the way the brain works and learning happens, we are not aware of many models in which plasticity takes place at the subsecond time scale, as it is hypothesized to do in assembly operations. Our use of assemblies as the basis of a computational system also departs from the usual discourse on assemblies, typically considered, implicitly, as fairly stable representations. In contrast, here we conjecture that assemblies can be formed and modified by the human brain at 4 Hz, the frequency of syllables in language (see the discussion on language below).
Our results and simulations assume uniformly random synaptic connectivity; however, experimental measurements (11, 12) suggest a departure from uniformity (but not from randomness). Our analytical results can be extended routinely to nonuniform random synaptic connectivity of this kind. In fact, our conclusions regarding the density and stability of assemblies can be further strengthened in such regimes. For example, suppose that, conditioned on the existence of two directed edges and , the presence of edge is much more likely, as concluded in ref. 11. Depending on the precise values of the parameters , this would likely trigger a “birthday paradox” phenomenon (existence of two assembly cells with synaptic connections from the same cell of the parent assembly) that would further enhance the synaptic density, and hence the stability, of assemblies.
The basic operations of the Assembly Calculus as presented here—projection, association, reciprocal projection, and merge—correspond to neural population events which 1) are plausible, in the sense that they can be reproduced in simulations and predicted by mathematical analysis, and 2) provide parsimonious explanations of experimental results (for the merge and reciprocal project operations, see the discussion of language below). In contrast, the read and control operations—read, fire, disinhibit—however simple and elementary, lack in such justification, and were added for the purpose of rendering the Assembly Calculus a programmable computational system. Replacing them with more biologically plausible control operations leading to the same effect would be very interesting.
Assemblies and Language.
We hypothesized that assemblies and their operations may be involved in mediating higher cognitive functions in humans. In ending, we speculate below on how assemblies may be implicated in language.
Linguists have long argued that the language faculty is unique to humans, and that the human brain’s capability for syntax (i.e., for mastering the rules that govern the structure of sentences in a language) lies at the core of this ability (23, 24). In particular, the linguistic primitive Merge has been proposed as a minimalistic basis of syntax. Merge combines two syntactic units to form one new syntactic unit; recursive applications of Merge can create a syntax tree, which captures the structure of a sentence (Fig. 3A). If this theory is correct, then which are the brain correlates of syntax and Merge? Recent experiments provide some clues; see ref. 38 for an extremely comprehensive and informative recent treatment of the subject.
1)
Two different brain areas of the left superior temporal gyrus (which contains Wernicke’s area, known since the 19th century to be implicated in the use of words in language) seem to respond to the subject vs. the object of a presented sentence (39).
2)
The completion of phrases and sentences presented is marked by activity in Broca’s area in the left hemisphere (40), known to be implicated in syntax.
3)
A sequence of sentences consisting of four monosyllabic words (such as “Bad cats eat fish”) presented at the rate of four words per second (the natural frequency of articulated syllables) elicits a pattern of brain responses with energy peaks at 1, 2, and 4 Hz, consistent with the hypothesis that syntax trees for the sentences are being constructed during the presentation (41).
Fig. 3.

If one accepts the hypothesis that, indeed, something akin to syntax trees is constructed in our brain during the parsing—and presumably also during the generation—of sentences, one must next ask, How is this accomplished? According to ref. 38, chapter 4, these and a plethora of other experimental results point to a functional cortical network for lexical and syntactic processing, involving the MTL, Broca’s areas BA 44 an BA 45, and Wernicke’s area in the superior temporal gyrus, as well as axon fibers connecting these four areas, all in the left hemisphere. Syntactic processing of language seems to entail a complex sequence of activations of these areas and transmission through these fibers. How is this orchestrated sequence of events carried out? Does it involve the generation and processing, on-the-fly, of representations of the constituents of language (words, phrases, sentences)? Angela Friederici (ref. 38, p. 134) proposes that, “for language there are what I call mirror neural ensembles through which two distinct brain regions communicate with each other.” Could it be that assemblies and their operations play this role?
We propose a dynamic brain architecture for the generation of a simple sentence, powered by assemblies and the operations and , and consistent with the experimental results and their interpretations outlined above (Fig. 3). In particular, we propose that the construction of the syntax tree of a simple sentence being generated by a subject can be implemented by the following program of the Assembly Calculus:
do in parallel:
find-verb,
find-subj,
find-obj;
do in parallel:
,
,
;
merge ;
merge .
The generation of a sentence such as “The boy kicks the ball” starts with a desire by the subject to assemble—and possibly articulate—a particular fact. The raw fact to be put together is denoted here by —an image, sensed or mental. In the first line, a brain area containing the lexicon is searched in order to identify the verb (the action in the fact relayed in ), the subject (the agent of the fact), and the object (the patient of the fact). Here we are assuming that the lexicon is a searchable collection of tens of thousands of assemblies in the left MTL of the subject, representing words in the language; this is in rough agreement with current views (42), even though much still remains to be understood about the representation of words and the associated information. We also assume that the corresponding brain functions find-verb, etc., are already in place. For the example “The boy kicks the ball,” the three assemblies , , and are identified after the first step, encoding the words “kick”, “boy”, and “ball”, respectively. In the second line, these three assemblies are projected to the three subareas of Wernicke’s area specializing in the verb, subject, and object of sentences, respectively. In fact, instead of ordinary projection, the operation is used, as first proposed in ref. 26, for reasons that will become clear soon. Next, an assembly is formed in the pars opercularis of Broca’s area (BA 44) representing the verb phrase “kicks the ball” through the merge operation applied to the two assemblies and encoding the constituent words of the phrase in Wernicke’s area. Finally, an assembly corresponding to the whole sentence is formed in the pars triangularis of Broca’s area (BA 45) via the merge of assemblies and , completing the construction of the rudimentary syntax tree of the whole sentence.
The Assembly Calculus program above accounts only for the first phase of sentence production, during which the syntax tree of the sentence is constructed. Next, the sentence formed may be articulated, and we can speculate on how this process is carried out: Assembly is activated, and this eventually causes the assemblies —the leaves of the syntax tree—to fire. The activation of these three assemblies is done in the order specific to the speaker’s language learned at infancy (for example, in English, subject–verb–object). Note that the first phase of building the syntax tree was largely language independent. Eventually, the lexicon assemblies will be activated in the correct order (this activation was the purpose of using the operation), and these, in turn, will activate the appropriate motor functions which will ultimately translate the word sequence into sounds.
The above narrative is only about the building of the basic syntactic structure—the “scaffold”—of extremely simple sentences, and does not account for many other facets: how the three find- tasks in the first line are implemented; how the noun phrases are adorned by determiners such as “the,” and how the verb is modified to reflect person and tense (“kicks” or “kicked”), and the important inverse processes of parsing and comprehension, among a myriad of other aspects of language that remain a mystery.
Data Availability
Data deposition: Our simulation system is available online, making it possible to reproduce and extend our simulation results: http://github.com/dmitropolsky/assemblies.
Acknowledgments
We are grateful to Larry Abbott for his insightful ideas and comments concerning this work. We also wish to thank several colleagues for many helpful discussions and useful feedback: Richard Axel, Bob Berwick, Costis Daskalakis, Boris Harizanov, Pentti Kanerva, Robert Legenstein, Chris Manning, Bruno Olshausen, Les Valiant, and Rafa Yuste. C.H.P.’s research was partially supported by NSF Awards CCF1763970 and CCF1910700, and by a research contract with Softbank; S.S.V.’s research was partially supported by NSF Awards 1717349, 1839323, and 1909756; and W.M.’s research was partially supported by the Human Brain Project Grant 785907 of the European Union.
Supporting Information
Appendix (PDF)
- Download
- 342.54 KB
References
1
R. Axel, Q & A. Neuron 99, 1110–1112 (2018).
2
S. T. Piantadosi, J. B. Tenenbaum, N. D. Goodman, Bootstrapping in a language of thought: A formal model of numerical concept learning. Cognition 123, 199–217 (2012).
3
S. T. Piantadosi, J. B. Tenenbaum, N. D. Goodman, The logical primitives of thought: Empirical foundations for compositional cognitive models. Psychol. Rev. 123, 392–424 (2016).
4
D. O. Hebb, The Organization of Behavior: A Neuropsychological Theory (Wiley, New York, NY, 1949).
5
M. Abeles, Corticonics: Neural Circuits of the Cerebral Cortex (Cambridge University Press, 1991).
6
K. D. Harris, Neural signatures of cell assembly organization. Nat. Rev. Neurosci. 6, 399–407 (2005).
7
G. Buzsaki, Neural syntax: Cell assemblies, synapsembles, and readers. Neuron 68, 362–385 (2010).
8
J.-e. K. Miller, I. Ayzenshtat, L. Carrillo-Reid, R. Yuste, Visual stimuli recruit intrinsically generated cortical ensembles. Proc. Natl. Acad. Sci. U.S.A. 111, E4053–E4061 (2014).
9
G. Buzsáki, The Brain from Inside Out (Oxford University Press, 2019).
10
P. Erdős, A. Renyi, On the evolution of random graphs. Publ. Math. Inst. Hungary. Acad. Sci. 5, 17–61 (1960).
11
S. Song, P. J. Sjöström, M. Reigl, S. Nelson, D. B. Chklovskii, Highly nonrandom features of synaptic connectivity in local cortical circuits. PLoS Biol. 3, e68 (2005).
12
S. Guzman, A. J. Schlö gl, M. Frotscher, P. Jonas, Synaptic mechanisms of pattern completion in the hippocampal ca3 network. Science 353, 1117–1123 (2016).
13
R. Quian Quiroga, Neuronal codes for visual perception and memory. Neuropsychologia 83, 227–241 (2016).
14
K. M. Franks et al., Recurrent circuitry dynamically shapes the activation of piriform cortex. Neuron 72, 49–56 (2011).
15
M. J. Ison, R. Quian Quiroga, I. Fried. Rapid encoding of new memories by individual neurons in the human brain. Neuron 87, 220–230 (2015).
16
R. Legenstein, W. Maass, C. H. Papadimitriou, S. S. Vempala, “Long-term memory and the densest k-subgraph problem” in Proceedings of 9th Innovations in Theoretical Computer Science (ITCS) Conference, A. R. Karlin, Ed. (Massachusetts Institute of Technology, Cambridge, MA, 2018) pp. 57:1–57:15.
17
C. H. Papadimitriou, S. S. Vempala, “Random projection in the brain and computation with assemblies of neurons” in 10th Innovations in Theoretical Computer Science Conference, ITCS 2019, A. Blum, Ed. (Leibniz-Zentrum fuer Informatik, 2019), pp. 57:1–57:19.
18
C. Pokorny et al., Associations between memory traces emerge in a generic neural circuit model through STDP. bioRxiv:188938 (14 September 2017).
19
R. Legenstein, C. H. Papadimitriou, S. Vempala, W. Maass, Assembly pointers for variable binding in networks of spiking neurons. arXiv:1611.03698 (11 November 2016).
20
U. Feige, D. Peleg, K. Guy, The dense k-subgraph problem. Algorithmica 29, 410–421 (2001).
21
S. Dasgupta, C. F. Stevens, S. Navlakha, A neural algorithm for a fundamental computing problem. Science 358, 793–796 (2017).
22
N. Anari et al., “Smoothed analysis of discrete tensor decomposition and assemblies of neurons” in Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, S. Bengio, H. M. Wallach, Eds. (Curran Associates, 2018), pp. 10880–10890.
23
R. C. Berwick, N. Chomsky, Why Only Us: Language and Evolution (MIT Press, 2016).
24
M. D. Hauser, N. Chomsky, W. T. Fitch, The faculty of language: What is it, who has it, and how did it evolve?Science 298, 1569–1579 (2002).
25
E. Zaccarella, L. Meyer, M. Makuuchi, A. D. Friederici, Building by syntax: The neural basis of minimal linguistic structures. Cerebr. Cortex 27, 411–421 (2017).
26
R. Legenstein, M. G. Müller, C. H. Papadimitriou, S. Vempala, W. Maass, Assembly projections support the assignment of thematic roles to concepts in networks of spiking neurons. arXiv:1611.03698 (11 November 2016).
27
K. D. Harris, G. M. G. Shepherd, The neocortical circuit: Themes and variations. Nat. Neurosci. 18, 170–181 (2015).
28
C. H. Papadimitriou, Computational Complexity (Wiley, 2003).
29
H. Eichenbaum, Barlow versus Hebb: When is it time to abandon the notion of feature detectors and adopt the cell assembly as the unit of cognition?Neurosci. Lett. 680, 88–93 (2018).
30
J. Ranhel, Neural assembly computing. IEEE Trans. Neural Netw. Learn. Syst. 23, 916–927 (2012).
31
T. Plate, “Holographic reduced representations: Convolution algebra for compositional distributed representations” in Proceedings of the 12th International Joint Conference on Artificial Intelligence (IJCAI), J. Mylopoulos, R. Reiter, Eds. (Morgan Kaufman, 1991), pp. 30–35.
32
P. Kanerva, Hyperdimensional computing: An introduction to computing in distributed representation with high-dimensional random vectors. Cogn. Comput. 1, 139–159 (2009).
33
J. Karlgren, P. Kanerva, High-dimensional distributed semantic spaces for utterances. Nat. Lang. Eng. 25, 503–517 (2019).
34
L. G. Valiant, Circuits of the Mind (Oxford University Press, 1994).
35
L. Carrillo-Reid, S. Han, W. Yang, A. Akrouh, R. Yuste, Triggering visually-guided behavior by holographic activation of pattern completion neurons in cortical ensembles. bioRxiv:394999 (17 August 2018).
36
H. G. Rey et al., Encoding of long-term associations through neural unitization in the human medial temporal lobe. Nat. Commun. 9, 4372 (2018).
37
E. De Falco, M. J. Ison, I. Fried, R. Quian Quiroga, Long-term coding of personal and universal associations underlying the memory web in the human brain. Nat. Commun. 7, 13408 (2016).
38
A. D. Friederici, Language in Our Brain: The Origins of a Uniquely Human Capacity (MIT Press, 2017).
39
S. M. Frankland, J. D. Greene, An architecture for encoding sentence meaning in left mid-superior temporal cortex. Proc. Natl. Acad. Sci. U.S.A. 112, 11732–11737 (2015).
40
E. Zaccarella, A. D. Friederici, Merge in the human brain: A sub-region based functional investigation in the left pars opercularis. Front. Psychol. 6, 1818 (2015).
41
N. Ding, L. Melloni, H. Zhang, X. Tian, D. Poeppel, Cortical tracking of hierarchical linguistic structures in connected speech. Nat. Neurosci. 19, 158–164 (2016).
42
E. F. Lau, C. Phillips, D. Poeppel, A cortical network for semantics:(de) constructing the n400. Nat. Rev. Neurosci. 9, 920–933 (2008).
Information & Authors
Information
Published in
Classifications
Copyright
Copyright © 2020 the Author(s). Published by PNAS. This open access article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND).
Data Availability
Data deposition: Our simulation system is available online, making it possible to reproduce and extend our simulation results: http://github.com/dmitropolsky/assemblies.
Submission history
Published online: June 9, 2020
Published in issue: June 23, 2020
Keywords
Acknowledgments
We are grateful to Larry Abbott for his insightful ideas and comments concerning this work. We also wish to thank several colleagues for many helpful discussions and useful feedback: Richard Axel, Bob Berwick, Costis Daskalakis, Boris Harizanov, Pentti Kanerva, Robert Legenstein, Chris Manning, Bruno Olshausen, Les Valiant, and Rafa Yuste. C.H.P.’s research was partially supported by NSF Awards CCF1763970 and CCF1910700, and by a research contract with Softbank; S.S.V.’s research was partially supported by NSF Awards 1717349, 1839323, and 1909756; and W.M.’s research was partially supported by the Human Brain Project Grant 785907 of the European Union.
Notes
*Our assumption that all areas contain the same number of neurons is one of expository, not mathematical, convenience, and provides the basic numerical parameter of our analysis. A slightly more complex version would have parameters and for each area .
†
Several experimental results (11, 12) suggest deviations of the synaptic connectivity of the animal brain from the uniformity of . We discuss how such deviations affect—mostly support—our framework in Discussion.
‡
In musical metaphor, an assembly is thought to be not an octave but a cord sequence.
Authors
Competing Interests
The authors declare no competing interest.
Metrics & Citations
Metrics
Altmetrics
Citations
Cite this article
Brain computation by assemblies of neurons, Proc. Natl. Acad. Sci. U.S.A.
117 (25) 14464-14472,
https://doi.org/10.1073/pnas.2001893117
(2020).
Copied!
Copying failed.
Export the article citation data by selecting a format from the list below and clicking Export.
Cited by
Loading...
View Options
View options
PDF format
Download this article as a PDF file
DOWNLOAD PDFLogin options
Check if you have access through your login credentials or your institution to get full access on this article.
Personal login Institutional LoginRecommend to a librarian
Recommend PNAS to a LibrarianPurchase options
Purchase this article to access the full text.