Abstract
Introduction
The identification of a suitable theoretical framework and appropriate engineering principles for collective intelligence are open problems. In this paper, we begin to address these gaps by developing a synthesis of perspectives usually considered to be quite distinct. To do this, we first dissolve a number of limiting misconceptions that cause collective intelligence and individual intelligence to be treated as separate topics; second, we introduce a speculative conceptual framework to unify them.
For an intelligence to belong properly to a collective, it must arise not from the cleverness of its members but from having the right kind of functional relationships between them. What kinds of functional relationships, and in what specific organisation, are required to turn a collective that is not intelligent into a collective that is? We use a specific understanding of cognition and learning that is already well-developed for individual intelligence to synthesise collective intelligence with aspects of development and evolution. In particular, we explore how connectionist models of cognition and learning, familiar in neural network models of
Individual and collective intelligence are distinct phenomena. Or are they?
At first glance, it might seem that models of individual intelligence are not relevant to collective intelligence. Individuals have brains that can cognise and learn, and although colonies and swarms might be composed of individuals with brains, the collective
All individuals are collectives
All individuals are collectives, made of parts that used to be individuals themselves. This is true not only for multicellular organisms derived from unicellular ancestors but also for eukaryotic cells with multiple organelles arising from bacterial ancestors, and for simpler cells that contain the first chromosomes arising from the union of previously free-living self-replicating molecules (Godfrey-Smith, 2009; Maynard Smith and Szathmáry, 1997; Michod, 2000; Okasha, 2006; West et al., 2015). Moreover, the proper functioning of organisms – their robustness, adaptability and evolvability – depends on the continued autonomy of their component parts (Levin, 2019; 2021a). Multicellular organisms exhibit multi-scale autonomy, a dynamic interplay of competition and cooperation, and coordinated collective action inherent to their development, function and behaviour, while being a society of cells (Fields and Levin, 2022; Levin, 2019; 2022; 2023; Sonnenschein and Soto, 1999). Thus, individuals like you and I, and collectives like swarms and colonies, are not as categorically different as they first appear.
All intelligences are collectives
Individual intelligence, in the familiar guise of a central nervous system or a brain, arises from the interaction of many unintelligent components (neurons) arranged in the right organisation with the right connections. This is the foundation of
Cognition and learning are substrate-independent
The principles of distributed cognition familiar in artificial neural networks can be implemented by any network of signals and non-linear responses to suitably weighted inputs (Evans et al., 2022; Stern and Murugan, 2022; Watson et al., 2016). Gene-regulation networks, ecological networks and social networks can all compute in the same sense as neural networks if the connections are suitably arranged (Biswas et al., 2021; Davies et al., 2011; Herrera-Delgado et al., 2018; Power et al., 2015; Szabó et al., 2012; Tareen and Kinney, 2020; Watson et al., 2014). In development and organismic biology, many different levels of adaptive networks exist aside from neural networks, including gene-regulation networks, protein networks, metabolic networks, morphogen diffusion networks and endocrine systems. In addition, it is clear that morphogenesis, physiological function and the adaptive processes of robustness and repair all require information integration and collective action that constitute cognition – in many cases without neurons. Each of these phenomena exhibits the same learning behaviours, including the storage and retrieval of multiple associative memories, effecting classification and recognition with generalisation capabilities, and learning to solve combinatorial optimisation problems better with experience (Watson et al., 2011a; Watson et al., 2011b, 2011c).
The credit assignment problems inherent in collective intelligence are fundamental in all cognition and learning, and in all biological individuality
It is true that collective intelligence is fundamentally about
Towards a unified theory of intelligence and cognition
In collectives, each component selects behaviours based on the rewards they receive for their own actions (Figure 1(a)). In intelligent systems, the reward feedback is effectively operating at a higher level – and the system as a whole selects behaviours based on the rewards received by the system as a whole (Figure 1(b)). Accordingly, it makes sense that the system selects behaviours that facilitate long-term collective reward. But operationally, each component within the intelligent system is still autonomous, selecting individual actions based on individual rewards given the relational context they find themselves in. The question is, Perspectives on individual intelligence and collective intelligence. Complex systems are composed of many interacting components. But where is the agent – at the component level or the system level? (a) Swarms are often characterised as collectives, but the agency (reward feedback and decision making) is generally attributed to each component. These are obviously collectives but not obviously intelligent. (b) Animal intelligence is often characterised as a single, system-level agent (exhibiting information integration and collective action), but the components are generally considered to be 'parts' without agency. These systems are obviously intelligent but not obviously collectives. (c) In reality, all intelligences are made out of components that act on local information based on individual feedbacks. In a multicellular organism, for example, the individual cells exhibit agency based on local information and rewards, and the system (cellular swarm) as a whole does also, exhibiting information integration and anatomical decision making at the system-scale.
Establishing these commonalities has significant consequences for understanding: since some of these questions have well-developed answers in the context of individual intelligences, those answers can be transferred to provide a framework for approaching collective intelligences. While connectionist models of cognition and learning do not have all the answers, they do identify the kind of relationships that turn a collection of unintelligent components into a collective intelligence, with cognitive and learning abilities that belong to the whole and not the parts. Additionally, connectionist models identify conditions where collective intelligence can arise bottom-up, using only distributed learning mechanisms without system-level or global feedback.
We do not attempt a comprehensive review of the many related topics involved. Rather, we have selected foundational points to clarify a vision of distributed cognition, bottom-up adaptation and, more generally, the ‘more than the sum of the parts’ conceptual territory.
A framework for interrogating collective intelligence
Our thinking builds on a core conjecture that the kind of relationships necessary to produce evolutionary individuality – the generation and heritability of fitness differences at the collective level (Watson et al., 2022) – are the same as those required to produce organismic individuality – the information integration and collective action characteristic of a self (Levin, 2019, 2022a). More specifically, we propose that these relationships are
Understanding the parallels between individual and collective intelligence via a connectionist framework
The curious thing about collective intelligence is that the more intelligent something is, the less it looks like a collective. When component members act in an efficiently coordinated manner, with behaviours that can be diverted from their short-term self-interest to serve long-term collective interest, a collective looks more like an individual at a higher level of organisation. Recognising this, biology is full of collective intelligence – not just in the weak sense of swarms and colonies with emergent behaviours but because any organism is actually an intelligent collective. Here we look at (1) collective behaviours and functions within organisms, especially development and the idea of ‘basal cognition’; (2) the complex relationship between organismal identity and evolutionary units, and how this has changed over evolutionary time; (3) the substrate-independence of intelligence and how cognition and learning can occur in various kinds of biological networks; and briefly, (4) how the problem of credit assignment arises as a core theme in these issues.
Organisms as collective intelligences: Development and basal cognition
All organisms are collectives at multiple levels: from collections of active molecules in a cell, to collections of cells in a multicellular organism or a tissue, to collections of tissues in an individual organism. What makes collectives individuals (as opposed to merely populations in containers) is their intelligence – their degree of competency in solving novel problems (Fields and Levin, 2022; Levin, 2023). The processes of development are the substrate of this intelligence – the ‘glue’ that makes the whole more than the sum of the parts (Levin, 2019). A large body of work indicates that development is not well-characterised as the execution of a pre-programmed genetic script but rather as an active, dynamic and adaptive process. Although all cells in most multicellular organisms share the same genome, the remarkable protein machinery that genome encodes, along with the cytoskeletal and lipid structures each cell inherits from its ancestors (Fields and Levin, 2018b), enables a collective of embryonic cells to develop differentiated roles and self-organise into a large-scale, functional machine. Development thus involves a multi-scale hierarchy of cooperating and competing subunits (Fields and Levin, 2020), each with local computational and goal-directed capacity, that enables the whole to function as a singular subject of memories and preferences – a unified locus of learning and homeostatic loops that harness its subunits towards goal states.
Morphogenesis as an instantiation of collective intelligence
Anatomical homeostasis – the ability to adjust anatomy despite injury or drastic rearrangement (Harris, 2018; Levin et al., 2019) – requires the collective to have a degree of autonomous problem-solving activity in morphospace, defined as the space of possible anatomical configurations (Stone, 1997). For example, eyes developed ectopically in the tails of frog embryos still allow the animals to see (Blackiston and Levin, 2013) because the eye primordia cells succeed not only in forming an eye and optic nerve in an abnormal environment but also in connecting the optic nerve to the nervous system (in this case, via synapse onto the spinal cord, rather than the brain). Another example is the development of the newt kidney tubule (Fankhauser, 1945a; 1945b): normally cell–cell communication among ∼8 cells produces the correct tubule diameter, but if the cells are made very large, they still produce the same diameter tubule by using fewer cells. Even when cell size gets very large, a single cell can achieve the same diameter tubule by bending around itself (this time using cytoskeletal mechanisms). Thus, genetically wild-type cells can harness distinct molecular components, depending on the novel circumstances, to reach the same high-level anatomical goal.
This disrupts a straightforward reductionist or bottom-up account of organismal morphology and function. Whilst natural selection provides the genetic hardware, this hardware has a very particular kind of plasticity, which implements robustness to both external and internal novelty. This derives from an architecture of multi-scale competency (Fields and Levin, 2020; Gawne et al., 2020), where many subsystems are themselves goal-directed and can pursue specific endpoints despite changes in their tissue environment, greatly potentiating evolvability. The idea of organisms as pre-specified machines, assembled by genetic scripts, fails in the context of these and other examples of developmental robustness. We therefore seek to understand these capacities in the context of a different and more flexible conceptual space.
Basal cognition in development: Morphological problem-solving
‘Basal cognition’ refers to information processing that occurs in an unconventional substrate and/or as a simpler evolutionary precursor to what we conventionally consider cognition (Baluška and Levin, 2016; Levin, 2019; Manicka and Levin, 2019a). This is not cognition that depends on neurons or necessarily involves second-order self-awareness (Levin, 2019). It refers to cognition in an algorithmic sense that is substrate-independent (Levin, 2019) and is observable as problem-solving across phylogenetic history (Keijzer et al., 2013; Levin et al., 2021; Lyon, 2015; Lyon et al., 2021). What is important in basal cognition is not the presence of neurons but the presence of functional and informational interactions that facilitate both information integration and the ability to orchestrate cued responses that coordinate action (Bechtel and Bich, 2021; Grossberg, 1978; Levin, 2019). This can be implemented by suitable interactions of any nature including gene regulatory networks, cell signalling, bio-electric networks and morphogenetic chemical feedbacks (Lyon et al., 2021).
For example, the process of growing a limb constitutes basal cognition, as it requires both integration of multi-dimensional information (e.g. to ‘decide’ appendage type or handedness, from context) and collective action to put this ‘basal decision’ (Bechtel and Bich, 2021) into action (e.g. to coordinate the timing, abundance and positioning of cellular differentiation and growth (Dinet et al., 2021; Fields and Levin, 2020; Moczek, 2019). More broadly, regulative development, regeneration and remodelling (such as morphogenesis) require collective decision making and memory at two scales: on the part of cells (collectives of molecular networks) and of tissues (collectives of cells). Limb regeneration, for example, requires a memory of the correct pattern, the ability to compare current state with the target state and the ability to traverse anatomical morphospace in different ways depending on context and perturbations (Pezzulo and Levin, 2016).
William James’ definition of intelligence – the ability to achieve the same goal in multiple ways (James, 1890) – provides context for considering the basal intelligence of cell collectives in morphogenesis. It has become clear that the large-scale morphological goals of an organism override and harness the local competencies of individual cells to adaptively navigate morphospace (Levin, 2022a). That navigation capacity is not hardwired but shows considerable problem-solving plasticity (reviewed in (Levin, 2023)). Numerous examples indicate that morphogenesis meets James’s definition of intelligence by achieving normal anatomy despite a wide range of serious perturbations. For example, developing Xenopus tadpoles can attain the same anatomical outcome despite starting with their craniofacial organs scrambled (Vandenberg et al., 2012) or with the wrong number of cells (Cooke, 1979, 1981). Even mammalian embryos can overcome drastic perturbations such as amputation; and early embryo splitting in humans results in normal monozygotic twins rather than partial bodies.
The ability of collectives of cells to pursue, with various degrees of competency, target states in anatomical morphospace (Levin, 2023; Stone, 1997) reveals an important aspect of being an individual: solving problems in a space different from that occupied by its parts (Fields and Levin, 2022; Levin, 2023). While individual cells cannot ascertain the right number, size or position of eyes or fingers, tissues do so routinely, that is, the tissue as a collective executes morphogenesis through differential cell reproduction and differentiation, stopping when the correct structure is complete (Birnbaum and Sánchez Alvarado, 2008). While cells navigate transcriptional and metabolic spaces, cellular collectives can navigate anatomical morphospaces and the conventional behavioural space (Fields and Levin, 2022).
Altered states: Basal cognition and manipulated target morphology
This framework makes a strong prediction: if intercellular signalling (not genes) is the cognitive medium of a morphogenetic individual, it should be possible to exploit the tools of behavioural and neuro-science and learn to read, interpret and re-write its information content in a way that allows predictive control over its behaviour (in this case, growth and form) without genetic changes. This prediction has been validated in several species. The bioelectric signatures that drive accurate regenerative reproduction/development in planaria have been identified (‘reading and interpreting’ anatomical target information, Durant et al., 2016; Durant et al., 2017; Pezzulo et al., 2021). Planaria normally have one head, but this is not genetically determined, merely a default: transient bioelectrical modulation of the body-wide pattern memory circuits can shift them to a persistent two-headed state, causing subsequent pieces of that planarian to regenerate into two-headed worms (‘re-writing’) (Durant et al., 2016). This induced phenotype then persists through future rounds of amputation until set back to normal with a different bioelectrical manipulation (Durant et al., 2017); it even exhibits features of advanced individual cognition such as bi-stability (Pezzulo et al., 2021). These target morphology shifts occur despite the fact that all of the individual cells have unaltered normal genomes, showing that competent subunits can be pushed to implement diverse organism-scale goals by physiological signals (experiences) without modification of their essential hardware. In addition, this can happen rapidly – not requiring evolutionary timeframes. Other examples of reading, interpreting and rewriting the bioelectric information dictating morphogenesis have been described in a range of model systems (Levin, 2021b). Consistent with the idea that cellular swarms can act as a consolidated cognitive agent, morphogenesis is known to be altered by prior experiences (e.g. amphibian limbs ceasing to regenerate after repeated amputation (Bryant et al., 2017)) and confused by exposure to classic cognitive modifier drugs (Sullivan and Levin, 2016).
Bioelectricity: A ‘cognitive glue’ common to collective and individual intelligence
The many parallels between behavioural control by nervous systems, and the ancestral capacity of morphogenetic control by all cell networks (Fields et al., 2020), are reviewed elsewhere (Pezzulo and Levin, 2015). But it’s crucial to note that the very same cognitive glue – bioelectrical networks implemented by ion channels and electrical synapses – operates to bind neurons into competent individuals in the 3D world of behaviour and to bind other cell types into competent individuals in the morphogenetic space of anatomical control. These insights are now driving computational models used to understand the tissue-level decision making that results in birth defects (Manicka and Levin, 2019b, 2022) and their repair (Pai et al., 2018; Pai et al., 2020; Pai and Levin, 2022), giving rise to promising therapeutics.
These capacities of morphogenetic cellular collectives are
Collective intelligence as a product of evolutionary selection, or evolutionary selection as a product of collective intelligence?
Biological individuality has traditionally been associated with the scope of an evolutionary unit (Clarke, 2016) – the unit that is subject to differential survival and reproduction. Within this orthodox view, whilst the processes of developmental basal cognition are certainly complicated and might have the appearance of collective behaviour, they are merely complex parts of a single individual. However, this view turns out to be wholly inadequate to understand and manipulate the multi-scale nature of life.
Genetic identity and biological individuality
The idea that biological individuality can be defined by genetic identity is clearly insufficient: the structural and functional demarcations of coherent individuals often diverge from their genetic information. Note that a colony of bacteria may be genetically homogeneous but not an individual, while planaria are biological individuals by any reasonable sense of the word but not genetically homogeneous (Fields and Levin, 2018a). Even though genetically identical, the tissues and cells within a classical organism (body) often compete with each other (Gawne et al., 2020); conversely, cells from distant species cooperate well within chimeric organisms (Nanos and Levin, 2022). In addition, genetic information does not always predict the structure and function of bioelectrically modified organisms (Levin, 2014, 2021a) or of self-organising synthetic living machines (Blackiston et al., 2021; Kriegman et al., 2020). Likewise, often it is the degree of bioelectrical coupling, not genetic differences, that determines whether cellular optimisation occurs at the single-cell level (cancer) vs. at the organ-level (normal morphogenesis) (Chernet and Levin, 2013).
Evolutionary units and biological individuality
Can a notion of evolutionary units beyond genetic relatedness rescue a meaningful concept of biological individuality? That is, the ability to exhibit heritable variation in reproductive success might obtain for a complex or composite whose components are not genetically related. For example, despite being of separate ancestral origins, the nuclear and mitochondrial DNA of eukaryotes can be considered a single evolutionary unit (under most conditions) by the virtue of their common vertical transmission. However, identifying what exactly constitutes an evolutionary unit in general is also non-trivial – especially because they change over evolutionary time and new units arise at new levels of organisation (Okasha, 2006).
To be a bona fide evolutionary unit, a collective must exhibit heritable variation in reproductive success that belongs properly to the collective level – over and above the sum of that exhibited by its component parts (Okasha, 2006; Watson et al., 2022; Watson and Thies, 2019). This requires organised functional relationships that cause short-sighted self-interested entities to behave in a manner that serves the long-term collective interest of the whole. In this light, the complex nature of functional relationships between component parts begins to look less like the product of selection at the system level, and more like the source of evolutionary individuality.
Practical implications: Beyond philosophy
Such considerations matter fundamentally to our understanding of the organismic, evolutionary and developmental biology (i.e., emergent functionality) and thus to our ability to predict, control, manage and manipulate multi-scale biological systems. Understanding what kind of relationships instantiate biological individuality is thus of great importance to synthetic bioengineering, regenerative medicine, exobiology, robotics and artificial intelligence.
For example, to intervene in the processes that coordinate component parts to create or regenerate an organ or a limb – or produce an entirely novel construct such as a self-assembling biobot (Ebrahimkhani and Levin, 2021) – we must be able to manipulate the very relationships that define individuality (Levin, 2021c). Such bioengineering goals therefore depend intimately on our knowledge of collective intelligence at multiple levels of biological organisation (Beane et al., 2013; Herrera-Rincon et al., 2018; Pezzulo and Levin, 2015).
Recent work has begun to apply the tools of collective intelligence and cognitive neuroscience to morphogenesis and its disorders, including cancer, a disease of dysregulated morphogenesis (Deisboeck and Couzin, 2009; Doursat et al., 2013; Friston et al., 2015; Pezzulo et al., 2021; Pezzulo and Levin, 2015, 2016; Rubenstein et al., 2014; Slavkov et al., 2018). Disconnection from the bioelectric network of tissues often gives rise to fragmenting of coherent anatomical individuals into invasive single cells and tumors; their release from higher level collective goals is readily apparent because they pursue anatomical, histological and physiological states quite different from those that the organism tries to maintain (Egeblad et al., 2010; Levin, 2021c; Radisky et al., 2001; Soto et al., 2008). This fragmentation can be reversed: despite strong oncogenic mutations, cancer phenotypes can be suppressed by forcing bioelectrical connections among cells, thus overriding single-cell level goals with large-scale morphogenetic ones (Chernet and Levin, 2013).
Cognition, learning and problem-solving in biological networks: Generalised principles of connectionism
The link between evolution and simple types of learning has often been noted (Skinner, 1981; Watson and Szathmáry, 2016) but sometimes interpreted in an uninteresting way:
Hebbian learning in networks
A simple example of such a neural model, demonstrating distributed computation and learning, is the Hopfield network (Hopfield, 1982) (Appendix Box 1). Given that the Hopfield network is inspired by neural dynamics and learning in cognitive systems, its learning and problem-solving abilities are perhaps not so surprising, despite their decentralised operation. However, the underlying principles are extremely simple and general: the same computational algorithms also apply in systems that we don’t normally expect to be capable of cognition or learning; gene-regulation networks, protein interaction networks and ecological community networks can all implement the same kinds of functions as neural networks if organised appropriately (Biswas et al., 2021; Herrera-Delgado et al., 2018; Szabó et al., 2012; Tareen and Kinney, 2020). However, cognition in different substrates may have very different spatio-temporal scales – from the cellular, to the familiar organismic scale, and perhaps to the ecological scale (Power et al., 2015; Watson et al., 2014). Can these kinds of networks also learn as neural networks do?
The answer is yes. Hebbian learning in a self-modelling dynamical system (Appendix Box 1) effects a positive feedback on correlations; the more things co-occur, the more the connection between them changes to make them more likely to co-occur in future. This positive feedback on correlations is quite natural. In some conditions, it does not require an active learning mechanism that strengthens connections, instead it is sufficient to differentially relax or weaken connections according to the frustration or stress experienced in that connection (Buckley et al., in prep). Thus, connectionist modes of cognitive learning can be instantiated in various kinds of non-neural networks (Davies et al., 2011; McCabe et al., 2011; Power et al., 2015; Watson et al., 2011b).
Importantly, the application of connectionist models also extends into the domain of evolutionary systems, where the connections of a network are changed by variation and selection, as seen in the evolution of interaction networks in development and ecology (Brun-Usan, Rago, et al., 2020; Brun-Usan, Thies, et al., 2020; Kouvaris et al., 2017; Rago et al., 2019; Watson et al., 2014; Watson et al., 2016; Watson and Szathmáry, 2016). In these ‘evolutionary connectionism’ models, ordinary processes of random variation and selection act on the functional interactions between components, altering their organisation in a way that positively reinforces correlations – functionally equivalent to connectionist learning models (Watson and Szathmáry, 2016). The algorithmic principles well-understood in neural networks, are equally demonstrable in gene-regulation networks (Brun-Usan et al., 2020; Brun-Usan et al., 2020; Kounios et al., 2016; Kouvaris et al., 2017; Rago et al., 2019; Watson et al., 2014), and ecological community networks (Power et al., 2015) and social networks (Davies et al., 2011; Watson et al., 2011a). This algorithmic unification between connectionist learning and evolution (Watson et al., 2016; Watson and Szathmáry, 2016) opens up the transfer of an extensive, well-developed toolset from machine learning into evolutionary theory to naturalistically explain evolutionary ‘intelligence’ (Kounios et al., 2016; Watson et al., 2022; Watson and Szathmáry, 2016).
In particular, it is important to recognise that connectionist models can exhibit learning bottom-up, without centralised control or an external teacher, and without any performance feedback applied at the system level, via fully distributed and unsupervised learning principles (Watson et al., 2011a; Watson et al., 2011b, 2011c). This means that the same learning behaviours can be exhibited by an ecological community without selection at the community level (Power et al., 2015). This is potentially important to understanding the evolution of intelligent collectives (and evolutionary transitions in individuality (ETIs)) because it identifies conditions where relationships between evolving entities can be organised via natural selection acting at the lower level
So, what kind of cognition can such networks exhibit?
We find it useful to operationalise
The ability of distributed learning to improve problem-solving ability in this way is now well-developed (Kounios et al., 2016; Mills, 2010; Mills et al., 2014; Watson et al., 2011a; Watson et al., 2011b, 2011c; Watson et al., 2016). In some conditions, a learning neural network can enable a sort of ‘chunking’, rescaling the search process to a higher level of organisation (Caldwell et al., 2018; Mills, 2010; Watson et al., in review; Watson et al., 2011c; Watson et al., 2016). Elsewhere, we hypothesise that this rescaling of the problem-solving search process is intrinsic to transitions in individuality (Watson et al., 2016), suggesting that ETIs constitute a form of deep model induction (Czegel et al., 2019; Vanchurin et al., 2021; Watson et al., 2022).
Credit assignment in individuals and collectives
Conventional accounts of intelligence and behavioural protocols (Watson, 1967) assume a singular subject of intelligence and of the goals that it can pursue. However, this is a significant over-simplification that obscures important questions about how centralised intelligences arise out of cellular components (Levin, 2019, 2021c). For example, one trains a rat to press a lever and receive a delicious reward, in instrumental or associative learning paradigms (Abramson, 1994; Best, 1965; Rescorla and Solomon, 1967). The rat is understood to be an intelligent agent solving an instrumental learning problem; but it is also a collection of cells. Indeed, the cells that perform the action (muscle and skin cells that interact with the lever) and the ones involved in sensing the environment (seeing the lever, feeling the lever and tasting the reward) are not the cells that immediately receive the nutritional benefit of the reward (intestinal lining). No individual cell has the entire experience of performing an action and reaping its benefits – that relationship only exists in the ‘group mind’ of the collective agent. How do the parts discern which of their actions should be reinforced? Problems of distributed credit assignment are a key aspect of intelligence, even in conventional organisms.
It is imperative to understand the developmental algorithms and signals by which tissue-level agents incentivise lower-level subunits (e.g. cells and molecular pathways), distorting their option space so that simple, local descent down free-energy paths (short-sighted self-interest) result in higher order adaptive activity (long-term collective interest). The key to being an individual is to have a functional structure in which diverse experiences across its components are bound together in a way that generates causal relationships and composite memories that belong to the higher space of the individual and not its components (Fields and Levin, 2022).
How does scaling of reward dynamics bind subunits into intelligent collectives that better navigate novel problem spaces? Lessons from machine learning
It is no accident that the issue of credit assignment, and the application of credit to parts or wholes, is a central one in evolutionary selection, developmental and organismic biology and cognitive science. It is a feature of many difficult learning tasks that they require sequences of actions that are many steps away from ultimate goals – making it intrinsically difficult to incentivise the component parts involved. This is what makes difficult tasks difficult; conversely, having feedbacks that are additive and individual, is what makes easy tasks easy. It is no coincidence then, that these issues of credit assignment have well-developed formalisms in the domain of machine learning (Watson et al., 2022). In particular, one of the touchstones of machine learning – the ability to represent non-linearly separable functions (such as XOR - Exclusive OR logical operator) – is distinguished from linearly separable functions exactly because improvements in the output cannot be ascribed to the independent contribution of individual inputs (Watson et al., 2022). Nonetheless, simple connectionist models can learn such functions if they have a suitable architecture (see below).
Connectionist models thus identify some basic criteria about the kind of relationships that turn a collection of unintelligent components into a non-decomposable intelligence with cognitive and learning abilities that belong properly to the whole and not the parts. Moreover, the ability of
Together, these observations show that the apparent distinction between individual intelligence and collective intelligence is not substantial: at a minimum, they exist on a continuum. Further, the connectionist models of cognition and learning developed for individual intelligence are not simply relevant to understanding what is required for a collective to be intelligent, it may be that it is precisely these cognitive capacities that are the fundamental difference-maker with respect to individuality itself; i.e. between ‘many individuals’ and ‘one individual’.
What kinds of interaction structures are necessary for what kind of (collective) intelligence and how can these structures emerge?
What kinds of relationships are necessary to turn a society into an individual? A central aspect of how intelligence arises from a collection of subunits is the specific communication and functional linkages between them, as well as the algorithms for updating those interactions in light of experience. It is thus essential to determine what kinds of architectures underlie different degrees of agency (which support memory, problem-solving, information integration and collective action, higher-level autonomy, etc.) across the continuum. Here, we leverage connectionist models of cognition and learning (top row) to specify known architectures that embody key waypoints along the collective intelligence spectrum (bottom row), as well as to identify knowledge gaps that highlight opportunity for next steps in this field. References indicate examples of potentially relevant models where available. Shading indicates speculative suggestions and opportunities for future research. In the final column, the biological examples are known but the relevant topology is not.
The structure of interactions
Naturally, the ability to represent relationships (e.g. correlations or associations) among variables, rather than a system of independent variables, is essential for any connectionist model of cognition, and requires components to have connections of one kind or another. For example, in development, gap-junctions between non-neural cells are physiologically tunable ‘synapses’ that communicate cellular behaviours (Mathews and Levin, 2017; Palacios-Prado and Bukauskas, 2009) and indeed can drive genetically wild-type cells to build body organs belonging to diverse species (Emmons-Bell et al., 2015). There are many other levels of biological organisation with different ‘signals and responses’ between components, or sensitivity to one another’s behaviours. Being connected is necessary but not sufficient for cognitive functions, however. Connectionist principles enable us to be more specific about what kinds of connection structure are important.
Instructive neural architectures from machine learning
This is by no means a survey of machine learning techniques or a comprehensive description of neural architectures; our aim is simply to highlight some of the key architectural issues and their significance with respect to different cognitive abilities. Three particular architectural issues have special significance: 1. 2. 3.
Implications for evolutionary intelligence and basal cognition
Naturally, for a collection of individuals to exhibit any kind of collective intelligence, it is, at the very least, necessary that the behaviour of one individual has some sensitivity to the behaviour of another. Such interactions can coordinate behaviours to take advantage of scenarios where the benefit/reward or fitness that one individual receives is sensitive to the behaviour of another. However, if this credit-assignment interaction (or fitness epistasis) constitutes a linearly separable function this is not really a difficult problem; although the benefit they receive will vary in different contexts, the behaviour that maximises their benefit is always the same. In contrast, when the credit that one individual receives has an interaction with the credit that another individual receives which constitutes a non-linearly separable function (Watson and Thies, 2019) (or reciprocal sign epistasis, (Weinreich et al., 2005)), this requires that one individual can change its behaviour (or ‘do the opposite’) depending on the context of what other individuals are doing. For a collective to coordinate behaviours to take advantage of such interactions, it must be able to represent non-linearly separable functions, which requires the interaction structure between individuals to have some depth (Watson et al., 2022).
These are just the kind of relationships that make the credit assignment or fitness of the whole not only different from the sum of the rewards/fitnesses of the parts but also a non-decomposable function. Intuitively, this changes our relationship from ‘how good this is for me depends on what you are doing’ to ‘what is best for me to do depends on what you are doing’. This is important because, when it is reciprocal, the fitness-affecting characteristics of one component only have meaning in the context of the other. In other words, it creates a ‘we’;
Deep representations also have a special significance in recurrent networks. In non-hierarchical networks, the many connections between components can cause the system to hold state over time (i.e. internal states can be maintained as dynamical attractors even when the inputs to the network are removed or have changed). This enables the network to exhibit temporally extended behaviours, but it also has the effect that it becomes difficult to change the system state and, therefore, to be sensitive to system inputs. Getting out of one dynamical basin of attraction and into another can require large and/or specific state perturbations. The system acts as a whole but cannot ‘change its mind’ easily (Hills et al., 2015; Nash et al., in prep; Watson et al., in review). This is problematic for organismic adaptability and evolutionary variability. In contrast, a hierarchical representation can cause coordinated behaviour in many downstream parts but retain the capacity for small changes to variables in the higher level representation to move all the downstream variables to a new state (Nash et al., in prep). A recent alternative model is provided by a network of neurons that have a ‘decision cycle’ that repeatedly re-decides which states to adopt with a timing based on learned connections (Watson et al., in review). By learning to synchronise the decision cycle of particular groups of components, this kind of network exhibits multi-scale problem-solving capabilities without having an explicit or pre-defined multi-layer structure.
Cascading control architectures – where a small number of variables cause large coordinated changes in the state of many downstream variables – are common in organisms through many scales from molecular to morphological. This takes explanatory focus away from the collective and onto the units at deeper levels of the causal chain, for example, a gene cues the coordination of other biomolecules within the cell, and the germ line cues coordination of other cells within the organism. However, natural organisms are neither single-layer recurrent networks (with every component connected equally to every other like the Hopfield network) nor strictly feed-forward multi-level hierarchies (with components in one layer only connected to components in the layer below like the MLP). They are not quite like deep-belief networks either, of course, but they do contain elements of both cascading control and recurrent control architectures. This means that different levels of organisation can both be influenced by higher level control variables and be collectives that co-define and sustain their own (non-decomposable) meaning. These considerations suggest that this kind of deep and also partially recurrent architecture is relevant to the multi-scale autonomy observed in complex organisms.
Learning the structure of interactions
The previous section discussed how the types of relationships, and their organisation, might influence the type of information integration and coordinated action that could be exhibited by a collective. But how do such organisations arise? For this, we turn our attention from connectionist architectures to models of connectionist
Gradient methods versus stochastic local search, supervised learning versus reinforcement learning
For many learning tasks, it is useful to express the error in the output (with respect to an input and a target) as a function of the connection strengths in the network. If this function is differentiable, then this can be used (in artificial machine learning methods) to define a gradient method which computes a change in the weights of the network that will systematically reduce the error. In biological evolution or emergent collective intelligence, there is no explicit target or desired output predetermined by an external agent or teacher. There is therefore no ‘error’ function, as such. The more relevant type of learning is
What makes learning systems smart, however, is not merely the ability to increase the fit of model parameters to data; what makes such systems interesting is that the parameters they adjust and the data to which they fit are
Generalisation is fundamental to learning and intelligence
Without it, a system can only respond to current inputs in a manner consistent with past rewards. At one extreme, if the future is going to be exactly like the past, this is fine. At the other extreme, if the future has nothing at all in common with the past, then there is not much that can be done about that. But, in other cases, the future is not the same as the past, but it shares some kind of underlying regularity in common with it. These are the cases where intelligence has some meaning. Specifically, a system that can generalise can act in a manner that is consistent with long-term rewards, even when this appears to oppose immediate or short-term interests. For individuals that interact with others in a collective, the ability to act in a manner that is consistent with
Unsupervised learning
It might seem curious that any kind of learning can occur without supervision or system-level reward feedback of some kind. How can a learning system know what to learn if nothing tells it what it is supposed to learn?
The level of credit assignment in reinforcement learning and collectives
Consideration of unsupervised learning has direct significance for the evolution and reward of collective intelligence. This is because reinforcement learning acting on the individual characteristics affecting their connections to others can result in dynamics that are equivalent to unsupervised learning at the system scale (Davies et al., 2011; Power et al., 2015; Watson et al., 2011a). Intuitively, if B is rewarded for being activated, then one of the ways it can increase its reward is to increase the strength of its connection from A (e.g. when A and the connection are positive). This increases the individual reward B receives right now, but it also makes the future activation of B correlated with the activation of A (the principle of Hebbian learning in another guise). The same considerations apply to A and its connection from B. Note that neither component is making the connection with the other because it is interested in the collective reward that A and B receive together, nor because it makes the future dynamics of the AB pair more consistent with their past correlation. Nonetheless, it does make the future dynamics of the AB pair more consistent with their past correlation (Watson et al., 2014).
This observation creates a fundamental linkage between the principles of individual learning or individual utility-maximisation and the principles of system-level or collective intelligence (Watson et al., 2011a). Note that the mechanism of Hebbian learning was identified by Donald Hebb to explain neural learning because it is the right way to modify synaptic connections if you want the network to model observed correlations. This equivalent mechanism, in contrast, is motivated bottom-up – it is a consequence of components that are incentivised only by
How does distributed learning effect system-level rewards and credit assignment?
This distributed learning is not
This bottom-up incremental adjustment of relationships can thus increase system-level welfare. It does so in a manner that is functionally equivalent to distributed learning mechanisms familiar in artificial neural networks, without presupposing system-level rewards or credit assignment. Moreover, in so doing, it creates a non-decomposable whole (attractors that are non-linearly separable functions of the inputs and depend on the system’s own internal history), which means that credit assignment or reward at the level of individual parts and their individual behaviours becomes ineffective. Instead, credit assignment (if it applies at all) and any possibility of effecting modified behaviours through reward become meaningful only at the higher level of organisation.
Modelling collective intelligence and basal cognition: Evolutionary individuality, organismic individuality and cognition are coextensive
As discussed above, the basic computational elements of such distributed learning are substrate-agnostic and common to a wide range of biological networks (Cervera et al., 2018; Pietak and Levin, 2016, 2017). However, the conditions for distributed learning are non-trivial; not all of these networks may meet them. The important thing to note is that there is no requirement for an incentive to model long-term or collective consequences of individual actions, or for a system-level incentive to model the structure or pattern of observations. We do not yet know which of these biological systems might meet these conditions and the extent to which this influences their collective intelligence. But it is known that organismic individuality evolved through a bottom-up process of collective intelligence, resulting in information integration and coordinated action so well-organised that we observe a new level of organismic and evolutionary individuality. The principles of connectionist cognition and learning described above provide a roadmap of gaps and opportunities that future research might explore to better understand how such emergent individuality occurs. In particular, the architecture of the interactions – whether they are feed-forward or recurrent, capable of representing non-linearly separable functions or not, shallow or deep or some mixture of these characteristics – has important consequences for the type of cognitive model they can represent.
The ecological models developed thus far demonstrate that connectionist learning principles are relevant to collective intelligence in systems that are not (yet) evolutionary units. They fall short, however, of demonstrating the spontaneous evolution of a new level of individuality. In algorithmic terms, such models cannot do the ‘chunking’ of the search space or rescaling of the search process that is facilitated by the induction of deep models (Caldwell et al., 2018; Mills, 2010; Mills et al., 2014; Watson et al., 2011b; Watson et al., 2016; Watson et al., 2009). We hypothesise that this is because they are single-level networks of symmetric interactions; our roadmap supports the idea that the evolutionary transitions in individuality correspond to deep interaction structures (Czegel et al., 2019; Watson et al., 2022) or perhaps other mechanisms of multi-scale dynamics (Watson, accepted; Watson et al., in review).
We propose that some of the gaps in this picture might be addressed by exploring the hypothesis that evolutionary individuality, organismic individuality and cognition are coextensive (Watson et al., 2022). The idea is that acting in a manner consistent with long-term collective interests, in particular when this conflicts with short-term self-interest, is not just a hallmark of collective intelligence but is in fact what constitutes cognition and individuality at the collective level. This can perhaps be formalised through the consideration of non-linearly separable functions. Specifically, if a system of functional interactions among the parts represents a non-linearly separable function, then the incentive of the whole
Conclusions
Commonalities between cognitive and evolutionary processes and those that shape growth and form have been hinted at in the past (Grossberg, 1978; Pezzulo and Levin, 2015; Spemann, 1967). We argue that conceptual advances in the links between machine learning and evolution now provide quantitative formalisms with which to begin to develop testable models of collective intelligence across scales. From subcellular processes, to cellular swarms during morphogenesis, to ecological dynamics on evolutionary timescales – all of these processes are driven by the scaling of reward dynamics that bind subunits into collectives that better navigate novel problem spaces.
In addition to shedding light on biological evolution, a better understanding of the origin and operation of collective intelligences would have a number of practical applications. Molecular medicine today is focused almost entirely on the micro-hardware of life – modifying DNA and rewiring molecular pathways – with limited success due to difficult inverse problems (Lobo et al., 2014). The capacity to manipulate the collective intelligence of cell groups might offer powerful ways to guide native and synthetic morphogenesis top-down (Pezzulo and Levin, 2016). Insights gleaned from biological systems could also significantly enhance the engineering of intelligent robots whose behaviour results from cooperation, competition and merging of subunits across multiple levels of organization.
Harnessing the native capability of collective intelligence in the service of biomedicine or bioengineering will require a much better understanding of how to identify, characterise and motivate emergent agents in anatomical, physiological and transcriptional spaces (Levin, 2022a; Pezzulo and Levin, 2015). As a starting point, we need to develop appropriate formalisms for top-down control of multi-scale intelligent agents of diverse composition. We argue that the tools and concepts of machine learning, behavioural neuroscience and evolutionary biology apply to problems of collective intelligence at multiple scales and offer a promising way forward.
There is a deep, fundamental symmetry between the origin of new evolutionary individuals from competent subunits and the assembly of an integrated cognitive agent as a collective intelligence composed of sub-agents. Future experimental and in silico work will quantitatively identify the necessary and sufficient relationships that effect such transitions. Such work has the potential to drive a flourishing sub-field of collective intelligence with implications ranging from basic evolutionary biology to regenerative medicine and artificial intelligence.
