Active projects

AI Society: Emergence of Communication amongst Autonomous Neural Networks

The next challenge in AI will probably not be about making faster computers, collecting more data, or designing adaptive robot embodiment. The key will be to allow for machines to communicate their internal states, in a process arguably similar to humans sharing about their emotions. The now very popular deep neural networks, even though extremely efficient at implementing complicated tasks, represent hundreds of thousands of parameters. Apart from looking at the outputs, no human can make sense anymore of how computations are really made inside those networks, or “how the AI thinks”. The next step will naturally be for the machines themselves to report the way they reach conclusions. In order for those reports to be understandable to humans and other machines, communication will need to be established, much like a natural language for AI. In this research, we connect a population of neural networks together, with the task of teaching each other relevant information to solve different sets of tasks, using a limited medium. Our research aims to understand the underlying principles of the spontaneous emergence of communication, from the interaction between autonomous agents. From the connectivity between different AIs, emerges a society that coevolves with its environment. This society may acquire its own swarm mind, transitioning to a phase in which it is controlled by new sets of phenomena, as a result making them more and more independent from their hardware.

full_architecture2 (1)


Evolvable Information: Investigating the Universal Principles of Communication with Deep Learning Multiagent Simulations

Why communicate? Why do scientists bother talking to each other? Anyone with an Internet connection already has access to all the information needed to conduct research, so in theory, scientists could do their work alone locked up in their office. Yet, there seems to be a huge intrinsic value to exchanging ideas with peers. Through repeated transfers from mind to mind, concepts seem to converge towards new theorems, philosophical concepts, and scientific theories.
Recent progress in deep learning, combined with social learning simulations, offers us new tools to model these transfers from the bottom up.
However, in order to do so, communication research needs to focus on the concept of evolvable information. The best communication systems not only serve as good information maps onto useful concepts (knowledge in mathematics, physics, etc.) but they are also shaped so as to be able to naturally evolve into even better maps in the future.
In this project, I propose to model evolvable communication, by integrating deep learning with multiagent simulations, to examine the dynamics of evolvability in communication codes.
This research has important implications for the design of an evolvable-communication-based AI capable of generalizing representations through social learning, i.e. an AI that can become wiser through self-reflection. It also has the potential to yield new theories on the evolution of language, insights for the planning of future communication technology, a novel characterization of evolvable information transfers in the origin of life, and new insights for communication with extraterrestrial intelligence.


Investigating the Future of Human Communication. Image credit: Mike Winkelmann

The Future of Collective Intelligence: Investigating the Impact of High-Dimensional Sphere Packing and Massively Multichannel Societies on Communication

While the biosphere is entering its third age, communication among a society of lifeforms has become a major factor of change. Collective intelligence will highly depend on the combination of artificial intelligence technology with artificial life insights on the fundamental principles of communication. In this context, I propose to look at language optimization from the angle of the sphere packing problem. I use an evolutionary toy model to explore the effects of increasing the dimensionality of channels of communication among a network of agents sending messages to each other. The agents need to optimize a fitness function equal to the sum of successfully transmitted messages, over a range of multidimensional noisy channels. Comparing preliminary results with a collision-driven packing generation algorithm derived from the Lubachevsky–Stillinger and Torquato-Jiao algorithms generalizable to n dimensions, rearranging and compressing an assembly of hard hyperspheres ito find their densest spatial arrangement within constraints, I showed that the solution reached by the evolutionary simulation was consistently suboptimal for the range of simulated experiments. Partial results seem to indicate that for higher dimensionality, the density ratio undergoes several transitions, in an irregular manner. Results also highlight the importance of jammed codes, where codewords are interlocked into multidimensional place with surprisingly both situations allowing them or not to reorganize. The preliminary simulation results suggest that future intelligent lifeforms, natural or artificial, from their interaction over largely broadband-channel networks, may invent novel linguistic structures in high-dimensional spaces. With new ways to communicate, future life may achieve unanticipated cognitive jumps in problem solving.

Screen Shot 2018-05-07 at 12.05.17 AM

Illustration of an error-correcting code C as a set of hyperspheres in 2 dimensions.



Visualization of collective communication opti- mization, 100 codewords in 2 dimensions, after 500 generations.

The Expansion of Intelligence in Emergent Systems

All forms of technology are tools invented or discovered by living beings which brought them a different – arguably more efficient – picture of reality, helping them to make choices through their lifetime. One example of such technology is language, which since it has become part of human cognition, has increased the human ability to learn about the regularities in their environment, and has eventually given rise to science (after the invention of writing). Mathematics and AI are other examples of technologies which increased the global cognitive capacity of human culture. Now, if one considers human cognition to be separable from the tool, one may worry about the danger of one cognitive entity taking over the computation made in another. For example, there is a danger that in the future humans may offload so much of their thinking to their smartphones that they become much less capable of performing the kind of thinking process they used to be capable of before. In this work, we first analyze the conditions of such separation, to then analyze the effects of local increases and decreases of intelligence, as computational processes, in artificial life models. Preliminary results demonstrate how agents can append technology to themselves in such a way that their own cognitive ability is increased, and not shrunk. Implications could result in a new theory of integration of so-called “relevant computation”, i.e. cooperative processing of information among groups of intelligent entities.

Screen Shot 2018-05-07 at 12.16.18 AM.png

Expansion of Self-Correcting Systems in Space and Time.


Reynolds’ Boids Swarming Behavior Evolved in Artificial Neural Networks

Swarming behavior is common in biology, from cell colonies to insect swarms and bird flocks. However, the conditions leading to the emergence of such behavior are still subject to research. Since Reynolds’ boids, many artificial models have reproduced swarming behavior, focusing on details ranging from obstacle avoidance to the introduction of fixed leaders. In this project, we are evolving artificial agents, able to develop a range of different swarming behaviors based on their communication systems. Instead of a centralized algorithm, the model simulates a population of agents each controlled by a deep neural network, which are evolved by an original asynchronous genetic algorithm.