Connective issue: AI learns by doing more with less

Sparsity and energy constraints guide learning and communications in silicon neuronal networks

Brains have evolved to do more with less. Take a tiny insect brain, which has less than a million neurons but shows a diversity of behaviors and is more energy efficient than current AI systems. These tiny brains serve as models for computing systems that are becoming more sophisticated as billions of silicon neurons can be implemented on hardware.

The secret to achieving energy efficiency lies in the silicon neurons’ ability to learn to communicate and form networks, as shown by new research from the lab of Shantanu Chakrabartty, the Clifford W. Murphy Professor in the Preston M. Green Department of Electrical & Systems Engineering at Washington University in St. Louis’ McKelvey School of Engineering.

Their results were published July 28, 2021 in the journal Frontiers in Neuroscience.

For several years, his research group studied dynamical systems approaches to address the neuron-to-network performance gap and provide a blueprint for AI systems as energy efficient as biological ones.

Shantanu Chakrabartty headshot
Chakrabartty

Previous work from his group showed that in a computational system, spiking neurons create perturbations which allow each neuron to “know” which others are spiking and which are responding. It’s as if the neurons were all embedded in a rubber sheet formed by energy constraints; a single ripple, caused by a spike, would create a wave that affects them all. Like all physical processes, systems of silicon neurons tend to self-optimize to their least-energetic states, while also being affected by the other neurons in the network. These constraints come together to form a kind of secondary communication network, where additional information can be communicated through the dynamic but synchronized topology of spikes. It’s as if the rubber sheet vibrates in a synchronized rhythm in response to multiple spikes.

In the latest research, Chakrabartty and doctoral student Ahana Gangopadhyay showed how the neurons learn to pick the most energy-efficient perturbations and wave patterns in the rubber sheet. They show that if the learning is guided by sparsity (less energy), it’s like the electrical stiffness of the rubber sheet is adjusted by each neuron so that the entire network vibrates in a most energy-efficient way. The neuron does this using only local information which is communicated more efficiently. Communications between the neurons then become an emergent phenomenon guided by the need to optimize energy use.

This result could have significant implications on how neuromorphic AI systems might be designed. “We want to learn from neurobiology,” Chakrabartty said. “But we want to be able to exploit the best principles from both neurobiology and silicon engineering.”

Historically, neuromorphic engineering — modeling AI systems on biology — has been based on a relatively straightforward model of the brain. Take some neurons, a few synapses, connect everything together and, voila, it’s… if not alive, at least able to perform a simple task (recognizing images, for example) as efficiently, or moreso, than a biological brain. These systems are built by connecting memory (synapses) and processors (neurons). Each performing its single task, as it was presumed to work in the brain. But this one-structure-to-one-function approach, though easy to understand and model, misses the full complexity and flexibility of the brain.

Recent brain research has shown tasks are not so neatly divided, and there may be instances in which the same function is being performed by different brain structures, or multiple structures working together. “There is more and more information showing that this reductionist approach we’ve followed might not be complete,” Chakrabartty said.

Sparsity makes the spiking activity and communications between the neurons more energy efficient as the neurons learn without using backpropagation. (Courtesy: Chakrabartty Lab)

The key to building an efficient system that can learn new things is the use of energy and structural constraints as a medium for computing and communications or, as Chakrabartty said, “Optimization using sparsity.”

The situation is reminiscent of the theory of six-degrees of Kevin Bacon: The challenge — or constraint — is to make connections to the actor by connecting six or fewer people.

For a neuron that is physically located on one chip to be its most efficient, the challenge — or constraint — is completing its task within the allotted amount of energy. It might be more efficient for one neuron to communicate through intermediaries to get to the destination neuron. The challenge is how to pick the right set of “friend” neurons among many choices that might be available. Enter energy constraints and sparsity. 

Like a tired professor, a system in which energy has been constrained also will seek the least resistant way to complete an assigned task. Unlike the professor, an AI system can test all of its options at once, thanks to the superposition techniques developed in Chakrabartty’s lab, which uses analog computing methods. In essence, a silicon neuron can attempt all communication routes at once, finding the most efficient way to connect in order to complete the assigned task.

The current paper shows that a network of 1,000 silicon neurons can accurately detect odors with very few training examples. The long-term goal is to look for analogs in the brain of a locust which has also been shown to be adept in classifying odors. Chakrabartty has been collaborating with Barani Raman, a professor in Department of Biomedical Engineering, and Srikanth Singamaneni, The Lilyan & E. Lisle Hughes Professor in the Department of Mechanical Engineering & Materials Science, to create a sort of cyborg locust — one with two brains, a silicon one connected to the biological one.

“This would be the most interesting and satisfactory aspect of this research if and when we can start connecting the two realms,” Chakrabartty said. “Not just physically, but also functionally.”


The McKelvey School of Engineering at Washington University in St. Louis promotes independent inquiry and education with an emphasis on scientific excellence, innovation and collaboration without boundaries. McKelvey Engineering has top-ranked research and graduate programs across departments, particularly in biomedical engineering, environmental engineering and computing, and has one of the most selective undergraduate programs in the country. With 140 full-time faculty, 1,387 undergraduate students, 1,448 graduate students and 21,000 living alumni, we are working to solve some of society’s greatest challenges; to prepare students to become leaders and innovate throughout their careers; and to be a catalyst of economic development for the St. Louis region and beyond.

This work was supported in part by a research grant from the National Science Foundation, ECCS: 1935073