Over its lifetime, the average car is responsible for emitting about 126,000 pounds of the greenhouse gas carbon dioxide (CO2).
Compare those emissions with the carbon footprint left behind by artificial intelligence (AI) technology. In 2019, training top-of-the-line artificial intelligence was responsible for more than 625,000 pounds of CO2 emissions. AI energy requirements have only gotten bigger since.
To reduce AI’s energy footprint, Shantanu Chakrabartty, the Clifford W. Murphy Professor at the McKelvey School of Engineering at Washington University in St. Louis, has reported a prototype of a new kind of computer memory. The findings were published March 29 in the journal Nature Communications.
The co-first authors on this article are Darshit Mehta and Mustafizur Rahman, both members of Chakrabartty’s research group.
A disproportionate amount of energy is consumed to train an AI, when the computer searches different configurations as it learns the best solution to a problem, such as correctly recognizing a face or translating a language.
Because of this energy use, most companies can’t afford to train a new AI from scratch. Instead, they train it “enough” and then maybe tweak some parameters for different applications. Or, if a company is big enough, Chakrabartty said, they’ll move their data centers to a more convenient, waterfront property.
All that energy use creates a lot of heat and needs a lot of water to keep cool. “They are boiling a lake, practically, to build a neural network,” he said.
Instead of boiling, Chakrabartty’s research group turned to quantum tunneling.
When a computer searches for an answer, the system is using electricity to flip billions of tiny switches “on” and “off” as it seeks the shortest path to the solution. Once a switch is flipped, the energy is dumped out and additional energy is used to hold the switch in position, or hold it as memory. It’s this use of electricity that creates such a large carbon footprint.
Instead of feeding a constant stream of energy into a memory array, Chakrabartty lets the physical memory do what it does in the wild.
“Electrons naturally want to move to the lowest stable state,” said Chakrabartty, who is also vice dean for research and graduate education in the Preston M. Green Department of Electrical & System Engineering.
And electrons do so using the least amount of energy possible. He uses that law of physics to his advantage. By setting the solution (say, the recognition of the word “water” as “agua,”) as a stable state, electrons tunnel toward the right answer mostly on their own with just a little guidance, i.e. direction from the training algorithm.
In this way, the laws of nature dictate that the electrons will find the fastest, most energy-efficient route to the answer on their own. At the ground state, they are surrounded by a barrier large enough that the electrons almost certainly will not tunnel through.
Whereas modern memory uses a more brute-force approach by recording its route into memory and using energy at each flipped switch, Chakrabartty’s learning-in-memory design simply lets electrons do what they do without much interference and hardly any additional energy.
Once they’ve reached the final barrier, the AI is said to have learned something.
“It’s like trying to remember a song,” he said. “In the beginning, you’re searching your memory, looking everywhere for the song. But once you’ve found it, the memory is now fixed, and then you can’t get it out of your head.”
This work was supported by the National Science Foundation (ECCS: 1935073); the Office of Naval Research (N00014-16-1-2426, N00014-19-1-2049); and the National Institutes of Health (1R21EY028362-01)
The McKelvey School of Engineering at Washington University in St. Louis promotes independent inquiry and education with an emphasis on scientific excellence, innovation and collaboration without boundaries. McKelvey Engineering has top-ranked research and graduate programs across departments, particularly in biomedical engineering, environmental engineering and computing, and has one of the most selective undergraduate programs in the country. With 140 full-time faculty, 1,387 undergraduate students, 1,448 graduate students and 21,000 living alumni, we are working to solve some of society’s greatest challenges; to prepare students to become leaders and innovate throughout their careers; and to be a catalyst of economic development for the St. Louis region and beyond.
Comments and respectful dialogue are encouraged, but content will be moderated. Please, no personal attacks, obscenity or profanity, selling of commercial products, or endorsements of political candidates or positions. We reserve the right to remove any inappropriate comments. We also cannot address individual medical concerns or provide medical advice in this forum.