Microsoft wants to bring Three Mile Island back online. It was a headline that really made people sit up and take notice. The decommissioned US nuclear power plant is best known for its partial meltdown in 1979. The project shows the lengths the tech giant is prepared to go to to cover the enormous energy requirements of its applications. And Microsoft is far from alone. AI applications are driving up the power requirements of data centers worldwide. The amount of energy consumed to train GPT-4, for example, would have been enough to power a small town. Moore’s Law, which states that the number of transistors in an integrated circuit doubles approximately every two years, leading to an increase in computing power, may be slowing down, but computers will continue to become more powerful. The ever-growing density of computer chips is part of the energy problem. Because the processor and memory are separate, data is constantly in motion.
The core disciplines of AI – pattern recognition, analysis and inference – require a particularly enormous volume of data transfer. This back and forth comes at the cost of latency and requires a great deal of energy. In order to resolve this classic bottleneck problem, current research approaches in microelectronics are focusing on “in-memory computing.” The idea is that computing operations should run in close proximity to the memory. The block-based architecture of storage and processor units is eliminated and energy-intensive data traffic is minimized. Nature shows us how information processing works in energy-saving mode. Our brain manages with a power of around 20 watts, which is less than some light bulbs. Inspired by the human brain, scientists are developing hardware structures that are based on biological neurons. The aim of neuromorphic computing is to create an AI infrastructure that is energy-efficient, fast and autonomous enough to move intelligence to the edge, i.e. as close as possible to the data source or to the end devices. Battery-powered sensors or wearables could benefit from low energy consumption. Instead of draining the battery in hours or days, edge AI applications could run for months or years, depending on the operating mode. Since the detour via the cloud would be eliminated, the data would remain local. This would reduce latency and improve robustness. The system would remain operational even without a network connection.