Researchers Cut AI Energy Use by 90% with Smarter, Leaner Models

A massive AI revolution is underway—scientists in Germany have cracked the code to slash AI’s energy consumption by up to 90%. By shrinking AI models without sacrificing performance, they’re paving the way for greener, more efficient technology that even small businesses can use.

Image Credit: Caureem / ShutterstockImage Credit: Caureem / Shutterstock

Powering artificial intelligence comes with a massive energy bill attached. Professor Wolfgang Maaß and his research team at Saarland University and the German Research Center for Artificial Intelligence (DFKI) want to make AI up to 90% percent more energy efficient. To improve AI's carbon footprint, the Saarbrücken team is rethinking data centers, large language models, and image analysis models – and their research is opening up access to powerful AI models for small and medium-sized companies. From 31 March to 4 April, the researchers will be at this year's Hannover Messe, showcasing their work at the stand of the Federal Ministry for Economic Affairs and Climate Action (Hall 2, Stand A18).

Data centers consume vast quantities of energy. According to Bitkom, the leading industry association in Germany's digital sector, the electricity requirements to power data centers have more than doubled over the past decade. And with digital transformation only just out of the starting blocks, this trend is really gathering pace. Storing, processing, transmitting, and retrieving data takes energy. Artificial intelligence, in particular, is a huge energy guzzler. Globally, multiple terawatt-hours are being used to train and run today's massive AI models. (One terawatt hour equals one billion kilowatt hours of electrical energy). Using these models to generate images and texts consumes vast energy. As a result, data centers have to get bigger and bigger, which means they need more and more electricity to power and cool the enormous numbers of processors involved, which in turn is causing a massive uptick in their carbon footprint. None of this is helping Europe achieve its goal of net-zero greenhouse gas emissions by 2050. Clearly, something has to change.

'AI can become far more energy efficient. With the right approach, we can make the data centres of the future much more sustainable,' says Professor Wolfgang Maaß, who conducts research at Saarland University and the German Research Center for Artificial Intelligence (DFKI). His research team is developing leaner, customized AI models to curb AI's hunger for energy and conserve resources. They also want to identify ways data centers can become more energy-smart.

'By making the models smaller and more efficient, we're helping to drive sustainability,' says Dr. Sabine Janzen, a senior research scientist in Wolfgang Maaß's team. 'Our work is also opening up access to powerful AI models for small and medium-sized businesses, because these smaller, leaner AI models don't need a large technical infrastructure. This will enable everyone – not just the big players – to leverage this new technology,' says Janzen.

Today's AI chatbots, such as ChatGPT and visual AI models, use trillions of parameters and utilize vast datasets to perform their tasks. The amount of energy they consume is correspondingly huge. The researchers in Saarbrücken are developing ways to reduce this energy consumption without compromising the output quality of these leaner AI models. 'A central element of our work is a technique known as knowledge distillation. It's a type of compression technique that enables us to make smaller and therefore more energy efficient models that perform just as well as the larger models,' explains Sabine Janzen.

The approach used by the research team could be described as follows: When looking for the answer to a specific question, you don't read an entire library; you focus only on those books that are relevant to your question. The researchers in Saarbrücken extract smaller, more focused, and more energy-efficient 'student' models from larger 'teacher' models. By distilling the knowledge needed to perform tasks in a specific area and reducing it to the essentials, they can reduce the size of the data models by up to ninety percent. Model parameters irrelevant to the area of interest are not touched. 'In terms of inference speed, i.e., how quickly the model can process input data and produce results, these student models perform at a level comparable to that of the larger teacher models, but require 90% less energy to do so,' explains Janzen.

By using another automated efficiency technique known as 'neural architecture search' (NAS), the team has also achieved impressive results with visual AI models, i.e., models that process digital image data. 'Our most recent results show that we can use the NAS method to reduce the models' size by around ninety percent,' says Sabine Janzen. In this work, the researchers focus on machine learning with artificial neural networks – a very energy-intensive AI method that can analyze large volumes of data. Artificial neural networks are designed to mimic the human brain. Our brains contain billions of nerve cells, called neurons, that are connected via trillions of synapses. A synapse is essentially the interface between two neurons across which the two nerve cells communicate. When we learn something new, neurons send electrical signals to each other across synapses; as we continue learning, the same neurons keep firing together, and the connections between them get stronger, whereas the connections between inactive neurons weaken.

Learning processes in artificial neural networks are similar, and by feeding these networks large amounts of data, they can be trained to recognize patterns in natural language or images. However, whereas the brain is a master of energy-efficient learning, training an extensive artificial neural network requires a lot of computing power and a lot of energy. Training an artificial neural network to yield meaningful results also involves much human input. Typically, these artificial networks are designed and configured manually, and the many parameters involved are adjusted and optimized by experts until they perform at the required level. This is where the Saarbrücken researchers bring 'neural architecture search' (NAS) into play. 'Instead of designing the neural networks manually, we automate the design optimization process using NAS,' explains Sabine Janzen. 'NAS allows us to examine different network architectures and optimize them to create a model that delivers high performance, efficiency, and reduced costs.'

Wolfgang Maaß's team is working with the steel company Stahl Holding Saar to test these compacter AI models in practice. The aim is to teach the artificial neural networks to sort steel scrap efficiently. In order to produce new steel from scrap steel, producers need a scrap of the right quality. Only certain types of scrap can be recycled to manufacture high-quality steel. However, the steel scrap delivered to the smelting plant is a mix of all types and must be sorted. Scrap sorting can be automated, but the AI model is too big to be practical. 'We have compressed the visual AI sorting model, making it compact and more energy efficient. In fact, the smaller model performs better on certain metrics, making the steel recycling process more efficient,' says Janzen. Where previously, a huge AI model would have required a lot of energy to operate, a small, customized, energy-efficient model can now perform the same task.

The researchers start by training their models with the full dataset that contains all the information. They then shrink the AI models using knowledge distillation and specially compiled neural networks so that the models only contain the parameters necessary for the task at hand. In this particular case, the aim is to create an AI with all the knowledge it needs to analyze camera images to classify the scrap steel being delivered to the site.

The Saarbrücken research team is also working with partners to outline a concept and compile recommendations for sustainable data centers and energy-efficient AI. Until now, estimating just how much energy is needed to create and operate an AI model has been difficult. That makes it harder for businesses to plan ahead,' explains PhD student Hannah Stein, who is researching these energy-efficient AI models. 'We're currently developing a tool that provides reliable forecasts of the energy consumed by and the costs associated with the different AI models,' says Stein. Data centers and AI users can then use this information to plan more effectively, identify inefficient processes, and take corrective action as necessary – for example, scheduling heavy computational loads at times when the price of electricity is low.

Professor Wolfgang Maaß and his team's research was selected for the Federal Ministry for Economic Affairs and Climate Action's stand at this year's Hannover Messe. The team will present the latest results from the federally funded ESCADE project, which is based at the German Research Centre for Artificial Intelligence DFKI.

Background: ESCADE ('Energy-Efficient Large-Scale Artificial Intelligence for Sustainable Data Centers') is a three-year project with a budget of around €5 million being financed by the Federal Ministry for Economic Affairs and Climate Action (BMWK).

The project will run until the end of April 2026. The ESCADE consortium is made up of the research team headed by Wolfgang Maaß (Saarland University and DFKI), NT Neue Technologie AG, Stahl-Holding-Saar GmbH & Co. KGaA, SEITEC GmbH, Dresden University of Technology, the University of Bielefeld and the Austrian applied research institute Salzburg Research.
https://escade-project.de

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of AZoAi.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.