According to a study published in the journal Nature, a run-of-the-mill training model used for natural language processing in 2019 emitted 300,000 kg of CO2, the equivalent of 125 round-trip flights between New York and Beijing. Five years down the line, all sectors of society are now eagerly embracing deep neural networks, and as Artificial Intelligence grows to unprecedented proportions, so does the toll on the planet.
In light of this, the main objective of the European project SustainML is to create a framework that will make it easier for AI designers to consider the energy consumption of their Machine Learning applications as they develop them. Janin Koch, scientist of the Ex-Situ project-team common to the Inria Saclay Centre, the Université Paris-Saclay, and the CNRS within the LISN, delves more specifically into how Human-Computer Interaction (HCI) can be leveraged to assist AI designers in making more sustainable decisions across the entire ML life-cycle and to raise awareness of the cost-benefit trade-off behind each of these choices.
The project started in October 2022 and is coordinated by the Spanish middleware company eProsima. The consortium further includes IBM, French semiconductor company UPMEN, Kaiserslautern-Landau University (RPTU), l'Université de Copenhague, the German Research Center for Artificial Intelligence (DFKI) and Inria.
Quantifying the Carbon Impact Behind Machine Learning Models
The project encompasses different areas of research. One cornerstone aspect is the quantification of the environmental impact of algorithms and, more precisely, of the consequences of each decision taken throughout the ML life-cycle. For instance, choosing to train a ML model in a cloud facility relying on non-fossil renewable hydroelectricity as opposed to a data center powered by a coal-fired plant will evidently make quite a difference in how much carbon is emitted.
In the end, it's not only about improving an algorithm, but also about improving the application's entire life cycle.
Inria researcher of the Ex-Situ project-team
However, that is not the end of the story. "It's a much broader problem than just picking a clean cloud," according to Janin Koch it's actually a matter of rethinking what we really need. “The trend in the AI community is to say: the more data, the more complex the model, the better the final results. Which, to some extent, is not entirely unfounded, especially when it comes to complex problems. However, many applications do not necessarily require this level of accuracy or this volume of data." So, before we even begin an AI project, scientists should ask themselves, ‘What do I really need?’
Are there more sustainable alternatives that need less data or require less run time? Instead of collecting large amount of data, couldn’t I just reuse/repurpose existing datasets that are available out there? Should I create and train a model from scratch, or can I reuse one that is already available in code repositories? Is it really necessary to run my model for a long time? “In the end, it's not only about improving an algorithm, but also about improving the application's entire life cycle."
Human-Centric Interactive Framework
Aside from raising awareness about sustainability trade-offs, the project intends to create an interactive tool to help developers make more sustainable decisions at every stage of the development process. And this is where Koch's contribution comes in. "My research area is Human-Computer Interaction. I am interested in how humans and systems can collaborate to explore new ideas. HCI includes both how users express their goals to a system, and how systems make suggestions and explain them iteratively."
"In the context of this project, this means: What do developers know before starting a project, and how could they describe the overall goal to a system? This can be quite vague at times. Hence, we consider how systems can assist in determining what is required to achieve a given goal and which approaches would be suitable for doing so."
Human-Computer Interaction includes both how users express their goals to a system, and how systems make suggestions and explain them iteratively.
Inria researcher of the Ex-Situ project-team
For such a tool to work, it must be able to explain to the user how a decision is reached, how a conclusion is made, how a constraint is enforced. “This process is actually quite challenging. If an algorithm claims that one particular decision is 80% better, what does it mean to the user? That’s not the way people understand things”. Instead, she suggest to contextualize explanations within the project’s goals and the process of the user, to make these explanations more meaningful.
Koch is also involved in two European projects closely related to this latter line of research. The first is ALMA, a project intended to offer a human-centric approach to Algebraic Machine Learning (AML). This relatively new ML technique displays mathematical properties that open a path to a more understandable AI. The method can deliver a better transparency as to why decisions are made and aims to give the user more latitude to interact with the algorithm.
As one of the five projects of the ICT 48 Excellence Network, the second initiative is HumanE AI. This gathering of 53 institutions aims, in essence, to shape the ML revolution in a direction that adheres to Europe’s ethical values. In this respect, one of the identified core challenges is to develop systems that could be described as understanding humans, adapting to complex real-world environments, and appropriately interacting in complex social settings.
Echoing this vision, the SustainML project is expected to have a significant impact on the so-called “democratisation of Green AI”, allowing not just tech giants but also SMEs, private enthusiasts, NGOs, and individual innovators to develop AI in a more sustainable way.