On the (im)possibility of sustainable artificial intelligence

Rainer Rehak, Weizenbaum Institute for the Networked Society, Berlin, Germany, rainer.rehak@weizenbaum-institute.de

PUBLISHED ON: 30 Sep 2024

This opinion piece is part of AI systems for the public interest, a special issue of Internet Policy Review guest-edited by Theresa Züger and Hadi Asghari.

Funding Note

This work was funded by the Federal Ministry of Education and Research of Germany (BMBF) under grant no. 16DII131 (“Deutsches Internet-Institut”).

Introduction

Two of the existential issues of our time are the climate catastrophe and the dramatic loss of biodiversity. They are existential in the way that our human livelihood depends on functioning ecosystems (WBGU, 2019). Therefore, calls for ecological sustainability are in fact calls for protecting human life, democratic societies, and their institutions, such as public interest infrastructures in a globally fair and equitable fashion. The decline of ecological systems directly entails the accelerating magnification of long known political problems like global inequality, poverty, hunger, or war. Increased scarcity of life-sustaining resources, even concerning habitable areas, fuel conflict and injustice (Abel et al., 2019).

In light of the herculean endeavour of shaping sustainable societies (e.g. see the UN Sustainable Development Goals), where the needs of all present human beings as well as future generations are met within the planetary ecological boundaries (Raworth, 2012), promising approaches and solutions are desperately needed. The search especially includes the use of digital tools like artificial intelligence (AI), which is currently considered a “game-changer” and “revolutionary” technology in the fight against climate change by many, even the UN itself (UN-OHCHR, 2024).

While AI is not one single technology but a large bouquet of related digital methods, there are indeed use cases for sustainability-oriented AI systems like circular economy optimisation (Wilson et al., 2022), reduction of resource and energy consumption (Himeur et al., 2021), local CO2 emission reduction (Alli et al., 2023), smart city optimisation (Cugurullo, 2020), sustainable mobility (Vermesan et al., 2021), waste separation and disposal (Wilts et al., 2021), and new ways of information gathering like environmental pollution detection (Pouyanfar et al., 2022) or radically improved approaches to earth observation (Bereta at al., 2018). Calling these and other concrete examples a sustainability game-changer discloses a very reduced understanding of the sustainability project, as well as of AI technology, because it ignores the political and economic context of AI applications and even its own resource demand. I therefore want to argue that centering any AI technology in the sustainability struggle will, if continuing on the current path, very likely do more harm than good and might even be a distraction from the actually relevant societal tasks of a sustainability transformation. I advocate for moving forward in a problem- and goal-oriented manner and not in a technology- or even AI-driven one, as AI might eventually play a role in societal sustainability transitions, although a very small one.

In this short opinion piece I extensively draw from insights by critical data and algorithm studies, STS, transformative sustainability science, critical computer science, and, more remotely, public interest theory. I will first outline some core properties of AI systems, then socially contextualise and critically discuss these regarding their sustainability implications. I will make constructive remarks concerning a sensible use of certain kinds of AI in combating problems like climate change and conclude the piece.

AI materialities

AI, like all digital tools, has specific common materialities. While the concept of materiality traditionally refers to physical properties, I will use it in a more abstract sense underlining its resource-ness. This includes physical characteristics, but also informational properties and limitations, and even inner and outer social dynamics (Rehak, 2021; Rehak, 2023).

On the physical level it is important to focus on the ecological implications. While there are many possible sustainability-related use cases of AI, the concrete technologies residing under this umbrella term have an astronomically high consumption of material resources like water, cobalt lithium, or energy (e.g., Li et al., 2023; de Vries, 2023), whose production and disposal has major ecological impacts, especially in the global south. Overall, the use of AI skyrockets all indicators of current and future estimated digital ecological impact (Taddeo et al., 2021). Major tech companies like Microsoft and Google just reported rises in resource uses previously unheard of, stopped carbon offsetting, and announced that they would miss their already loose (Hoffmann, 2022) sustainability pledges, precisely because of their large scale AI roll-out (e.g. Marx, 2024; Metz, 2024; Rathi & Bass, 2024). Although there are attempts to minimise AI resource use (Wu et al., 2022; Nenno, 2024), their success is very limited, and efficiency gains will very likely be eaten up by the usual digital rebound effects (Freitag et al., 2021; Lange et al. 2020; Bergman & Foxon, 2022).

Regarding the informational materiality, AI systems can be differentiated between discriminative and generative systems. The former can detect patterns in input data (e.g. categorisation, grouping), the latter can produce patterns (e.g., text, images) related to input data. However both are based on advanced statistics and provide heuristic (not precise) results. This makes AI systems applicable for specific use cases where clearly defined and precise results are not needed, but AI is in no way a “general purpose“ technology (Bender et al., 2021). In addition, AI systems require large amounts of up-to-date valid data to be preconfigured for tasks. This implies a tendency towards centralisation as well as fundamental limits regarding practicability and usefulness (Bender et al., 2021).

And concerning the aspect of social materiality, the global AI supply chain has to be taken into focus. AI inherits the general implications of the digital supply chain, but in a much more intense way. Not only do AI technologies consume vast amounts of materials, lots of them from conflict-torn regions (e.g. conflict minerals, with the mining mainly commissioned by western corporations) and are disposed of in equally problematic e-waste sites heavily affecting the health and freedom of the workers and communities involved. Furthermore, many AI systems require complex and exploitative global networks (Mühlhoff, 2020) of data workers gathering, sorting, categorising, and labelling data in precarious and disempowered work arrangements (Miceli & Posada, 2022), again especially in the global south.

The purpose of (sustainable) AI

After having briefly outlined some key characteristics of AI, we can now reflect further on its current use and its general potential, beyond the ecology-oriented applications mentioned above. A current hot topic in public and academic AI discourse is how to use “AI technology” for the common good. While the transition from focussing on the values being inscribed into AI systems to focussing more on AI governance itself by applying public interest theory seems very promising for specific public service use cases (Züger & Asghari, 2023), a sustainability perspective has to take an overall net impact stance. From this point of view of limited resources, we need to choose wisely where to (not) use AI.

The current overpromises regarding sustainable AI could not be explained by small claims like sorting trash better or optimising the length of machinery use by predictive maintenance alone, but rather by impressive claims like mega city traffic optimisation or global fair resource distribution, meaning: really hard societal problems. But while AI is actually quite useful for sorting trash and predictive maintenance (Wilts et al., 2021; Kamel, 2022), the latter two are, in fact, no technical problems to be calculated, but social problems of collectively agreeing on the very meaning of optimality in the given case.

The example of city traffic optimisation exemplifies that there first has to be a decision on what to optimise for, before then using AI based optimisation systems. Traffic could be optimised for cars, bikes or pedestrians, for low emissions, short commute time, good citizen health, or any other parameter or mix of potentially contradicting parameters. AI cannot make this decision, since it has neither agency nor is it neutral (Rehak, 2021; Prietl, 2019). It cannot produce objectively good solutions agreeable by everyone. I leave the task to the reader to apply this thought experiment to the case of global resource distribution.

The main claim of AI, that it could technically produce a result, which is in fact the social precondition necessary to meaningfully apply AI, is clearly just circular reasoning. This misjudgement explains why AI can not be the “game changer” being able to break the glass ceiling of transformation (Hausknost, 2019): AI systems can only implement what we as society have decided so far and are always created under the current, fundamentally unsustainable framework conditions. AI therefore currently, and under current conditions, only helps to carry on with business as usual (cf. Kwet, 2019), and sustainable AI only works as change placebo resulting in placebo change; or to put it more concisely: “sustainable AI is the technical solution to the climate crisis from a techno-solutionist vantage point simply reproducing the status quo. The enthusiasm for sustainable AI primarily serves hegemonic interests” (Schütze, 2024, p. 1).

The necessary societal and political negotiations cannot be automated (Eyert & Lopez, 2023), and for the sake of sustainability it is crucial to decenter technological thinking and focus on questions of power and social change (Creutzig et al., 2022).

A path forward in the public interest

If we do not want to employ sustainability reasons to ban AI altogether, because we still want to harness the impressive results it can have in some areas, two guiding principles should be seriously discussed and taken into account in policy discourse, better sooner than later.

Stop unformation gathering – Many new sustainability-related AI applications promise to generate new insights, may it be regarding forest health, biodiversity indicators or crop quality. Yet, in most cases of possible protective climate or biodiversity action we as humanity already have sufficient scientific actionable insight. Getting more information is never bad, but the grave implications of AI use might not justify mere theoretical curiosity. Such research could even have a delaying effect, if decision-makers use it to wait for more detailed but practically meaningless results. I call these kinds of results, i.e. the information produced by an obsessive focus on getting more data – without taking action on the sufficient knowledge already available – unformation, which must be called out and prevented, especially when used as pretext for inaction.

Small is beautiful – The larger an AI system, the more devastating are the sustainability-related effects as described above. Although there is much talk currently regarding large language models (LLMs) like ChatGPT, Bart, and the like, the field of AI itself is already nearly 70 years old and has also, together with adjacent disciplines such as statistics, produced all kinds of rather lightweight approaches (e.g. Hansen et al., 2020). Those small systems are usually really good for one specific use case, like water usage optimisation or visual bike detection. They are not only environmentally lightweight, but also do not promote power centralisation like LLMs necessarily do (Rehak, 2023, p. 30); small AI is beautiful AI. Furthermore, currently all kinds of purposes are being pursued utilising powerful AI systems. However, in many areas it has been proven that with a bit of development effort, similarly usable results can be achieved using conventional and lightweight non-AI data processing methods, such as traditional statistics (e.g. Cerasa et al., 2022). If this is possible, AI must not be used. So let’s focus on lightweight systems and call for digital degrowth (Aanestad, 2023; Kwet 2024).

Collective conclusion

Currently hyped kinds of AI like LLMs are amongst the resource hungriest digital technologies we humans have ever developed and they depend on exceptionally exploitative and destructive AI supply chains, e.g. the broad introduction of AI drives the construction of one hyperscaler data centre after the other (Marx, 2024) while taking renewable energy away from other uses. From a sustainability point of view such technology needs an extremely well argued case for still being deployed today. Merely increasing profit for a few is no such case and given the fact that LLMs are largely used for misinformation, surveillance, and desire creation (Castor & Gerard, 2023), it is hard to see why such systems should exist at all. There are indeed some benefits, but are they really worth it?

No single person can answer that, so this clearly is a topic for collective negotiation, preferably as a public interest issue and preferably without AI stakeholders.

References

Aanestad, M. (2023). Digital degrowth – Beyond solutionism. In M. R. Jones, A. S. Mukherjee, D. Thapa, & Y. Zheng (Eds.), After Latour: Globalisation, inequity and climate change (Vol. 696, pp. 55–60). Springer. https://doi.org/10.1007/978-3-031-50154-8_6

Abel, G. J., Brottrager, M., Crespo Cuaresma, J., & Muttarak, R. (2019). Climate, conflict and forced migration. Global Environmental Change, 54, 239–249. https://doi.org/10.1016/j.gloenvcha.2018.12.003

Alli, Y. A., Oladoye, P. O., Onawole, A. T., Anuar, H., Adewuyi, S., Ogunbiyi, O. D., & Philippot, K. (2023). Photocatalysts for CO2 reduction and computational insights. Fuel, 344. https://doi.org/10.1016/j.fuel.2023.128101

Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). On the dangers of stochastic parrots: Can language models be too big? Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 610–623. https://doi.org/10.1145/3442188.3445922

Bereta, K., Koubarakis, M., Manegold, S., Stamoulis, G., & Demir, B. (2018). From Big Data to Big Information and Big Knowledge: The case of Earth observation data. Proceedings of the 27th ACM International Conference on Information and Knowledge Management, 2293–2294. https://doi.org/10.1145/3269206.3274270

Bergman, N., & Foxon, T. J. (2022). Drivers and effects of digitalization on energy demand in low-carbon scenarios. Climate Policy, 23(3), 329–342. https://doi.org/10.1080/14693062.2022.2145260

Castor, A., & Gerard, D. (2023, September 12). Pivot to AI: Pay no attention to the man behind the curtain. Amy Castor. https://amycastor.com/2023/09/12/pivot-to-ai-pay-no-attention-to-the-man-behind-the-curtain/.

Cerasa, A., Tartarisco, G., Bruschetta, R., Ciancarelli, I., Morone, G., Calabrò, R. S., Pioggia, G., Tonin, P., & Iosa, M. (2022). Predicting outcome in patients with brain injury: Differences between machine learning versus conventional statistics. Biomedicines, 10(9). https://doi.org/10.3390/biomedicines10092267

Creutzig, F., Acemoglu, D., Bai, X., Edwards, P. N., Hintz, M. J., Kaack, L. H., Kilkis, S., Kunkel, S., Luers, A., Milojevic-Dupont, N., Rejeski, D., Renn, J., Rolnick, D., Rosol, C., Russ, D., Turnbull, T., Verdolini, E., Wagner, F., Wilson, C., … Zumwald, M. (2022). Digitalization and the anthropocene. Annual Review of Environment and Resources, 47(1), 479–509. https://doi.org/10.1146/annurev-environ-120920-100056

Cugurullo, F. (2020). Urban artificial intelligence: From automation to autonomy in the smart city. Frontiers in Sustainable Cities, 2. https://doi.org/10.3389/frsc.2020.00038

de Vries, A. (2023). The growing energy footprint of artificial intelligence. Joule, 7(10), 2191–2194. https://doi.org/10.1016/j.joule.2023.09.004

Eyert, F., & Lopez, P. (2023). Rethinking transparency as a communicative constellation. Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, 444–454. https://doi.org/10.1145/3593013.3594010

Freitag, C., Berners-Lee, M., Widdicks, K., Knowles, B., Blair, G. S., & Friday, A. (2021). The real climate and transformative impact of ICT: A critique of estimates, trends, and regulations. Patterns, 2(9). https://doi.org/10.1016/j.patter.2021.100340

German Advisory Council on Global Change (WBGU). (2019). Towards our common digital future [Flagship Report]. https://www.wbgu.de/en/publications/publication/towards-our-common-digital-future

Hansen, E. B., Iftikhar, N., & Bøgh, S. (2020). Concept of easy-to-use versatile artificial intelligence in industrial small & medium-sized enterprises. Procedia Manufacturing, 51, 1146–1152. https://doi.org/10.1016/j.promfg.2020.10.161

Hausknost, D. (2020). The environmental state and the glass ceiling of transformation. Environmental Politics, 29(1), 17–37. https://doi.org/10.1080/09644016.2019.1680062

Himeur, Y., Ghanem, K., Alsalemi, A., Bensaali, F., & Amira, A. (2021). Artificial intelligence based anomaly detection of energy consumption in buildings: A review, current trends and new perspectives. Applied Energy, 287. https://doi.org/10.1016/j.apenergy.2021.116601

Hoffmann, M. L. (2022). Climate pledge rating A hands-on evaluation and visualization of companies’ responses to climate change (GCA Working Paper Series) [Working paper]. Green Consumption Assistant (GCA) Research Project. https://doi.org/10.14279/DEPOSITONCE-18624

Kamel, H. (2022). Artificial intelligence for predictive maintenance. Journal of Physics: Conference Series, 2299. https://doi.org/10.1088/1742-6596/2299/1/012001

Kwet, M. (2019). Digital colonialism: South Africa’s education transformation in the shadow of Silicon Valley [Doctoral dissertation, Rhodes University]. http://dx.doi.org/10.2139/ssrn.3496049

Kwet, M. (2024). Digital degrowth: Technology in the age of survival. Pluto Press.

Lange, S., Pohl, J., & Santarius, T. (2020). Digitalization and energy consumption. Does ICT reduce energy demand? Ecological Economics, 176. https://doi.org/10.1016/j.ecolecon.2020.106760

Li, P., Yang, J., Islam, M. A., & Ren, S. (2023). Making AI less ‘thirsty’: Uncovering and addressing the secret water footprint of AI models (Version 3). arXiv. https://doi.org/10.48550/arXiv.2304.03271

Marx, P. (2024, July 5). Generative AI is a climate disaster – Tech companies are abandoning emissions pledges to chase AI market share. Disconnect. https://disconnect.blog/generative-ai-is-a-climate-disaster/.

Metz, R. (2024, July 2). Google’s emissions shot up 48% over five years due to AI. Bloomberg. https://www.bloomberg.com/news/articles/2024-07-02/google-s-emissions-shot-up-48-over-five-years-due-to-ai

Miceli, M., & Posada, J. (2022). The data-production dispositif. Proceedings of the ACM on Human-Computer Interaction, 6(CSCW2), 1–37. https://doi.org/10.1145/3555561

Mühlhoff, R. (2020). Human-aided artificial intelligence: Or, how to run large computations in human brains? Toward a media sociology of machine learning. New Media & Society, 22(10), 1868–1884. https://doi.org/10.1177/1461444819885334

Nenno, S. (2024). Potentials and limitations of active learning: For the reduction of energy consumption during model training. Weizenbaum Journal of the Digital Society, 4(1). https://doi.org/10.34669/WI.WJDS/4.1.3

Pouyanfar, N., Harofte, S. Z., Soltani, M., Siavashy, S., Asadian, E., Ghorbani-Bidkorbeh, F., Keçili, R., & Hussain, C. M. (2022). Artificial intelligence-based microfluidic platforms for the sensitive detection of environmental pollutants: Recent advances and prospects. Trends in Environmental Analytical Chemistry, 34. https://doi.org/10.1016/j.teac.2022.e00160

Prietl, B. (2019). Big data: Inequality by design? Proceedings of the Weizenbaum Conference 2019: Challenges of Digital Inequality, 75–84. https://doi.org/10.34669/WI.CP/2.11

Rathi, A., & Bass, D. (2024, May 15). Microsoft’s AI push imperils climate goal as carbon emissions jump 30%. Bloomberg. https://www.bloomberg.com/news/articles/2024-05-15/microsoft-s-ai-investment-imperils-climate-goal-as-emissions-jump-30.

Raworth, K. (2012). A safe and just space for humanity: Can we live within the doughnut? (Oxfam Discussion Papers) [Discussion paper]. Oxfam. https://www-cdn.oxfam.org/s3fs-public/file_attachments/dp-a-safe-and-just-space-for-humanity-130212-en_5.pdf

Rehak, R. (2021). The language labyrinth: Constructive critique on the terminology used in the AI discourse. In P. Verdegem (Ed.), AI for everyone? Critical perspectives (pp. 87–102). University of Westminster Press. https://doi.org/10.16997/book55.f

Rehak, R. (2023). Artificial intelligence for real sustainability. In P. Jankowski, A. Höfner, M. L. Hoffmann, F. Rohde, R. Rehak, & J. Graf (Eds.), Shaping digital transformation for a sustainable society: Contributions from Bits & Bäume (pp. 26–31). Technische Universität Berlin. https://depositonce.tu-berlin.de/handle/11303/18717

Schütze, P. (2024). The problem of sustainable AI: A critical assessment of an emerging phenomenon. Weizenbaum Journal of the Digital Society, 4(1). https://doi.org/10.34669/WI.WJDS/4.1.4

Taddeo, M., Tsamados, A., Cowls, J., & Floridi, L. (2021). Artificial intelligence and the climate emergency: Opportunities, challenges, and recommendations. One Earth, 4(6), 776–779. https://doi.org/10.1016/j.oneear.2021.05.018

United Nations. (2015). Transforming our world: The 2030 agenda for sustainable development [Agenda]. https://sdgs.un.org/2030agenda

United Nations Office of the High Commissioner for Human Rights (UN-OHCHR). (2024, June 14). Artificial intelligence: A game-changer for sustainable development. https://www.ohchr.org/en/stories/2024/06/artificial-intelligence-game-changer-sustainable-development

Vermesan, O., John, R., Pype, P., Daalderop, G., Kriegel, K., Mitic, G., Lorentz, V., Bahr, R., Sand, H. E., Bockrath, S., & Waldhör, S. (2021). Automotive intelligence embedded in electric connected autonomous and shared vehicles technology for sustainable green mobility. Frontiers in Future Transportation, 2. https://doi.org/10.3389/ffutr.2021.688482

Wilson, M., Paschen, J., & Pitt, L. (2022). The circular economy meets artificial intelligence (AI): Understanding the opportunities of AI for reverse logistics. Management of Environmental Quality: An International Journal, 33(1), 9–25. https://doi.org/10.1108/MEQ-10-2020-0222

Wilts, H., Garcia, B. R., Garlito, R. G., Gómez, L. S., & Prieto, E. G. (2021). Artificial intelligence in the sorting of municipal waste as an enabler of the circular economy. Resources, 10(4). https://doi.org/10.3390/resources10040028

Wu, C.-J., Raghavendra, R., Gupta, U., Acun, B., Ardalani, N., Maeng, K., Chang, G., Behram, F. A., Huang, J., Bai, C., Gschwind, M., Gupta, A., Ott, M., Melnikov, A., Candido, S., Brooks, D., Chauhan, G., Lee, B., Lee, H.-H. S., … Hazelwood, K. (2021). Sustainable AI: Environmental implications, challenges and opportunities (Version 2). arXiv. https://doi.org/10.48550/ARXIV.2111.00364

Züger, T., & Asghari, H. (2023). AI for the public. How public interest theory shifts the discourse on AI. AI & Society, 38(2), 815–828. https://doi.org/10.1007/s00146-022-01480-5