Abstract
The ongoing global race for bigger and better artificial intelligence (AI) systems is expected to have a profound societal and environmental impact by altering job markets, disrupting business models, and enabling new governance and societal welfare structures that can affect global consensus for climate action pathways. However, the current AI systems are trained on biased datasets that could destabilize political agencies impacting climate change mitigation and adaptation decisions and compromise social stability, potentially leading to societal tipping events. Thus, the appropriate design of a less biased AI system that reflects both direct and indirect effects on societies and planetary challenges is a question of paramount importance. In this paper, we tackle the question of data-centric knowledge generation for climate action in ways that minimize biased AI. We argue for the need to co-align a less biased AI with an epistemic web on planetary health challenges for more trustworthy decision-making. A human-in-the-loop AI can be designed to align with three goals. First, it can contribute to a planetary epistemic web that supports climate action. Second, it can directly enable mitigation and adaptation interventions through knowledge of social tipping elements. Finally, it can reduce the data injustices associated with AI pretraining datasets.
Similar content being viewed by others
Introduction
The age of artificial intelligence (AI) has begun and is filled with opportunities and responsibilities. It is yet to be clearly understood how AI or machine intelligence can help address present global challenges, including climate change.
A global digital transformation would need an unprecedented level of machine intelligence. Making this machine intelligence sustainable and aligning it with planetary health challenges is a grand challenge on its own, starting with the rapid reduction of GHG emissions associated with the internet and currently carbon-intensive data centers1,2. The literature emphasizes several ways in which AI can play a crucial role in addressing climate change. It can provide innovative solutions to mitigate the negative impacts of greenhouse gas emissions, increase energy efficiency, and promote sustainable development3 (discussed later in detail).
Addressing climate change through AI is extremely challenging because of the enormous number of variables associated with this complex system. For instance, climate datasets are vast and take a significant amount of time to collect, analyze and use to make informed decisions that can translate into climate action. Using AI to account for the continually changing factors of climate change allows us to generate better-informed predictions about environmental changes, allowing us to deploy mitigation strategies earlier. This remains one of the most promising applications of AI in climate action planning. However, while explaining the potential of AI tools in physics-driven modeling of earth systems for predicting climate change, Irrgang et al.4 emphasize the need to rely on clear, physically meaningful research hypotheses, the geophysical determinism of process-based modeling and careful human evaluation against domain-specific knowledge to develop a meaningful AI that can address the challenges of climate science with classical earth system models.
Moreover, as the embodied impact of some of the current machine intelligence and AI systems associated with cryptocurrency mining, cloud computing, and large-scale machine learning models is just beginning to be understood, the accelerating impact of digitalization on consumption and resource extraction appears to be an increasingly troubling problem. As a result, our current trajectory of digitalization seems like a double-edged sword that may increase greenhouse gas emissions, worsening overall planetary health2.
Furthermore, digitalization’s influence on the natural environment and social systems is unknown and will require careful public policy design in many domains5. The desirable design of an accountable machine intelligence system, reflecting both direct and indirect effects on societies and planetary health, is a question of paramount importance that we expand on in this paper from a design thinking lens. We emphasize the need to co-align an epistemic web of planetary health challenges with the goals of a less-biased climate action AI, and debiasing large-scale pretraining datasets can pave the initial path. Key concepts and definitions used in this paper are illustrated in Box 1.
An epistemic web of planetary health challenges for climate action
Climate action through machine intelligence must mean supporting climate mitigation and adaptation decisions at a global scale while avoiding excess emissions. However, the current generation of machine intelligence systems that drive digitalization has embedded biases and data justice issues, making them less trustworthy for transparent decision-making. Thus, for effective climate action, there is a need for a less-biased and collaborative AI that works not in competition with humans but with them to address such urgent planetary health challenges6,7 —emphasizing a human-centric/human-in-the-loop AI. Different people must bring their perspectives and knowledge to developing a less biased AI. Such a knowledge system could constitute what Jürgen Renn8 calls an ‘epistemic web’.
In this perspective, we investigate data-centric knowledge generation for climate action in the context of biased (or less biased) AI. For this, we envision the co-production of an epistemic web of planetary health challenges based on Renn’s epistemology8, relying on social, semiotic, and semantic networks for aligning desirable machine intelligence that captures the closely intertwined dimensions of the present human knowledge that informs current generation machine intelligence models (see Fig. 1). Individual or collective actors form the basis of a social network where these actors possess knowledge and are involved in producing, exchanging, disseminating, or appropriating knowledge. The process of social and communicative exchange manifests in the form of traditions, rules, conventions, and norms, as well as in terms of constraints and power structures that strengthen, weaken, facilitate, or impede ties within social networks8. These form the basis of the ‘contextualization’ of global challenges, where, for example, local knowledge and social norms can derive relevant climate mitigation and adaptation approaches9. However, existing data injustices represent a meaningful deterrent to realizing more inclusive knowledge and experience when it comes to climate action1,10,11,12.
The semiotic network, which communicates meaning, includes the entire material context of the action, including technological artifacts generated based on the technological knowledge of producers8. A recent example is OpenAI’s ChatGPT13, which uses billions of text parameters from Wikipedia and other internet sources as its pre-trained dataset. It produces ‘new knowledge’ in the form of a dialog format. According to Renn8, historically, semiotic networks are often the starting point for reconstructing other aspects of the epistemic web. This shapes the motivation of this article. As more significant and better AI models emerge, we can align accountable machine intelligence with an epistemic web of planetary health challenges.
Semantic networks must be reconstructed from external representations, such as written texts. It uses the fact that concepts have expressions in language. However, semantic networks have no one-to-one relation to either concepts or other cognitive building blocks; one and the same concept may correspond to different terms in language, while the same term may represent different concepts8. This has been the basis of many pre-trained large language models (LLMs) for AI systems, including foundational models like GPTs. Theoretically, Renn8 argues for deductively organized semantic networks to form highly organized knowledge systems. We expand Renn’s framework to create a way forward for co-producing a less biased machine intelligence with an AI-driven epistemic web of planetary challenges through digitalization actions.
Using the theoretical basis of the epistemic web of planetary health challenges, we take a deconstructivist approach to analyze how current pre-trained machine intelligence influences relationships among digitalization, equity, political agency, societal stability, and climate action. In doing so, we first define the scope of present machine intelligence systems in climate action, especially in relation to mitigation and adaptation. Next, we show that for an epistemic web of planetary challenges, there is a need to overcome accountability risks associated with biased AI systems for climate action (see Table 1). This is where the social network dimension of Renn’s epistemic web (see Fig. 1) becomes important as a foundation for collective consensus generation and societal stability. We emphasize that a human-in-the-loop AI design is critical to such an epistemic web that is free of biased datasets, biased programming, and biased algorithms (see Fig. 2). Finally, we emphasize removing the barriers to diversity and inclusion in the machine intelligence community to create grounded and reliable training datasets that can sustain an AI-driven epistemic loop of ‘data from knowledge’ and ‘knowledge from data’ (see Fig. 3).
Machine intelligence accountability risks for climate action
At present, machine intelligence systems for climate action are at an early stage of development, and their impact is just beginning to be understood, which embeds biases in their entire value chain, making these AI systems less trustworthy for climate action decision-making14. Drawing inference from the application of AI in climate modeling (as discussed above), biases can influence prediction accuracy, reliability, and interpretability, which can seriously affect decisions for climate mitigation and adaptation actions. For example, if a biased climate model is trained on data that excludes certain regions or time periods, the predictions may not accurately reflect the complete scope of climate change. Similarly, such models may underrepresent certain variables or factors and provide inaccurate estimations of carbon emissions from particular industries, leading to an underestimation of the actual impact of those variables on the climate. Furthermore, biased climate models can worsen climate impact and response inequities. If AI models only consider how climate change will affect a small number of regions or populations, those regions and people may experience disproportionately negative effects.
Biases can arise from overlapping classes like biased databases, biased algorithms, and biased programming. For instance, Rich and Gureckis15 point out three causes of bias in present machine intelligence systems: small and incomplete datasets, learning from the results of our decisions, and biased inference and evaluation processes. These biases reduce the accuracy and reliability of many present-generation machine intelligence systems for climate action, making them less accountable, interpretable, and explainable (see a comprehensive survey of black box models here16). Fairness, Accountability, and Transparency (FAccT) researchers present five algorithmic accountability risks17 that can emerge from such biases. We synthesize these risks with respect to the design of climate action AI (see Table 1).
While FAccT and AI ethics researchers are beginning to discuss the potential role of AI in mitigating the aforementioned accountability issues in climate action3,14,18, the paths to developing a less biased AI for climate assessment remain uncertain. From this paper’s scoping of the epistemic web in Fig. 1, we focus on the need for quality training datasets that represent the diverse grounded reality of human perspective and experiences (epistemic knowledge) with as little bias as possible, which becomes highly critical in making AI less biased19,20. Also, this feature of representing the different kinds of human knowledge in the machine intelligence system is needed when these models will be relied upon to make decisions about how to deal with climate change as a planetary health challenge. For example, Gupta et al.21 define and implement Earth system justice to link the physical and social aspects of climate change and make sure that planetary health actions reduce harm, improve well-being, and reflect both substantive and procedural justice.
Thus, a desirable feature of machine intelligence and AI systems for climate action is the embedding of epistemic knowledge, which can be achieved through diverse and representative pre-training datasets. However, literature shows embedding epistemic knowledge is not simple, as even the most advanced present generation AI systems must be made more transparent, explainable, and interpretable22. Interpretability means that cause and effect can be determined in machine learning (ML) models. Explainability is crucial for piercing the black box of ML models and understanding how the model works16,23. These two characteristics are critical for reducing algorithmic accountability risks (see Table 1) and making AI safer for high-stakes climate action decision-making23.
Another significant impediment is the need for more precise uncertainty quantification in existing AI systems24. It makes many current-generation machine intelligence systems overconfident, and they make mistakes25. Therefore, the current epistemic base (like the world wide web) for machine intelligence embeds these limitations and biases that make it less practical for individual and collective decision-making for climate action. As a result, current AI systems are less useful for direct applications in climate mitigation and adaptation.
Aligning human-in-the-loop AI design with climate action
Recent advances in human-in-the-loop machine learning approaches using large-language models (LLMs) show a way forward to integrate epistemic feedback loops into black box models. Human-in-the-loop models refer to machine learning systems or algorithms that involve human interaction and input to improve their accuracy. A recent example is Open AI’s ChatGPT13 which uses a human-in-the-loop26 system. In ChatGPT, the AI system interacts conversationally with the human user. This dialog format makes it possible to answer follow-up questions, admit mistakes, challenge incorrect premises, and reject inappropriate requests13. The machine intelligence element of ChatGPT is in its model training using reinforcement learning from human feedback (RLHF)27 driven by the proximal policy optimization (PPO) algorithm.
As a starting point for conceptualizing an epistemic web, Creutzig et al.2 demonstrate a relationship among digitalization, equity, political agency, climate change and planetary stability. It emphasizes AI’s direct and indirect impact on climate action. For example, a direct impact like energy demand for training large machine learning models in data centers. Indirect impacts like machine intelligence applications that reduce greenhouse gas emissions and environmental impact. The digitalization of social networks via algorithms (i.e., social media platforms) is instrumental in creating polarization (through misinformation and disinformation)28 and shaping political opinion that affects social equity within and between countries. High levels of inequity and polarization reduce the feasibility of consensual climate actions, leading to irreversible social tipping scenarios. They are thus indirectly relevant for machine intelligence design and its reward models.
We connect the epistemic interdependencies of machine intelligence with political agency and democracy, equity, and social tipping elements (discussed in the next section). We illustrate an epistemic web basis to define desirable machine intelligence for digitalization that balances social equity, stabilizes political agency (and therefore democracy), and ensures climate mitigation and adaptation goals are met through sustained climate action, thus, potentially preventing irreversible social tipping3,29,30. Digitalization should enable collective action (as data for knowledge) to be transferred from the epistemic web of planetary challenges for training the AI systems. Thus, enabling knowledge generation from the data. This will define the true scale of a human-in-the-loop system for planetary-level digitalization.
Increasingly, the context of human-in-the-loop AI is gaining critical importance, as it is beneficial to reduce biases when a diverse group of humans with different identities, backgrounds, and viewpoints, using collective intelligence31, participate in machine intelligence system design26. Under the best circumstances, utilizing such collective intelligence for human-machine collaboration and co-creation results in knowledge generation as an epistemic web.
We present this design framework in Fig. 2, which emphasizes that the epistemic web contains grounded and diverse knowledge of social structures that are critical social tipping elements29. In the social epistemic network (see Fig. 1), AI systems trained through such an epistemic web can help reduce misinformation, remove skepticism, and restore trust32. Thereby ensuring the stability of socio-political institutions that are critical for determining consensus for climate action.
When envisioning a climate action AI system, we establish that such systems are driven by the need for accountability risk reduction (see Table 2) that delivers a less biased AI, coupled with the drive for planetary-scale digitalization that enables collective climate action, which by itself is influenced by the epistemic web. This epistemic web creates a more robust foundation for collective decision-making and individual action, efforts that could come to play a more important role in accelerating climate policies for mitigation and adaptation that could contribute towards minimizing the risks of irreversible social collapse.
Human-in-the-loop AI designed on social tipping points
The most pressing challenge in this context is the need for more diverse and reliable datasets to build different and reliable algorithms that represent grounded reality, as well as deliberate decision-making on these algorithms, which is shaping current debates on the urgent need for data justice and its agencies10.
For example, Schramowski et al.33 have shown that large language models (LLMs) such as BERT, GPT-2, and GPT-3 trained on unfiltered text corpora suffer from degenerated and biased behavior. Nonetheless, the authors successfully demonstrated that the human-corrected, pre-trained LLM could mitigate the associated risks of toxic degeneration. They used a questionnaire survey of 117 questions to create a human-in-the-loop design to give the AI system a moral direction of what is right and wrong to do. This characteristic of climate action AI systems is critical to shaping the epistemic web that embeds knowledge layers of social structures critical to social tipping points.
Such interactive learning of machine intelligence with human intelligence is desirable to foster AI-driven climate action. This human-machine interactivity is at the core of accountable AI systems34 that can reason about social, cultural and moral norms as critical social structure datasets for enabling climate mitigation and adaptation consensus which do not exist currently35. An attempt was made by Forbes et al.36 through the creation of a large-scale social corpora called Social-Chem-101. Similarly, Colas et al.37 conceptualized the immersion of autotelic agents into rich sociocultural worlds for making AI systems more trustworthy (Autotelic agents are intrinsically motivated learning agents that can learn to represent, generate, select and solve their own problems).
As a timely case study for generative AI, reinforcement learning through human feedback (RLHF) in ChatGPT shows that human intelligence can be integrated with machine intelligence to design specific reward models for the AI system. This can be leveraged to improve the trustworthiness of machine intelligence systems through a human-centered design approach38 for fine-tuning RLHF that asks what is desirable in the current megatrend of digital transformation of economies and societies. Such applications of human-in-the-loop design show opportunities for contextualizing machine intelligence for system-scale behavioral climate action that prevents social tipping points (STPs, see Box 1)32. In Box 2, we present a theoretical scoping of data representing social structural layers of social tipping elements (STEs).
It is not yet known how LLM-based human-in-the-loop AI systems like ChatGPT can support climate mitigation and adaptation, especially across the STEs. One recent example is the creation of ClimateBERT, where researchers fine-tuned a LLM using the IPCC reports for improving public understanding of climate risks and making climate information more accessible to the wider community39. However, caution should be taken to remove existing biases and uncertainties in machine intelligence design and operations. For instance, researchers at DeepMind recently unveiled a taxonomy of risks posed by LLMs, which are used to train generative AI40, including: i) discrimination, hate speech, and exclusion; ii) information hazards; iii) misinformation harms; iv) malicious uses; v) human-computer interaction harms; and vi) environmental and socio-economic harms. A part of this problem is that such LLM-based AI systems are far from trustworthy AI systems, as their interpretability and explainability are still exclusively dependent on their pre-training datasets from internet sources like Wikipedia, News, and BookCorpus databases. This triangulates our focus on the need to align with an epistemic web that represents reliable and diverse human knowledge and perspectives.
Co-producing knowledge for climate action and a less biased AI
Present-day AI is less trustworthy for decision-making, especially when it relates to climate mitigation and adaptation efforts. We synthesized the accountability risks associated with such systems in Table 2, as well as the need to act on biased programming, biased algorithms, and biased datasets for a more trustworthy climate action AI. For example, in the previous section, we emphasized that biased AI-led climate modeling can lead to inaccurate forecasting and impact assessment, which will affect decision-making. To correct such biases, the AI systems require humans-in-the-loop, especially to produce and feed the training data into the algorithms at the initial stage of model development26. This calls for sincere efforts towards embedding data justice in the pretraining datasets.
In this purview, we argue that a human-in-the-loop design of the climate action AI is critical that embraces diversity in perspectives and knowledge from engineers, social scientists, philosophers, industry practitioners, policymakers and the public26. For example, the concept of trustworthy AI that is humanistic, just, and ethical is at the core of a desirable machine intelligence system’s design19. We expand this argument for a human-in-the-loop climate action AI.
However, the notion of algorithmic trust is subjective to the context (as illustrated in Table 2) and emphasizes the need for the ML/AI experts to relate how their metrics of trust impact trust by individuals and the public41,42. This makes fairness-aware machine learning (ML) a difficult challenge as the measurement parameters of fairness in terms of bias and related notions are still not clearly understood. For instance, Narayanan43 has identified 21 definitions of fairness in the literature, which cannot necessarily all be obtained at the same time.
Two widely used definitions that have been widely incorporated into ML pipelines are those of individual fairness, which states that individuals who are similar should be treated similarly, and, group fairness, which states that demographic groups should, on the whole, receive similar decisions44,45. It is important to understand what assumptions are reasonable in a given context before developing and deploying fair mechanisms (i.e., contextualization); without this work, incorrect assumptions could lead to unfair mechanisms44,46.
Bridging such an epistemic gap associated with the meaning of fairness is critical if climate action AI systems are to design and implement climate mitigation and adaptation strategies. FAccT scholars have proposed various solutions to establish a subset of fairness notions41,42,47. One approach that is most relevant to our paper’s scoping is to reduce the bias in the pre-training dataset, known as debiasing data, which can fulfill the objectives of data justice10.
A critical step towards debiasing pretraining datasets is creating data-centric AI that emphasizes more accountable and context-specific datasets related to climate action. FAccT literature stresses the need for ‘social transparency’ in making AI systems trustworthy by aligning efforts needed to establish organizational and regulatory ecosystems for the assurance of trustworthy AI in the public domain (e.g., the MATCH model)48,49,50. Moreover, literature at the intersection of social science and data science shows that a data justice lens is instrumental in bringing social transparency (which improves trustworthiness)10,11,12.
In Fig. 3, we highlight that for debiasing the pretraining datasets of a climate action AI, we must create a self-sustaining and interactive mechanism of ‘data from knowledge’ and ‘knowledge from data’. The ‘data’ in this case must contain multi-layered information on climate change impacts, mitigation, and adaptation strategies at an anthropocene scale, which can then generate the needed ‘knowledge’ base for appropriate and contextualized climate action for the avoidance of irreversible social tipping, as discussed earlier.
Here, we synthesize the desired data justice characteristics of a less biased climate action AI that embeds the knowledge of societal structural layers (see Fig. 2) by leveraging existing social justice dimensions of instrumentality (fair ends), distribution (fair allocation), and procedure (fair means) that correspond to the data diversity and inclusion needs through recognition, ownership, structure and space, as illustrated in Table 3. For example, a lack of data justice and its contextualization in most climate-vulnerable regions of the world pose a significant risk of training a biased AI that can virtually hallucinate during decision-making applications. We are already experiencing hallucinatory results with ChatGPT.
In Table 3, we connect these characteristics to the specific requirements of a less biased climate action AI that have applications in climate modeling, energy management, city planning, carbon capture and storage, and collective intelligence (as illustrated in Box 2).
Researchers are finding innovative ways to produce data-centric infrastructure to support this goal. For example, African AI researchers have established a common AI dataset repository for their local context, COCO-Africa51. MasakhaNER provides a large curated dataset of ten African languages with named-entity annotations52. Such initiatives are still very early, and more effort is needed to mainstream them, especially along the five data justice characteristics levers.
Without diversity and the inclusion of a full range of populations, we risk the development of biased algorithms20 and, subsequently, a biased epistemic web. Moreover, there is an added risk of failing to fulfill the AI talent pool and missing its broader societal benefit towards solving planetary challenges like climate change, as discussed above. Using AI for climate action (mitigation and adaptation) is especially challenging as it can be a double-edged sword that may increase greenhouse gas emissions and worsen overall planetary health, which we discuss from a social tipping point lens. This effect is due to embedded biases and injustices in the training datasets used in the design of present-day generative AI systems.
Conclusion
We conceptualized the co-production of an epistemic web of planetary challenges based on Renn’s epistemology8 for aligning desirable machine intelligence that captures the closely intertwined social, semiotic, and semantic network dimensions of present human knowledge that inform current generation pre-trained AI models. This epistemic web can help reduce accountability risks17 associated with machine learning and AI models while correcting algorithm-driven polarization for climate action32, leading to collective consensus for climate adaptation and mitigation policies. We envisaged leveraging the recent advances in human-in-the-loop AI through political agencies and democratic decision-making2. However, existing embedded inequalities associated with AI system fairness, ethics, and biases must be addressed.
The need of the hour is to be sensitive to digital inequalities and injustices within the machine intelligence community, especially when AI is used as an instrument for addressing planetary health challenges like climate change. That is where the role of social science, philosophy and the humanities becomes even more critical. A recent AI community review53 touched on this theme, specifically focusing on academic data science researchers. Bridging such a divide and debiasing datasets should be a core component of a desirable machine intelligence-driven digitalization system. Otherwise, the pre-trained datasets will remain biased and overrepresent certain groups. Thus, leading to a biased climate action AI.
Similarly, there is strong evidence of structural inequalities in climate action (mitigation and adaptation) between and within countries. It is even more prominent in vulnerable and resource-constrained communities in the Global South. Such inequalities, if sustained, are estimated to have catastrophic outcomes impacting societal collapse and planetary stability, including not fulfilling any climate mitigation pathways29.
Better aligning less-biased AI with climate protection and respecting planetary boundaries also creates an unprecedented opportunity to act on global injustices and embed positive data justice thinking in the current wave of digitalization. This encompasses ensuring that the benefits of digitalization redress existing injustices, that vulnerable groups are more involved in standard-setting and policymaking and that disadvantaged groups, in particular, have access to open data. It also suggests that ownership and exploitation of data expand to include civil society and communities themselves (e.g., via cooperatives and trust arrangements), the active participation of users, and the promotion of broadband internet access as a public good rather than a private commodity.
As machine intelligence increasingly impacts society, diversity and inclusion are growing concerns. Therefore, the co-creation and co-production of relevant knowledge, infrastructure, and human resources must be a desirable machine intelligence design priority that defines an epistemic web of collective action for addressing planetary challenges.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
References
Sovacool, B. K., Monyei, C. G. & Upham, P. Making the internet globally sustainable: Technical and policy options for improved energy management, governance and community acceptance of Nordic datacenters. Renew. Sustain. Energy Rev. 154, 111793 (2022).
Creutzig, F. et al. Digitalization and the Anthropocene. Annu. Rev. Environ. Resour. 47, 479–509 (2022).
Rolnick, D. et al. Tackling climate change with machine learning. ACM Comput. Surv. 55, 1–96 (2022).
Irrgang, C. et al. Towards neural Earth system modelling by integrating artificial intelligence in Earth system science. Nat. Mach. Intell. 3, 667–674 (2021).
Creutzig, F. et al. Leveraging digitalization for sustainability in urban transport. Global Sustain. 2, e14 (2019).
Whitmee, S. et al. Safeguarding human health in the Anthropocene epoch: report of The Rockefeller Foundation-Lancet Commission on planetary health. Lancet 386, 1973–2028 (2015).
Ripple, W. J. et al. Many risky feedback loops amplify the need for climate action. One Earth 6, 86–91 (2023).
Renn, J. The Evolution of Knowledge. Princeton University Press, 2020. https://doi.org/10.1515/9780691185675.
Pörtner, H.-O. et al. Climate Change 2022: Impacts, Adaptation, and Vulnerability. Contribution of Working Group II to the Sixth Assessment Report of the Intergovernmental Panel on Climate Change. (2022).
Taylor, L. What is data justice? The case for connecting digital rights and freedoms globally. Big Data Soc. 4, https://doi.org/10.1177/2053951717736335 (2017).
Heeks, R. & Renken, J. Data justice for development: What would it mean? Inf. Dev. 34, 90–102 (2018).
Heeks, R. & Shekhar, S. Datafication, development and marginalised urban communities: an applied data justice framework. Inf. Commun. Soc. 22, 992–1011 (2019).
OpenAI. ChatGPT: Optimizing language models for dialogue. OpenAI https://openai.com/blog/chatgpt/ (2022).
Kaack, L. H. et al. Aligning artificial intelligence with climate change mitigation. Nat. Clim. Chang. 12, 518–527 (2022).
Rich, A. S. & Gureckis, T. M. Lessons for artificial intelligence from the study of natural stupidity. Nat. Mach. Intel. 1, 174–180 (2019).
Guidotti, R. et al. A Survey of Methods for Explaining Black Box Models. ACM Comput. Surv. 51, 1–42 (2018).
Wieringa, M. What to account for when accounting for algorithms: a systematic literature review on algorithmic accountability. in Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency 1–18 https://doi.org/10.1145/3351095.3372833 (Association for Computing Machinery, 2020).
Leal Filho, W. et al. Deploying artificial intelligence for climate change adaptation. Technol. Forecast. Soc. Change 180, 121662 (2022).
Liang, W. et al. Advances, challenges and opportunities in creating data for trustworthy AI. Nat. Mach. Intell. 4, 669–677 (2022).
Wongvibulsin, S. Educational strategies to foster diversity and inclusion in machine intelligence. Nat. Mach. Intell. 1, 70–71 (2019).
Gupta, J. et al. Earth system justice needed to identify and live within Earth system boundaries. Nat. Sust. 1–9 https://doi.org/10.1038/s41893-023-01064-1 (2023).
Gunning, D. et al. XAI-Explainable artificial intelligence. Sci. Robot. 4, (2019).
Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1, 206–215 (2019).
Bécue, A., Praça, I. & Gama, J. Artificial intelligence, cyber-threats and Industry 4.0: challenges and opportunities. Artif. Intell. Rev. 54, 3849–3886 (2021).
Franzoni, V., Vallverdù, J. & Milani, A. Errors, Biases and Overconfidence in Artificial Emotional Modeling. in IEEE/WIC/ACM International Conference on Web Intelligence - Companion Volume 86–90 https://doi.org/10.1145/3358695.3361749 (Association for Computing Machinery, 2019).
Mosqueira-Rey, E., Hernández-Pereira, E., Alonso-Ríos, D., Bobes-Bascarán, J. & Fernández-Leal, Á. Human-in-the-loop machine learning: a state of the art. Artif. Intell. Rev. https://doi.org/10.1007/s10462-022-10246-w (2022).
Christiano, P. et al. Deep reinforcement learning from human preferences. arXiv [stat.ML] (2017).
Hornsey, M. J. & Lewandowsky, S. A toolkit for understanding and addressing climate scepticism. Nat. Hum. Behav. 6, 1454–1464 (2022).
Otto, I. M. et al. Social tipping dynamics for stabilizing Earth’s climate by 2050. Proc. Nat. Acad. Sci. 117, 2354–2365 (2020).
Beckage, B., Moore, F. C. & Lacasse, K. Incorporating human behaviour into Earth system modelling. Nat. Hum. Behav. 6, 1493–1502 (2022).
Yang, V. C. & Sandberg, A. Collective Intelligence as Infrastructure for Reducing Broad Global Catastrophic Risks. arXiv [nlin.AO] (2022).
Debnath, R., van der Linden, S., Sovacool, B. K. & Alvarez, R. M. Facilitating system-level behavioral climate action using computational social science. Nat. Human Behav. https://doi.org/10.1038/s41562-023-01527-7 (2023).
Schramowski, P., Turan, C., Andersen, N., Rothkopf, C. A. & Kersting, K. Large pre-trained language models contain human-like biases of what is right and wrong to do. Nat. Mach. Intell. 4, 258–268 (2022).
Donahue, K., Chouldechova, A. & Kenthapadi, K. Human-Algorithm Collaboration: Achieving Complementarity and Avoiding Unfairness. in 2022 ACM Conference on Fairness, Accountability, and Transparency 1639–1656. https://doi.org/10.1145/3531146.3533221 (Association for Computing Machinery, 2022).
Lepri, B., Oliver, N. & Pentland, A. Ethical machines: The human-centric use of artificial intelligence. iScience 24, 102249 (2021).
Forbes, M., Hwang, J. D., Shwartz, V., Sap, M. & Choi, Y. Social Chemistry 101: Learning to Reason about Social and Moral Norms. arXiv [cs.CL] (2020).
Colas, C., Karch, T., Moulin-Frier, C. & Oudeyer, P.-Y. Language and culture internalization for human-like autotelic AI. Nat. Mach. Intell. 4, 1068–1076 (2022).
Verganti, R., Vendraminelli, L. & Iansiti, M. Innovation and design in the age of artificial intelligence. J. Prod. Innov. Manage. 37, 212–227 (2020).
Webersinke, N., Kraus, M., Bingler, J. A. & Leippold, M. ClimateBert: A Pretrained Language Model for Climate-Related Text. arXiv [cs.CL] (2021).
Weidinger, L. et al. Taxonomy of Risks posed by Language Models. in 2022 ACM Conference on Fairness, Accountability, and Transparency 214–229 https://doi.org/10.1145/3531146.3533088 (Association for Computing Machinery, 2022).
Toreini, E. et al. The relationship between trust in AI and trustworthy machine learning technologies. in Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency 272–283 https://doi.org/10.1145/3351095.3372834 (Association for Computing Machinery, 2020).
Birhane, A. et al. The Forgotten Margins of AI Ethics. in 2022 ACM Conference on Fairness, Accountability, and Transparency 948–958 https://doi.org/10.1145/3531146.3533157 (Association for Computing Machinery, 2022).
Narayanan, A. Translational tutorial: 21 fairness definitions and their politics. in Fairness, Accountability and Transparency (2018).
Friedler, S. A., Scheidegger, C. & Venkatasubramanian, S. The (Im)possibility of fairness: different value systems require different mechanisms for fair decision making. Commun. ACM 64, 136–143 (2021).
Zafar, M. B., Valera, I., Gomez-Rodriguez, M. & Gummadi, K. P. Fairness constraints: a flexible approach for fair classification. J. Mach. Learn. Res. 20, 1–42 (2019).
Barocas, S., Hardt, M. & Narayanan, A. Fairness and Machine Learning: Limitations and Opportunities. (fairmlbook.org, 2019).
Deng, W. H. et al. Exploring How Machine Learning Practitioners (Try To) Use Fairness Toolkits. in 2022 ACM Conference on Fairness, Accountability, and Transparency 473–484 https://doi.org/10.1145/3531146.3533113 (Association for Computing Machinery, 2022).
Liao, Q. V. & Sundar, S. S. Designing for Responsible Trust in AI Systems: A Communication Perspective. in 2022 ACM Conference on Fairness, Accountability, and Transparency 1257–1268 https://doi.org/10.1145/3531146.3533182 (Association for Computing Machinery, 2022).
Knowles, B. & Richards, J. T. The Sanction of Authority: Promoting Public Trust in AI. in Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency 262–271 https://doi.org/10.1145/3442188.3445890 (Association for Computing Machinery, 2021).
Ehsan, U., Liao, Q. V., Muller, M., Riedl, M. O. & Weisz, J. D. Expanding Explainability: Towards Social Transparency in AI systems. in Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems https://doi.org/10.1145/3411764.3445188 1–19 (Association for Computing Machinery, 2021).
Dibia, V. cocoafrica: A Curation Tool and Dataset of Common Objects in the Context of Africa. (Github).
Adelani, D. I. et al. MasakhaNER: Named Entity Recognition for African Languages. arXiv [cs.CL] (2021).
Greene, T., Martens, D. & Shmueli, G. Barriers to academic data science research in the new realm of algorithmic behaviour modification by digital platforms. Nat. Mach. Intell. 4, 323–330 (2022).
Ravuri, S. et al. Skilful precipitation nowcasting using deep generative models of radar. Nature 597, 672–677 (2021).
Ng, N. Y. Y., Gopalan, H., Raghavan, V. S. G. & Ooi, C. C. Model-Agnostic Hybrid Numerical Weather Prediction and Machine Learning Paradigm for Solar Forecasting in the Tropics. arXiv [cs.LG] (2021).
Using machine learning to ‘nowcast’ precipitation in high resolution. https://ai.googleblog.com/2020/01/using-machine-learning-to-nowcast.html?m=1.
Dwivedi, R. et al. Explainable AI (XAI): core ideas, techniques, and solutions. ACM Comput. Surv. 55, 1–33 (2023).
Cheong, S.-M., Sankaran, K. & Bastani, H. Artificial intelligence for climate change adaptation. WIREs Data Mining Knowl. Discov. 12, e1459 (2022).
Agrawal, S. et al. Machine Learning for Precipitation Nowcasting from Radar Images. arXiv [cs.CV] (2019).
Yan, Y. et al. Harnessing the power of machine learning for carbon capture, utilisation, and storage (CCUS) – a state-of-the-art review. Energy Environ. Sci. 14, 6122–6157 (2021).
Wagner, F. et al. Using explainable machine learning to understand how urban form shapes sustainable mobility. Transp. Res. Part D: Trans. Environ. 111, 103442 (2022).
Callaghan, M. et al. Machine-learning-based evidence and attribution mapping of 100,000 climate impact studies. Nat. Clim. Chang. 11, 966–972 (2021).
Alizadeh, T. The spatial justice implications of telecommunication infrastructure: The socio-economic status of early National Broadband Network rollout in Australia. Int. J. Crit. Infrastruc. 11, 278–296 (2015).
Delacroix, S. & Lawrence, N. D. Bottom-up data Trusts: disturbing the ‘one size fits all’ approach to data governance. Int. Data Priv. Law 9, 236–252 (2019).
Komljenovic, J. The future of value in digitalised higher education: why data privacy should not be our biggest concern. High. Educ. 83, 119–135 (2022).
Langley, P. & Leyshon, A. Platform capitalism : the intermediation and capitalization of digital economic circulation. Fin. Soc. 3, 0 (2017).
Mills, S. Who Owns the Future? Data Trusts, Data Commons, and the Future of Data Ownership. https://doi.org/10.2139/ssrn.3437936 (2019)
Pick, J. B. & Nishida, T. Digital divides in the world and its regions: A spatial and multivariate analysis of technological utilization. Technol. Forecast. Soc. Change 91, 1–17 (2015).
Sareen, S., Saltelli, A. & Rommetveit, K. Ethics of quantification: illumination, obfuscation and performative legitimation. Palgrave Commun. 6, https://doi.org/10.1057/s41599-020-0396-5 (2020).
Savona, M. The Value of Data: Towards a Framework to Redistribute It. https://doi.org/10.2139/ssrn.3476668 (2019).
Brock, A., Sovacool, B. K. & Hook, A. Volatile photovoltaics: green industrialization, sacrifice zones, and the political ecology of solar energy in Germany. Ann. Assoc. Am. Geogr. 111, 1756–1778 (2021).
Soja, E. W. Seeking Spatial Justice. (University of Minnesota Press, 2010).
Yenneti, K., Day, R. & Golubchikov, O. Spatial justice and the land politics of renewables: Dispossessing vulnerable communities through solar energy mega-projects. Geoforum 76, 90–99 (2016).
Everitt, T., Lea, G. & Hutter, M. AGI Safety Literature Review. arXiv [cs.AI] (2018).
Goertzel, B. Artificial general intelligence: concept, state of the art, and future prospects. J. Artif. Gen. Intell. 5, 1–48 (2014).
Sovacool, B. K. & Furszyfer Del Rio, D. D. Smart home technologies in Europe: A critical review of concepts, benefits, risks and policies. Ren. Sust. Energy Rev. 120, 109663 (2020).
Furszyfer Del Rio, D. D., Sovacool, B. K. & Martiskainen, M. Controllable, frightening, or fun? Exploring the gendered dynamics of smart home technology preferences in the United Kingdom. Energy Res. Soc. Sci. 77, 102105 (2021).
Sovacool, B., Furszyfer-Del Rio, D. D. & Martiskainen, M. Can Prosuming Become Perilous? Exploring Systems of Control and Domestic Abuse in the Smart Homes of the Future. Frontiers in Energy Research 9, (2021).
Furszyfer Del Rio, D. D., Sovacool, B. K. & Griffiths, S. Culture, energy and climate sustainability, and smart home technologies: A mixed methods comparison of four countries. Energy Clim. Change 2, 100035 (2021).
Milkoreit, M. Social tipping points everywhere?—Patterns and risks of overuse. Wiley Interdiscip. Rev. Clim. Change https://doi.org/10.1002/wcc.813 (2022).
Juhola, S. et al. Social tipping points and adaptation limits in the context of systemic risk: Concepts, models and governance. Front. Clim. 4, (2022).
Farmer, J. D. et al. Sensitive intervention points in the post-carbon transition. Science 364, 132–134 (2019).
Lenton, T. M. et al. Tipping elements in the Earth’s climate system. Proc. Natl. Acad. Sci. USA. 105, 1786–1793 (2008).
Milkoreit, M. et al. Defining tipping points for social-ecological systems scholarship—an interdisciplinary literature review. Environ. Res. Lett. 13, 033005 (2018).
Nyborg, K. et al. Social norms as solutions. Science 354, 42–43 (2016).
van Ginkel, K. C. H. et al. Climate change induced socio-economic tipping points: review and stakeholder consultation for policy relevant research. Environ. Res. Lett. 15, 023001 (2020).
Acknowledgements
R.D. acknowledges the support from the Quadrature Climate Foundation (01-21-000149), Keynes Fund (JHVH) and Google Cloud Climate Innovation Challenge (2022). BKS gratefully acknowledges support from UK Research and Innovation as well as the JPI SOLSTICE 2020 scheme through the “Responsive Organizing for Low Emission Societies (ROLES)” Project, Grant Agreement No. ES/V01403X/1.The authors are also thankful to Professor Adrian Smith and Dr. Max Lacey-Barnacle from the University of Sussex for inspiring the analysis on data justice.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
R.D. and F.C. conceptualized the study. All authors contributed to the drafting and editing of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Debnath, R., Creutzig, F., Sovacool, B.K. et al. Harnessing human and machine intelligence for planetary-level climate action. npj Clim. Action 2, 20 (2023). https://doi.org/10.1038/s44168-023-00056-3
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s44168-023-00056-3