On the Carbon Footprint of Economic Research in the Age of Generative AI

arXiv cs.AI / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that Green AI research should measure the carbon footprint of end-to-end computational workflows using GenAI tools, not just the emissions of training or model inference.
  • It models prompts as decision policies that determine what gets executed and when iteration stops, framing researcher–system discretion as a controllable lever for emissions.
  • The authors map recent Green AI work into seven themes, noting that training footprint remains the largest area while inference efficiency and system-level optimization are accelerating.
  • Benchmarking a modern economic survey workflow (LDA-based mapping) with GenAI-assisted coding shows that generic “green” prompt wording has no reliable effect, but operational constraints and decision-rule prompts produce large, stable CO2e reductions.
  • The results suggest human-in-the-loop governance and rule-based prompt constraints can align GenAI productivity with environmental efficiency without changing the decision-equivalent outputs.

Abstract

Generative artificial intelligence (AI) is increasingly used to write and refactor research code, expanding computational workflows. At the same time, Green AI research has largely measured the footprint of models rather than the downstream workflows in which GenAI is a tool. We shift the unit of analysis from models to workflows and treat prompts as decision policies that allocate discretion between researcher and system, governing what is executed and when iteration stops. We contribute in two ways. First, we map the recent Green AI literature into seven themes: training footprint is the largest cluster, while inference efficiency and system level optimisation are growing rapidly, alongside measurement protocols, green algorithms, governance, and security and efficiency trade-offs. Second, we benchmark a modern economic survey workflow, an LDA-based literature mapping implemented with GenAI assisted coding and executed in a fixed cloud notebook, measuring runtime and estimated CO2e with CodeCarbon. Injecting generic green language into prompts has no reliable effect, whereas operational constraints and decision rule prompts deliver large and stable footprint reductions while preserving decision equivalent topic outputs. The results identify human in the loop governance as a practical lever to align GenAI productivity with environmental efficiency.