Joint Energy Management and Coordinated AIGC Workload Scheduling for Distributed Data Centers: A Diffusion-Aided Reward Shaping Approach

arXiv cs.LG / 5/6/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a joint framework for energy management and coordinated AIGC workload scheduling across distributed data centers to cut energy costs without sacrificing content generation quality.
  • It explicitly models service quality to enable job transfers between ASPs and to support fine-grained control of inference processes, while jointly leveraging multiple energy resources for greater power flexibility.
  • The authors formulate a system utility maximization problem that balances AIGC service revenue against operational penalties and energy-related costs.
  • Because scheduling decisions are strongly coupled—causing sparse rewards that hinder standard deep reinforcement learning—the paper introduces a diffusion-model-aided reward shaping method that generates denser, complementary reward signals via multi-step denoising.
  • Experiments using real-world models and datasets show improved learning convergence and higher overall system utility, including robustness to electricity price fluctuations and heterogeneity among AIGC models across ASPs.

Abstract

Artificial intelligence-generated content (AIGC) has emerged as a transformative paradigm for automating the creation of diverse and customized content, giving rise to rapidly growing computational workloads in cloud data centers. It is imperative for AIGC service providers (ASPs) to strategically schedule AIGC workloads to reduce data center energy costs while guaranteeing high-quality content generation. However, the distinctive characteristics of AIGC services pose critical challenges, including model heterogeneity across ASPs, implicit service quality evaluation, and complex inference process control. To tackle these challenges, we propose a joint energy management and coordinated AIGC workload scheduling framework, which introduces an explicit mathematical characterization of service quality to promote both job transfer among ASPs and fine-grained inference process configuration. Moreover, various energy resources within data centers are jointly considered to enhance power usage flexibility. Subsequently, a system utility maximization problem is formulated to balance AIGC service revenue with operational penalties and costs. Nevertheless, the strong coupling among job scheduling decisions induces severe reward sparsity, which limits the effectiveness of existing deep reinforcement learning (DRL) algorithms. To address this issue, we develop a diffusion model-aided reward shaping approach to synthesize complementary reward signals through a multi-step denoising process. This approach is seamlessly integrated with DRL to enable efficient learning of scheduling policies under sparse environmental feedback. Experiments based on real-world models and datasets demonstrate that our scheme effectively accommodates electricity price fluctuations and AIGC model heterogeneity, while achieving superior learning convergence and system utility compared with benchmark methods.