DOE PAGES title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Approximating Nash Equilibrium in Day-ahead Electricity Market Bidding with Multi-agent Deep Reinforcement Learning

Abstract

In this paper, a day-ahead electricity market bidding problem with multiple strategic generation company (GEN-CO) bidders is studied. The problem is formulated as a Markov game model, where GENCO bidders interact with each other todevelop their optimal day-ahead bidding strategies. Considering unobservable information in the problem, a model-free and data-driven approach, known as multi-agent deep deterministic policy gradient (MADDPG), is applied for approximating the Nash equilibrium (NE) in the above Markov game. The MADDPG algorithm has the advantage of generalization due to the automatic feature extraction ability of the deep neural networks. The algorithm is tested on an IEEE 30-bus system with three competitive GENCO bidders in both an uncongested caseand a congested case. Comparisons with a truthful bidding strategy and state-of-the-art deep reinforcement learning methods including deep Q network and deep deterministic policy gradient (DDPG) demonstrate that the applied MADDPG algorithm can find a superior bidding strategy for all the market participants with increased profit gains. In addition, the comparison with a conventional model-based method shows that the MADDPG algorithm has higher computational efficiency, which is feasible for real-world applications.

Authors:
 [1];  [1]; ORCiD logo [2]; ORCiD logo [2]
  1. Univ. of Tennessee, Knoxville, TN (United States)
  2. Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)
Publication Date:
Research Org.:
Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States)
Sponsoring Org.:
USDOE Office of Electricity (OE)
OSTI Identifier:
1782068
Grant/Contract Number:  
AC05-00OR22725
Resource Type:
Accepted Manuscript
Journal Name:
Journal of Modern Power Systems and Clean Energy
Additional Journal Information:
Journal Volume: 9; Journal Issue: 3; Journal ID: ISSN 2196-5625
Publisher:
Springer
Country of Publication:
United States
Language:
English
Subject:
24 POWER TRANSMISSION AND DISTRIBUTION; Games; electricity supply industry; Markov processes; computational modeling; training; Nash equilibrium; mathematical model; bidding strategy; day-ahead electricity market; deep reinforcement learning; Markov game; multi-agent deterministic policy gradient (MADDPG)

Citation Formats

Du, Yan, Li, Fangxing, Zandi, Helia, and Xue, Yaosuo. Approximating Nash Equilibrium in Day-ahead Electricity Market Bidding with Multi-agent Deep Reinforcement Learning. United States: N. p., 2021. Web. doi:10.35833/mpce.2020.000502.
Du, Yan, Li, Fangxing, Zandi, Helia, & Xue, Yaosuo. Approximating Nash Equilibrium in Day-ahead Electricity Market Bidding with Multi-agent Deep Reinforcement Learning. United States. https://doi.org/10.35833/mpce.2020.000502
Du, Yan, Li, Fangxing, Zandi, Helia, and Xue, Yaosuo. Mon . "Approximating Nash Equilibrium in Day-ahead Electricity Market Bidding with Multi-agent Deep Reinforcement Learning". United States. https://doi.org/10.35833/mpce.2020.000502. https://www.osti.gov/servlets/purl/1782068.
@article{osti_1782068,
title = {Approximating Nash Equilibrium in Day-ahead Electricity Market Bidding with Multi-agent Deep Reinforcement Learning},
author = {Du, Yan and Li, Fangxing and Zandi, Helia and Xue, Yaosuo},
abstractNote = {In this paper, a day-ahead electricity market bidding problem with multiple strategic generation company (GEN-CO) bidders is studied. The problem is formulated as a Markov game model, where GENCO bidders interact with each other todevelop their optimal day-ahead bidding strategies. Considering unobservable information in the problem, a model-free and data-driven approach, known as multi-agent deep deterministic policy gradient (MADDPG), is applied for approximating the Nash equilibrium (NE) in the above Markov game. The MADDPG algorithm has the advantage of generalization due to the automatic feature extraction ability of the deep neural networks. The algorithm is tested on an IEEE 30-bus system with three competitive GENCO bidders in both an uncongested caseand a congested case. Comparisons with a truthful bidding strategy and state-of-the-art deep reinforcement learning methods including deep Q network and deep deterministic policy gradient (DDPG) demonstrate that the applied MADDPG algorithm can find a superior bidding strategy for all the market participants with increased profit gains. In addition, the comparison with a conventional model-based method shows that the MADDPG algorithm has higher computational efficiency, which is feasible for real-world applications.},
doi = {10.35833/mpce.2020.000502},
journal = {Journal of Modern Power Systems and Clean Energy},
number = 3,
volume = 9,
place = {United States},
year = {Mon Apr 19 00:00:00 EDT 2021},
month = {Mon Apr 19 00:00:00 EDT 2021}
}