2007 IEEE Lausanne Power Tech 2007
DOI: 10.1109/pct.2007.4538442
|View full text |Cite
|
Sign up to set email alerts
|

A Reinforcement Learning Algorithm for Market Participants in FTR Auctions

Abstract: This paper presents a Q-Learning algorithm for the development of bidding strategies for market participants in FTR auctions. Each market participant is represented by an autonomous adaptive agent capable of developing its own bidding behavior based on a Q-learning algorithm. Initially, a bilevel optimization problem is formulated. At the first level, a market participant tries to maximize his expected profit under the constraint that, at the second level, an independent system operator tries to maximize the r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
7
0

Year Published

2009
2009
2020
2020

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 14 publications
0
7
0
Order By: Relevance
“…Q-learning is certainly one of the most studied Reinforcement Learning (RL) algorithm and has been applied with success in several domains, from relatively simple toy problems, such as Cliff-Walking (Sutton & Barto, 1998), to more complex ones, such as web-based education (Iglesias et al, 2008) and face recognition (Harandi et al, 2008). Initially proposed for single-agent environments, the simplicity and effectiveness of this algorithm has led to its application also in multiagent configurations, for example Galstyan et al (2004) and Ziogos et al (2007). In this case, however, its supporting theoretical framework and convergence guarantees are lost.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Q-learning is certainly one of the most studied Reinforcement Learning (RL) algorithm and has been applied with success in several domains, from relatively simple toy problems, such as Cliff-Walking (Sutton & Barto, 1998), to more complex ones, such as web-based education (Iglesias et al, 2008) and face recognition (Harandi et al, 2008). Initially proposed for single-agent environments, the simplicity and effectiveness of this algorithm has led to its application also in multiagent configurations, for example Galstyan et al (2004) and Ziogos et al (2007). In this case, however, its supporting theoretical framework and convergence guarantees are lost.…”
Section: Introductionmentioning
confidence: 99%
“…For example: Galstyan et al (2004) applies the algorithm to develop a decentralized resource allocation mechanism; Gomes and Kowalczyk (2007) study the problem of learning demand functions; and Ziogos et al (2007) investigate the development of bidding strategies. Therefore, in this paper we present a framework to model the dynamics of Multiagent Q-learning with the ǫ-greedy exploration mechanism.…”
Section: Introductionmentioning
confidence: 99%
“…The first FTR auction took place in 1999, in the PJM Interconnection in the U.S. In the auctions, ISOs have a goal of maximizing FTR revenues, subject to the constraints of transmission capacity and contingencies [10]. Electric suppliers calculate FTR values of the paths to bid, based on their own forecasts of future LMP prices in the interested locations.…”
Section: Financial Transmission Rightsmentioning
confidence: 99%
“…Generators submit offers, and customer loads submit bids to the ISO with hourly MWs for each hour of the next day. The ISO calculates a nodal price, or a locational marginal price (LMP) of a location, based on all the submitted offers and bids, subject to the Lagrange multipliers, or constraints of active power balance and transmission [10]. FTR auction results in ISO-NE provide the magnitude of the auctions and major FTR participants [14].…”
Section: Financial Transmission Rightsmentioning
confidence: 99%
“…The importance of obtaining a Q-learning algorithm with ε -greedy exploration is also justified through a large number of applications. For example, Galstyan et al [ 34 ] applied a Q-learning algorithm with ε -greedy exploration to develop a decentralised resource allocation mechanism; Gomes and Kowalczyk [ 35 ] studied the problem of learning demand functions; and Ziogos et al [ 36 ] investigated the development of bidding strategies.…”
Section: The Multi-agent Frameworkmentioning
confidence: 99%