TY - GEN
T1 - Dynamic pricing for smart grid with reinforcement learning
AU - Kim, Byung Gook
AU - Zhang, Yu
AU - Van Der Schaar, Mihaela
AU - Lee, Jang Won
PY - 2014
Y1 - 2014
N2 - In the smart grid system, dynamic pricing can be an efficient tool for the service provider which enables efficient and automated management of the grid. However, in practice, the lack of information about the customers' time-varying load demand and energy consumption patterns and the volatility of electricity price in the wholesale market make the implementation of dynamic pricing highly challenging. In this paper, we study a dynamic pricing problem in the smart grid system where the service provider decides the electricity price in the retail market. In order to overcome the challenges in implementing dynamic pricing, we develop a reinforcement learning algorithm. To resolve the drawbacks of the conventional reinforcement learning algorithm such as high computational complexity and low convergence speed, we propose an approximate state definition and adopt virtual experience. Numerical results show that the proposed reinforcement learning algorithm can effectively work without a priori information of the system dynamics.
AB - In the smart grid system, dynamic pricing can be an efficient tool for the service provider which enables efficient and automated management of the grid. However, in practice, the lack of information about the customers' time-varying load demand and energy consumption patterns and the volatility of electricity price in the wholesale market make the implementation of dynamic pricing highly challenging. In this paper, we study a dynamic pricing problem in the smart grid system where the service provider decides the electricity price in the retail market. In order to overcome the challenges in implementing dynamic pricing, we develop a reinforcement learning algorithm. To resolve the drawbacks of the conventional reinforcement learning algorithm such as high computational complexity and low convergence speed, we propose an approximate state definition and adopt virtual experience. Numerical results show that the proposed reinforcement learning algorithm can effectively work without a priori information of the system dynamics.
UR - http://www.scopus.com/inward/record.url?scp=84904504757&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84904504757&partnerID=8YFLogxK
U2 - 10.1109/INFCOMW.2014.6849306
DO - 10.1109/INFCOMW.2014.6849306
M3 - Conference contribution
AN - SCOPUS:84904504757
SN - 9781479930883
T3 - Proceedings - IEEE INFOCOM
SP - 640
EP - 645
BT - 2014 IEEE Conference on Computer Communications Workshops, INFOCOM WKSHPS 2014
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2014 IEEE Conference on Computer Communications Workshops, INFOCOM WKSHPS 2014
Y2 - 27 April 2014 through 2 May 2014
ER -