
You have already added 0 works in your ORCID record related to the merged Research product.
You have already added 0 works in your ORCID record related to the merged Research product.
<script type="text/javascript">
<!--
document.write('<div id="oa_widget"></div>');
document.write('<script type="text/javascript" src="https://beta.openaire.eu/index.php?option=com_openaire&view=widget&format=raw&projectId=undefined&type=result"></script>');
-->
</script>
Deep Reinforcement Learning for Economic Dispatch of Virtual Power Plant in Internet of Energy

Deep Reinforcement Learning for Economic Dispatch of Virtual Power Plant in Internet of Energy
With the high penetration of large-scale distributed renewable energy generation, the power system is facing enormous challenges in terms of the inherent uncertainty of power generation of renewable energy resources. In this regard, virtual power plants (VPPs) can play a crucial role in integrating a large number of distributed generation units (DGs) more effectively to improve the stability of the power systems. Due to the uncertainty and nonlinear characteristics of DGs, reliable economic dispatch in VPPs requires timely and reliable communication between DGs, and between the generation side and the load side. The online economic dispatch optimizes the cost of VPPs. In this article, we propose a deep reinforcement learning (DRL) algorithm for the optimal online economic dispatch strategy in VPPs. By utilizing DRL, our proposed algorithm reduced the computational complexity while also incorporating large and continuous state space due to the stochastic characteristics of distributed power generation. We further design an edge computing framework to handle the stochastic and large-state space characteristics of VPPs. The DRL-based real-time economic dispatch algorithm is executed online. We utilize real meteorological and load data to analyze and validate the performance of our proposed algorithm. The experimental results show that our proposed DRL-based algorithm can successfully learn the characteristics of DGs and industrial user demands. It can learn to choose actions to minimize the cost of VPPs. Compared with the deterministic policy gradient algorithm and DDPG, our proposed method has lower time complexity.
- Heilongjiang University China (People's Republic of)
- University of Oslo Norway
- Heilongjiang University China (People's Republic of)
- SIMULA METROPOLITAN CENTER FOR DIGITAL ENGINEERING AS Norway
- SIMULA METROPOLITAN CENTER FOR DIGITAL ENGINEERING AS Norway
4 Research products, page 1 of 1
- 2011IsAmongTopNSimilarDocuments
- 2017IsAmongTopNSimilarDocuments
- 2019IsAmongTopNSimilarDocuments
citations This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).99 popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.Top 1% influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).Top 10% impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.Top 1%
