Safe reinforcement learning versus classical controllers for voltage regulation and power quality in the IEEE 33-bus distribution system
1 viewDOI:
https://doi.org/10.54939/1859-1043.j.mst.110.2026.12-21Keywords:
Safe reinforcement learning; Voltage regulation; Power quality; Distribution networks; Controller performance.Abstract
Modern distribution networks face increasing challenges due to the integration of inverter-based resources and stochastic load variations. This study aims to evaluate whether advanced controllers, including a reinforcement learning based safe controller, can provide superior voltage regulation, stability, and power quality compared with conventional proportional–integral, droop, and predictive optimal strategies. The IEEE 33-bus feeder was simulated using realistic 24-hour demand data and sensitivity matrices derived from feeder impedances. Four controllers were implemented under the linearized DistFlow formulation with explicit constraints on voltage deviation and reactive power limits. The analysis focused on both control quality indices, such as settling time, overshoot, steady-state error, and integral squared error, and power system indices, including voltage deviation, regulation, total harmonic distortion, power factor, and network losses. Results show that the safe reinforcement learning controller achieved the fastest settling time of 16.8 hours, the lowest overshoot of 3.2 percent, and the smallest steady-state error of 0.0065 per unit, while also reducing power losses to 0.145 megawatts and maintaining voltage stability above 0.95 per unit. By contrast, the proportional–integral controller exhibited overshoot of 12 percent, harmonic distortion of 8.18 percent, and voltage drops below 0.90 per unit. These findings indicate that embedding safety guarantees into reinforcement learning yields controllers that not only surpass classical strategies but also meet utility-relevant criteria, providing a promising pathway for future deployment in smart distribution networks.
References
[1]. A. Sajadi, R. W. Kenyon, and B.-M. Hodge. “Synchronization in electric power networks with inherent heterogeneity up to 100% inverter-based renewable generation”. Nature Communications, vol. 13, no. 1, p. 2490, (2022).
[2]. S. Wagle, A. K. Gupta, and K. O. Nygård. “Co-simulation-based optimal reactive power control in smart distribution network”. Electrical Engineering, vol. 106, pp. 2441–2464, (2024).
[3]. W. Cui, Y. Li, and B. Zhang. “Decentralized Safe Reinforcement Learning for Voltage Control”. IEEE Transactions on Smart Grid, vol. 13, no. 6, pp. 4957–4968, (2022).
[4]. R. R. Hossain, Y. Zhou, and G. Hug. “Efficient learning of power grid voltage control strategies via model-based deep reinforcement learning”. Machine Learning, vol. 113, pp. 2675–2700, (2024).
[5]. C. Hu, Y. Sun, and Q. Li. “A soft actor-critic deep reinforcement learning method for multi-timescale coordinated operation of microgrids”. Protection and Control of Modern Power Systems, vol. 7, p. 29, (2022).
[6]. A. El-Fergany. “Reviews, Challenges, and Insights on Computational Methods for Network Reconfigurations in Smart Electricity Distribution Networks”. Archives of Computational Methods in Engineering, vol. 31, pp. 1233–1253, (2024).
[7]. V. Waghmare, V. P. Singh, and T. Varshney. “A systematic review of reinforcement learning-based control for microgrids: trends, challenges, and emerging algorithms”. Discover Applied Sciences, vol. 7, p. 939, (2025).
[8]. W. Cui, Y. Li, and B. Zhang. “Safe Reinforcement Learning for Decentralized Voltage Control with Stability Guarantees”. IEEE Transactions on Power Systems, vol. 38, no. 2, pp. 1456–1468, (2023).
[9]. H. Liu and W. Wu. “Online Multi-Agent Reinforcement Learning for Decentralized Inverter-Based Volt-VAR Control”. IEEE Transactions on Smart Grid, vol. 12, no. 4, pp. 2980–2990, (2021).
[10]. P. Yu, M. Anghel, and J. W. Kimball. “Safe Reinforcement Learning for Power System Control: A Review”. IEEE Transactions on Power Systems, vol. 39, no. 1, pp. 2–18, (2024).
[11]. P. Li, F. Xu, and H. Sun. “Optimal real-time Voltage/Var control for distribution network: Droop-control based multi-agent deep reinforcement learning”. International Journal of Electrical Power & Energy Systems, vol. 153, p. 109370, (2023).
[12]. X. Ding, L. Jiang, and Y. Zhang. “Multi-time-scale voltage control of the distribution network”. Frontiers in Energy Research, vol. 12, p. 1445623, (2024).
[13]. S. Bhowmik, Z. Wang, J. M. Guerrero, K. Xie, and J. C. Vázquez. “Hybrid-compatible grid-forming inverters based on topological migration-optimized virtual synchronous control”. Scientific Reports, vol. 15, p. 26194, (2025).
[14]. X. Zhang, L. Yang, and M. Chen. “Impedance-shaping-based stabilization control method for grid-forming converters”. Scientific Reports, vol. 15, p. 14523, (2025).
[15]. M. Sravani, P. D. P. Reddy, and M. S. Bhaskar. “Deep reinforcement learning-based controller for DC-link voltage stabilization in grid-connected PV systems”. Scientific Reports, vol. 15, p. 8729, (2025).
