Smart grid stability prediction using Adaptive Aquila Optimizer and ensemble stacked BiLSTM
Background: Smart grids, characterized by their ability to integrate renewable energy sources and manage the dynamic balance between supply and demand, require sophisticated prediction models to maintain stability. Traditional machine learning (ML) models often fall short in predicting the highly va...
Saved in:
| Main Authors: | , , , , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Elsevier
2024-12-01
|
| Series: | Results in Engineering |
| Subjects: | |
| Online Access: | http://www.sciencedirect.com/science/article/pii/S2590123024015159 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | Background: Smart grids, characterized by their ability to integrate renewable energy sources and manage the dynamic balance between supply and demand, require sophisticated prediction models to maintain stability. Traditional machine learning (ML) models often fall short in predicting the highly variable nature of smart grid operations. Methods: This study introduces an ensemble stacked bidirectional Long Short-Term Memory model enhanced by a proposed Adaptive Aquila Optimizer (AAO). The AAO uses a Sigmoid Factor to balance exploration and exploitation, adapting the transition from broad searches to focused ones based on iteration progress. It is utilized for feature selection by identifying and excluding irrelevant and redundant features and methodically evaluates seven key hyperparameters to fine-tune the model's performance. Additionally, a weighted voting mechanism is employed to aggregate predictions in the ensemble model. Results: Multiple rounds of empirical experiments using different sets of optimizers and configurations, supported by the visualization capabilities of TensorBoard, demonstrate significant improvements in the performance of the AAO-BiLSTM model. The results show profound potential with accuracy, precision, recall, and F1-score rates of 99.55%, surpassing both traditional ML algorithms and state-of-the-art approaches. |
|---|---|
| ISSN: | 2590-1230 |