Optimizing Bidirectional LSTM for Energy Consumption Prediction Using Chaotic Particle Swarm Optimization and Hyperparameter Tuning

(1) * Candra Juni Cahyo Kusuma Mail (Universitas PGRI Yogyakarta, Indonesia)
(2) Khairunnisa Khairunnisa Mail (National Taiwan University of Science and Technology, Taiwan, Province of China)
*corresponding author

Abstract


This study aims to improve the accuracy of energy consumption prediction using the Bidirectional Long Short-Term Memory (BLSTM) model which is known to be able to handle temporal dependencies in time series data. However, the performance of BLSTM is greatly affected by the hyperparameter configuration, which often requires manual tuning which is inefficient. To address this, this study proposes an optimization framework that combines BLSTM with Chaotic Particle Swarm Optimization (CPSO) to automatically adjust hyperparameters such as the number of hidden units and learning rate. Experiments show that BLSTM optimized with CPSO produces higher prediction accuracy compared to traditional methods such as grid search and random search. By utilizing the chaos map, CPSO improves exploration and exploitation capabilities, accelerates convergence, and finds more optimal solutions. The integration of CPSO and BLSTM shows promising results for improving the performance of time series prediction models, especially in energy consumption forecasting.

   

DOI

https://doi.org/10.33292/ijarlit.v2i2.37
      

Article metrics

10.33292/ijarlit.v2i2.37 Abstract views : 86 | PDF views : 20

   

Cite

   

Full Text

Download

References


S. K. Hora, R. Poongodan, R. P. de Prado, M. Wozniak, and P. B. Divakarachari, “Long Short-Term Memory Network-Based Metaheuristic for Effective Electric Energy Consumption Prediction,” Appl. Sci., vol. 11, no. 23, 2021, doi: 10.3390/APP112311263.

B. J. Van Ruijven, E. De Cian, and I. Sue Wing, “Amplification of future energy demand growth due to climate change,” Nat. Commun., vol. 10, no. 1, p. 2762, 2019.

G. N. Kouziokas, “Long Short-Term Memory (LSTM) Deep Neural Networks in Energy Appliances Prediction,” 2019, vol. 2019, pp. 1–5. doi: 10.1109/PACET48583.2019.8956252.

S. H. Lee, T. Lee, S. Kim, and S. Park, “Energy Consumption Prediction System Based on Deep Learning with Edge Computing,” 2019, pp. 473–477. doi: 10.1109/ELTECH.2019.8839589.

E. Balraj, T. Pugalendhi, M. R. Sureshkumar, and K. Vijayasarathi, “Optimized LSTM Model for Electric Load Forecasting using Deep Learning with Genetic Algorithm,” 2023, pp. 301–304. doi: 10.1109/ICICCS56967.2023.10142580.

A. S. Varghese, S. Sarang, V. Yadav, B. Karotra, and N. Gandhi, “Bidirectional LSTM Joint Model for Intent Classification and Named Entity Recognition in Natural Language Understanding,” Int. J. Hybrid Intell. Syst., vol. 16, no. 1, pp. 13–23, 2020, doi: 10.3233/HIS-190275.

A. Chao, H. Zhijie, Z. Yan, C. Dunliang, and Q. Yinchuan, “Bidirectional LSTM neural network-based statement identification method and apparatus.” 2016.

E. Kiperwasser and Y. Goldberg, “Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations,” arXiv: Computation and Language. 2016. [Online]. Available: https://www.arxiv.org/pdf/1603.04351

J. Cross and L. Huang, “Incremental Parsing with Minimal Features Using Bi-Directional LSTM,” 2016, vol. 2, pp. 32–37. doi: 10.18653/V1/P16-2006.

M. M. Rahman, Y. Watanobe, and K. Nakamura, “A Bidirectional LSTM Language Model for Code Evaluation and Repair,” Symmetry (Basel)., vol. 13, no. 2, 2021, doi: 10.3390/SYM13020247.

A. B. Gumelar, E. M. Yuniarno, D. P. Adi, A. G. Sooai, I. Sugiarto, and M. H. Purnomo, “BiLSTM-CNN Hyperparameter Optimization for Speech Emotion and Stress Recognition,” 2021. doi: 10.1109/IES53407.2021.9594024.

A. Sinha, T. Khandait, and R. Mohanty, “A Gradient-based Bilevel Optimization Approach for Tuning Hyperparameters in Machine Learning.,” arXiv: Learning. 2020. [Online]. Available: https://arxiv.org/pdf/2007.11022.pdf

M. Munem, T. M. Rubaith Bashar, M. H. Roni, M. Shahriar, T. B. Shawkat, and H. Rahaman, “Electric power load forecasting based on multivariate LSTM neural network using bayesian optimization,” 2020 IEEE Electr. Power Energy Conf. EPEC 2020, vol. 3, 2020, doi: 10.1109/EPEC48502.2020.9320123.

J. J. Cherian et al., “Efficient hyperparameter optimization by way of PAC-Bayes bound minimization.,” arXiv: Machine Learning. 2020. [Online]. Available: https://arxiv.org/pdf/2008.06431

K. C. Ukaoha and E. C. Igodan, “Architecture Optimization Model for the Deep Neural Network For Binary Classification Problems,” p. 0, 2020, doi: 10.21608/IJICIS.2020.18509.1008.

Q. Lyu and J. Zhu, “Revisit Long Short-Term Memory: An Optimization Perspective.” 2015. [Online]. Available: https://ml.cs.tsinghua.edu.cn/~jun/pub/lstm-parallel.pdf

F. Ye, “Particle swarm optimization-based automatic parameter selection for deep neural networks and its applications in large-scale and high-dimensional data.,” PLoS One, vol. 12, no. 12, 2017, doi: 10.1371/JOURNAL.PONE.0188746.

H. Xing and X. Pan, “Application of improved particle swarm optimization in system identification,” 2018. doi: 10.1109/CCDC.2018.8407336.

K. E. Parsopoulos, “Particle Swarm Methods,” 2018, pp. 639–685. doi: 10.1007/978-3-319-07124-4_22.

M. Kaminski, “Neural Network Training Using Particle Swarm Optimization - a Case Study,” 2019, pp. 115–120. doi: 10.1109/MMAR.2019.8864679.

G. Shang and Y. Jing, “Research on Chaos Particle Swarm Optimization Algorithm,” Pattern Recognit. Artif. Intell., 2006.

D. Tian and Z. Shi, “MPSO: Modified particle swarm optimization and its applications,” Swarm Evol. Comput., vol. 41, pp. 49–68, 2018, doi: 10.1016/j.swevo.2018.01.011.

H. Zhao, Z. Ru, X. Chang, and S. Li, “Reliability Analysis Using Chaotic Particle Swarm Optimization,” Qual. Reliab. Eng. Int., vol. 31, no. 8, pp. 1537–1552, 2015, doi: 10.1002/QRE.1689.

L. Candanedo, “Appliances Energy Prediction.” UCI Machine Learning Repository, 2017. [Online]. Available: https://doi.org/10.24432/C5VC8G


Refbacks

  • There are currently no refbacks.


View My Stats IJARLIT