Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.

Surrogate optimisation holds a big promise for building energy optimisation studies due to its goal to replace the use of lengthy building energy simulations within an optimisation step with expendable local surrogate models that can quickly predict simulation results. To be useful for such purpose,...

Full description

Saved in:
Bibliographic Details
Main Authors: Sanja Stevanović, Husain Dashti, Marko Milošević, Salem Al-Yakoob, Dragan Stevanović
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2024-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0312573
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850033705613000704
author Sanja Stevanović
Husain Dashti
Marko Milošević
Salem Al-Yakoob
Dragan Stevanović
author_facet Sanja Stevanović
Husain Dashti
Marko Milošević
Salem Al-Yakoob
Dragan Stevanović
author_sort Sanja Stevanović
collection DOAJ
description Surrogate optimisation holds a big promise for building energy optimisation studies due to its goal to replace the use of lengthy building energy simulations within an optimisation step with expendable local surrogate models that can quickly predict simulation results. To be useful for such purpose, it should be possible to quickly train precise surrogate models from a small number of simulation results (10-100) obtained from appropriately sampled points in the desired part of the design space. Two sampling methods and two machine learning models are compared here. Latin hypercube sampling (LHS), widely accepted in building energy community, is compared to an exploratory Monte Carlo-based sequential design method mc-intersite-proj-th (MIPT). Artificial neural networks (ANN), also widely accepted in building energy community, are compared to gradient-boosted tree ensembles (XGBoost), model of choice in many machine learning competitions. In order to get a better understanding of the behaviour of these two sampling methods and two machine learning models, we compare their predictions against a large set of generated synthetic data. For this purpose, a simple case study of an office cell model with a single window and a fixed overhang, whose main input parameters are overhang depth and height, while climate type, presence of obstacles, orientation and heating and cooling set points are additional input parameters, was extensively simulated with EnergyPlus, to form a large underlying dataset of 729,000 simulation results. Expendable local surrogate models for predicting simulated heating, cooling and lighting loads and equivalent primary energy needs of the office cell were trained using both LHS and MIPT and both ANN and XGBoost for several main hyperparameter choices. Results show that XGBoost models are more precise than ANN models, and that for both machine learning models, the use of MIPT sampling leads to more precise surrogates than LHS.
format Article
id doaj-art-caa059e3ac5e406486d830819ed7903f
institution DOAJ
issn 1932-6203
language English
publishDate 2024-01-01
publisher Public Library of Science (PLoS)
record_format Article
series PLoS ONE
spelling doaj-art-caa059e3ac5e406486d830819ed7903f2025-08-20T02:58:07ZengPublic Library of Science (PLoS)PLoS ONE1932-62032024-01-011910e031257310.1371/journal.pone.0312573Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.Sanja StevanovićHusain DashtiMarko MiloševićSalem Al-YakoobDragan StevanovićSurrogate optimisation holds a big promise for building energy optimisation studies due to its goal to replace the use of lengthy building energy simulations within an optimisation step with expendable local surrogate models that can quickly predict simulation results. To be useful for such purpose, it should be possible to quickly train precise surrogate models from a small number of simulation results (10-100) obtained from appropriately sampled points in the desired part of the design space. Two sampling methods and two machine learning models are compared here. Latin hypercube sampling (LHS), widely accepted in building energy community, is compared to an exploratory Monte Carlo-based sequential design method mc-intersite-proj-th (MIPT). Artificial neural networks (ANN), also widely accepted in building energy community, are compared to gradient-boosted tree ensembles (XGBoost), model of choice in many machine learning competitions. In order to get a better understanding of the behaviour of these two sampling methods and two machine learning models, we compare their predictions against a large set of generated synthetic data. For this purpose, a simple case study of an office cell model with a single window and a fixed overhang, whose main input parameters are overhang depth and height, while climate type, presence of obstacles, orientation and heating and cooling set points are additional input parameters, was extensively simulated with EnergyPlus, to form a large underlying dataset of 729,000 simulation results. Expendable local surrogate models for predicting simulated heating, cooling and lighting loads and equivalent primary energy needs of the office cell were trained using both LHS and MIPT and both ANN and XGBoost for several main hyperparameter choices. Results show that XGBoost models are more precise than ANN models, and that for both machine learning models, the use of MIPT sampling leads to more precise surrogates than LHS.https://doi.org/10.1371/journal.pone.0312573
spellingShingle Sanja Stevanović
Husain Dashti
Marko Milošević
Salem Al-Yakoob
Dragan Stevanović
Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.
PLoS ONE
title Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.
title_full Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.
title_fullStr Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.
title_full_unstemmed Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.
title_short Comparison of ANN and XGBoost surrogate models trained on small numbers of building energy simulations.
title_sort comparison of ann and xgboost surrogate models trained on small numbers of building energy simulations
url https://doi.org/10.1371/journal.pone.0312573
work_keys_str_mv AT sanjastevanovic comparisonofannandxgboostsurrogatemodelstrainedonsmallnumbersofbuildingenergysimulations
AT husaindashti comparisonofannandxgboostsurrogatemodelstrainedonsmallnumbersofbuildingenergysimulations
AT markomilosevic comparisonofannandxgboostsurrogatemodelstrainedonsmallnumbersofbuildingenergysimulations
AT salemalyakoob comparisonofannandxgboostsurrogatemodelstrainedonsmallnumbersofbuildingenergysimulations
AT draganstevanovic comparisonofannandxgboostsurrogatemodelstrainedonsmallnumbersofbuildingenergysimulations