Detailed Record



How good is the advice from ChatGPT for building science? Comparison of four scenarios


Abstract This paper resulted from several questions discussed between its human authors shortly after the public launch of OpenAI’s ChatGPT: Can a language model, trained on an unimaginably vast database, be able to resolve fundamental data inference and data-driven forecasting problems which have been ’typical’ research fare in the building science domain? Is it possible that research problems which ’typically’ require user-intensive tools, such as building performance simulation and problem-specific machine learning models, can today be addressed by ChatGPT in a manner of seconds? If so, what does this mean for the future of building science, let alone the writing of novel research contributions in academia? The entirety of this paper was produced with significant use of ChatGPT. Four arbitrarily-selected case studies were extracted from recent peer-reviewed journals and reputable sources. ChatGPT was tasked with attempting to infer the same results as the publications using only each case study’s input data. Not only were ChatGPT’s results found to be relatively credible, ChatGPT was able to communicate its results instantly and in an academic language. From start to finish, the entirety of this paper, from initial brainstorming to final editing, was completed in no more than 8 human-hours by the study’s (human) authors. The content of this paper is original and has not been published previously.
Authors Adam Rysanek ORCID , Zoltán Nagy ORCID , Clayton Miller ORCID , Aysegul Demir Dilsiz University of WyomingORCID
Journal Info IOP Publishing | Journal of Physics: Conference Series , vol: 2600 , iss: 8 , pages: 082006 - 082006
Publication Date 11/1/2023
ISSN 1742-6588
TypeKeyword Image article
Open Access bronze Bronze Access
DOI https://doi.org/10.1088/1742-6596/2600/8/082006
KeywordsKeyword Image Building Performance (Score: 0.507071) , Short-Term Forecasting (Score: 0.506502)