Optimized code generation in BIM with retrieval-augmented LLMs

Loading...
Thumbnail Image

URL

Journal Title

Journal ISSN

Volume Title

School of Science | Master's thesis

Department

Major/Subject

Mcode

Language

en

Pages

64

Series

Abstract

The growing interest in Large Language Models (LLMs), such as Generative Pre-Trained Transformers (GPT), has spanned various industries, including software development, construction, and medicine. In the construction sector, LLMs hold potential for enhancing Building Information Modeling (BIM) software, which is widely used to create digital representations of physical structures for planning and design purposes. This study explores how LLMs can improve BIM by enabling users to implement architectural features through text-based prompts. To achieve this, LLMs must query relevant knowledge bases to generate macros (code files) that modify BIM templates. Given that LLMs are trained on generalpurpose data, Retrieval-Augmented Generation (RAG) techniques can be employed to enhance their ability to access domain-specific information. While LLMs show promise, further research is required to determine whether GPT-generated macros can be effectively integrated into BIM workflows to improve productivity. Additionally, there is a shortage of studies assessing the reliability of RAG in generating accurate code for BIM applications. This research aims to address these gaps by exploring various RAG implementation methods within GPT models and benchmarking the performance of the resulting macros. The study begins with a review of the relevant background on LLMs and RAG, followed by the research methodology, analysis of results, and conclusions. Ultimately, this work lays the foundation for leveraging LLMs in BIM software to enhance efficiency and accuracy.

Description

Supervisor

Jung, Alex

Thesis advisor

Haikola, Ville

Other note

Citation