Domain adaptation methodology using synthetic images for computer vision-based automated construction progress monitoring
Seong-Mo Choi & Hee Sung
Abstract
Purpose This study aims to improve the generalization of computer vision models used for automated construction progress monitoring by addressing the domain shift problem between synthetic and real images. The objective is to enable more reliable deployment of synthetic data-driven models in actual construction environments. Design/methodology/approach A hybrid optimization algorithm is proposed to minimize domain discrepancy, measured by maximum mean discrepancy (MMD), between synthetic and real images. The method uses a YOLOv8-based feature extractor to compute class-specific distributions and applies both gradient-based and Bayesian optimization techniques to tune image transformation parameters. These transformations are selectively applied to synthetic images to improve alignment with real-world feature distributions. Findings Experimental results show that the proposed method effectively reduces MMD values for both concrete and insulation classes, enhancing visual and structural similarity to real images. Moreover, the use of class-wise single-transformation optimization demonstrated superior performance and stability compared to full-parameter optimization. The hybrid data set combining raw and transformed synthetic images yielded the highest detection accuracy across all test scenarios. Originality/value This study introduces a domain adaptation framework based on MMD optimization specifically tailored for the construction industry. Unlike conventional approaches, the proposed method provides a class-specific, quantitatively driven adaptation strategy that enhances model reliability without requiring extensive real-image data sets. The framework offers practical applicability for progress inference and forms a foundational module for future integration with BIM-based monitoring systems.
Evidence weight
Balanced mode · F 0.40 / M 0.15 / V 0.05 / R 0.40
| F · citation impact | 0.50 × 0.4 = 0.20 |
| M · momentum | 0.50 × 0.15 = 0.07 |
| V · venue signal | 0.50 × 0.05 = 0.03 |
| R · text relevance † | 0.50 × 0.4 = 0.20 |
† Text relevance is estimated at 0.50 on the detail page — for your query’s actual relevance score, open this paper from a search result.