Le, Thai-HoaThai-HoaLeChang, Ju-ChiJu-ChiChangHsu, Wei-YiWei-YiHsuLin, Tzu-YangTzu-YangLinChang, Ting-weiTing-weiChangJACOB JE-CHIAN LIN2025-11-202025-11-202025https://www.scopus.com/record/display.uri?eid=2-s2.0-105016625828&origin=resultslisthttps://scholars.lib.ntu.edu.tw/handle/123456789/733869Image-based indoor localization is a promising approach to enhancing facility management efficiency. However, ensuring localization accuracy and improving data accessibility remain key challenges. Therefore, this research aims to automatedly localize images captured during facility inspections by matching the viewpoint of the camera with a corresponding viewpoint in a Building Information Modeling-based (BIM) simulated environment. In this paper, we present a framework that generates photorealistic synthetic images and trains a deep learning model for camera pose estimation. Synthetic datasets are generated in a simulation environment, allowing precise control over scene parameters, camera positions, and lighting conditions. This allows the creation of diverse and realistic training data tailored to specific facility environments. The deep learning model takes RGB images, semantic segmented maps, and corresponding camera poses as inputs to predict six-degree-of-freedom (6DOF) camera poses, including position and orientation. Experimental results demonstrate that the proposed approach can enable indoor image localization with an average translation error of 5.8 meters and a rotation error of 69.05 degrees.trueBIMFacility managementImage localizationPose estimationSynthetic dataLearning-based 6 DOF Camera Pose Estimation Using BIM-generated Virtual Scene for Facility Managementconference paper10.22260/ISARC2025/00072-s2.0-105016625828