Abstract for presentation (Poster or Podium) with a Paper in the Conference Proceedings
Data Sensing and Analytics
Sultan Al Shafian (he/him/his)
Graduate Research Assistant
Kennesaw State University
Marietta, GA, United States
Sultan Al Shafian (he/him/his)
Graduate Research Assistant
Kennesaw State University
Marietta, GA, United States
Sultan Al Shafian (he/him/his)
Graduate Research Assistant
Kennesaw State University
Marietta, GA, United States
Da Hu
Assistant Professor
Kennesaw State University
Marietta, GA, United States
Da Hu
Assistant Professor
Kennesaw State University
Marietta, GA, United States
Da Hu
Kennesaw State University
Marietta, Georgia, United States
In the fast-paced world of global construction activities, there is an increasing focus on improving safety measures at construction sites. Achieving a comprehensive understanding of work zones is essential to achieving this goal. This involves recognizing potential hazards and understanding the complete dynamics of the construction environment, including equipment placement and worker activities. Such an understanding can greatly enhance worker safety and operational efficiency. The use of state-of-the-art technologies, such as machine learning and deep learning, combined with visual data capture methods like on-site cameras, provides a transformative solution for achieving this. They enable real-time, continuous monitoring and detailed scene analysis, significantly mitigating risks and optimizing workflows. However, the application of these technologies is often limited by the need for vast, varied, and accurately annotated datasets depicting different construction site scenarios.
Our study introduces a pioneering approach that employs synthetic image datasets for training deep learning models tailored specifically for work zone scene understanding. This approach simplifies the data acquisition phase and guarantees a dataset rich in diverse construction situations, encompassing both hazards and typical work zone activities. By using the Unity game engine platform, we synthesized 10,000 images with precise ground truth semantic information. This precision ensured consistency and high-quality annotations, vital during the model training and validation processes, thus reducing the ambiguities frequently associated with real-world datasets. Moreover, Unity's virtual environment allowed us to create diverse scenarios, capturing the breadth and depth of activities and challenges often observed in real-life work zones. Preliminary results are noteworthy, as image segmentation networks such as UNet and DeepLab V3+ trained on our synthetic dataset achieved an average precision of 91% with synthetic images and 75% in real-world applications. This highlights the effectiveness of synthetic datasets in training models to understand complex work zone scenes. By bringing together synthetic imagery and real-world site photos, our study holds the promise of reshaping scene understanding in construction zones, thereby significantly enhancing safety and efficiency standards in the construction sector.