Semester of Graduation

Summer 2025

Degree

Master of Construction Science and Management (MCSM)

Department

The Bert S. Turner Department of Construction Management

Document Type

Thesis

Abstract

Construction sites are dynamic and inherently hazardous environments, where small hand tools—although essential—pose serious safety risks due to their frequent use, portability, and tendency to be misplaced or dropped. This study introduces a novel and lightweight deep learning-based architecture, Lightweight Small Tool Detection (LSTD), specifically designed for fast detection of small tools in unstructured and challenging construction environments. Recognizing that small object detection remains a persistent limitation in existing computer vision models, particularly under poor lighting or cluttered backgrounds, LSTD integrates advanced modules for enhanced feature extraction, fusion, and classification. It achieves notable improvements in accuracy, recall, and computational efficiency compared to existing methods. The core architecture of LSTD is built upon YOLOv5 but is augmented with three significant components: Dynamic Feature Extraction (DFE), Integrated Feature Fusion (IFF), and Accurate Separated Head (ASH). These modules are tailored to improve the detection of small tools by capturing fine-grained details with fewer parameters. The DFE module focuses on preserving critical visual cues despite the limited size of target objects, while the IFF module ensures robust multi-scale feature representation without excessive computation. The ASH module decouples regression and classification to enhance convergence and precision. Together, these innovations lead to a substantial 73% reduction in model parameters and a 28% drop in computational load, while achieving a mean Average Precision (mAP) of 87.3%. To evaluate the robustness and generalizability of the LSTD model, the research utilized a comprehensive dataset of over 34,700 images of 12 commonly used construction tools. The dataset incorporated diverse conditions such as occlusions, varying illumination levels, and realistic construction site backgrounds. Experiments included ablation studies, comparisons with state-of-the-art models, and stress testing under misty and low-light scenarios. The LSTD model consistently outperformed other lightweight detectors like YOLOv6-Small, YOLOv7-Tiny,and YOLOv8-Small, demonstrating high detection accuracy across different environmental challenges. Moreover, the integration of Convolutional Block Attention Module (CBAM) proved critical in enabling the model to effectively distinguish tools from visually complex backgrounds. Ultimately, this research contributes a compact yet powerful detection model that supports real-time applications on edge devices, paving the way for smarter and safer construction practices. By enabling accurate detection of misplaced, dropped, or unauthorized tools, LSTD offers a proactive mechanism to reduce tripping hazards, improve inventory tracking, and support robotic monitoring systems. Its lightweight design makes it particularly suited for deployment in mobile platforms and embedded construction site monitoring systems. Future work may explore edge implementation, integration with autonomous robotics, and real-time adaptive learning in dynamic construction environments. This work marks a significant step forward in aligning modern computer vision technologies with the urgent safety needs of the construction industry.

Date

7-24-2025

Committee Chair

Wang, Chao

Share

COinS