Project tagline:From helmets to boots — AI-powered PPE compliance for safer construction worksites.
Overview (Problem → MVP solution)
Problem:
In construction and industrial environments, workers often forget or misuse PPE (helmets, boots, gloves, vests). Current inspection workflows are manual, slow, and error-prone, increasing risk of accidents and compliance issues.
Solution / MVP:
SafeSight AI Assistant is an image-based PPE compliance system. Users upload images of workers or site conditions, and the system:
Detects PPE items using YOLOv8.
Classifies worn vs. not-worn PPE.
Generates human-readable compliance reasoning using BERT + NER.
Provides a retrieval-augmented chatbot (FAISS + LLM) for PPE-specific questions.
Note: This MVP currently only works with images, not live camera feeds. Accuracy is limited due to small dataset and GPU constraints. Adding more labeled images will make the system more robust and reliable. Future deployment to embedded systems (Raspberry Pi / edge devices) is planned.
Single-sentence MVP description
An image-based PPE compliance assistant that detects PPE items in photos, classifies presence/absence, and provides explainable compliance reasoning and corrective recommendations.
What we built (current features)
YOLOv8 object detection for helmets, boots, gloves, and vests.
Segmentation & classification for partial or incorrectly worn PPE.
BERT-based OSHA-style question classifier for compliance reasoning.
NER module to extract PPE-related entities from reports.
LLM-powered chatbot (RAG) that answers only PPE compliance questions using FAISS retrieval.
Streamlit dashboard to visualize images, detections, explanations, and compliance reports.
Note: Future improvements include adding more PPE classes, expanding chatbot knowledge, and integrating real-time video monitoring.
🗂 Project Structure (Main Skeleton)
This is the main structure; the production codebase contains additional modules/configs beyond this outline.
flowchart LR A[User uploads Image] --> B[Preprocessing - resize & augment] B --> C[YOLOv8 Detection] C --> D{Detected PPE Items} D -->|Helmet| E[Helmet classifier & segmentation] D -->|Boots| F[Boots classifier] D -->|Gloves| G[Gloves classifier] D -->|Vest| H[Vest classifier] E --> I[Compliance rules & thresholds] F --> I G --> I H --> I I --> J[Compliance reasoning - NER & BERT] J --> K[LLM Assistant / Chatbot] K --> L[Dashboard / Exportable Report - CSV/PDF]
Sequence: Image → Decision → Report
Loading
sequenceDiagram participant U as User participant I as Image participant M as Model (YOLO + Classifiers) participant R as Reasoner (BERT + NER) participant L as LLM participant D as Dashboard
flowchart LR A[User Query / PPE Question] --> B[FAISS Vector Search] B --> C{Relevant PPE Context Found?} C -->|Yes| D[Retrieve Context & Metadata] C -->|No| E[Return Cannot Answer Outside PPE Scope] D --> F[LLM Generates Response Using Retrieved Context] F --> G[Provide Human-Friendly Answer / Recommendation] G --> H[Dashboard / Chat Interface] H --> A[User Sees Answer]