Sedentary office work increases musculoskeletal risk, underscoring the need for non-intrusive, real-time posture monitoring. This study presents a computer vision approach that classifies ergonomic versus non-ergonomic sitting postures using upper body key points extracted by MoveNet thunder. Images from 30 participants were captured from frontal and side views, and labeled according to SNI 9011:2021 criteria. Seventeen key points were detected, with head-to-hip landmarks retained, then normalized and centered. Three classifiers—adaptive boosting (AdaBoost), extreme gradient boosting (XGBoost), and a multi-layer perceptron (MLP)—were trained and evaluated with 10-fold stratified cross-validation. XGBoost achieved the best performance, with accuracy 93.0%±1.9%, precision 94.6%, recall 91.4%, F1-score 92.9%, and area under the receiver operating characteristic curve (ROC-AUC) 0.974±0.010, outperforming MLP and AdaBoost. The method supports privacy-preserving, on-device inference and is suitable for integration into smart office systems to reduce exposure to high-risk postures. Limitations include controlled capture conditions and an upper body focus; future work will expand posture taxonomy and real-world deployment.
Copyrights © 2026