KKG PUBLICATIONS
  • Home
  • Journals
    • BUSINESS & ADMINISTRATIVE STUDIES
    • HUMANITIES, ARTS & SOCIAL SCIENCES
    • TECHNOLOGY & ENGINEERING STUDIES
    • APPLIED SCIENCES
    • MEDICAL SCIENCES
  • Publishing Ethics
  • Privacy Policy
  • Crossmark Policy
  • Contact Us
  • Home
  • Journals
    • BUSINESS & ADMINISTRATIVE STUDIES
    • HUMANITIES, ARTS & SOCIAL SCIENCES
    • TECHNOLOGY & ENGINEERING STUDIES
    • APPLIED SCIENCES
    • MEDICAL SCIENCES
  • Publishing Ethics
  • Privacy Policy
  • Crossmark Policy
  • Contact Us
  • https://evolua.ispcaala.com/
  • https://portal-indonesia.id/
  • https://sipena.rsjrw.id/
  • https://www.ijecom.org/
  • https://askimahciwimandiri.co.id/
  • http://lapassumbawa.com/
  • https://ijecom.org
  • https://spartan.mizoram.gov.in/
  • https://ejournal.neurona.web.id/
  • https://e-journal.staibta.ac.id/
  • http://ilim.not.kg/
  • http://journals.ieu.kiev.ua/

Real-Time 3D Human Pose Estimation Using Deep Learning Model for Ergonomics



Volume 8, Issue 1
Jayabhaduri Radhakrishnan, Aadesh Vijayaraghavan, Ajay Karthik R, Ramana Prasath G, Mohamed Arshath S

Published online: 28 December 2022
Article Views: 25

Abstract

In the recent days, most of the people stay in a hunchback position for a long time, due to usage of electronic gadgets like smartphones, personal computers, laptops and tablets, which causes neck and back pain in large numbers, which causes Text neck syndrome, Musculoskeletal disorders, Carpal Tunnel Syndrome and Computer Vision Syndrome. Hence it has become mandatory for people to be mindful of their posture while sitting for long hours. Human pose estimation has gained a lot of attention amongst researchers in a wide range of applications including computer vision, video analytics and motion analysis. To address these risk factors, attempts have been made to develop a 3D Human Pose Estimation (3D-HPE) model for detecting and correcting frontal plane (anterior) sitting postures in computer workstation ergonomics using MediaPipe, and various variants of YOLO, algorithms. The proposed model locates landmarks and analyzes kinematic points from the input video captured through a web camera. From these kinematic points, the 3D-HPE model analyzes whether the human postures are good or bad based on the temporal duration of prolonged time of poor posture. YOLOv8 gives the highest accuracy which is determined by mAP (Mean Average Precision) value found 91.2 in terms of computational time and human pose estimation. Hence, YOLOv8x-pose is the most suited deep learning algorithm for Real-time 3D Human Pose Estimation in Ergonomics. The proposed model notifies the human by sending an alert message to the device for posture correction.

Reference

  1. S. Kwak, B. Han, and J. H. Han, “Scenario-based video event recognition by constraint flow,” in CVPR 2011. IEEE, 2011, pp. 3345–3352.
  2. U. Gaur, Y. Zhu, B. Song, and A. Roy-Chowdhury, “A string of feature graphs model for recognition of complex activities in natural videos,” in 2011 International conference on computer vision. IEEE, 2011, pp. 2595–2602.
  3. S. Park and J. Aggarwal, “Recognition of twoperson interactions using a hierarchical bayesian network,” in First ACM SIGMM international workshop on Video surveillance, 2003, pp. 65–76.
  4. I. N. Junejo, E. Dexter, I. Laptev, and P. Perez, “View-independent action recognition from temporal self-similarities,” IEEE transactions on pattern analysis and machine intelligence, vol. 33, no. 1, pp. 172–185, 2010.
  5. Z. Duric, W. D. Gray, R. Heishman, F. Li, A. Rosenfeld, M. J. Schoelles, C. Schunn, and H. Wechsler, “Integrating perceptual and cognitive modeling for adaptive and intelligent human-computer interaction,” Proceedings of the IEEE, vol. 90, no. 7, pp. 1272–1289, 2002.
  6. J.-D. Huang, “Kinerehab: a kinect-based system for physical rehabilitation: a pilot study for young adults with motor disabilities,” in The proceedings of the 13th international ACM SIGACCESS conference on Computers and accessibility, 2011, pp. 319–320.
  7. A. Thangali, J. P. Nash, S. Sclaroff, and C. Neidle, “Exploiting phonological constraints for handshape inference in asl video,” in CVPR 2011. IEEE, 2011, pp. 521–528.
  8. A. T. Varadaraju, “Exploiting phonological constraints for handshape recognition in sign language video,” Ph.D. dissertation, Boston University, 2013.
  9. H. Cooper and R. Bowden, “Large lexicon detection of sign language,” in Human–Computer Interaction: IEEE International Workshop, HCI 2007 Rio de Janeiro, Brazil, October 20, 2007 Proceedings 4. Springer, 2007, pp. 88–97.
  10. J. Rehg, G. Abowd, A. Rozga, M. Romero, M. Clements, S. Sclaroff, I. Essa, O. Ousley, Y. Li, C. Kim et al., “Decoding children’s social behavior,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2013, pp. 3414–3421.
  11. L. Presti, S. Sclaroff, and A. Rozga, “Joint alignment and modeling of correlated behavior streams,” in Proceedings of the IEEE International Conference on Computer Vision Workshops, 2013, pp. 730–737.
  12. H. Moon, R. Sharma, and N. Jung, “Method and system for measuring shopper response to products based on behavior and facial expression,” Jul. 10 2012, uS Patent 8,219,438.
  13. M. Andriluka, S. Roth, and B. Schiele, “Pictorial structures revisited: People detection and articulated pose estimation,” in 2009 IEEE conference on computer vision and pattern recognition. IEEE, 2009, pp. 1014–1021.
  14. Y. Yang and D. Ramanan, “Articulated pose estimation with flexible mixtures-of-parts,” in CVPR 2011. IEEE, 2011, pp. 1385–1392.
  15. D. Ramanan, D. A. Forsyth, and A. Zisserman, “Strike a pose: Tracking people by finding stylized poses,” in 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), vol. 1. IEEE, 2005, pp. 271–278.
  16. L. Bourdev and J. Malik, “Poselets: Body part detectors trained using 3d human pose annotations,” in 2009 IEEE 12th international conference on computer vision. IEEE, 2009, pp. 1365–1372.
  17. D. Tran and D. Forsyth, “Improved human parsing with a full relational model,” in Computer Vision– ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, September 5-11, 2010, Proceedings, Part IV 11. Springer, 2010, pp. 227–240.
  18. C.-H. Chen and D. Ramanan, “3d human pose estimation= 2d pose estimation+ matching,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 7035–7043.
  19. X. Zhang, P. Zheng, T. Peng, Q. He, C. K. Lee, and R. Tang, “Promoting employee health in smart office: A survey,” Advanced Engineering Informatics, vol. 51, p. 101518, 2022.
  20. X. Zhang, J. Fan, T. Peng, P. Zheng, C. K. Lee, and R. Tang, “A privacy-preserving and unobtrusive sitting posture recognition system via pressure array sensor and infrared array sensor for office workers,” Advanced Engineering Informatics, vol. 53, p. 101690, 2022.
  21. X. Zhang, J. Fan, T. Peng, P. Zheng, X. Zhang, and R. Tang, “Multimodal data-based deep learning model for sitting posture recognition toward office workers health promotion,” Sensors and Actuators A: Physical, vol. 350, p. 114150, 2023.
  22. W. Kim, J. Sung, D. Saakes, C. Huang, and S. Xiong, “Ergonomic postural assessment using a new open-source human pose estimation technology (openpose),” International Journal of Industrial Ergonomics, vol. 84, p. 103164, 2021.
  23. L. Barks, S. L. Luther, L. M. Brown, B. Schulz, M. E. Bowen, and G. Powell-Cope, “Development and initial validation of the seated posture scale,” JRRD: Journal of Rehabilitation Research and Development, vol. 52, no. 2, p. 201, 2015.
  24. A. R. Pathak, M. Pandey, and S. Rautaray, “Application of deep learning for object detection,” Procedia computer science, vol. 132, pp. 1706–1717, 2018.
  25. S. S. A. Zaidi, M. S. Ansari, A. Aslam, N. Kanwal, M. Asghar, and B. Lee, “A survey of modern deep learning based object detection models,” Digital Signal Processing, vol. 126, p. 103514, 2022.
  26. H. Fu, J. Gao, and H. Liu, “Human pose estimation and action recognition for fitness movements,” Computers & Graphics, vol. 116, pp. 418–426, 2023.
  27. N. Sarafianos, B. Boteanu, B. Ionescu, and I. A. Kakadiaris, “3d human pose estimation: A review of the literature and analysis of covariates,” Computer Vision and Image Understanding, vol. 152, pp. 1–20, 2016.
  28. J. Wang, S. Tan, X. Zhen, S. Xu, F. Zheng, Z. He, and L. Shao, “Deep 3d human pose estimation: A review,” Computer Vision and Image Understanding, vol. 210, p. 103225, 2021.
  29. L. L. Presti and M. La Cascia, “3d skeleton-based human action classification: A survey,” Pattern Recognition, vol. 53, pp. 130–147, 2016.
  30. Z. Liu, J. Zhu, J. Bu, and C. Chen, “A survey of human pose estimation: the body parts parsing based methods,” Journal of Visual Communication and Image Representation, vol. 32, pp. 10–19, 2015.
  31. B. Liu, Y. Li, S. Zhang, and X. Ye, “Healthy human sitting posture estimation in rgb-d scenes using object context,” Multimedia Tools and Applications, vol. 76, pp. 10 721–10 739, 2017.
  32. D. Maji, S. Nagori, M. Mathew, and D. Poddar, “Yolo-pose: Enhancing yolo for multi person pose estimation using object keypoint similarity loss,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 2637–2646.

To Cite this article

J. Radhakrishnan , A. Vijayaraghavan, A. Karthik, R. Prasath and M. Arshath “Real-Time 3D Human Pose Estimation Using Deep Learning Model for Ergonomics” International Journal of Technology and Engineering Studies, vol. 8, no. 1, pp. 34–40, 2022.



© 2020. KKG Publications
Calle Alarcon 66, Sant Adrian De Besos 08930, Barcelona Spain | 00 34 610 911 348
About Us | Contact Us | Feedback

Search