Technical Papers
Oct 11, 2022

3D Excavator Pose Estimation Using Projection-Based Pose Optimization for Contact-Driven Hazard Monitoring

Publication: Journal of Computing in Civil Engineering
Volume 37, Issue 1

Abstract

Contact-driven accidents involving actuated excavators have led to a significant number of fatalities in the construction industry. The revolving mechanical arm of excavators poses a major risk of contact-driven accidents for workers in its proximity due to its articulated pose. Detecting the 3D pose of excavator arms is thus essential to prevent contact-driven accidents near excavators. Previous works have attempted to estimate 3D excavator poses using sensor-based or computer vision-based methods. However, existing methods require extensive preparation work, such as attaching physical sensors, calibrating stereo cameras, or collecting 3D training data. As a result, existing methods cannot be easily integrated into the current construction workflow and are seldom applied in real-world situations. The authors propose a projection-based 3D pose optimization method that utilizes excavator kinematic constraints to infer 3D excavator poses from monocular image sequences with no dependency on 3D training data. The proposed method first extracts the 2D excavator pose from images using a keypoint region-based convolution neural network. Then, the 2D pose is reconstructed into 3D by enforcing the rigid excavator kinematic constraints (e.g., arm length) and minimizing the 2D reprojection error of the excavator pose. Tests using a 1:14 miniature excavator model showed a 3D position error of 7.3 cm (or 1.03 m when scaled up to real-world dimensions) for keypoints on the excavator pose, demonstrating the capabilities of the proposed method in estimating 3D excavator poses from monocular images. The proximity measuring capacity of the estimated 3D pose was also evaluated, achieving a mean absolute distance error of 4.7 cm (or 0.66 m scaled). The proposed method offers a 3D excavator pose estimation method using only a monocular camera and without relying on 3D training data. The estimated 3D excavator pose enables safety managers to monitor potential contact-driven accidents near excavators and alert workers of unsafe situations and promotes safer working environments for construction workers near excavators.

Practical Applications

The authors present a monocular vision-based method for 3D excavator pose estimation, which can serve as the groundwork for monitoring contact-driven accidents near excavators. The method allows safety managers to monitor the 3D pose of an excavator using one single camera (e.g., smartphone, site surveillance camera, or action camera). Unlike previous methods, the proposed method does not require construction professionals to conduct challenging preparation work such as setting up multiple stereo cameras or collecting custom 3D excavator training datasets. For instance, for an excavator 21 m away, the proposed method can estimate the 3D excavator’s keypoint positions with an expected 3D position error of 1.03 m. Paired with other onsite information (e.g., utility line or worker location) collected from existing drawings or using other vision-based methods, the proposed method can measure the proximity between the excavator arm and surrounding objects with an expected error of 0.66 m, enabling proactive safety interventions (e.g., proximity alerts). Such safety interventions can make it safer for workers to conduct construction work near excavators. Though originally devised for excavators, the proposed method can be further adapted to monitor the 3D pose of other construction equipment such as backhoes or crawler cranes, enabling a wider range of safety applications for monitoring contact-driven accidents.

Get full access to this article

View all available purchase options and get full access to this article.

Data Availability Statement

All data, models, or codes that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

This research was supported financially by a National Science Foundation Award (No. IIS-1734266, ‘Scene Understanding and Predictive Monitoring for Safe Human-Robot Collaboration in Unstructured and Dynamic Construction Environments’).

References

Andriluka, M., L. Pishchulin, P. Gehler, and B. Schiele. 2014. “2D human pose estimation: New benchmark and state of the art analysis.” In Proc., IEEE Conf. on Computer Vision and Pattern Recognition (CVPR). New York: IEEE.
Apple. 2022. “iPhone 13 pro and 13 pro max - technical specifications.” Accessed May 18, 2022. https://www.apple.com/iphone-13-pro/specs/.
Artacho, B., and A. Savakis. 2020. “UniPose: Unified human pose estimation in single images and videos.” In Proc., IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), 7033–7042. New York: IEEE.
Assadzadeh, A., M. Arashpour, I. Brilakis, T. Ngo, and E. Konstantinou. 2022. “Vision-based excavator pose estimation using synthetically generated datasets with domain randomization.” Autom. Constr. 134 (Feb): 104089. https://doi.org/10.1016/j.autcon.2021.104089.
Bang, S., Y. Hong, and H. Kim. 2021. “Proactive proximity monitoring with instance segmentation and unmanned aerial vehicle-acquired video-frame prediction.” Comput.-Aided Civ. Infrastruct. Eng. 36 (6): 800–816. https://doi.org/10.1111/mice.12672.
BLS (Bureau of Labor Statistics). 2021. “Census of fatal occupational injuries - archived data.” Accessed February 3, 2022. https://www.bls.gov/iif/oshcfoiarchive.htm#RATES.
Bobadilla, L., A. Mostafavi, T. Carmenate, and S. Bista. 2014. “Predictive assessment and proactive monitoring of struck-by safety hazards in construction sites: An information space approach.” In Computing in civil and building engineering (2014), 989–996. Reston, VA: ASCE.
Bradski, G. 2000. “The OpenCV library.” Dr. Dobb’s J. 25 (11): 120–123.
Brown, S., W. Harris, R. D. Brooks, and X. S. Dong. 2021. Fatal and nonfatal struck-by injuries in the construction industry, 2011-2019. Silver Spring, MD: The Center for Construction Research and Training.
Built Robotics. 2022. “How construction automation works.” Accessed January 8, 2022. https://www.builtrobotics.com/how-it-works.
Caterpillar. 2016. “Cat MineStar detect.” Accessed February 23, 2022. https://www.cat.com/en_US/by-industry/mining/surface-mining/surface-technology/detect1.html.
Chen, Y., Z. Wang, Y. Peng, Z. Zhang, G. Yu, and J. Sun. 2018. “Cascaded pyramid network for multi-person pose estimation.” In Proc., IEEE Conf. on Computer Vision and Pattern Recognition, 7103–7112. New York: IEEE.
Cho, Y. K., X. Yang, and J. Park. 2017. Improving work zone safety utilizing a new mobile proximity sensing technology. Silver Spring, MD: The Center for Construction Research and Training.
Cooau. 2021. “Cool sries SPC06—COOAU.” Accessed January 26, 2022. https://www.cooau.com/products/cool-series-spc06.
Deng, J., W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. 2009. “Imagenet: A large-scale hierarchical image database.” In Proc., IEEE Conf. on Computer Vision and Pattern Recognition, 248–255. New York: IEEE.
He, K., G. Gkioxari, P. Dollár, and R. Girshick. 2020. “Mask R-CNN.” IEEE Trans. Pattern Anal. Mach. Intell. 42 (2): 386–397. https://doi.org/10.1109/TPAMI.2018.2844175.
He, K., X. Zhang, S. Ren, and J. Sun. 2016. “Deep residual learning for image recognition.” In Proc., IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 770–778. New York: IEEE.
IHSA (Infrastructure Health and Safety Association). 2015. Struck-by incidents and heavy equipment. Etobicoke, ON, Canada: IHSA.
Jahne, B. 2004. Practical handbook on image processing for scientific and technical applications. Boca Raton, FL: CRC Press.
Kim, D., M. Liu, S. Lee, and V. R. Kamat. 2019. “Remote proximity monitoring between mobile construction resources using camera-mounted UAVs.” Autom. Constr. 99 (Mar): 168–182. https://doi.org/10.1016/j.autcon.2018.12.014.
Kim, H., K. Kim, and H. Kim. 2016. “Vision-based object-centric safety assessment using fuzzy inference: Monitoring struck-by accidents with moving objects.” J. Comput. Civ. Eng. 30 (4): 04015075. https://doi.org/10.1061/(ASCE)CP.1943-5487.0000562.
Konstantinou, E., J. Lasenby, and I. Brilakis. 2019. “Adaptive computer vision-based 2D tracking of workers in complex environments.” Autom. Constr. 103 (Jul): 168–184. https://doi.org/10.1016/j.autcon.2019.01.018.
KUKA Robotics Corporation (Keller und Knappich Augsburg). 2018. “KR QUANTEC pro.” Accessed March 4, 2022. https://www.kuka.com/en-us/products/robotics-systems/industrial-robots/kr-quantec.
Liang, C.-J., K. M. Lundeen, W. McGee, C. C. Menassa, S. Lee, and V. R. Kamat. 2019. “A vision-based marker-less pose estimation system for articulated construction robots.” Autom. Constr. 104 (Aug): 80–94. https://doi.org/10.1016/j.autcon.2019.04.004.
Lin, T.-Y., P. Dollar, R. Girshick, K. He, B. Hariharan, and S. Belongie. 2017. “Feature pyramid networks for object detection.” In Proc., IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 936–944. New York: IEEE.
Lin, T.-Y., M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick. 2014. “Microsoft COCO: Common objects in context.” In Computer vision—ECCV 2014, lecture notes in computer science, edited by D. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, 740–755. Cham, Switzerland: Springer.
Luo, H., M. Wang, P. K.-Y. Wong, and J. C. P. Cheng. 2020. “Full body pose estimation of construction equipment using computer vision and deep learning techniques.” Autom. Constr. 110 (Feb): 103016. https://doi.org/10.1016/j.autcon.2019.103016.
Mahmood, B., S. Han, and J. Seo. 2022. “Implementation experiments on convolutional neural network training using synthetic images for 3D pose estimation of an excavator on real images.” Autom. Constr. 133 (Jan): 103996. https://doi.org/10.1016/j.autcon.2021.103996.
Martinez, J., R. Hossain, J. Romero, and J. J. Little. 2017. “A simple yet effective baseline for 3D human pose estimation.” In Proc., IEEE Int. Conf. on Computer Vision, 2659–2668. New York: IEEE. https://doi.org/10.1109/ICCV.2017.288.
Newell, A., K. Yang, and J. Deng. 2016. “Stacked hourglass networks for human pose estimation.” In Proc., European Conf. on Computer Vision, 483–499. New York: Springer.
NIOSH (National Institute for Occupational Safety and Health). 2003. Preventing injuries when working with hydraulic excavators and backhoe loaders. Washington, DC: NIOSH.
OSHA (Occupational Safety and Health Administration). 1994. Safety procedures associated with the use of wheel or track propelled excavators. Washington, DC: OSHA.
Rebbitt, D. 2016. The spotting dilemma. Toronto: Canadian Occupational Safety.
Ren, S., K. He, R. Girshick, and J. Sun. 2015. “Faster R-CNN: Towards real-time object detection with region proposal networks.” In Advances in neural information processing systems. Red Hook, NY: Curran Associates.
Richardson, A., J. Strom, and E. Olson. 2013. “AprilCal: Assisted and repeatable camera calibration.” In Proc., IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, 1814–1821. New York: IEEE.
Schops, T., V. Larsson, M. Pollefeys, and T. Sattler. 2020. “Why having 10,000 parameters in your camera model is better than twelve.” In Proc., IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), 2532–2541. New York: IEEE.
ShockWatch. 2018. “ShockWatch intelligent proximity alert system (IPAS) by KIGIS.” Accessed February 23, 2022. https://www.shockwatch.com.au/product/intelligent-proximity-alert-system-ipas/.
Soltani, M. M., Z. Zhu, and A. Hammad. 2018. “Framework for location data fusion and pose estimation of excavators using stereo vision.” J. Comput. Civ. Eng. 32 (6): 04018045. https://doi.org/10.1061/(ASCE)CP.1943-5487.0000783.
Son, H., H. Seong, H. Choi, and C. Kim. 2019. “Real-time vision-based warning system for prevention of collisions between workers and heavy equipment.” J. Comput. Civ. Eng. 33 (5): 04019029. https://doi.org/10.1061/(ASCE)CP.1943-5487.0000845.
Strecha, C., W. von Hansen, L. Van Gool, P. Fua, and U. Thoennessen. 2008. “On benchmarking camera calibration and multi-view stereo for high resolution imagery.” In Proc., IEEE Conf. on Computer Vision and Pattern Recognition, 1–8. New York: IEEE.
SynTech. 2020. “Hit-not proximity detection.” Accessed February 23, 2022. https://syntechsales.com/wp-content/uploads/2020/10/hit-not-brochure.pdf.
Tang, J., H. Luo, P. K. Y. Wong, and J. C. P. Cheng. 2020. “Study of IMU installation position for posture estimation of excavators.” In Proc., Int. Conf. on Computing in Civil and Building Engineering. New York: Springer.
Teizer, J., B. S. Allread, and U. Mantripragada. 2010. “Automating the blind spot measurement of construction equipment.” Autom. Constr. 19 (4): 491–501. https://doi.org/10.1016/j.autcon.2009.12.012.
Vahdatikhaki, F., A. Hammad, and H. Siddiqui. 2015. “Optimization-based excavator pose estimation using real-time location systems.” Autom. Constr. 56 (Aug): 76–92. https://doi.org/10.1016/j.autcon.2015.03.006.
Wang, J., et al. 2021a. “Deep high-resolution representation learning for visual recognition.” IEEE Trans. Pattern Anal. Mach. Intell. 43 (10): 3349–3364. https://doi.org/10.1109/TPAMI.2020.2983686.
Wang, J., S. Tan, X. Zhen, S. Xu, F. Zheng, Z. He, and L. Shao. 2021b. “Deep 3D human pose estimation: A review.” Comput. Vision Image Understanding 210 (Sep): 103225. https://doi.org/10.1016/j.cviu.2021.103225.
Wang, X., R. Katz, S. Schneider, B. Memarian, and X. S. Dong. 2017. Struck-by injuries and prevention in the construction industry. Silver Spring, MD: The Center for Construction Research and Training.
WCCTV (Wireless Closed-Circuit Television). 2022. “Construction site security surveillance cameras—WCCTV.” Accessed May 18, 2022. https://www.wcctv.com/construction-job-site-security/.
Wu, Y., A. Kirillov, F. Massa, W.-Y. Lo, and R. Girshick. 2019. “Detectron2.” Accessed November 16, 2021. https://github.com/facebookresearch/detectron2.
Xiang, Y., D. Y. Sun, W. Fan, and X. G. Gong. 1997. “Generalized simulated annealing algorithm and its application to the Thomson model.” Phys. Lett. A 233 (3): 216–220. https://doi.org/10.1016/S0375-9601(97)00474-X.
Xiong, Y., and L. Matthies. 1997. “Error analysis of a real-time stereo system.” In Proc., IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, 1087–1093. New York: IEEE.
Yuan, C., S. Li, and H. Cai. 2017. “Vision-based excavator detection and tracking using hybrid kinematic shapes and key nodes.” J. Comput. Civ. Eng. 31 (1): 04016038. https://doi.org/10.1061/(ASCE)CP.1943-5487.0000602.
Zhou, X., Q. Huang, X. Sun, X. Xue, and Y. Wei. 2017. “Towards 3D human pose estimation in the wild: A weakly-supervised approach.” In Proc., IEEE Int. Conf. on Computer Vision (ICCV), 398–407. New York: IEEE.
Zhu, C., R. H. Byrd, P. Lu, and J. Nocedal. 1997. “Algorithm 778: L-BFGS-B: Fortran subroutines for large-scale bound-constrained optimization.” ACM Trans. Math. Software 23 (4): 550–560. https://doi.org/10.1145/279232.279236.
Zhuang, F., Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong, and Q. He. 2021. “A comprehensive survey on transfer learning.” Proc. IEEE 109 (1): 43–76. https://doi.org/10.1109/JPROC.2020.3004555.

Information & Authors

Information

Published In

Go to Journal of Computing in Civil Engineering
Journal of Computing in Civil Engineering
Volume 37Issue 1January 2023

History

Received: Mar 15, 2022
Accepted: Aug 2, 2022
Published online: Oct 11, 2022
Published in print: Jan 1, 2023
Discussion open until: Mar 11, 2023

Permissions

Request permissions for this article.

Authors

Affiliations

Leyang Wen, S.M.ASCE [email protected]
Ph.D. Student, Dept. of Civil and Environmental Engineering, Univ. of Michigan, 2350 Hayward St., Ann Arbor, MI 48109. Email: [email protected]
Assistant Professor, Dept. of Civil and Mineral Engineering, Univ. of Toronto, 35 St. George St., Toronto, ON, Canada M5S 1A4. ORCID: https://orcid.org/0000-0002-7381-9805. Email: [email protected]
Meiyin Liu, Ph.D., A.M.ASCE [email protected]
Assistant Professor, Dept. of Civil and Environmental Engineering, Rutgers Univ., 500 Bartholomew Rd., Piscataway, NJ 08854. Email: [email protected]
SangHyun Lee, Ph.D., M.ASCE [email protected]
Professor, Dept. of Civil and Environmental Engineering, Univ. of Michigan, 2350 Hayward St., Ann Arbor, MI 48109 (corresponding author). Email: [email protected]

Metrics & Citations

Metrics

Citations

Download citation

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited by

  • An Application of Cycle GAN for Creating Generated Real Training Images with 3D Excavator Pose Labels from a Synthetic Model, Construction Research Congress 2024, 10.1061/9780784485262.068, (670-678), (2024).
  • Feedback System for Enhancing Human-Robot Interaction Performance in Construction Robots, Construction Research Congress 2024, 10.1061/9780784485262.009, (79-88), (2024).
  • Excavator 3D pose estimation using deep learning and hybrid datasets, Advanced Engineering Informatics, 10.1016/j.aei.2023.101875, 55, (101875), (2023).

View Options

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Article
$35.00
Add to cart

Get Access

Access content

Please select your options to get access

Log in/Register Log in via your institution (Shibboleth)
ASCE Members: Please log in to see member pricing

Purchase

Save for later Information on ASCE Library Cards
ASCE Library Cards let you download journal articles, proceedings papers, and available book chapters across the entire ASCE Library platform. ASCE Library Cards remain active for 24 months or until all downloads are used. Note: This content will be debited as one download at time of checkout.

Terms of Use: ASCE Library Cards are for individual, personal use only. Reselling, republishing, or forwarding the materials to libraries or reading rooms is prohibited.
ASCE Library Card (5 downloads)
$105.00
Add to cart
ASCE Library Card (20 downloads)
$280.00
Add to cart
Buy Single Article
$35.00
Add to cart

Media

Figures

Other

Tables

Share

Share

Copy the content Link

Share with email

Email a colleague

Share