Pedestrian Tracking at Signalized Intersections Leveraging Multi-Camera Field of Views Using Covolutional Neural Network-Based Pose Estimation Algorithm
Publication: International Conference on Transportation and Development 2024
ABSTRACT
Pedestrian detection poses significant challenges for traffic safety researchers, given diverse motion patterns, clothing colors, partial occlusions, and varying positions concerning detection devices. Surveillance cameras, light detection and ranging (LiDAR), and microwaves beams have been utilized in pedestrians’ detection, leading to robust safety assessment methodologies. Despite their efficacy, widespread deployment encounters constraints. This study introduces an optimized approach by combining surveillance cameras with a distinctive convolutional neural network (CNN)-based pose estimation algorithm for precise pedestrian detection at signalized intersections. A geometrical spatial proximity method, grounded in linear and curvilinear perspectives, restores pedestrian joint coordinates from the image plane to a top-down view. These coordinates are clustered and integrated to construct pedestrian trajectories. The proposed framework serves as an efficient and accurate tool for assessing pedestrian motions, addressing challenges in widespread deployment and contributing significantly to advancing traffic safety practices at signalized intersections.
Get full access to this chapter
View all available purchase options and get full access to this chapter.
REFERENCES
Alahi, A., M. Bierlaire, and P. Vandergheynst. (2014). “Robust Real-Time Pedestrians Detection in Urban Environments with Low-Resolution Cameras”. Transportation Research Part C: Emerging Technologies, Vol. 39, pp. 113–128. https://doi.org/10.1016/j.trc.2013.11.019.
Alfred Daniel, J., C. Chandru Vignesh, B. A. Muthu, R. Senthil Kumar, C. Sivaparthipan, and C. E. M. Marin. (2023). “Fully Convolutional Neural Networks for LIDAR–Camera Fusion for Pedestrian Detection in Autonomous Vehicle”. Multimedia Tools and Applications, Vol. 82, No. 16, pp. 25107–25130. https://doi.org/10.1007/s11042-023-14417-x.
Antonio, J., and M. Romero. (2018). “Pedestrians’ Detection Methods in Video Images: A Literature Review”. 2018 International Conference on Computational Science and Computational Intelligence (CSCI), 354–360. https://doi.org/10.1109/CSCI46756.2018.00074.
Brownson, C. D. (1981). “Euclid’s Optics and Its Compatibility with Linear Perspective. Vol. 24, No. 3 pp. 165–194, 1981. https://www.jstor.org/stable/41133617.
Brunetti, A., D. Buongiorno, G. F. Trotta, and V. Bevilacqua. (2018). “Computer vision and deep learning techniques for pedestrian detection and tracking: A survey”. Neurocomputing, 300, 17–33. https://doi.org/10.1016/j.neucom.2018.01.092.
Cavallaro, A., O. Steiger, and T. Ebrahimi. (2005). “Tracking Video Objects in Cluttered Background”. IEEE Transactions on Circuits and Systems for Video Technology, Vol. 15, No. 4, pp. 575–584. https://doi.org/10.1109/TCSVT.2005.844447.
Cicchino, J. B. (2022). “Effects of Automatic Emergency Braking Systems on Pedestrian Crash Risk”. Accident Analysis and Prevention, Vol. 172. https://doi.org/10.1016/j.aap.2022.106686.
Dahlkamp, H., A. E. C. Pece, A. Ottlik, and H.-H. Nagel. (2004). “Differential Analysis of Two Model-Based Vehicle Tracking Approaches”. DAGM-Symposium.
Dollár, P., C. Wojek, B. Schiele, and P. Perona. (2012). “Pedestrian Detection: An Evaluation of the State of the Art”. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 34, No. 4, pp. 743–761. https://doi.org/10.1109/TPAMI.2011.155.
Enzweiler, M., and D. M. Gavrila. (2010). “Monocular pedestrian detection: Survey and experiments”. IEEE Trans. on PAMI 31(12), 2179–2195. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 31. 2179-2195. https://doi.org/10.1109/TPAMI.2008.260.
Essa, M., and T. Sayed. (2015). “Simulated Traffic Conflicts: Do They Accurately Represent Field-Measured 3 Conflicts?”. Transportation Research Record, Vol. 2514, pp. 48–57. 4 https://doi.org/10.3141/2514-06.
Federal Highway Administration, US DOT. (2023). Traffic Safety Facts. Retrieved from [https://highways.dot.gov/research/research-programs/safety/intersection-safety].
Fu, H., J. Gao, and H. Liu. (2023). “Human Pose Estimation and Action Recognition for Fitness Movements”. Computers & Graphics, https://doi.org/10.1016/j.cag.2023.09.008.
Hou, J., G. F. List, and X. Guo. (2014). “New Algorithms for Computing the Time-to-Collision in Freeway 6 Traffic Simulation Models”. Computational Intelligence and Neuroscience, Vol. 2014, 7 https://doi.org/10.1155/2014/761047.
Ismail, K. A. (2010). Application of computer vision techniques for automated road safety analysis and traffic data collection. (Doctoral dissertation, University of British Columbia). https://doi.org/10.14288/1.0062871.
Kim, D.-G., S. Washington, and J. Oh. (2006). “Modeling Crash Types: New Insights into the Effects of 4 Covariates on Crashes at Rural Intersections”. Journal Of Transportation Engineering, pp. 282–292, https://doi.org/10.1061/(ASCE)0733-947X(2006)132:4(282).
Koller, D., J. Weber, and J. Malik. (1994). Multiple Car Tracking with Occlusion Reasoning. Lecture Notes in Computer Science, vol 800. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-57956-7_22.
Magee, D. R. (2004). “Tracking Multiple Vehicles Using Foreground, Background and Motion Models”. Image and Vision Computing, Vol. 22, No. 2, pp. 143–155. https://doi.org/10.1016/S0262-8856(03)00145-8.
Maurin, B., O. Masoud, and N. P. Papanikolopoulos. (2004). “Tracking All Traffic: Computer Vision 13 Algorithms for Monitoring Vehicles Individuals, and Crowds”. IEEE Robotics and Automation 14 Magazine, Vol. 12, No. 1, pp. 29–36. https://doi.org/10.1109/MRA.2005.1411416.
Santoyo, J., and M. A. Santoyo. (2021). “The 360° Curvilinear Perspective: A Hybrid Hypercubic Angular Space Grid Based on the 1968 Barre and Flocon Proposal”. Nexus Network Journal, Vol. 23, No. 3, pp. 717–735. https://doi.org/10.1007/s00004-021-00552-1.
Saunders, J. A., and B. T. Backus. (2006). “The Accuracy and Reliability of Perceived Depth from Linear Perspective as a Function of Image Size”. Journal of Vision, Vol. 6, No. 9. https://doi.org/10.1167/6.9.7.
Saunier, N., and T. Sayed. (2006). “A Feature-Based Tracking Algorithm for Vehicles in Intersections”. The 3rd Canadian Conference on Computer and Robot Vision (CRV'06), Quebec, Canada, 2006, pp. 59–59, https://doi.org/10.1109/CRV.2006.3.
Stauffer, C., and W. E. L. Grimson. (2000). “Learning Patterns of Activity Using Real-Time Tracking”. IEEE 16 Transactions on Pattern Analysis and Machine Intelligence, Vol. 22, No. 8, pp. 747–757. 17 https://doi.org/10.1109/34.868677.
Veeraraghavan, H., O. Masoud, and N. P. Papanikolopoulos. (2003). “Computer Vision Algorithms for 19 Intersection Monitoring”. IEEE Transactions on Intelligent Transportation Systems, Vol. 4, No. 2, pp. 78–89. https://doi.org/10.1109/TITS.2003.821212.
Wang, C.-Y., A. Bochkovskiy, and H.-Y. M. Liao. (2022). “YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors”. https://doi.org/10.48550/arXiv.2207.02696.
Wu, B., Z. J. He, and T. L. Ooi. (2007). “The linear perspective information in ground surface representation and distance judgment”. Perception & Psychophysics 69, 654–672. https://doi.org/10.3758/BF03193769.
Zhang, S., R. Benenson, and B. Schiele. (2017). “CityPersons: A Diverse Dataset for Pedestrian Detection”. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017): 4457–4465. https://doi.org/10.48550/arXiv.1702.05693.
Information & Authors
Information
Published In
History
Published online: Jun 13, 2024
Authors
Metrics & Citations
Metrics
Citations
Download citation
If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.