Home   >   CSC-OpenAccess Library   >    Manuscript Information
Full Text Available

This is an Open Access publication published under CSC-OpenAccess Policy.
Pedestrian Counting in Video Sequences based on Optical Flow Clustering
Sizuka Fujisawa, Go Hasegawa, Yoshiaki Taniguchi, Hirotaka Nakano
Pages - 1 - 16     |    Revised - 15-01-2013     |    Published - 28-02-2013
Volume - 7   Issue - 1    |    Publication Date - February 2013  Table of Contents
Pedestrian Counting, Video Processing, Optical Flow, Clustering.
The demand for automatic counting of pedestrians at event sites, buildings, or streets has been increased. Existing systems for counting pedestrians in video sequences have a problem that counting accuracy degrades when many pedestrians coexist and occlusion occurs frequently. In this paper, we introduce a method of clustering optical flows extracted from pedestrians in video frames to improve the counting accuracy. The proposed method counts the number of pedestrians by using pre-learned statistics, based on the strong correlation between the number of optical flow clusters and the actual number of pedestrians. We evaluate the accuracy of the proposed method using several video sequences, focusing in particular on the effect of parameters for optical flow clustering. We find that the proposed method improves the counting accuracy by up to 25% as compared with a non-clustering method. We also report that using a clustering threshold of angles less than 1 degree is effective for enhancing counting accuracy. Furthermore, we compare the performance of two algorithms that use feature points and lattice points when optical flows are detected. We confirm that the counting accuracy using feature points is higher than that using lattice points especially when the number of occluded pedestrians increases.
CITED BY (10)  
1 Taniguchi, Y., Mizushima, M., Hasegawa, G., Nakano, H., & Matsuoka, M. (2016). Counting Pedestrians Passing Through a Line in Crowded Scenes by Extracting Optical Flows. International Information Institute (Tokyo). Information, 19(1), 303.
2 Yoshida, T., & Taniguchi, Y. Estimating the number of people using existing WiFi access point in indoor environment.
3 Kim, J., Jang, G., Kim, G., & Kim, M. H. (2015). Crowd Activity Recognition using Optical Flow Orientation Distribution. KSII Transactions on Internet and Information Systems (TIIS), 9(8), 2948-2963.
4 Taniguchi, Y., & Nakano, H. (2014). Modeling and evaluation of a ceiling-mounted compound-eye sensor. International Information Institute (Tokyo). Information, 17(2), 663.
5 Jatropha Compro, Shigeru Matsuoka & Gordon. (2014). H. 264 / AVC with i ta wo ho Walker ? ri a number Hikaru Tatari ? Rousseau presumption. Portrait Society of Electronics, 43 (4), 599-604.
6 History of Tao, Cheng & party. (2014) based on multi-feature fusion pedestrian detection and tracking statistical methods. Television Technology, 38 (19), 179-183.
7 Mo Jianwen, Zhao Pu, & Yuan Hua. (2014). Based on Gabor features and enhancements RSC face recognition algorithm. Television Technology, 38 (19), 170-174.
8 Hashimoto, k., taniguchi, y., hasegawa, g., nakano, h., & matsuoka, m. pedestrian counting based on the number of salient points considering non-linear effect of occlusions.
9 MIZUSHIMA, M., TANIGUCHI, Y., Hasegawa, G., NAKANO, H., & MATSUOKA, M. (2013). Counting Pedestrians Passing through a Line in Video Sequences based on Optical Flow Extraction. Proceedings of CSECS 2013, 26-33.
10 Lin, D. T., & Huang, C. Robust detection for moving objects based on edge information and frame differences.
1 Google Scholar
2 CiteSeerX
3 refSeek
4 Scribd
5 SlideShare
6 PdfSR
1 F. Bu, R. Greene-Roesel, M. Diogenes, and D. Ragland, “Estimating pedestrian accident exposure: Automated pedestrian counting devices report,” UC Berkeley Traffic Safety Center,Tech. Rep., Mar. 2007.
2 X. Liu, P. H. Tu, J. Rittscher, A. G. A. Perera, and N. Krahnstoever, “Detecting and counting people in surveillance applications,” in Proc. AVSS 2005, pp. 306–311, Sep. 2005.
3 H. Zhao and R. Shibasaki, “A novel system for tracking pedestrians using multiple single-row laser-range scanners,” Systems, Man and Cybernetics, vol. 35, pp. 283–291, Mar. 2005.
4 A. Fod, A. Howard, and M. J. Mataric, “Laser-based people tracking,” in Proc. IEEE ICRA 2002, pp. 3024–3029, May 2002.
5 J. Cui, H. Zha, H. Zhao, and R. Shibasaki, “Laser-based detection and tracking of multiple people in crowds,” Computer Vision and Image Understanding, vol. 106, pp. 300–312, May 2007.
6 C. Harris and M. Stephen, “A real-time system for monitoring pedestrians,” in Proc. WACVMOTION 2005, vol. 1, pp. 378–385, Jan. 2005.
7 A. Leykin and R. Hammoud, “Robust multi-pedestrian tracking in thermal-visible surveillance videos,” in Proc. Conference on CVPR Workshop 2006, p. 136, Jun. 2006.
8 E. Goubet, J. Katz, and F. Porikli, “Pedestrian tracking using thermal infrared imaging,” in Proc.SPIE 2006, no. 6206, pp. 797–808, Apr. 2006.
9 K. Hashimoto, M. Yoshinomoto, S. Matsueda, K. Morinaka, and N. Yoshiike, “Development of people-counting system with human-information sensor using multi-element pyroelectric infrared array detector,” Sensors and Actuators A: Physical, no. 58, pp. 165–171, Feb. 1998.
10 SenSource Inc, “PCW-2BX03 Directional People Counter.” Internet:www.sensourceinc.com/PDF/PCW-Wired-Directional%20-%20Brochure.pdf. [Nov. 22, 2012].
11 Eco counter, “People counters - Eco-counter.” Internet: www.eco-compteur.com. [Nov. 22,2012].
12 InfraRed Integrated Systems Ltd, “IRISYS people counter,” Internet: irisys.co.uk/peoplecounting.[Nov. 22, 2012].
13 K. Hashimoto, M. Yoshinomoto, S. Matsueda, K. Morinaka, and N. Yoshiike, “Development of people-counting system with human-information sensor using multielement pyroelectric infrared array detector,” Sensors and Actuators, vol. 58, no. 2, pp. 165–171, Feb. 1997.
14 Q. Chen, M. Gao, J. Ma, D. Zhang, L. M. Ni, and Y. Liu, “Mocus: Moving object counting using ultrasonic sensor networks,” International Journal of Sensor Networks, vol. 3, no. 1, pp. 55–65, Dec. 2007.
15 T. Zhao and R. Nevatia, “Bayesian human segmentation in crowded situations,” in Proc.CVPR 2003, pp. 459-466, Jun. 2003.
16 D. Conte, P. Foggia, G. Percannella, and M. Vento, “Performance evaluation of a people tracking system on pets2009 database,” in Proc. ACSS 2010, pp. 119–126, Aug. 2010.
17 P. Kilambi, E. Ribnick, A. J. Joshi, O. Masoud, and N. Papanikolopoulos, “Estimating pedestrian counts in groups,” Computer Vision and Image Understanding, vol. 110, pp. 43-59,Apr. 2008.
18 James Ferryman, “PETS 2010.” Internet: www.cvg.rdg.ac.uk/PETS2010. [Nov. 22, 2012].
19 T. Ojala, M. Pietikainen, and T. Maenpaa, “Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,” IEEE Transactions on Neural Networks, vol. 24,no. 7, pp. 971–989, Jul. 2002.
20 S. Bileschi and L. Wolf, “A unified system for object detection, texture recognition, and context analysis based on the standard model feature set,” in Proc. BMVC 2005, Sep. 2005.
21 M. Heeikkila and M. Pietikainen, “A texture-based method for modeling the background and detection moving objects,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.28, no. 4, pp. 657–662, Apr. 2006.
22 V. Mahadevan and N. Vasconcelos, “Background subtraction in highly dynamic scenes,” in Proc. CVPR 2008, pp. 1–6, Jun. 2008.
23 T. Zhao and B. Wu, “Segmentation and tracking of multiple humans in crowded environments,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, no. 7,pp. 1198–1211, Jul. 2008.
24 H. Kitazawa, Z. Li, and K. Yabuta, “Moving object extraction and tracking based on the exclusive block matching,” IEICE technical report, vol. 108, no. 4, pp. 49–54, Apr. 2008.
25 N. Baba, T. Ejima, and H. Matsuo, “Headfinder: A real-time robust head detection and tracking system,” in Proc. SPIE Electronic Imaging Conference on Real-Time Imaging 2002, vol.4666, pp. 42–51, Feb. 2002.
26 A. Ess, B. Leibe, K. Schindler, and L. V. Gool, “A mobile vision system for robust multiperson tracking,” in Proc. CVPR 2008, pp. 1–8, Jun. 2008.
27 R. Eshel and Y. Moses, “Homography based multiple camera detection and tracking of people in a dense crowd,” in Proc. CVPR 2008, pp. 1–8, Jun. 2008.
28 T. Teixeira and A. Savvide, “Lightweight people counting and localizing in indoor spaces using camera sensor nodes,” in Proc. ICDSC 2007, pp. 36–43, Sep. 2007.
29 V. Kettnaker and R. Zabih, “Counting people from multiple cameras,” in Proc. MCS 1999, vol.2, pp. 267–271, Jul. 1999.
30 A. Mittal and L. S. Davis, “M2tracker: A multi-view approach to segmenting and tracking people in a cluttered scene using region-based stereo,” Computer Vision, vol. 2350, pp. 189–203,May 2002.
31 J. P. Batista, “Tracking pedestrians under occlusion using multiple cameras,” in Proc. ICIAR 2004, vol. 3212, pp. 555–562, Oct. 2004.
32 N. Friedman and S. Russell, “Image segmentation in video sequences: A probabilistic approach,” in Proc. UAI 1997, Aug. 1997.
33 A. Elegammal, D. Harwood, and L. Davis, “Non-parametric model for background subtraction,” in Proc. ECCV 2000, vol. 2, pp. 751–767, Jun. 2000.
34 T. Horprasert, D. Harwood, and L. Davis, “A statistical approach for real-time robust background subtraction and shadow detection,” in Proc. IEEE ECCV 1999 Frame-Rate Workshop,Sep. 1999.
35 J.-S. Hu, T.-M. Su, and S.-C. Jeng, “Robust background subtraction with shadow and highlight removal for indoor surveillance,” in Proc. IROS 2006, pp. 4545–4550, Oct. 2006.
36 N. Friedman and S. Russell, “Image segmentation in video sequences: A probabilistic approach,” in Proc. UAI 1997, pp. 175–181, Aug. 1997.
37 C. Harris and M. Stephen, “A combined corner and edge detector,” in Proc. AVC 1988, pp.147–152, Aug. 1988.
38 R. Mehrotra and S. Nichani, “Corner detection,” Pattern Recognition, vol. 23, pp. 1223–1233,Mar. 1990.
39 B. D. Lucas and T. Kanade, “An iterative image registration technique with an application to stereo vision,” in Proc. the 1981 DARPA Image Understanding Workshop, pp. 121–130, Apr.1981.
40 James Ferryman, “Video sequence take from PETS workshop,” available Internet:www.cvg.rdg.ac.uk/PETS2010/a.html. [Nov. 22, 2012].
Miss Sizuka Fujisawa
Graduate School of Information Science and Technology Osaka University Suita, Osaka, 565-0871 - Japan
Associate Professor Go Hasegawa
Cybermedia Center Osaka University Toyonaka, Osaka, 560-0043 - Japan
Dr. Yoshiaki Taniguchi
Cybermedia Center Osaka University Toyonaka, Osaka, 560-0043 - Japan
Professor Hirotaka Nakano
Cybermedia Center Osaka University Toyonaka, Osaka, 560-0043 - Japan