TY - JOUR
T1 - PCC Net
T2 - Perspective crowd counting via spatial convolutional network
AU - Gao, Junyu
AU - Wang, Qi
AU - Li, Xuelong
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2020/10
Y1 - 2020/10
N2 - Crowd counting from a single image is a challenging task due to high appearance similarity, perspective changes, and severe congestion. Many methods only focus on the local appearance features and they cannot handle the aforementioned challenges. In order to tackle them, we propose a perspective crowd counting network (PCC Net), which consists of three parts: 1) density map estimation (DME) focuses on learning very local features of density map estimation; 2) random high-level density classification (R-HDC) extracts global features to predict the coarse density labels of random patches in images; and 3) fore-/background segmentation (FBS) encodes mid-level features to segments the foreground and background. Besides, the Down, Up, Left, and Right (DULR) module is embedded in PCC Net to encode the perspective changes on four directions (DULR). The proposed PCC Net is verified on five mainstream datasets, which achieves the state-of-the-art performance on the one and attains the competitive results on the other four datasets. The source code is available at https://github.com/gjy3035/PCC-Net.
AB - Crowd counting from a single image is a challenging task due to high appearance similarity, perspective changes, and severe congestion. Many methods only focus on the local appearance features and they cannot handle the aforementioned challenges. In order to tackle them, we propose a perspective crowd counting network (PCC Net), which consists of three parts: 1) density map estimation (DME) focuses on learning very local features of density map estimation; 2) random high-level density classification (R-HDC) extracts global features to predict the coarse density labels of random patches in images; and 3) fore-/background segmentation (FBS) encodes mid-level features to segments the foreground and background. Besides, the Down, Up, Left, and Right (DULR) module is embedded in PCC Net to encode the perspective changes on four directions (DULR). The proposed PCC Net is verified on five mainstream datasets, which achieves the state-of-the-art performance on the one and attains the competitive results on the other four datasets. The source code is available at https://github.com/gjy3035/PCC-Net.
KW - Crowd counting
KW - background segmentation
KW - crowd analysis
KW - multi-task learning
KW - spatial convolutional network
UR - http://www.scopus.com/inward/record.url?scp=85092446345&partnerID=8YFLogxK
U2 - 10.1109/TCSVT.2019.2919139
DO - 10.1109/TCSVT.2019.2919139
M3 - 文章
AN - SCOPUS:85092446345
SN - 1051-8215
VL - 30
SP - 3486
EP - 3498
JO - IEEE Transactions on Circuits and Systems for Video Technology
JF - IEEE Transactions on Circuits and Systems for Video Technology
IS - 10
M1 - 8723079
ER -