TY - GEN
T1 - Text-based Person Search in Full Images via Semantic-Driven Proposal Generation
AU - Zhang, Shizhou
AU - Cheng, De
AU - Luo, Wenlong
AU - Xing, Yinghui
AU - Long, Duo
AU - Li, Hao
AU - Niu, Kai
AU - Liang, Guoqiang
AU - Zhang, Yanning
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/11/2
Y1 - 2023/11/2
N2 - Finding target persons in full scene images with a query of text description has important practical applications in intelligent video surveillance. However, different from the real-world scenarios where the bounding boxes are not available, existing text-based person re- trieval methods mainly focus on the cross modal matching between the query text descriptions and the gallery of cropped pedestrian images. To close the gap, we study the problem of text-based person search in full images by proposing a new end-to-end learning framework which jointly optimize the pedestrian detection, identification and visual-semantic feature embedding tasks. To take full advantage of the query text, the semantic features are leveraged to instruct the Region Proposal Network to pay more attention to the text-described proposals. Besides, a cross-scale visual-semantic embedding mechanism is utilized to improve the performance. To validate the proposed method, we collect and annotate two large-scale benchmark datasets based on the widely adopted image-based person search datasets CUHK-SYSU and PRW. Comprehensive experiments are conducted on the two datasets and compared with the baseline methods, our method achieves the state-of-the-art performance.
AB - Finding target persons in full scene images with a query of text description has important practical applications in intelligent video surveillance. However, different from the real-world scenarios where the bounding boxes are not available, existing text-based person re- trieval methods mainly focus on the cross modal matching between the query text descriptions and the gallery of cropped pedestrian images. To close the gap, we study the problem of text-based person search in full images by proposing a new end-to-end learning framework which jointly optimize the pedestrian detection, identification and visual-semantic feature embedding tasks. To take full advantage of the query text, the semantic features are leveraged to instruct the Region Proposal Network to pay more attention to the text-described proposals. Besides, a cross-scale visual-semantic embedding mechanism is utilized to improve the performance. To validate the proposed method, we collect and annotate two large-scale benchmark datasets based on the widely adopted image-based person search datasets CUHK-SYSU and PRW. Comprehensive experiments are conducted on the two datasets and compared with the baseline methods, our method achieves the state-of-the-art performance.
KW - cross scale alignment.
KW - semantic-driven rpn
KW - text-based person search
UR - http://www.scopus.com/inward/record.url?scp=85178584785&partnerID=8YFLogxK
U2 - 10.1145/3606041.3618058
DO - 10.1145/3606041.3618058
M3 - 会议稿件
AN - SCOPUS:85178584785
T3 - HCMA 2023 - Proceedings of the 4th International Workshop on Human-centric Multimedia Analysis, Co-located with: MM 2023
SP - 5
EP - 14
BT - HCMA 2023 - Proceedings of the 4th International Workshop on Human-centric Multimedia Analysis, Co-located with
PB - Association for Computing Machinery, Inc
T2 - 4th International Workshop on Human-centric Multimedia Analysis, HCMA 2023
Y2 - 2 November 2023
ER -