PropertyValue
?:abstract
  • Crowd counting is a fundamental yet challenging problem, which desires rich information to generate pixel-wise crowd density maps. However, most previous methods only utilized the limited information of RGB images and may fail to discover the potential pedestrians in unconstrained environments. In this work, we find that incorporating optical and thermal information can greatly help to recognize pedestrians. To promote future researches in this field, we introduce a large-scale RGBT Crowd Counting (RGBT-CC) benchmark, which contains 2,030 pairs of RGB-thermal images with 138,389 annotated people. Furthermore, to facilitate the multimodal crowd counting, we propose a cross-modal collaborative representation learning framework, which consists of multiple modality-specific branches, a modality-shared branch, and an Information Aggregation-Distribution Module (IADM) to fully capture the complementary information of different modalities. Specifically, our IADM incorporates two collaborative information transfer components to dynamically enhance the modality-shared and modality-specific representations with a dual information propagation mechanism. Extensive experiments conducted on the RGBT-CC benchmark demonstrate the effectiveness of our framework for RGBT crowd counting. Moreover, the proposed approach is universal for multimodal crowd counting and is also capable to achieve superior performance on the ShanghaiTechRGBD dataset.
is ?:annotates of
?:arxiv_id
  • 2012.04529
?:creator
?:externalLink
?:license
  • arxiv
?:pdf_json_files
  • document_parses/pdf_json/9447d554c0dd7607fdf1bafbee58dce05cebd4bd.json
?:publication_isRelatedTo_Disease
?:sha_id
?:source
  • ArXiv
?:title
  • Cross-Modal Collaborative Representation Learning and a Large-Scale RGBT Benchmark for Crowd Counting
?:type
?:year
  • 2020-12-08

Metadata

Anon_0  
expand all