Deep Active Learning From Multispectral Data Through Cross-Modality Prediction Inconsistency
Heng Zhang, Elisa Fromont, Sebastien Lefevre, Bruno Avignon
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:07:15
Data from multiple sensors provide independent and complementary information, which may improve the robustness and reliability of scene analysis applications. While there exist many large-scale labelled benchmarks acquired by a single sensor, collecting labelled multi-sensor data is more expensive and time-consuming. In this work, we explore the construction of an accurate multispectral (here, visible & thermal cameras) scene analysis system with minimal annotation efforts via an active learning strategy based on the cross-modality prediction inconsistency. Experiments on multispectral datasets and vision tasks demonstrate the effectiveness of our method. In particular, with only 10% of labelled data on KAIST multispectral pedestrian detection dataset, we obtain comparable performance as other fully supervised State-of-the-Art methods.