Deep Learning Classification of Large-Scale Point Clouds: A Case Study On Cuneiform Tablets
Frederik Hagelskjaer
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:11:20
Image retrieval consists in the selection of a set of images from a database, based on their visual similarity to a given query image. To compute this visual similarity, a global or multiple local features are required per image. in this work, we present a system that uses multiple high-level semantic feature maps from a single image for global feature generation. Our method uses feature maps extracted from a backbone architecture at different spatial resolutions, enhancing the semantic information relevant at each scale using a transformer-based approach. Then, these feature maps are combined using a self-attention mechanism generating a global feature. The main contribution of this work is a novel global feature generation method, which outperforms current state-of-the-art techniques that use either global-only or a combination of global and local features. This result is assessed using the Google Landmarks v2 resulting in an improvement of 3.8% in mAP@100. Also, we assess our system using Revisiting Oxford and Paris datasets, obtaining an improvement up to 3.7% in mAP compared with current methods.