Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Few-shot Classification of Aerial Scene Images via Meta-learning

Version 1 : Received: 1 October 2020 / Approved: 2 October 2020 / Online: 2 October 2020 (09:24:19 CEST)
Version 2 : Received: 14 December 2020 / Approved: 15 December 2020 / Online: 15 December 2020 (13:21:49 CET)

A peer-reviewed article of this Preprint also exists.

Journal reference: Remote Sensing 2021
DOI: 10.3390/rs13010108


CNN-based methods have dominated the field of aerial scene classification for the past few years. While achieving remarkable success, CNN-based methods suffer from excessive parameters and notoriously rely on large amounts of training data. In this work, we introduce few-shot learning to the aerial scene classification problem. Few-shot learning aims to learn a model on base-set that can quickly adapt to unseen categories in novel-set, using only a few labeled samples. To this end, we proposed a meta-learning method for few-shot classification of aerial scene images. First, we train a feature extractor on all base categories to learn a representation of inputs. Then in the meta-training stage, the classifier is optimized in the metric space by cosine distance with a learnable scale parameter. At last, in the meta-testing stage, the query sample in the unseen category is predicted by the adapted classifier given a few support samples. We conduct extensive experiments on two challenging datasets: NWPU-RESISC45 and RSD46-WHU. The experimental results show that our method outperforms three state-of-the-art few-shot algorithms and one typical CNN-based method, D-CNN. Furthermore, several ablation experiments are conducted to investigate the effects of dataset scale and support shots; the experiment results confirm that our model is specifically effective in few-shot settings.


aerial scene classification; remote-sensing image classification; few-shot learning; meta-learning


MATHEMATICS & COMPUTER SCIENCE, General & Theoretical Computer Science

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our diversity statement.

Leave a public comment
Send a private comment to the author(s)
Views 0
Downloads 0
Comments 0
Metrics 0

Notify me about updates to this article or when a peer-reviewed version is published.

We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.