Semantic Segmentation of Remote Sensing Images With Sparse Annotations

Training convolutional neural networks (CNNs) for very high-resolution images requires a large quantity of high-quality pixel-level annotations, which is extremely labor-intensive and time-consuming to produce. Moreover, professional photograph interpreters might have to be involved in guaranteeing the correctness of annotations. To alleviate such a burden, we propose a framework for semantic segmentation of aerial images based on incomplete annotations, where annotators are asked to label a few pixels with easy-to-draw scribbles. To exploit these sparse scribbled annotations, we propose the FEature and Spatial relaTional regulArization (FESTA) method to complement the supervised task with an unsupervised learning signal that accounts for neighborhood structures both in spatial and feature terms. For the evaluation of our framework, we perform experiments on two remote sensing image segmentation data sets involving aerial and satellite imagery, respectively. Experimental results demonstrate that the exploitation of sparse annotations can significantly reduce labeling costs, while the proposed method can help improve the performance of semantic segmentation when training on such annotations. The sparse labels and codes are publicly available for reproducibility purposes.1

Saved in:
Bibliographic Details
Main Authors: Hua, Yuansheng, Marcos, Diego, Mou, Lichao, Zhu, Xiao Xiang, Tuia, Devis
Format: Article/Letter to editor biblioteca
Language:English
Subjects:Aerial image, Annotations, Image color analysis, Image segmentation, Kernel, Remote sensing, Semantics, Training, convolutional neural networks (CNNs), semantic segmentation, semisupervised learning, sparse scribbled annotation.,
Online Access:https://research.wur.nl/en/publications/semantic-segmentation-of-remote-sensing-images-with-sparse-annota
Tags: Add Tag
No Tags, Be the first to tag this record!
id dig-wur-nl-wurpubs-579357
record_format koha
spelling dig-wur-nl-wurpubs-5793572025-01-14 Hua, Yuansheng Marcos, Diego Mou, Lichao Zhu, Xiao Xiang Tuia, Devis Article/Letter to editor IEEE Geoscience and Remote Sensing Letters 19 (2022) ISSN: 1545-598X Semantic Segmentation of Remote Sensing Images With Sparse Annotations 2022 Training convolutional neural networks (CNNs) for very high-resolution images requires a large quantity of high-quality pixel-level annotations, which is extremely labor-intensive and time-consuming to produce. Moreover, professional photograph interpreters might have to be involved in guaranteeing the correctness of annotations. To alleviate such a burden, we propose a framework for semantic segmentation of aerial images based on incomplete annotations, where annotators are asked to label a few pixels with easy-to-draw scribbles. To exploit these sparse scribbled annotations, we propose the FEature and Spatial relaTional regulArization (FESTA) method to complement the supervised task with an unsupervised learning signal that accounts for neighborhood structures both in spatial and feature terms. For the evaluation of our framework, we perform experiments on two remote sensing image segmentation data sets involving aerial and satellite imagery, respectively. Experimental results demonstrate that the exploitation of sparse annotations can significantly reduce labeling costs, while the proposed method can help improve the performance of semantic segmentation when training on such annotations. The sparse labels and codes are publicly available for reproducibility purposes.1 en application/pdf https://research.wur.nl/en/publications/semantic-segmentation-of-remote-sensing-images-with-sparse-annota 10.1109/LGRS.2021.3051053 https://edepot.wur.nl/542099 Aerial image Annotations Image color analysis Image segmentation Kernel Remote sensing Semantics Training convolutional neural networks (CNNs) semantic segmentation semisupervised learning sparse scribbled annotation. https://creativecommons.org/licenses/by/4.0/ https://creativecommons.org/licenses/by/4.0/ Wageningen University & Research
institution WUR NL
collection DSpace
country Países bajos
countrycode NL
component Bibliográfico
access En linea
databasecode dig-wur-nl
tag biblioteca
region Europa del Oeste
libraryname WUR Library Netherlands
language English
topic Aerial image
Annotations
Image color analysis
Image segmentation
Kernel
Remote sensing
Semantics
Training
convolutional neural networks (CNNs)
semantic segmentation
semisupervised learning
sparse scribbled annotation.
Aerial image
Annotations
Image color analysis
Image segmentation
Kernel
Remote sensing
Semantics
Training
convolutional neural networks (CNNs)
semantic segmentation
semisupervised learning
sparse scribbled annotation.
spellingShingle Aerial image
Annotations
Image color analysis
Image segmentation
Kernel
Remote sensing
Semantics
Training
convolutional neural networks (CNNs)
semantic segmentation
semisupervised learning
sparse scribbled annotation.
Aerial image
Annotations
Image color analysis
Image segmentation
Kernel
Remote sensing
Semantics
Training
convolutional neural networks (CNNs)
semantic segmentation
semisupervised learning
sparse scribbled annotation.
Hua, Yuansheng
Marcos, Diego
Mou, Lichao
Zhu, Xiao Xiang
Tuia, Devis
Semantic Segmentation of Remote Sensing Images With Sparse Annotations
description Training convolutional neural networks (CNNs) for very high-resolution images requires a large quantity of high-quality pixel-level annotations, which is extremely labor-intensive and time-consuming to produce. Moreover, professional photograph interpreters might have to be involved in guaranteeing the correctness of annotations. To alleviate such a burden, we propose a framework for semantic segmentation of aerial images based on incomplete annotations, where annotators are asked to label a few pixels with easy-to-draw scribbles. To exploit these sparse scribbled annotations, we propose the FEature and Spatial relaTional regulArization (FESTA) method to complement the supervised task with an unsupervised learning signal that accounts for neighborhood structures both in spatial and feature terms. For the evaluation of our framework, we perform experiments on two remote sensing image segmentation data sets involving aerial and satellite imagery, respectively. Experimental results demonstrate that the exploitation of sparse annotations can significantly reduce labeling costs, while the proposed method can help improve the performance of semantic segmentation when training on such annotations. The sparse labels and codes are publicly available for reproducibility purposes.1
format Article/Letter to editor
topic_facet Aerial image
Annotations
Image color analysis
Image segmentation
Kernel
Remote sensing
Semantics
Training
convolutional neural networks (CNNs)
semantic segmentation
semisupervised learning
sparse scribbled annotation.
author Hua, Yuansheng
Marcos, Diego
Mou, Lichao
Zhu, Xiao Xiang
Tuia, Devis
author_facet Hua, Yuansheng
Marcos, Diego
Mou, Lichao
Zhu, Xiao Xiang
Tuia, Devis
author_sort Hua, Yuansheng
title Semantic Segmentation of Remote Sensing Images With Sparse Annotations
title_short Semantic Segmentation of Remote Sensing Images With Sparse Annotations
title_full Semantic Segmentation of Remote Sensing Images With Sparse Annotations
title_fullStr Semantic Segmentation of Remote Sensing Images With Sparse Annotations
title_full_unstemmed Semantic Segmentation of Remote Sensing Images With Sparse Annotations
title_sort semantic segmentation of remote sensing images with sparse annotations
url https://research.wur.nl/en/publications/semantic-segmentation-of-remote-sensing-images-with-sparse-annota
work_keys_str_mv AT huayuansheng semanticsegmentationofremotesensingimageswithsparseannotations
AT marcosdiego semanticsegmentationofremotesensingimageswithsparseannotations
AT moulichao semanticsegmentationofremotesensingimageswithsparseannotations
AT zhuxiaoxiang semanticsegmentationofremotesensingimageswithsparseannotations
AT tuiadevis semanticsegmentationofremotesensingimageswithsparseannotations
_version_ 1822265926274252800