Détail de l'auteur
Auteur Gui-Song Xia |
Documents disponibles écrits par cet auteur (1)
![](./images/expand_all.gif)
![](./images/collapse_all.gif)
![Tris disponibles](./images/orderby_az.gif)
Parsing very high resolution urban scene images by learning deep ConvNets with edge-aware loss / Xianwei Zheng in ISPRS Journal of photogrammetry and remote sensing, vol 170 (December 2020)
![]()
[article]
Titre : Parsing very high resolution urban scene images by learning deep ConvNets with edge-aware loss Type de document : Article/Communication Auteurs : Xianwei Zheng, Auteur ; Linxi Huan, Auteur ; Gui-Song Xia, Auteur ; Jianya Gong, Auteur Année de publication : 2020 Article en page(s) : pp 15-28 Note générale : Bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] classification basée sur les régions
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] contour
[Termes IGN] image à très haute résolution
[Termes IGN] méthode fondée sur le noyau
[Termes IGN] scène urbaine
[Termes IGN] segmentation sémantiqueRésumé : (Auteur) Parsing very high resolution (VHR) urban scene images into regions with semantic meaning, e.g. buildings and cars, is a fundamental task in urban scene understanding. However, due to the huge quantity of details contained in an image and the large variations of objects in scale and appearance, the existing semantic segmentation methods often break one object into pieces, or confuse adjacent objects and thus fail to depict these objects consistently. To address these issues uniformly, we propose a standalone end-to-end edge-aware neural network (EaNet) for urban scene semantic segmentation. For semantic consistency preservation inside objects, the EaNet model incorporates a large kernel pyramid pooling (LKPP) module to capture rich multi-scale context with strong continuous feature relations. To effectively separate confusing objects with sharp contours, a Dice-based edge-aware loss function (EA loss) is devised to guide the EaNet to refine both the pixel- and image-level edge information directly from semantic segmentation prediction. In the proposed EaNet model, the LKPP and the EA loss couple to enable comprehensive feature learning across an entire semantic object. Extensive experiments on three challenging datasets demonstrate that our method can be readily generalized to multi-scale ground/aerial urban scene images, achieving 81.7% in mIoU on Cityscapes Test set and 90.8% in the mean F1-score on the ISPRS Vaihingen 2D Test set. Code is available at: https://github.com/geovsion/EaNet. Numéro de notice : A2020-703 Affiliation des auteurs : non IGN Thématique : IMAGERIE/INFORMATIQUE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2020.09.019 Date de publication en ligne : 14/10/2020 En ligne : https://doi.org/10.1016/j.isprsjprs.2020.09.019 Format de la ressource électronique : URL Article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=96228
in ISPRS Journal of photogrammetry and remote sensing > vol 170 (December 2020) . - pp 15-28[article]Exemplaires(1)
Code-barres Cote Support Localisation Section Disponibilité 081-2020121 RAB Revue Centre de documentation En réserve L003 Disponible