IEEE Transactions on geoscience and remote sensing / IEEE Geoscience and remote sensing society (Etats-Unis) . vol 60 n° 6Paru le : 01/06/2022 |
[n° ou bulletin]
est un bulletin de IEEE Transactions on geoscience and remote sensing / IEEE Geoscience and remote sensing society (Etats-Unis) (1986 -)
[n° ou bulletin]
|
Dépouillements
Ajouter le résultat dans votre panierFeature-selection high-resolution network with hypersphere embedding for semantic segmentation of VHR remote sensing images / Hanwen Xu in IEEE Transactions on geoscience and remote sensing, vol 60 n° 6 (June 2022)
[article]
Titre : Feature-selection high-resolution network with hypersphere embedding for semantic segmentation of VHR remote sensing images Type de document : Article/Communication Auteurs : Hanwen Xu, Auteur ; Xinming Tang, Auteur ; Bo Ai, Auteur ; et al., Auteur Année de publication : 2022 Article en page(s) : n° 4411915 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] apprentissage profond
[Termes IGN] architecture de réseau
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] entropie
[Termes IGN] extraction de traits caractéristiques
[Termes IGN] image à très haute résolution
[Termes IGN] segmentation multi-échelle
[Termes IGN] segmentation sémantiqueRésumé : (auteur) Very-high-resolution (VHR) remote sensing images contain various multiscale objects, such as large-scale buildings and small-scale cars. However, these multiscale objects cannot be considered simultaneously in the widely used backbones with a large downsampling factor (e.g., VGG-like and ResNet-like), resulting in the appearance of various context aggregation approaches, such as fusing low-level features and attention-based modules. To alleviate this problem caused by backbones with a large downsampling factor, we propose a feature-selection high-resolution network (FSHRNet) based on an observation: if the features maintain high resolution throughout the network, a high precision segmentation result can be obtained by only using a 1× 1 convolution layer with no need for complex context aggregation modules. Specifically, the backbone of FSHRNet is a multibranch structure similar to HRNet where the high-resolution branch is the principal line. Then, a lightweight dynamic weight module, named the feature-selection convolution (FSConv) layer, is presented to fuse multiresolution features, allowing adaptively feature selection based on the characteristic of objects. Finally, a specially designed 1× 1 convolution layer derived from hypersphere embedding is used to produce the segmentation result. Experiments with other widely used methods show that the proposed FSHRNet obtains competitive performance on the ISPRS Vaihingen dataset, the ISPRS Potsdam dataset, and the iSAID dataset. Numéro de notice : A2022-559 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1109/TGRS.2022.3183144 En ligne : https://doi.org/10.1109/TGRS.2022.3183144 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=101184
in IEEE Transactions on geoscience and remote sensing > vol 60 n° 6 (June 2022) . - n° 4411915[article]Invariant structure representation for remote sensing object detection based on graph modeling / Zicong Zhu in IEEE Transactions on geoscience and remote sensing, vol 60 n° 6 (June 2022)
[article]
Titre : Invariant structure representation for remote sensing object detection based on graph modeling Type de document : Article/Communication Auteurs : Zicong Zhu, Auteur ; Xian Sun, Auteur ; Wenhui Diao, Auteur ; et al., Auteur Année de publication : 2022 Article en page(s) : n° 5625217 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] détection d'objet
[Termes IGN] données d'entrainement sans étiquette
[Termes IGN] extraction de traits caractéristiques
[Termes IGN] filtrage numérique d'image
[Termes IGN] granularité d'image
[Termes IGN] graphe
[Termes IGN] invariantRésumé : (auteur) Due to the characteristics of vertical orthophoto imaging, the apparent structural features of the object in the remote sensing (RS) image are relatively stable, such as the cross-shaped structure of the aircraft and the rectangular structure of the vehicle. Compared with the traditional visual features, using these features is conducive to improving the accuracy of object detection. However, there are few studies on such characteristics. In this article, we systematically study the invariant structural features of remote sensing objects and propose a graph focusing aggregation network (GFA-Net) to represent the structural features of remote sensing objects. Among them, in view of the problem that traditional convolutional neural networks (CNNs) are sensitive to the changes in rotation, scale, and other factors, which makes it difficult to extract structural features, we propose the graph focusing process (GFP) based on the idea of graph convolution. Analysis and experiments show that graph structure has significant advantages over Euclidean feature space under CNN in expressing such structural features. In order to realize the end-to-end efficient training of the above model, we design a graph aggregation network (GAN) to update the weight of nodes. We verify the effectiveness of our method on the proposed multitask datasets aircraft component segmentation dataset (ACSD) and the large-scale Fine-grAined object recognItion in high-Resolution RS imagery (FAIR1M). Experiments conducted on the object detection datasets of large-scale Dataset for Object deTection in Aerial images (DOTA) and HRSC2016 prove that the proposed method is superior to the current state-of-the-art (SOTA) method. Numéro de notice : A2022-560 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1109/TGRS.2022.3181686 Date de publication en ligne : 09/06/2022 En ligne : https://doi.org/10.1109/TGRS.2022.3181686 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=101186
in IEEE Transactions on geoscience and remote sensing > vol 60 n° 6 (June 2022) . - n° 5625217[article]Context-aware network for semantic segmentation toward large-scale point clouds in urban environments / Chun Liu in IEEE Transactions on geoscience and remote sensing, vol 60 n° 6 (June 2022)
[article]
Titre : Context-aware network for semantic segmentation toward large-scale point clouds in urban environments Type de document : Article/Communication Auteurs : Chun Liu, Auteur ; Doudou Zeng, Auteur ; Akram Akbar, Auteur ; et al., Auteur Année de publication : 2022 Article en page(s) : n° 5703915 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Lasergrammétrie
[Termes IGN] agrégation de détails
[Termes IGN] apprentissage profond
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] données lidar
[Termes IGN] données localisées 3D
[Termes IGN] extraction de traits caractéristiques
[Termes IGN] graphe
[Termes IGN] prise en compte du contexte
[Termes IGN] segmentation sémantique
[Termes IGN] semis de points
[Termes IGN] zone urbaineRésumé : (auteur) Point cloud semantic segmentation in urban scenes plays a vital role in intelligent city modeling, autonomous driving, and urban planning. Point cloud semantic segmentation based on deep learning methods has achieved significant improvement. However, it is also challenging for accurate semantic segmentation in large scenes due to complex elements, variety of scene classes, occlusions, and noise. Besides, most methods need to split the original point cloud into multiple blocks before processing and cannot directly deal with the point clouds on a large scale. We propose a novel context-aware network (CAN) that can directly deal with large-scale point clouds. In the proposed network, a local feature aggregation module (LFAM) is designed to preserve rich geometric details in the raw point cloud and reduce the information loss during feature extraction. Then, in combination with a global context aggregation module (GCAM), capture long-range dependencies to enhance the network feature representation and suppress the noise. Finally, a context-aware upsampling module (CAUM) is embedded into the proposed network to capture the global perception from a broad perspective. The ensemble of low-level and high-level features facilitates the effectiveness and efficiency of 3-D point cloud feature refinement. Comprehensive experiments were carried out on three large-scale point cloud datasets in both outdoor and indoor environments to evaluate the performance of the proposed network. The results show that the proposed method outperformed the state-of-the-art representative semantic segmentation networks, and the overall accuracy (OA) of Tongji-3D, Semantic3D, and Stanford large-scale 3-D indoor spaces (S3DIS) is 96.01%, 95.0%, and 88.55%, respectively. Numéro de notice : A2022-561 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1109/TGRS.2022.3182776 Date de publication en ligne : 13/06/2022 En ligne : https://doi.org/10.1109/TGRS.2022.3182776 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=101188
in IEEE Transactions on geoscience and remote sensing > vol 60 n° 6 (June 2022) . - n° 5703915[article]