Descripteur
Termes IGN > sciences naturelles > physique > traitement d'image > compréhension de l'image
compréhension de l'imageVoir aussi |
Documents disponibles dans cette catégorie (44)
Ajouter le résultat dans votre panier
Visionner les documents numériques
Affiner la recherche Interroger des sources externes
Etendre la recherche sur niveau(x) vers le bas
X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data / Danfeng Hong in ISPRS Journal of photogrammetry and remote sensing, vol 167 (September 2020)
[article]
Titre : X-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data Type de document : Article/Communication Auteurs : Danfeng Hong, Auteur ; Naoto Yokoya, Auteur ; Gui-Song Sia, Auteur ; et al., Auteur Année de publication : 2020 Article en page(s) : pp 12 - 23 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image mixte
[Termes IGN] apprentissage profond
[Termes IGN] apprentissage semi-dirigé
[Termes IGN] bruit blanc
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] compréhension de l'image
[Termes IGN] image hyperspectrale
[Termes IGN] image multibande
[Termes IGN] image radar moirée
[Termes IGN] image Sentinel-MSI
[Termes IGN] scène urbaine
[Termes IGN] transmission de donnéesRésumé : (auteur) This paper addresses the problem of semi-supervised transfer learning with limited cross-modality data in remote sensing. A large amount of multi-modal earth observation images, such as multispectral imagery (MSI) or synthetic aperture radar (SAR) data, are openly available on a global scale, enabling parsing global urban scenes through remote sensing imagery. However, their ability in identifying materials (pixel-wise classification) remains limited, due to the noisy collection environment and poor discriminative information as well as limited number of well-annotated training images. To this end, we propose a novel cross-modal deep-learning framework, called X-ModalNet, with three well-designed modules: self-adversarial module, interactive learning module, and label propagation module, by learning to transfer more discriminative information from a small-scale hyperspectral image (HSI) into the classification task using a large-scale MSI or SAR data. Significantly, X-ModalNet generalizes well, owing to propagating labels on an updatable graph constructed by high-level features on the top of the network, yielding semi-supervised cross-modality learning. We evaluate X-ModalNet on two multi-modal remote sensing datasets (HSI-MSI and HSI-SAR) and achieve a significant improvement in comparison with several state-of-the-art methods. Numéro de notice : A2020-544 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2020.06.014 Date de publication en ligne : 11/07/2020 En ligne : https://doi.org/10.1016/j.isprsjprs.2020.06.014 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=95770
in ISPRS Journal of photogrammetry and remote sensing > vol 167 (September 2020) . - pp 12 - 23[article]Réservation
Réserver ce documentExemplaires(3)
Code-barres Cote Support Localisation Section Disponibilité 081-2020091 RAB Revue Centre de documentation En réserve L003 Disponible 081-2020093 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2020092 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Automated terrain feature identification from remote sensing imagery: a deep learning approach / Wenwen Li in International journal of geographical information science IJGIS, vol 34 n° 4 (April 2020)
[article]
Titre : Automated terrain feature identification from remote sensing imagery: a deep learning approach Type de document : Article/Communication Auteurs : Wenwen Li, Auteur ; Chia-Yu Hsu, Auteur Année de publication : 2020 Article en page(s) : pp 637 - 660 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image
[Termes IGN] analyse d'image orientée objet
[Termes IGN] analyse du paysage
[Termes IGN] apprentissage profond
[Termes IGN] base de données d'images
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] compréhension de l'image
[Termes IGN] détection automatique
[Termes IGN] détection d'objet
[Termes IGN] extraction de traits caractéristiques
[Termes IGN] intelligence artificielleRésumé : (auteur) Terrain feature detection is a fundamental task in terrain analysis and landscape scene interpretation. Discovering where a specific feature (i.e. sand dune, crater, etc.) is located and how it evolves over time is essential for understanding landform processes and their impacts on the environment, ecosystem, and human population. Traditional induction-based approaches are challenged by their inefficiency for generalizing diverse and complex terrain features as well as their performance for scalable processing of the massive geospatial data available. This paper presents a new deep learning (DL) approach to support automatic detection of terrain features from remotely sensed images. The novelty of this work lies in: (1) a terrain feature database containing 12,000 remotely sensed images (1,000 original images and 11,000 derived images from data augmentation) that supports data-driven model training and new discovery; (2) a DL-based object detection network empowered by ensemble learning and deep and deeper convolutional neural networks to achieve high-accuracy object detection; and (3) fine-tuning the model’s characteristics and behaviors to identify the best combination of hyperparameters and other network factors. The introduction of DL into geospatial applications is expected to contribute significantly to intelligent terrain analysis, landscape scene interpretation, and the maturation of spatial data science. Numéro de notice : A2020-108 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1080/13658816.2018.1542697 Date de publication en ligne : 07/11/2018 En ligne : https://doi.org/10.1080/13658816.2018.1542697 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=94708
in International journal of geographical information science IJGIS > vol 34 n° 4 (April 2020) . - pp 637 - 660[article]Pyramid scene parsing network in 3D: Improving semantic segmentation of point clouds with multi-scale contextual information / Hao Fang in ISPRS Journal of photogrammetry and remote sensing, vol 154 (August 2019)
[article]
Titre : Pyramid scene parsing network in 3D: Improving semantic segmentation of point clouds with multi-scale contextual information Type de document : Article/Communication Auteurs : Hao Fang, Auteur ; Florent Lafarge, Auteur Année de publication : 2019 Article en page(s) : pp 246 - 258 Note générale : Bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image
[Termes IGN] apprentissage profond
[Termes IGN] compréhension de l'image
[Termes IGN] données localisées 3D
[Termes IGN] prise en compte du contexte
[Termes IGN] représentation multiple
[Termes IGN] scène
[Termes IGN] scène intérieure
[Termes IGN] segmentation sémantique
[Termes IGN] semis de pointsRésumé : (Auteur) Analyzing and extracting geometric features from 3D data is a fundamental step in 3D scene understanding. Recent works demonstrated that deep learning architectures can operate directly on raw point clouds, i.e. without the use of intermediate grid-like structures. These architectures are however not designed to encode contextual information in-between objects efficiently. Inspired by a global feature aggregation algorithm designed for images (Zhao et al., 2017), we propose a 3D pyramid module to enrich pointwise features with multi-scale contextual information. Our module can be easily coupled with 3D semantic segmentation methods operating on 3D point clouds. We evaluated our method on three large scale datasets with four baseline models. Experimental results show that the use of enriched features brings significant improvements to the semantic segmentation of indoor and outdoor scenes. Numéro de notice : A2019-271 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.06.010 Date de publication en ligne : 01/07/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.06.010 Format de la ressource électronique : URL Article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=93089
in ISPRS Journal of photogrammetry and remote sensing > vol 154 (August 2019) . - pp 246 - 258[article]Réservation
Réserver ce documentExemplaires(3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019081 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019083 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019082 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Exploring semantic elements for urban scene recognition: Deep integration of high-resolution imagery and OpenStreetMap (OSM) / Wenzhi Zhao in ISPRS Journal of photogrammetry and remote sensing, vol 151 (May 2019)
[article]
Titre : Exploring semantic elements for urban scene recognition: Deep integration of high-resolution imagery and OpenStreetMap (OSM) Type de document : Article/Communication Auteurs : Wenzhi Zhao, Auteur ; Yanchen Bo, Auteur ; Jiage Chen, Auteur ; et al., Auteur Année de publication : 2019 Article en page(s) : pp 237 - 250 Note générale : Bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] apprentissage profond
[Termes IGN] classe sémantique
[Termes IGN] compréhension de l'image
[Termes IGN] fusion de données
[Termes IGN] image à haute résolution
[Termes IGN] reconnaissance d'objets
[Termes IGN] scène urbaineRésumé : (Auteur) Urban scenes refer to city blocks which are basic units of megacities, they play an important role in citizens’ welfare and city management. Remote sensing imagery with largescale coverage and accurate target descriptions, has been regarded as an ideal solution for monitoring the urban environment. However, due to the heterogeneity of remote sensing images, it is difficult to access their geographical content at the object level, let alone understanding urban scenes at the block level. Recently, deep learning-based strategies have been applied to interpret urban scenes with remarkable accuracies. However, the deep neural networks require a substantial number of training samples which are hard to satisfy, especially for high-resolution images. Meanwhile, the crowed-sourced Open Street Map (OSM) data provides rich annotation information about the urban targets but may encounter the problem of insufficient sampling (limited by the places where people can go). As a result, the combination of OSM and remote sensing images for efficient urban scene recognition is urgently needed. In this paper, we present a novel strategy to transfer existing OSM data to high-resolution images for semantic element determination and urban scene understanding. To be specific, the object-based convolutional neural network (OCNN) can be utilized for geographical object detection by feeding it rich semantic elements derived from OSM data. Then, geographical objects are further delineated into their functional labels by integrating points of interest (POIs), which contain rich semantic terms, such as commercial or educational labels. Lastly, the categories of urban scenes are easily acquired from the semantic objects inside. Experimental results indicate that the proposed method has an ability to classify complex urban scenes. The classification accuracies of the Beijing dataset are as high as 91% at the object-level and 88% at the scene level. Additionally, we are probably the first to investigate the object level semantic mapping by incorporating high-resolution images and OSM data of urban areas. Consequently, the method presented is effective in delineating urban scenes that could further boost urban environment monitoring and planning with high-resolution images. Numéro de notice : A2019-209 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.03.019 Date de publication en ligne : 29/03/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.03.019 Format de la ressource électronique : URL Article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=92675
in ISPRS Journal of photogrammetry and remote sensing > vol 151 (May 2019) . - pp 237 - 250[article]Réservation
Réserver ce documentExemplaires(3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019051 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019053 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019052 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt BIM-PoseNet: Indoor camera localisation using a 3D indoor model and deep learning from synthetic images / Debaditya Acharya in ISPRS Journal of photogrammetry and remote sensing, vol 150 (April 2019)
[article]
Titre : BIM-PoseNet: Indoor camera localisation using a 3D indoor model and deep learning from synthetic images Type de document : Article/Communication Auteurs : Debaditya Acharya, Auteur ; Kourosh Khoshelham, Auteur ; Stephan Winter, Auteur Année de publication : 2019 Article en page(s) : pp 245 - 258 Note générale : Bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Applications photogrammétriques
[Termes IGN] apprentissage profond
[Termes IGN] compréhension de l'image
[Termes IGN] estimation de pose
[Termes IGN] image de synthèse
[Termes IGN] modélisation 3D du bâti BIM
[Termes IGN] positionnement en intérieur
[Termes IGN] réseau neuronal convolutif
[Termes IGN] structure-from-motionRésumé : (Auteur) The ubiquity of cameras built in mobile devices has resulted in a renewed interest in image-based localisation in indoor environments where the global navigation satellite system (GNSS) signals are not available. Existing approaches for indoor localisation using images either require an initial location or need first to perform a 3D reconstruction of the whole environment using structure-from-motion (SfM) methods, which is challenging and time-consuming for large indoor spaces. In this paper, a visual localisation approach is proposed to eliminate the requirement of image-based reconstruction of the indoor environment by using a 3D indoor model. A deep convolutional neural network (DCNN) is fine-tuned using synthetic images obtained from the 3D indoor model to regress the camera pose. Results of the experiments indicate that the proposed approach can be used for indoor localisation in real-time with an accuracy of approximately 2 m. Numéro de notice : A2019-142 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.02.020 Date de publication en ligne : 05/03/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.02.020 Format de la ressource électronique : URL Article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=92480
in ISPRS Journal of photogrammetry and remote sensing > vol 150 (April 2019) . - pp 245 - 258[article]Réservation
Réserver ce documentExemplaires(3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019041 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019043 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019042 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Patch-based detection of dynamic objects in CrowdCam images / Gagan Kanojia in The Visual Computer, vol 35 n° 4 (April 2019)PermalinkVehicle detection in aerial images / Michael Ying Yang in Photogrammetric Engineering & Remote Sensing, PERS, vol 85 n° 4 (avril 2019)PermalinkSemantic understanding of scenes through the ADE20K dataset / Bolei Zhou in International journal of computer vision, vol 127 n° 3 (March 2019)PermalinkComplete 3D scene parsing from an RGBD image / Chuhang Zou in International journal of computer vision, vol 127 n° 2 (February 2019)PermalinkDétection et localisation d'objets 3D par apprentissage profond en topologie capteur / Pierre Biasutti (2019)PermalinkEstimation de profondeur à partir d'images monoculaires par apprentissage profond / Michel Moukari (2019)PermalinkMultimodal scene understanding: algorithms, applications and deep learning, ch. 8. Multimodal localization for embedded systems: a survey / Imane Salhi (2019)PermalinkThe necessary yet complex evaluation of 3D city models: a semantic approach / Oussama Ennafii (2019)PermalinkTowards visual urban scene understanding for autonomous vehicle path tracking using GPS positioning data / Citlalli Gamez Serna (2019)PermalinkConfigurable 3D scene synthesis and 2D image rendering with per-pixel ground truth using stochastic grammars / Chenfanfu Jiang in International journal of computer vision, vol 126 n° 9 (September 2018)Permalink