Descripteur
Documents disponibles dans cette catégorie (763)
Ajouter le résultat dans votre panier
Visionner les documents numériques
Affiner la recherche Interroger des sources externes
Etendre la recherche sur niveau(x) vers le bas
Combining Sentinel-1 and Sentinel-2 Satellite image time series for land cover mapping via a multi-source deep learning architecture / Dino Lenco in ISPRS Journal of photogrammetry and remote sensing, Vol 158 (December 2019)
[article]
Titre : Combining Sentinel-1 and Sentinel-2 Satellite image time series for land cover mapping via a multi-source deep learning architecture Type de document : Article/Communication Auteurs : Dino Lenco, Auteur ; Roberto Interdonato, Auteur ; Raffaele Gaetano, Auteur ; Ho Tong Minh Dinh, Auteur Année de publication : 2019 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image mixte
[Termes IGN] apprentissage profond
[Termes IGN] Burkina Faso
[Termes IGN] carte de la végétation
[Termes IGN] classification par forêts d'arbres décisionnels
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] fusion d'images
[Termes IGN] image à haute résolution
[Termes IGN] image multibande
[Termes IGN] image radar moirée
[Termes IGN] image Sentinel-MSI
[Termes IGN] image Sentinel-SAR
[Termes IGN] occupation du sol
[Termes IGN] Réunion, île de la
[Termes IGN] série temporelle
[Termes IGN] utilisation du solRésumé : (auteur) The huge amount of data currently produced by modern Earth Observation (EO) missions has allowed for the design of advanced machine learning techniques able to support complex Land Use/Land Cover (LULC) mapping tasks. The Copernicus programme developed by the European Space Agency provides, with missions such as Sentinel-1 (S1) and Sentinel-2 (S2), radar and optical (multi-spectral) imagery, respectively, at 10 m spatial resolution with revisit time around 5 days. Such high temporal resolution allows to collect Satellite Image Time Series (SITS) that support a plethora of Earth surface monitoring tasks. How to effectively combine the complementary information provided by such sensors remains an open problem in the remote sensing field. In this work, we propose a deep learning architecture to combine information coming from S1 and S2 time series, namely TWINNS (TWIn Neural Networks for Sentinel data), able to discover spatial and temporal dependencies in both types of SITS. The proposed architecture is devised to boost the land cover classification task by leveraging two levels of complementarity, i.e., the interplay between radar and optical SITS as well as the synergy between spatial and temporal dependencies. Experiments carried out on two study sites characterized by different land cover characteristics (i.e., the Koumbia site in Burkina Faso and Reunion Island, a overseas department of France in the Indian Ocean), demonstrate the significance of our proposal. Numéro de notice : A2019-544 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.09.016 Date de publication en ligne : 27/09/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.09.016 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=94186
in ISPRS Journal of photogrammetry and remote sensing > Vol 158 (December 2019)[article]Réservation
Réserver ce documentExemplaires (3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019121 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019123 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019122 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Deep learning for conifer/deciduous classification of airborne LiDAR 3D point clouds representing individual trees / Hamid Hamraz in ISPRS Journal of photogrammetry and remote sensing, Vol 158 (December 2019)
[article]
Titre : Deep learning for conifer/deciduous classification of airborne LiDAR 3D point clouds representing individual trees Type de document : Article/Communication Auteurs : Hamid Hamraz, Auteur ; Nathan B. Jacobs, Auteur ; Marco A. Contreras, Auteur ; Chase H. Clark, Auteur Année de publication : 2019 Article en page(s) : pp 219 - 230 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Lasergrammétrie
[Termes IGN] apprentissage profond
[Termes IGN] arbre caducifolié
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] données d'entrainement (apprentissage automatique)
[Termes IGN] données lidar
[Termes IGN] données localisées 3D
[Termes IGN] houppier
[Termes IGN] modèle numérique de surface
[Termes IGN] Pinophyta
[Termes IGN] semis de pointsRésumé : (auteur) The purpose of this study was to investigate the use of deep learning for coniferous/deciduous classification of individual trees segmented from airborne LiDAR data. To enable processing by a deep convolutional neural network (CNN), we designed two discrete representations using leaf-off and leaf-on LiDAR data: a digital surface model with four channels (DSM × 4) and a set of four 2D views (4 × 2D). A training dataset of tree crowns was generated via segmentation of tree crowns, followed by co-registration with field data. Potential mislabels due to GPS error or tree leaning were corrected using a statistical ensemble filtering procedure. Because the training data was heavily unbalanced (~8% conifers), we trained an ensemble of CNNs on random balanced sub-samples. Benchmarked against multiple traditional shallow learning methods using manually designed features, the CNNs improved accuracies up to 14%. The 4 × 2D representation yielded similar classification accuracies to the DSM × 4 representation (~82% coniferous and ~90% deciduous) while converging faster. Further experimentation showed that early/late fusion of the channels in the representations did not affect the accuracies in a significant way. The data augmentation that was used for the CNN training improved the classification accuracies, but more real training instances (especially coniferous) likely results in much stronger improvements. Leaf-off LiDAR data were the primary source of useful information, which is likely due to the perennial nature of coniferous foliage. LiDAR intensity values also proved to be useful, but normalization yielded no significant improvement. As we observed, large training data may compensate for the lack of a subset of important domain data. Lastly, the classification accuracies of overstory trees (~90%) were more balanced than those of understory trees (~90% deciduous and ~65% coniferous), which is likely due to the incomplete capture of understory tree crowns via airborne LiDAR. In domains like remote sensing and biomedical imaging, where the data contain a large amount of information and are not friendly to human visual system, human-designed features may become suboptimal. As exemplified by this study, automatic, objective derivation of optimal features via deep learning can improve prediction tasks in such domains. Numéro de notice : A2019-547 Affiliation des auteurs : non IGN Thématique : FORET/IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.10.011 Date de publication en ligne : 03/11/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.10.011 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=94192
in ISPRS Journal of photogrammetry and remote sensing > Vol 158 (December 2019) . - pp 219 - 230[article]Réservation
Réserver ce documentExemplaires (3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019121 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019123 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019122 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Half a percent of labels is enough: efficient animal detection in UAV imagery using deep CNNs and active learning / Benjamin Kellenberger in IEEE Transactions on geoscience and remote sensing, vol 57 n° 12 (December 2019)
[article]
Titre : Half a percent of labels is enough: efficient animal detection in UAV imagery using deep CNNs and active learning Type de document : Article/Communication Auteurs : Benjamin Kellenberger, Auteur ; Diego Marcos, Auteur ; Sylvain Lobry, Auteur ; Devis Tuia, Auteur Année de publication : 2019 Article en page(s) : pp 9524 - 9533 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] analyse d'image orientée objet
[Termes IGN] apprentissage profond
[Termes IGN] classification orientée objet
[Termes IGN] classification par réseau neuronal
[Termes IGN] détection d'objet
[Termes IGN] données localisées
[Termes IGN] échantillonnage de données
[Termes IGN] faune locale
[Termes IGN] image captée par drone
[Termes IGN] Namibie
[Termes IGN] objet mobile
[Termes IGN] réalité de terrain
[Termes IGN] recensementRésumé : (auteur) We present an Active Learning (AL) strategy for reusing a deep Convolutional Neural Network (CNN)-based object detector on a new data set. This is of particular interest for wildlife conservation: given a set of images acquired with an Unmanned Aerial Vehicle (UAV) and manually labeled ground truth, our goal is to train an animal detector that can be reused for repeated acquisitions, e.g., in follow-up years. Domain shifts between data sets typically prevent such a direct model application. We thus propose to bridge this gap using AL and introduce a new criterion called Transfer Sampling (TS). TS uses Optimal Transport (OT) to find corresponding regions between the source and the target data sets in the space of CNN activations. The CNN scores in the source data set are used to rank the samples according to their likelihood of being animals, and this ranking is transferred to the target data set. Unlike conventional AL criteria that exploit model uncertainty, TS focuses on very confident samples, thus allowing quick retrieval of true positives in the target data set, where positives are typically extremely rare and difficult to find by visual inspection. We extend TS with a new window cropping strategy that further accelerates sample retrieval. Our experiments show that with both strategies combined, less than half a percent of oracle-provided labels are enough to find almost 80% of the animals in challenging sets of UAV images, beating all baselines by a margin. Numéro de notice : A2019-598 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1109/TGRS.2019.2927393 Date de publication en ligne : 20/08/2019 En ligne : http://doi.org/10.1109/TGRS.2019.2927393 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=94592
in IEEE Transactions on geoscience and remote sensing > vol 57 n° 12 (December 2019) . - pp 9524 - 9533[article]Matching of TerraSAR-X derived ground control points to optical image patches using deep learning / Tatjana Bürgmann in ISPRS Journal of photogrammetry and remote sensing, Vol 158 (December 2019)
[article]
Titre : Matching of TerraSAR-X derived ground control points to optical image patches using deep learning Type de document : Article/Communication Auteurs : Tatjana Bürgmann, Auteur ; Wolfgang Koppe, Auteur ; Michael Schmitt, Auteur Année de publication : 2019 Article en page(s) : pp 241 - 248 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image mixte
[Termes IGN] appariement d'images
[Termes IGN] apprentissage profond
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] géolocalisation
[Termes IGN] image multicapteur
[Termes IGN] image optique
[Termes IGN] image Pléiades
[Termes IGN] image radar moirée
[Termes IGN] image Sentinel-MSI
[Termes IGN] image Sentinel-SAR
[Termes IGN] image TerraSAR-X
[Termes IGN] point d'appuiRésumé : (auteur) High resolution synthetic aperture radar (SAR) satellites like TerraSAR-X are capable of acquiring images exhibiting an absolute geolocation accuracy within a few centimeters, mainly because of the availability of precise orbit information and by compensating range delay errors due to atmospheric conditions. In contrast, satellite images from optical missions generally exhibit comparably low geolocation accuracies because of the propagation of errors in angular measurements over large distances. However, a variety of remote sensing applications, such as change detection, surface movement monitoring or ice flow measurements, require precisely geo-referenced and co-registered satellite images. By using Ground Control Points (GCPs) derived from TerraSAR-X, the absolute geolocation accuracy of optical satellite images can be improved. For this purpose, the corresponding matching points in the optical images need to be localized. In this paper, a deep learning based approach is investigated for an automated matching of SAR-derived GCPs to optical image elements. Therefore, a convolutional neural network is pretrained with medium resolution Sentinel-1 and Sentinel-2 imagery and fine-tuned on precisely co-registered TerraSAR-X and Pléiades training image pairs to learn a common descriptor representation. By using these descriptors, the similarity of SAR and optical image patches can be calculated. This similarity metric is then used in a sliding window approach to identify the matching points in the optical reference image. Subsequently, the derived points can be utilized for co-registration of the underlying images. The network is evaluated over nine study areas showing airports and their rural surroundings from several different countries around the world. The results show that based on TerraSAR-X-derived GCPs, corresponding points in the optical image can automatically and reliably be identified with a pixel-level localization accuracy. Numéro de notice : A2019-548 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.09.010 Date de publication en ligne : 05/11/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.09.010 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=94194
in ISPRS Journal of photogrammetry and remote sensing > Vol 158 (December 2019) . - pp 241 - 248[article]Réservation
Réserver ce documentExemplaires (3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019121 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019123 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019122 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Comparison between convolutional neural networks and random forest for local climate zone classification in mega urban areas using Landsat images / Cheolhee Yoo in ISPRS Journal of photogrammetry and remote sensing, vol 157 (November 2019)
[article]
Titre : Comparison between convolutional neural networks and random forest for local climate zone classification in mega urban areas using Landsat images Type de document : Article/Communication Auteurs : Cheolhee Yoo, Auteur ; Daehyeon Han, Auteur ; Jungho Im, Auteur ; Benjamin Bechtel, Auteur Année de publication : 2019 Article en page(s) : pp 155 - 170 Note générale : Bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] analyse comparative
[Termes IGN] apprentissage automatique
[Termes IGN] apprentissage profond
[Termes IGN] Chicago (Illinois)
[Termes IGN] classification par forêts d'arbres décisionnels
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] climat urbain
[Termes IGN] Hong-Kong
[Termes IGN] ilot thermique urbain
[Termes IGN] image Landsat-8
[Termes IGN] Madrid (Espagne)
[Termes IGN] Rome
[Termes IGN] World Urban Database and Access Portal Tools
[Termes IGN] zone urbaine denseRésumé : (Auteur) The Local Climate Zone (LCZ) scheme is a classification system providing a standardization framework to present the characteristics of urban forms and functions, especially for urban heat island (UHI) research. Landsat-based 100 m resolution LCZ maps have been classified by the World Urban Database and Portal Tool (WUDAPT) method using a random forest (RF) machine learning classifier. Some studies have proposed modified RF and convolutional neural network (CNN) approaches. This study aims to compare CNN with an RF classifier for LCZ mapping in great detail. We designed five schemes (three RF-based schemes (S1–S3) and two CNN-based ones (S4–S5)), which consist of various combinations of input features from bitemporal Landsat 8 data over four global mega cities: Rome, Hong Kong, Madrid, and Chicago. Among the five schemes, the CNN-based one with the incorporation of a larger neighborhood information showed the best classification performance. When compared to the WUDAPT workflow, the overall accuracies for entire land cover classes (OA) and for urban LCZ types (i.e., LCZ1-10; OAurb) increased by about 6–8% and 10–13%, respectively, for the four cities. The transferability of LCZ models for the four cities were evaluated, showing that CNN consistently resulted in higher accuracy (increased by about 7–18% and 18–29% for OA and OAurb, respectively) than RF. This study revealed that the CNN classifier classified particularly well for the specific LCZ classes in which buildings were mixed with trees or buildings or plants were sparsely distributed. The research findings can provide a basis for guidance of future LCZ classification using deep learning. Numéro de notice : A2019-495 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1016/j.isprsjprs.2019.09.009 Date de publication en ligne : 19/09/2019 En ligne : https://doi.org/10.1016/j.isprsjprs.2019.09.009 Format de la ressource électronique : URL Article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=93728
in ISPRS Journal of photogrammetry and remote sensing > vol 157 (November 2019) . - pp 155 - 170[article]Réservation
Réserver ce documentExemplaires (3)
Code-barres Cote Support Localisation Section Disponibilité 081-2019111 RAB Revue Centre de documentation En réserve L003 Disponible 081-2019113 DEP-RECP Revue LASTIG Dépôt en unité Exclu du prêt 081-2019112 DEP-RECF Revue Nancy Dépôt en unité Exclu du prêt Context pyramidal network for stereo matching regularized by disparity gradients / Junhua Kang in ISPRS Journal of photogrammetry and remote sensing, vol 157 (November 2019)PermalinkDeep learning for multi-modal classification of cloud, shadow and land cover scenes in PlanetScope and Sentinel-2 imagery / Yuri Shendryk in ISPRS Journal of photogrammetry and remote sensing, vol 157 (November 2019)PermalinkA double-strategy-check active learning algorithm for hyperspectral image classification / Ying Cui in Photogrammetric Engineering & Remote Sensing, PERS, vol 85 n° 11 (November 2019)PermalinkSig-NMS-based faster R-CNN combining transfer learning for small target detection in VHR optical remote sensing imagery / Ruchan Dong in IEEE Transactions on geoscience and remote sensing, vol 57 n° 11 (November 2019)PermalinkCombining machine learning and compact polarimetry for estimating soil moisture from C-Band SAR data / Emanuele Santi in Remote sensing, Vol 11 n° 20 (October-2 2019)PermalinkComparative analysis of the accuracy of surface soil moisture estimation from the C- and L-bands / Mohammad El Hajj in International journal of applied Earth observation and geoinformation, vol 82 (October 2019)PermalinkA machine learning approach to detect crude oil contamination in a real scenario using hyperspectral remote sensing / Ran Pelta in International journal of applied Earth observation and geoinformation, vol 82 (October 2019)PermalinkMapping dead forest cover using a deep convolutional neural network and digital aerial photography / Jean-Daniel Sylvain in ISPRS Journal of photogrammetry and remote sensing, vol 156 (October 2019)PermalinkMulti-sensor prediction of Eucalyptus stand volume: A support vector approach / Guilherme Silverio Aquino de Souza in ISPRS Journal of photogrammetry and remote sensing, vol 156 (October 2019)PermalinkOptimal segmentation of high spatial resolution images for the classification of buildings using random forests / James Bialas in International journal of applied Earth observation and geoinformation, vol 82 (October 2019)Permalink