Descripteur
Termes IGN > sciences naturelles > physique > traitement d'image > compréhension de l'image
compréhension de l'imageVoir aussi |
Documents disponibles dans cette catégorie (27)
Ajouter le résultat dans votre panier
Visionner les documents numériques
Affiner la recherche Interroger des sources externes
Etendre la recherche sur niveau(x) vers le bas
Assessing the cognition of movement trajectory visualizations: interpreting speed and direction / Crystal J. Bae in Cartography and Geographic Information Science, Vol 50 n° 2 (March 2023)
[article]
Titre : Assessing the cognition of movement trajectory visualizations: interpreting speed and direction Type de document : Article/Communication Auteurs : Crystal J. Bae, Auteur ; Somayeh Dodge, Auteur Année de publication : 2023 Article en page(s) : pp 143 - 161 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Termes IGN] analyse visuelle
[Termes IGN] cognition
[Termes IGN] compréhension de l'image
[Termes IGN] données spatiotemporelles
[Termes IGN] objet mobile
[Termes IGN] visualisation cartographique
[Vedettes matières IGN] GéovisualisationRésumé : (auteur) This paper evaluates cognitively plausible geovisualization techniques for mapping movement data. With the widespread increase in the availability and quality of space-time data capturing movement trajectories of individuals, meaningful representations are needed to properly visualize and communicate trajectory data and complex movement patterns using geographic displays. Many visualization and visual analytics approaches have been proposed to map movement trajectories (e.g. space-time paths, animations, trajectory lines, etc.). However, little is known about how effective these complex visualizations are in capturing important aspects of movement data. Given the complexity of movement data which involves space, time, and context dimensions, it is essential to evaluate the communicative efficiency and efficacy of various visualization forms in helping people understand movement data. This study assesses the effectiveness of static and dynamic movement displays as well as visual variables in communicating movement parameters along trajectories, such as speed and direction. To do so, a web-based survey is conducted to evaluate the understanding of movement visualizations by a nonspecialist audience. This and future studies contribute fundamental insights into the cognition of movement visualizations and inspire new methods for the empirical evaluation of geovisualizations. Numéro de notice : A2023-221 Affiliation des auteurs : non IGN Thématique : GEOMATIQUE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1080/15230406.2022.2157879 Date de publication en ligne : 23/01/2023 En ligne : https://doi.org/10.1080/15230406.2022.2157879 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=103167
in Cartography and Geographic Information Science > Vol 50 n° 2 (March 2023) . - pp 143 - 161[article]Characteristics of augmented map research from a cartographic perspective / Yi Cheng in Cartography and Geographic Information Science, Vol 49 n° 5 (September 2022)
[article]
Titre : Characteristics of augmented map research from a cartographic perspective Type de document : Article/Communication Auteurs : Yi Cheng, Auteur ; Guochuang Zhu, Auteur ; Cong Yang, Auteur ; et al., Auteur Année de publication : 2022 Article en page(s) : pp 426 - 442 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Termes IGN] carte interactive
[Termes IGN] cognition
[Termes IGN] compréhension de l'image
[Termes IGN] lecture de carte
[Termes IGN] réalité augmentée
[Termes IGN] représentation cartographique 3D
[Termes IGN] visualisation cartographique
[Vedettes matières IGN] CartologieRésumé : (auteur) “Why,” “what,” and “how” are basic questions to be answered in augmented map research as an intersecting direction. This study summarizes dynamic visual representations and true 3D interactions as characteristics of augmented maps from the cartographic perspective through analysis of the research progress in different disciplines. From secondary viewpoints of cognition and design, the augmented map cube is presented to establish the research framework in three directions: cognitive purposes, information dimensions, and interactive devices, in which map-based spatial cognition theory, augmented visualization, and interactive features are considered. The research evaluation is carried out to determine the reasonableness of the cube and then identify different research statuses in any one or two of the directions under the cube. Based on a literature search and classification, 30 typical studies were used for structural analysis to discover research trends and new directions that can be mined. The results show that the cube can be used to evaluate the coverage of an article or provide researchers with research trends and new possibilities. Our conclusions include but are not limited to the following: Research for retrieval purposes deserves attention, augmented visualization of specific individual elements is key to understanding maps, and interactive devices become more intangible. Numéro de notice : A2022-634 Affiliation des auteurs : non IGN Thématique : GEOMATIQUE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.1080/15230406.2022.2059571 Date de publication en ligne : 22/04/2022 En ligne : https://doi.org/10.1080/15230406.2022.2059571 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=101401
in Cartography and Geographic Information Science > Vol 49 n° 5 (September 2022) . - pp 426 - 442[article]
Titre : Scene understanding and gesture recognition for human-machine interaction Type de document : Thèse/HDR Auteurs : Naina Dhingra, Auteur Editeur : Zurich : Eidgenossische Technische Hochschule ETH - Ecole Polytechnique Fédérale de Zurich EPFZ Année de publication : 2022 Note générale : Bibliographie
A dissertation submitted to attain the degree of Doctor of Sciences of ETH ZurichLangues : Français (fre) Descripteur : [Vedettes matières IGN] Intelligence artificielle
[Termes IGN] apprentissage profond
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] classification orientée objet
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] classification par séparateurs à vaste marge
[Termes IGN] compréhension de l'image
[Termes IGN] image RVB
[Termes IGN] interaction homme-machine
[Termes IGN] oculométrie
[Termes IGN] reconnaissance automatique
[Termes IGN] reconnaissance de formes
[Termes IGN] reconnaissance de gestes
[Termes IGN] réseau neuronal récurrent
[Termes IGN] scène
[Termes IGN] vision par ordinateurRésumé : (auteur) Scene understanding and gesture recognition are useful for a myriad of applications such as human-robotic interaction, assisting blind and visually impaired people, advanced driver assistance systems, and autonomous driving. To work autonomously in real-world environments, automatic systems need to deliver non-verbal information to enhance the verbal communication in particular for blind people. We are exploring the holistic approach for providing the scene as well as gesture related information. We propose that incorporating attention mechanisms in neural networks which behave similarly to attention in the human brain, and conducting an integrated study using neural networks in real-time can yield significant improvements in the scene and gesture understanding, thereby enhancing the user experience. In this thesis, we investigate the understanding of visual scenes and gestures. We explore these two areas, in particular, by proposing novel architectures, training methods, user studies, and thorough evaluations. We show that, for deep learning approaches, attention or self attention mechanisms improve and push the boundaries of network performance for different tasks in consideration. We suggest that the various kinds of gestures can complement and supplement each other’s information to better understand non-verbal conversation; hence integrated gestures comprehension is useful. First, we focus on visual scene understanding using scene graph generation. We propose, BGT-Net, a new network that uses an object detection model with 1) bidirectional gated recurrent units for object-object communication and 2) transformer encoders including self attention to classify the objects and their relationships. We address the problem of bias caused by the long tailed distribution in the dataset. This enables the network to perform even for the unseen objects or relationships in the dataset. Second, we propose to learn hand gesture recognition from RGB and RGB-D videos using attention learning. We present a novel architecture based on residual connections and an attention mechanism. Our approach successfully detects hand gestures when evaluated on three open-source datasets. Third, we explore pointing gesture recognition and localization using open-source software, i.e. OpenPtrack which uses a deep learning based iii network to track multi-persons in the scene. We use a Kinect sensor as an input device and conduct a user study with 26 users to evaluate the system using two setup types. Fourth, we propose a technique to perform eye gaze tracking using OpenFace which is based on a deep learning model and RGB webcam. We use support vector machine regression to estimate the position of eye gaze on the screen. In a study, we evaluate the system with 28 users and show that this system can perform similarly to commercially expensive eye trackers. Finally, we focus on 3D head pose estimation using two models: 1)headPosr includes residual connections for the base network followed by a transformer encoder. It outperforms existing models but has a drawback of being computationally expensive; 2) lwPosr uses depthwise separable convolutions and transformer encoders. It is a two stream network in fine-grained fashion to estimate the three angles of the head pose. We demonstrate that this method is able to predict head poses better than state-of-the-art lightweight networks. Note de contenu : 1- Introduction
2- Background
3- State of the art
4- Scene graph generation
5- 3D hand gesture recognition
6- Pointing gesture recognition
7- Eye-gaze tracking
8- Head pose estimation
9- Lightweight head pose estimation
10- SummaryNuméro de notice : 24039 Affiliation des auteurs : non IGN Thématique : IMAGERIE/INFORMATIQUE Nature : Thèse étrangère Note de thèse : PhD Thesis : Sciences : ETH Zurich :2022 DOI : sans En ligne : https://www.research-collection.ethz.ch/handle/20.500.11850/559347 Format de la ressource électronique : URL Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=101876 Evaluating narrative in geoportals for territorial public policies / Luis Manuel Batista in Cartographica, vol 56 n° 4 (Winter 2021)
[article]
Titre : Evaluating narrative in geoportals for territorial public policies Type de document : Article/Communication Auteurs : Luis Manuel Batista, Auteur ; Ana Figueiras, Auteur Année de publication : 2021 Article en page(s) : pp 303 - 319 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Géomatique web
[Termes IGN] analyse comparative
[Termes IGN] carte thématique
[Termes IGN] collectivité territoriale
[Termes IGN] communication cartographique
[Termes IGN] compréhension de l'image
[Termes IGN] conception cartographique
[Termes IGN] géoportail
[Termes IGN] plan local d'urbanisme
[Termes IGN] politique territoriale
[Termes IGN] Portugal
[Termes IGN] web mappingRésumé : (auteur) To do territorial planning, we need several maps referring to different layers of information necessary to represent the territory according to a vast set of variables. At the Portuguese municipal level, the municipal master plan (PDM – plano diretor municipal) is the territorial management tool responsible for long-term territorial planning and ordering. Since the PDM will constrain the citizens’ lives, it should be of easy access and interpretation. However, due to its large amount of information, it is often hard for them to understand what is presented using only static cartographic elements. Comparing 60 Web sites that present geographical information for a wider public, we found that narrative, visual, and interactive elements and data presentation in flexible portions with the necessary information for partitioning, browsing, or querying made the data more engaging. This flexibility is particularly important with large data sets, where decreasing the level of complexity is vital for proper understanding and analysis. We found that geoportals dedicated to territorial public policies lack narrative elements that would improve comprehension. Since it is difficult for the general public to understand the strategy the municipality wants to implement, the desired public participation, although possible, will return poor results. Numéro de notice : A2021-887 Affiliation des auteurs : non IGN Thématique : GEOMATIQUE Nature : Article nature-HAL : ArtAvecCL-RevueIntern DOI : 10.3138/cart-2021-0023 Date de publication en ligne : 02/12/2021 En ligne : https://doi.org/10.3138/cart-2021-0023 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=99227
in Cartographica > vol 56 n° 4 (Winter 2021) . - pp 303 - 319[article]Exemplaires(1)
Code-barres Cote Support Localisation Section Disponibilité 031-2021041 SL Revue Centre de documentation Revues en salle Disponible Unsupervised self-adaptive deep learning classification network based on the optic nerve microsaccade mechanism for unmanned aerial vehicle remote sensing image classification / Ming Cong in Geocarto international, vol 36 n° 18 ([01/10/2021])
[article]
Titre : Unsupervised self-adaptive deep learning classification network based on the optic nerve microsaccade mechanism for unmanned aerial vehicle remote sensing image classification Type de document : Article/Communication Auteurs : Ming Cong, Auteur ; Zhiye Wang, Auteur ; Yiting Tao, Auteur ; et al., Auteur Année de publication : 2021 Article en page(s) : pp 2065 - 2084 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] analyse de groupement
[Termes IGN] chromatopsie
[Termes IGN] classification non dirigée
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] compréhension de l'image
[Termes IGN] échantillonnage d'image
[Termes IGN] filtrage numérique d'image
[Termes IGN] image captée par drone
[Termes IGN] vision
[Termes IGN] vision par ordinateurRésumé : (auteur) Unmanned aerial vehicle remote sensing images need to be precisely and efficiently classified. However, complex ground scenes produced by ultra-high ground resolution, data uniqueness caused by multi-perspective observations, and need for manual labelling make it difficult for current popular deep learning networks to obtain reliable references from heterogeneous samples. To address these problems, this paper proposes an optic nerve microsaccade (ONMS) classification network, developed based on multiple dilated convolution. ONMS first applies a Laplacian of Gaussian filter to find typical features of ground objects and establishes class labels using adaptive clustering. Then, using an image pyramid, multi-scale image data are mapped to the class labels adaptively to generate homologous reliable samples. Finally, an end-to-end multi-scale neural network is applied for classification. Experimental results show that ONMS significantly reduces sample labelling costs while retaining high cognitive performance, classification accuracy, and noise resistance—indicating that it has significant application advantages. Numéro de notice : A2021-707 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article DOI : 10.1080/10106049.2019.1687593 Date de publication en ligne : 07/11/2019 En ligne : https://doi.org/10.1080/10106049.2019.1687593 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=98602
in Geocarto international > vol 36 n° 18 [01/10/2021] . - pp 2065 - 2084[article]Semantic hierarchy emerges in deep generative representations for scene synthesis / Ceyuan Yang in International journal of computer vision, vol 129 n° 5 (May 2021)PermalinkRecognition of varying size scene images using semantic analysis of deep activation maps / Shikha Gupta in Machine Vision and Applications, vol 32 n° 2 (March 2021)PermalinkSpatial multi-criteria evaluation in 3D context: suitability analysis of urban vertical development / Kendra Munn in Cartography and Geographic Information Science, vol 48 n° 2 (March 2021)PermalinkActivity recognition in residential spaces with Internet of things devices and thermal imaging / Kshirasagar Naik in Sensors, vol 21 n° 3 (February 2021)PermalinkDeep convolutional neural networks for scene understanding and motion planning for self-driving vehicles / Abdelhak Loukkal (2021)PermalinkPermalinkPermalinkFrom point clouds to high-fidelity models - advanced methods for image-based 3D reconstruction / Audrey Richard (2021)PermalinkPermalinkPermalinkX-ModalNet: A semi-supervised deep cross-modal network for classification of remote sensing data / Danfeng Hong in ISPRS Journal of photogrammetry and remote sensing, vol 167 (September 2020)PermalinkAutomated terrain feature identification from remote sensing imagery: a deep learning approach / Wenwen Li in International journal of geographical information science IJGIS, vol 34 n° 4 (April 2020)PermalinkPyramid scene parsing network in 3D: Improving semantic segmentation of point clouds with multi-scale contextual information / Hao Fang in ISPRS Journal of photogrammetry and remote sensing, vol 154 (August 2019)PermalinkExploring semantic elements for urban scene recognition: Deep integration of high-resolution imagery and OpenStreetMap (OSM) / Wenzhi Zhao in ISPRS Journal of photogrammetry and remote sensing, vol 151 (May 2019)PermalinkBIM-PoseNet: Indoor camera localisation using a 3D indoor model and deep learning from synthetic images / Debaditya Acharya in ISPRS Journal of photogrammetry and remote sensing, vol 150 (April 2019)PermalinkPatch-based detection of dynamic objects in CrowdCam images / Gagan Kanojia in The Visual Computer, vol 35 n° 4 (April 2019)PermalinkVehicle detection in aerial images / Michael Ying Yang in Photogrammetric Engineering & Remote Sensing, PERS, vol 85 n° 4 (avril 2019)PermalinkSemantic understanding of scenes through the ADE20K dataset / Bolei Zhou in International journal of computer vision, vol 127 n° 3 (March 2019)PermalinkComplete 3D scene parsing from an RGBD image / Chuhang Zou in International journal of computer vision, vol 127 n° 2 (February 2019)PermalinkDétection et localisation d'objets 3D par apprentissage profond en topologie capteur / Pierre Biasutti (2019)PermalinkEstimation de profondeur à partir d'images monoculaires par apprentissage profond / Michel Moukari (2019)PermalinkMultimodal scene understanding: algorithms, applications and deep learning, ch. 8. Multimodal localization for embedded systems: a survey / Imane Salhi (2019)Permalink