Descripteur
Termes IGN > 1-Candidats > attention (apprentissage automatique)
attention (apprentissage automatique) |
Documents disponibles dans cette catégorie (45)
Ajouter le résultat dans votre panier Affiner la recherche Interroger des sources externes
Etendre la recherche sur niveau(x) vers le bas
A graph attention network for road marking classification from mobile LiDAR point clouds / Lina Fang in International journal of applied Earth observation and geoinformation, vol 108 (April 2022)
[article]
Titre : A graph attention network for road marking classification from mobile LiDAR point clouds Type de document : Article/Communication Auteurs : Lina Fang, Auteur ; Tongtong Sun, Auteur ; Shuang Wang, Auteur ; et al., Auteur Année de publication : 2022 Article en page(s) : n° 102735 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Lasergrammétrie
[Termes IGN] apprentissage profond
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] classification par Perceptron multicouche
[Termes IGN] données lidar
[Termes IGN] données localisées 3D
[Termes IGN] noeud
[Termes IGN] réseau neuronal de graphes
[Termes IGN] réseau routier
[Termes IGN] semis de points
[Termes IGN] signalisation routièreRésumé : (auteur) The category of road marking is a crucial element in Mobile laser scanning systems’ (MLSs) applications such as intelligent traffic systems, high-definition maps, location and navigation services. Due to the complexity of road scenes, considerable and various categories, occlusion and uneven intensities in MLS point clouds, finely road marking classification is considered as the challenging work. This paper proposes a graph attention network named GAT_SCNet to simultaneously group the road markings into 11 categories from MLS point clouds. Concretely, the proposed GAT_SCNet model constructs serial computable subgraphs and fulfills a multi-head attention mechanism to encode the geometric, topological, and spatial relationships between the node and neighbors to generate the distinguishable descriptor of road marking. To assess the effectiveness and generalization of the GAT_SCNet model, we conduct extensive experiments on five test datasets of about 100 km in total captured by different MLS systems. Three accuracy evaluation metrics: average Precision, Recall, and of 11 categories on the test datasets exceed 91%, respectively. Accuracy evaluations and comparative studies show that our method has achieved a new state-of-the-art work on road marking classification, especially on similar linear road markings like stop lines, zebra crossings, and dotted lines. Numéro de notice : A2022-234 Affiliation des auteurs : non IGN Thématique : IMAGERIE/INFORMATIQUE Nature : Article DOI : 10.1016/j.jag.2022.102735 Date de publication en ligne : 10/03/2022 En ligne : https://doi.org/10.1016/j.jag.2022.102735 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=100124
in International journal of applied Earth observation and geoinformation > vol 108 (April 2022) . - n° 102735[article]Graph neural network based model for multi-behavior session-based recommendation / Bo Yu in Geoinformatica, vol 26 n° 2 (April 2022)
[article]
Titre : Graph neural network based model for multi-behavior session-based recommendation Type de document : Article/Communication Auteurs : Bo Yu, Auteur ; Ruoqian Zhang, Auteur ; Wei Chen, Auteur ; Junhua Fang, Auteur Année de publication : 2022 Article en page(s) : pp 429 - 447 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Intelligence artificielle
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] comportement
[Termes IGN] consommation
[Termes IGN] modèle de simulation
[Termes IGN] réseau neuronal de graphes
[Termes IGN] réseau sémantique
[Termes IGN] service fondé sur la positionMots-clés libres : session Résumé : (auteur) Multi-behavior session-based recommendation aims to predict the next item, such as a location-based service (LBS) or a product, to be interacted by a specific behavior type (e.g., buy or click) in a session involving multiple types of behaviors. State-of-the-art methods generally model multi-behavior dependencies in item-level, but ignore the potential of discovering useful patterns of multi-behavior transition through feature-level representation learning. Besides, sequential and non-sequential patterns should be properly fused in session modeling to capture dynamic interests within the session. To this end, this paper proposes a Graph Neural Network based Hybrid Model GNNH, which enables feature-level deeper representations of multi-behavior interaction sequences for session-based recommendation. Specifically, we first construct multi-relational item graph (MRIG) and feature graph (MRFG) based on session sequences. On top of the MRIG and MRFG, our model takes advantage of GNN to capture item and feature representations, such that global item-to-item and feature-to-feature relations are fully preserved. Afterwards, each multi-behavior session is modeled by a seamless fusion of interacted item and feature representations, where self-attention and mean-pooling are used to obtain sequential and non-sequential patterns simultaneously. Experiments on two real datasets show that the GNNH model significantly outperforms the state-of-the-art methods. Numéro de notice : A2022-326 Affiliation des auteurs : non IGN Thématique : GEOMATIQUE/INFORMATIQUE Nature : Article DOI : 10.1007/s10707-021-00439-w Date de publication en ligne : 29/05/2021 En ligne : https://doi.org/10.1007/s10707-021-00439-w Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=100489
in Geoinformatica > vol 26 n° 2 (April 2022) . - pp 429 - 447[article]Visual vs internal attention mechanisms in deep neural networks for image classification and object detection / Abraham Montoya Obeso in Pattern recognition, vol 123 (March 2022)
[article]
Titre : Visual vs internal attention mechanisms in deep neural networks for image classification and object detection Type de document : Article/Communication Auteurs : Abraham Montoya Obeso, Auteur ; Jenny Benois-Pineau, Auteur ; Mireya S. García Vázquez, Auteur ; et al., Auteur Année de publication : 2022 Article en page(s) : n° 108411 Note générale : bibliographie Langues : Anglais (eng) Descripteur : [Vedettes matières IGN] Traitement d'image optique
[Termes IGN] analyse visuelle
[Termes IGN] apprentissage profond
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] détection d'objet
[Termes IGN] données d'entrainement (apprentissage automatique)
[Termes IGN] extraction de traits caractéristiques
[Termes IGN] oculométrie
[Termes IGN] saillance
[Termes IGN] segmentation sémantique
[Termes IGN] visualisation de donnéesRésumé : (auteur) The so-called “attention mechanisms” in Deep Neural Networks (DNNs) denote an automatic adaptation of DNNs to capture representative features given a specific classification task and related data. Such attention mechanisms perform both globally by reinforcing feature channels and locally by stressing features in each feature map. Channel and feature importance are learnt in the global end-to-end DNNs training process. In this paper, we present a study and propose a method with a different approach, adding supplementary visual data next to training images. We use human visual attention maps obtained independently with psycho-visual experiments, both in task-driven or in free viewing conditions, or powerful models for prediction of visual attention maps. We add visual attention maps as new data alongside images, thus introducing human visual attention into the DNNs training and compare it with both global and local automatic attention mechanisms. Experimental results show that known attention mechanisms in DNNs work pretty much as human visual attention, but still the proposed approach allows a faster convergence and better performance in image classification tasks. Numéro de notice : A2022-197 Affiliation des auteurs : non IGN Thématique : IMAGERIE Nature : Article DOI : 10.1016/j.patcog.2021.108411 Date de publication en ligne : 12/11/2021 En ligne : https://doi.org/10.1016/j.patcog.2021.108411 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=99988
in Pattern recognition > vol 123 (March 2022) . - n° 108411[article]Analysis of pedestrian movements and gestures using an on-board camera to predict their intentions / Joseph Gesnouin (2022)
Titre : Analysis of pedestrian movements and gestures using an on-board camera to predict their intentions Titre original : Analyse des mouvements et gestes des piétons via caméra embarquée pour la prédiction de leurs intentions Type de document : Thèse/HDR Auteurs : Joseph Gesnouin, Auteur ; Fabien Moutarde, Directeur de thèse Editeur : Paris : Université Paris Sciences et Lettres Année de publication : 2022 Importance : 171 p. Format : 21 x 30 cm Note générale : bibliographie
Thèse de doctorat de l'Université Paris Sciences et Lettres, Préparée à MINES ParisTech, Spécialité
Informatique temps réel, robotique et automatiqueLangues : Anglais (eng) Descripteur : [Vedettes matières IGN] Intelligence artificielle
[Termes IGN] apprentissage profond
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] classification par réseau neuronal convolutif
[Termes IGN] classification par réseau neuronal récurrent
[Termes IGN] estimation de pose
[Termes IGN] image RVB
[Termes IGN] instrument embarqué
[Termes IGN] navigation autonome
[Termes IGN] piéton
[Termes IGN] reconnaissance de gestes
[Termes IGN] réseau neuronal de graphes
[Termes IGN] squelettisation
[Termes IGN] trajectoire (véhicule non spatial)
[Termes IGN] vision par ordinateurIndex. décimale : THESE Thèses et HDR Résumé : (auteur) The autonomous vehicle (AV) is a major challenge for the mobility of tomorrow. Progress is being made every day to achieve it; however, many problems remain to be solved to achieve a safe outcome for the most vulnerable road users (VRUs). One of the major challenge faced by AVs is the ability to efficiently drive in urban environments. Such a task requires interactions between autonomous vehicles and VRUs to resolve traffic ambiguities. In order to interact with VRUs, AVs must be able to understand their intentions and predict their incoming actions. In this dissertation, our work revolves around machine learning technology as a way to understand and predict human behaviour from visual signals and more specifically pose kinematics. Our goal is to propose an assistance system to the AV that is lightweight, scene-agnostic that could be easily implemented in any embedded devices with real-time constraints. Firstly, in the gesture and action recognition domain, we study and introduce different representations for pose kinematics, based on deep learning models as a way to efficiently leverage their spatial and temporal components while staying in an euclidean grid-space. Secondly, in the autonomous driving domain, we show that it is possible to link the posture, the walking attitude and the future behaviours of the protagonists of a scene without using the contextual information of the scene (zebra crossing, traffic light...). This allowed us to divide by a factor of 20 the inference speed of existing approaches for pedestrian intention prediction while keeping the same prediction robustness. Finally, we assess the generalization capabilities of pedestrian crossing predictors and show that the classical train-test sets evaluation for pedestrian crossing prediction, i.e., models being trained and tested on the same dataset, is not sufficient to efficiently compare nor conclude anything about their applicability in a real-world scenario. To make the research field more sustainable and representative of the real advances to come. We propose new protocols and metrics based on uncertainty estimates under domain-shift in order to reach the end-goal of pedestrian crossing behavior predictors: vehicle implementation. Note de contenu : 1- Introduction
2- Human activity recognition with pose-driven deep learning models
3- From action recognition to pedestrian discrete intention prediction
4- Assessing the generalization of pedestrian crossing predictors
5- ConclusionNuméro de notice : 24066 Affiliation des auteurs : non IGN Thématique : INFORMATIQUE Nature : Thèse française Note de thèse : Thèse de Doctorat : Informatique temps réel, robotique et automatique : Paris Sciences et Lettres : 2022 DOI : sans En ligne : https://tel.hal.science/tel-03813520 Format de la ressource électronique : URL Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=102091
Titre : Attention-based vandalism detection in OpenStreetMap Type de document : Article/Communication Auteurs : Nicolas Tempelmeier, Auteur ; Elena Demidova, Auteur Editeur : New York [Etats-Unis] : Association for computing machinery ACM Année de publication : 2022 Conférence : WWW 2022, ACM Web Conference 2022 25/04/2022 29/04/2022 Lyon online France Proceedings ACM Importance : pp 643 - 651 Note générale : bibliographie Langues : Français (fre) Descripteur : [Vedettes matières IGN] Bases de données localisées
[Termes IGN] attention (apprentissage automatique)
[Termes IGN] détection d'anomalie
[Termes IGN] fiabilité des données
[Termes IGN] historique des données
[Termes IGN] OpenStreetMapMots-clés libres : vandalisme Résumé : (auteur) OpenStreetMap (OSM), a collaborative, crowdsourced Web map, is a unique source of openly available worldwide map data, increasingly adopted in Web applications. Vandalism detection is a critical task to support trust and maintain OSM transparency. This task is remarkably challenging due to the large scale of the dataset, the sheer number of contributors, various vandalism forms, and the lack of annotated data. This paper presents Ovid - a novel attention-based method for vandalism detection in OSM. Ovid relies on a novel neural architecture that adopts a multi-head attention mechanism to summarize information indicating vandalism from OSM changesets effectively. To facilitate automated vandalism detection, we introduce a set of original features that capture changeset, user, and edit information. Furthermore, we extract a dataset of real-world vandalism incidents from the OSM edit history for the first time and provide this dataset as open data. Our evaluation conducted on real-world vandalism data demonstrates the effectiveness of Ovid. Numéro de notice : C2022-008 Affiliation des auteurs : non IGN Thématique : GEOMATIQUE Nature : Communication DOI : 10.1145/3485447.3512224 Date de publication en ligne : 25/04/2022 En ligne : https://doi.org/10.1145/3485447.3512224 Format de la ressource électronique : URL article Permalink : https://documentation.ensg.eu/index.php?lvl=notice_display&id=100493 Learning spatio-temporal representations of satellite time series for large-scale crop mapping / Vivien Sainte Fare Garnot (2022)PermalinkPermalinkSelf-attention and generative adversarial networks for algae monitoring / Nhut Hai Huynh in European journal of remote sensing, vol 55 n° 1 (2022)PermalinkSemantic segmentation of high-resolution remote sensing images based on a class feature attention mechanism fused with Deeplabv3+ / Zhimin Wang in Computers & geosciences, vol 158 (January 2022)PermalinkTowards expressive graph neural networks : Theory, algorithms, and applications / Georgios Dasoulas (2022)PermalinkPermalinkUnsupervised pansharpening based on self-attention mechanism / Ying Qu in IEEE Transactions on geoscience and remote sensing, vol 59 n° 4 (April 2021)PermalinkMultiscale CNN with autoencoder regularization joint contextual attention network for SAR image classification / Zitong Wu in IEEE Transactions on geoscience and remote sensing, vol 59 n° 2 (February 2021)PermalinkPermalinkLightweight temporal self-attention for classifying satellite images time series / Vivien Sainte Fare Garnot (2020)Permalink