DE eng

Search in the Catalogues and Directories

Page: 1 2 3
Hits 1 – 20 of 41

1
Cross-Situational Learning Towards Robot Grounding
In: https://hal.archives-ouvertes.fr/hal-03628290 ; 2022 (2022)
BASE
Show details
2
Cross-Situational Learning Towards Robot Grounding
In: https://hal.archives-ouvertes.fr/hal-03628290 ; 2022 (2022)
BASE
Show details
3
What does the Canary Say? Low-Dimensional GAN Applied to Birdsong
In: https://hal.inria.fr/hal-03244723 ; 2021 (2021)
BASE
Show details
4
What does the Canary Say? Low-Dimensional GAN Applied to Birdsong
In: https://hal.inria.fr/hal-03244723 ; 2021 (2021)
BASE
Show details
5
Which Hype for my New Task? Hints and Random Search for Reservoir Computing Hyperparameters
In: ICANN 2021 - 30th International Conference on Artificial Neural Networks ; https://hal.inria.fr/hal-03203318 ; ICANN 2021 - 30th International Conference on Artificial Neural Networks, Sep 2021, Bratislava, Slovakia (2021)
BASE
Show details
6
Canary Song Decoder: Transduction and Implicit Segmentation with ESNs and LTSMs
In: https://hal.inria.fr/hal-03203374 ; 2021 (2021)
BASE
Show details
7
Which Hype for my New Task? Hints and Random Search for Reservoir Computing Hyperparameters
In: https://hal.inria.fr/hal-03203318 ; 2021 (2021)
BASE
Show details
8
Canary Song Decoder: Transduction and Implicit Segmentation with ESNs and LTSMs
In: ICANN 2021 - 30th International Conference on Artificial Neural Networks ; https://hal.inria.fr/hal-03203374 ; ICANN 2021 - 30th International Conference on Artificial Neural Networks, Sep 2021, Bratislava, Slovakia. pp.71--82, ⟨10.1007/978-3-030-86383-8_6⟩ ; https://link.springer.com/chapter/10.1007/978-3-030-86383-8_6 (2021)
BASE
Show details
9
Hierarchical-Task Reservoir for Online Semantic Analysis from Continuous Speech
In: ISSN: 2162-237X ; IEEE Transactions on Neural Networks and Learning Systems ; https://hal.inria.fr/hal-03031413 ; IEEE Transactions on Neural Networks and Learning Systems, IEEE, 2021, ⟨10.1109/TNNLS.2021.3095140⟩ ; https://ieeexplore.ieee.org/abstract/document/9548713/metrics#metrics (2021)
BASE
Show details
10
Editorial: Language and Robotics
In: ISSN: 2296-9144 ; Frontiers in Robotics and AI ; https://hal.inria.fr/hal-03533733 ; Frontiers in Robotics and AI, Frontiers Media S.A., 2021, 8, ⟨10.3389/frobt.2021.674832⟩ (2021)
BASE
Show details
11
Learning to Parse Sentences with Cross-Situational Learning using Different Word Embeddings Towards Robot Grounding ...
BASE
Show details
12
Cross-Situational Learning with Reservoir Computing for Language Acquisition Modelling
In: 2020 International Joint Conference on Neural Networks (IJCNN 2020) ; https://hal.inria.fr/hal-02594725 ; 2020 International Joint Conference on Neural Networks (IJCNN 2020), Jul 2020, Glasgow, Scotland, United Kingdom ; https://wcci2020.org/ (2020)
BASE
Show details
13
Hierarchical-Task Reservoir for Anytime POS Tagging from Continuous Speech
In: 2020 International Joint Conference on Neural Networks (IJCNN 2020) ; https://hal.inria.fr/hal-02594495 ; 2020 International Joint Conference on Neural Networks (IJCNN 2020), Jul 2020, Glasgow, Scotland, United Kingdom ; https://wcci2020.org/ (2020)
BASE
Show details
14
Language Acquisition with Echo State Networks: Towards Unsupervised Learning
In: ICDL 2020 - IEEE International Conference on Development and Learning ; https://hal.inria.fr/hal-02926613 ; ICDL 2020 - IEEE International Conference on Development and Learning, Oct 2020, Valparaiso / Virtual, Chile (2020)
BASE
Show details
15
A Journey in ESN and LSTM Visualisations on a Language Task
In: https://hal.inria.fr/hal-03030248 ; 2020 (2020)
Abstract: Echo States Networks (ESN) and Long-Short Term Memory networks (LSTM) are two popular architectures of Recurrent Neural Networks (RNN) to solve machine learning task involving sequential data. However, little have been done to compare their performances and their internal mechanisms on a common task. In this work, we trained ESNs and LSTMs on a Cross-Situationnal Learning (CSL) task. This task aims at modelling how infants learn language: they create associations between words and visual stimuli in order to extract meaning from words and sentences. The results are of three kinds: performance comparison, internal dynamics analyses and visualization of latent space. (1) We found that both models were able to successfully learn the task: the LSTM reached the lowest error for the basic corpus, but the ESN was quicker to train. Furthermore, the ESN was able to outperform LSTMs on datasets more challenging without any further tuning needed. (2) We also conducted an analysis of the internal units activations of LSTMs and ESNs. Despite the deep differences between both models (trained or fixed internal weights), we were able to uncover similar inner mechanisms: both put emphasis on the units encoding aspects of the sentence structure. (3) Moreover, we present Recurrent States Space Visualisations (RSSviz), a method to visualize the structure of latent state space of RNNs, based on dimension reduction (using UMAP). This technique enables us to observe a fractal embedding of sequences in the LSTM. RSSviz is also useful for the analysis of ESNs (i) to spot difficult examples and (ii) to generate animated plots showing the evolution of activations across learning stages. Finally, we explore qualitatively how the RSSviz could provide an intuitive visualisation to understand the influence of hyperparameters on the reservoir dynamics prior to ESN training.
Keyword: [INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG]; [INFO.INFO-NE]Computer Science [cs]/Neural and Evolutionary Computing [cs.NE]; [INFO.INFO-RB]Computer Science [cs]/Robotics [cs.RO]; [SDV.NEU]Life Sciences [q-bio]/Neurons and Cognition [q-bio.NC]; Cross-Situational Learning; Dimension Reduction; ESN; LSTM; UMAP; Visualisation
URL: https://hal.inria.fr/hal-03030248
https://hal.inria.fr/hal-03030248/file/Comparison_between_LSTM_and_ESN%2812%29.pdf
https://hal.inria.fr/hal-03030248/document
BASE
Hide details
16
Recurrent Neural Networks Models for Developmental Language Acquisition: Reservoirs Outperform LSTMs
In: SNL 2020 - 12th Annual Meeting of the Society for the Neurobiology of Language ; https://hal.inria.fr/hal-03146558 ; SNL 2020 - 12th Annual Meeting of the Society for the Neurobiology of Language, Oct 2020, Virtual Edition, Canada (2020)
BASE
Show details
17
Learning to Parse Grounded Language using Reservoir Computing
In: ICDL-Epirob 2019 - Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics ; https://hal.inria.fr/hal-02422157 ; ICDL-Epirob 2019 - Joint IEEE 9th International Conference on Development and Learning and Epigenetic Robotics, Aug 2019, Olso, Norway. ⟨10.1109/devlrn.2019.8850718⟩ ; https://ieeexplore.ieee.org/abstract/document/8850718 (2019)
BASE
Show details
18
Teach Your Robot Your Language! Trainable Neural Parser for Modelling Human Sentence Processing: Examples for 15 Languages
In: ISSN: 2379-8920 ; EISSN: 2379-8939 ; IEEE Transactions on Cognitive and Developmental Systems ; https://hal.inria.fr/hal-01964541 ; IEEE Transactions on Cognitive and Developmental Systems, Institute of Electrical and Electronics Engineers, Inc, 2019, ⟨10.1109/TCDS.2019.2957006⟩ ; https://doi.org/10.1109/tcds.2019.2957006 (2019)
BASE
Show details
19
A Reservoir Model for Intra-Sentential Code-Switching Comprehension in French and English
In: CogSci'19 - 41st Annual Meeting of the Cognitive Science Society ; https://hal.inria.fr/hal-02432831 ; CogSci'19 - 41st Annual Meeting of the Cognitive Science Society, Jul 2019, Montréal, Canada ; https://cognitivesciencesociety.org/cogsci-2019/ (2019)
BASE
Show details
20
Replication of Laje & Mindlin's model producing synthetic syllables
In: European Birdsong Meeting ; https://hal.inria.fr/hal-01964522 ; European Birdsong Meeting, Apr 2018, Odense, Denmark. 2018 (2018)
BASE
Show details

Page: 1 2 3

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
41
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern