DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 26

1
Trajectory Prediction with Linguistic Representations
Kuo, Yen-Ling; Huang, Xin; Barbu, Andrei. - : Center for Brains, Minds and Machines (CBMM), International Conference on Robotics and Automation (ICRA), 2022
BASE
Show details
2
Compositional Networks Enable Systematic Generalization for Grounded Language Understanding
Kuo, Yen-Ling; Katz, Boris; Barbu, Andrei. - : Center for Brains, Minds and Machines (CBMM), Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021
BASE
Show details
3
Trajectory Prediction with Linguistic Representations ...
BASE
Show details
4
Measuring Social Biases in Grounded Vision and Language Embeddings ...
NAACL 2021 2021; Barbu, Andrei; Katz, Boris. - : Underline Science Inc., 2021
BASE
Show details
5
Compositional Networks Enable Systematic Generalization for Grounded Language Understanding ...
BASE
Show details
6
Measuring Social Biases in Grounded Vision and Language Embeddings
Ross, Candace; Barbu, Andrei; Katz, Boris. - : Center for Brains, Minds and Machines (CBMM), Annual Conference of the North American Chapter of the Association for Computational Linguistics (HLT/NAACL), 2021
BASE
Show details
7
Learning a natural-language to LTL executable semantic parser for grounded robotics
Wang, Christopher; Ross, Candace; Kuo, Yen-Ling. - : Center for Brains, Minds and Machines (CBMM), Conference on Robot Learning (CoRL), 2020
BASE
Show details
8
Compositional Networks Enable Systematic Generalization for Grounded Language Understanding ...
BASE
Show details
9
Learning a natural-language to LTL executable semantic parser for grounded robotics ...
BASE
Show details
10
Measuring Social Biases in Grounded Vision and Language Embeddings ...
BASE
Show details
11
Encoding formulas as deep networks: Reinforcement learning for zero-shot execution of LTL formulas
Kuo, Yen-Ling; Katz, Boris; Barbu, Andrei. - : Center for Brains, Minds and Machines (CBMM), The Ninth International Conference on Learning Representations (ICLR), 2020
BASE
Show details
12
Temporal Grounding Graphs for Language Understanding with Accrued Visual-Linguistic Context ...
BASE
Show details
13
Temporal Grounding Graphs for Language Understanding with Accrued Visual-Linguistic Context
In: Other repository (2018)
BASE
Show details
14
Do You See What I Mean? Visual Resolution of Linguistic Ambiguities
Berzak, Yevgeni; Barbu, Andrei; Harari, Daniel. - : Center for Brains, Minds and Machines (CBMM), arXiv, 2016
BASE
Show details
15
Anchoring and Agreement in Syntactic Annotations
Berzak, Yevgeni; Huang, Yan; Barbu, Andrei. - : Center for Brains, Minds and Machines (CBMM), arXiv, 2016
BASE
Show details
16
Do You See What I Mean? Visual Resolution of Linguistic Ambiguities ...
BASE
Show details
17
Seeing What You’re Told: Sentence-Guided Activity Recognition In Video
Siddharth, Narayanaswamy; Barbu, Andrei; Siskind, Jeffrey Mark. - : Center for Brains, Minds and Machines (CBMM), arXiv, 2014
BASE
Show details
18
The Compositional Nature of Event Representations in the Human Brain
Abstract: How does the human brain represent simple compositions of constituents: actors, verbs, objects, directions, and locations? Subjects viewed videos during neuroimaging (fMRI) sessions from which sentential descriptions of those videos were identified by decoding the brain representations based only on their fMRI activation patterns. Constituents (e.g., fold and shirt) were independently decoded from a single presentation. Independent constituent classification was then compared to joint classification of aggregate concepts (e.g., fold -shirt); results were similar as measured by accuracy and correlation. The brain regions used for independent constituent classification are largely disjoint and largely cover those used for joint classification. This allows recovery of sentential descriptions of stimulus videos by composing the results of the independent constituent classifiers. Furthermore, classifiers trained on the words one set of subjects think of when watching a video can recognize sentences a different subject thinks of when watching a different video. ; This work was supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF-1231216.
Keyword: Computer Language; Language; Linguistics; Neuroscience; Vision and Language
URL: http://hdl.handle.net/1721.1/100175
BASE
Hide details
19
Video In Sentences Out ...
BASE
Show details
20
Saying What You're Looking For: Linguistics Meets Video Search ...
BASE
Show details

Page: 1 2

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
26
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern