1 |
RefVOS: A Closer Look at Referring Expressions for Video Object Segmentation ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Recurrent Instance Segmentation using Sequences of Referring Expressions ...
|
|
|
|
BASE
|
|
Show details
|
|
3 |
What do Entity-Centric Models Learn? Insights from Entity Linking in Multi-Party Dialogue ...
|
|
|
|
BASE
|
|
Show details
|
|
5 |
UHD: Cross-lingual word sense disambiguation using multilingual co-occurrence graphs
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Building a Multilingual Lexical Resource for Named Entity Disambiguation, Translation and Transliteration
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Casting implicit role linking as an anaphora resolution task
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Building a multilingual lexical resource for named entity disambiguation, translation and transliteration
|
|
|
|
BASE
|
|
Show details
|
|
10 |
UHD: cross-lingual word sense disambiguation using multilingual co-occurrence graphs
|
|
|
|
BASE
|
|
Show details
|
|
11 |
Object naming in language and vision: a survey and a new dataset
|
|
|
|
BASE
|
|
Show details
|
|
13 |
AMORE-UPF at SemEval-2018 Task 4: BiLSTM with entity library
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Grounding semantic roles in images
|
|
|
|
Abstract:
Comunicació presentada a la 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP 2018), celebrada del 31 d'octubre al 4 de novembre de 2018 a Brussel·les, Bèlgica. ; We address the task of visual semantic role labeling (vSRL), the identification of the participants of a situation or event in a visual scene, and their labeling with their semantic relations to the event or situation. We render candidate participants as image regions of objects, and train a model which learns to ground roles in the regions which depict the corresponding participant. Experimental results demonstrate that we can train a vSRL model without reliance on prohibitive image-based role annotations, by utilizing noisy data which we extract automatically from image captions using a linguistic SRL system. Furthermore, our model induces frame–semantic visual representations, and their comparison to previous work on supervised visual verb sense disambiguation yields overall better results. ; We thank the anonymous reviewers, Leonie Harter, Christine Sch¨afer, Michael Roth, Gemma Boleda, Anna Rohrbach and Bernt Schiele. This research was supported by the German Research Foundation (DFG EXC 285), by the European Research Council (ERC Horizon 2020 grant agreement No 715154), and the Spanish Ramon y Cajal programme (grant RYC-2015-18907).
|
|
URL: http://hdl.handle.net/10230/36137
|
|
BASE
|
|
Hide details
|
|
15 |
Humans meet models on object naming: a new dataset and analysis
|
|
|
|
BASE
|
|
Show details
|
|
17 |
What do entity-centric models learn? Insights from entity linking in multi-party dialogue
|
|
|
|
BASE
|
|
Show details
|
|
|
|