DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 22

1
IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages ...
BASE
Show details
2
MDAPT: Multilingual Domain Adaptive Pretraining in a Single Model ...
BASE
Show details
3
Vision-and-Language or Vision-for-Language? On Cross-Modal Influence in Multimodal Transformers ...
BASE
Show details
4
Multimodal Pretraining Unmasked: A Meta-Analysis and a Unified Framework of Vision-and-Language BERTs ...
BASE
Show details
5
Multimodal pretraining unmasked: A meta-analysis and a unified framework of vision-and-language berts ...
BASE
Show details
6
mDAPT: Multilingual Domain Adaptive Pretraining in a Single Model ...
BASE
Show details
7
Visually Grounded Reasoning across Languages and Cultures ...
BASE
Show details
8
Visually Grounded Reasoning across Languages and Cultures ...
Abstract: The design of widespread vision-and-language datasets and pre-trained encoders directly adopts, or draws inspiration from, the concepts and images of ImageNet. While one can hardly overestimate how much this benchmark contributed to progress in computer vision, it is mostly derived from lexical databases and image queries in English, resulting in source material with a North American or Western European bias. Therefore, we devise a new protocol to construct an ImageNet-style hierarchy representative of more languages and cultures. In particular, we let the selection of both concepts and images be entirely driven by native speakers, rather than scraping them automatically. Specifically, we focus on a typologically diverse set of languages, namely, Indonesian, Mandarin Chinese, Swahili, Tamil, and Turkish. On top of the concepts and images obtained through this new protocol, we create a multilingual dataset for {M}ulticultur{a}l {R}easoning over {V}ision and {L}anguage (MaRVL) by eliciting statements from ... : EMNLP 2021; Fangyu and Emanuele contributed equally; MaRVL website: https://marvl-challenge.github.io ...
Keyword: Artificial Intelligence cs.AI; Computation and Language cs.CL; Computer Vision and Pattern Recognition cs.CV; FOS Computer and information sciences
URL: https://dx.doi.org/10.48550/arxiv.2109.13238
https://arxiv.org/abs/2109.13238
BASE
Hide details
9
Multimodal pretraining unmasked: A meta-analysis and a unified framework of vision-and-language berts
In: Transactions of the Association for Computational Linguistics, 9 (2021)
BASE
Show details
10
The Role of Syntactic Planning in Compositional Image Captioning ...
BASE
Show details
11
Visually Grounded Reasoning across Languages and Cultures ...
BASE
Show details
12
CompGuessWhat?!: A Multi-task Evaluation Framework for Grounded Language Learning ...
BASE
Show details
13
The Sensitivity of Language Models and Humans to Winograd Schema Perturbations ...
BASE
Show details
14
Bootstrapping Disjoint Datasets for Multilingual Multimodal Representation Learning ...
BASE
Show details
15
Cross-lingual Visual Verb Sense Disambiguation ...
BASE
Show details
16
Lessons learned in multilingual grounded language learning ...
BASE
Show details
17
Findings of the Second Shared Task on Multimodal Machine Translation and Multilingual Image Description ...
BASE
Show details
18
Cross-linguistic differences and similarities in image descriptions ...
BASE
Show details
19
Multi30K: Multilingual English-German Image Descriptions
Elliott, Desmond [Verfasser]; Frank, Stella [Verfasser]; Sima'an, Khalil [Verfasser]. - Aachen : Universitätsbibliothek der RWTH Aachen, 2016
DNB Subject Category Language
Show details
20
Multi30K: Multilingual English-German Image Descriptions ...
BASE
Show details

Page: 1 2

Catalogues
0
0
0
0
1
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
21
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern