Page: 1 2 3 4 5 6 7 8 9... 68
81 |
The Dark Side of the Language: Pre-trained Transformers in the DarkNet ...
|
|
|
|
BASE
|
|
Show details
|
|
82 |
Discontinuous Constituency and BERT: A Case Study of Dutch ...
|
|
|
|
BASE
|
|
Show details
|
|
83 |
Cross-Platform Difference in Facebook and Text Messages Language Use: Illustrated by Depression Diagnosis ...
|
|
|
|
BASE
|
|
Show details
|
|
84 |
Improving Word Translation via Two-Stage Contrastive Learning ...
|
|
|
|
BASE
|
|
Show details
|
|
85 |
nigam@COLIEE-22: Legal Case Retrieval and Entailment using Cascading of Lexical and Semantic-based models ...
|
|
|
|
BASE
|
|
Show details
|
|
86 |
Learning grammar with a divide-and-concur neural network ...
|
|
|
|
BASE
|
|
Show details
|
|
87 |
Self-Supervised Representation Learning for Speech Using Visual Grounding and Masked Language Modeling ...
|
|
|
|
Abstract:
In this paper, we describe our submissions to the ZeroSpeech 2021 Challenge and SUPERB benchmark. Our submissions are based on the recently proposed FaST-VGS model, which is a Transformer-based model that learns to associate raw speech waveforms with semantically related images, all without the use of any transcriptions of the speech. Additionally, we introduce a novel extension of this model, FaST-VGS+, which is learned in a multi-task fashion with a masked language modeling objective in addition to the visual grounding objective. On ZeroSpeech 2021, we show that our models perform competitively on the ABX task, outperform all other concurrent submissions on the Syntactic and Semantic tasks, and nearly match the best system on the Lexical task. On the SUPERB benchmark, we show that our models also achieve strong performance, in some cases even outperforming the popular wav2vec2.0 model. ... : SAS workshop at AAAI2022, code and model weights available at https://github.com/jasonppy/FaST-VGS-Family ...
|
|
Keyword:
Audio and Speech Processing eess.AS; Computation and Language cs.CL; Computer Vision and Pattern Recognition cs.CV; FOS Computer and information sciences; FOS Electrical engineering, electronic engineering, information engineering; Machine Learning cs.LG; Sound cs.SD
|
|
URL: https://arxiv.org/abs/2202.03543 https://dx.doi.org/10.48550/arxiv.2202.03543
|
|
BASE
|
|
Hide details
|
|
88 |
Accurate Online Posterior Alignments for Principled Lexically-Constrained Decoding ...
|
|
|
|
BASE
|
|
Show details
|
|
89 |
Introducing Neural Bag of Whole-Words with ColBERTer: Contextualized Late Interactions using Enhanced Reduction ...
|
|
|
|
BASE
|
|
Show details
|
|
91 |
Improving Time Sensitivity for Question Answering over Temporal Knowledge Graphs ...
|
|
|
|
BASE
|
|
Show details
|
|
92 |
HistBERT: A Pre-trained Language Model for Diachronic Lexical Semantic Analysis ...
|
|
|
|
BASE
|
|
Show details
|
|
93 |
Towards Explainable Evaluation Metrics for Natural Language Generation ...
|
|
|
|
BASE
|
|
Show details
|
|
94 |
ASL Video Corpora & Sign Bank: Resources Available through the American Sign Language Linguistic Research Project (ASLLRP) ...
|
|
|
|
BASE
|
|
Show details
|
|
95 |
How do lexical semantics affect translation? An empirical study ...
|
|
|
|
BASE
|
|
Show details
|
|
97 |
How Effective is Incongruity? Implications for Code-mix Sarcasm Detection ...
|
|
|
|
BASE
|
|
Show details
|
|
98 |
Learning Meta Word Embeddings by Unsupervised Weighted Concatenation of Source Embeddings ...
|
|
|
|
BASE
|
|
Show details
|
|
99 |
COLD Decoding: Energy-based Constrained Text Generation with Langevin Dynamics ...
|
|
|
|
BASE
|
|
Show details
|
|
100 |
LaPraDoR: Unsupervised Pretrained Dense Retriever for Zero-Shot Text Retrieval ...
|
|
|
|
BASE
|
|
Show details
|
|
Page: 1 2 3 4 5 6 7 8 9... 68
|
|