번역(60)
-
[2022-02-11] 오늘의 자연어처리
Generating Training Data with Language Models: Towards Zero-Shot Language Understanding Pretrained language models (PLMs) have demonstrated remarkable performance in various natural language processing tasks: Unidirectional PLMs (e.g., GPT) are well known for their superior text generation capabilities; bidirectional PLMs (e.g., BERT) have been the prominent choice for natural language understan..
2022.02.11 -
[2022-02-10] 오늘의 자연어처리
Selecting Seed Words for Wordle using Character Statistics Wordle, a word guessing game rose to global popularity in the January of 2022. The goal of the game is to guess a five-letter English word within six tries. Each try provides the player with hints by means of colour changing tiles which inform whether or not a given character is part of the solution as well as, in cases where it is part ..
2022.02.10 -
[2022-02-09] 오늘의 자연어처리
Adaptive Fine-Tuning of Transformer-Based Language Models for Named Entity Recognition The current standard approach for fine-tuning transformer-based language models includes a fixed number of training epochs and a linear learning rate schedule. In order to obtain a near-optimal model for the given downstream task, a search in optimization hyperparameter space is usually required. In particular..
2022.02.09 -
[2022-02-08] 오늘의 자연어처리
mSLAM: Massively multilingual joint pre-training for speech and text We present mSLAM, a multilingual Speech and LAnguage Model that learns cross-lingual cross-modal representations of speech and text by pre-training jointly on large amounts of unlabeled speech and text in multiple languages. mSLAM combines w2v-BERT pre-training on speech with SpanBERT pre-training on character-level text, along..
2022.02.08 -
[2022-02-07] 오늘의 자연어처리
Joint Speech Recognition and Audio Captioning Speech samples recorded in both indoor and outdoor environments are often contaminated with secondary audio sources. Most end-to-end monaural speech recognition systems either remove these background sounds using speech enhancement or train noise-robust models. For better model interpretability and holistic understanding, we aim to bring together the..
2022.02.07 -
[2022-02-04] 오늘의 자연어처리
A Semi-Supervised Deep Clustering Pipeline for Mining Intentions From Texts Mining the latent intentions from large volumes of natural language inputs is a key step to help data analysts design and refine Intelligent Virtual Assistants (IVAs) for customer service. To aid data analysts in this task we present Verint Intent Manager (VIM), an analysis platform that combines unsupervised and semi-su..
2022.02.04