ID: 2306.10964

Multilingual Few-Shot Learning via Language Model Retrieval

June 19, 2023

View on ArXiv

Similar papers 2

Improving Few-Shot Performance of Language Models via Nearest Neighbor Calibration

December 5, 2022

92% Match
Feng Nie, Meixi Chen, ... , Cheng Xu
Computation and Language

Pre-trained language models (PLMs) have exhibited remarkable few-shot learning capabilities when provided a few examples in a natural language prompt as demonstrations of test instances, i.e., in-context learning. However, the performance of in-context learning is susceptible to the choice of prompt format, training examples and the ordering of the training examples. In this paper, we propose a novel nearest-neighbor calibration framework for in-context learning to ease this ...

Find SimilarView on arXiv

Is Prompt-Based Finetuning Always Better than Vanilla Finetuning? Insights from Cross-Lingual Language Understanding

July 15, 2023

92% Match
Bolei Ma, Ercong Nie, ... , Schütze Hinrich
Computation and Language

Multilingual pretrained language models (MPLMs) have demonstrated substantial performance improvements in zero-shot cross-lingual transfer across various natural language understanding tasks by finetuning MPLMs on task-specific labelled data of a source language (e.g. English) and evaluating on a wide range of target languages. Recent studies show that prompt-based finetuning surpasses regular finetuning in few-shot scenarios. However, the exploration of prompt-based learning...

Find SimilarView on arXiv

LLMs Are Few-Shot In-Context Low-Resource Language Learners

March 25, 2024

92% Match
Samuel Cahyawijaya, Holy Lovenia, Pascale Fung
Computation and Language
Artificial Intelligence

In-context learning (ICL) empowers large language models (LLMs) to perform diverse tasks in underrepresented languages using only short in-context information, offering a crucial avenue for narrowing the gap between high-resource and low-resource languages. Nonetheless, there is only a handful of works explored ICL for low-resource languages with most of them focusing on relatively high-resource languages, such as French and Spanish. In this work, we extensively study ICL and...

Find SimilarView on arXiv

Discrete and Soft Prompting for Multilingual Models

September 8, 2021

92% Match
Mengjie Zhao, Hinrich Schütze
Computation and Language

It has been shown for English that discrete and soft prompting perform strongly in few-shot learning with pretrained language models (PLMs). In this paper, we show that discrete and soft prompting perform better than finetuning in multilingual cases: Crosslingual transfer and in-language training of multilingual natural language inference. For example, with 48 English training examples, finetuning obtains 33.74% accuracy in crosslingual transfer, barely surpassing the majorit...

Find SimilarView on arXiv

Few-shot Learning with Multilingual Language Models

December 20, 2021

92% Match
Xi Victoria Lin, Todor Mihaylov, Mikel Artetxe, Tianlu Wang, Shuohui Chen, Daniel Simig, Myle Ott, Naman Goyal, Shruti Bhosale, Jingfei Du, Ramakanth Pasunuru, Sam Shleifer, Punit Singh Koura, Vishrav Chaudhary, Brian O'Horo, Jeff Wang, Luke Zettlemoyer, Zornitsa Kozareva, Mona Diab, ... , Li Xian
Computation and Language
Artificial Intelligence

Large-scale generative language models such as GPT-3 are competitive few-shot learners. While these models are known to be able to jointly represent many different languages, their training data is dominated by English, potentially limiting their cross-lingual generalization. In this work, we train multilingual generative language models on a corpus covering a diverse set of languages, and study their few- and zero-shot learning capabilities in a wide range of tasks. Our larg...

Find SimilarView on arXiv

Prompt-Tuning Can Be Much Better Than Fine-Tuning on Cross-lingual Understanding With Multilingual Language Models

October 22, 2022

92% Match
Lifu Tu, Caiming Xiong, Yingbo Zhou
Computation and Language
Artificial Intelligence

Pre-trained multilingual language models show significant performance gains for zero-shot cross-lingual model transfer on a wide range of natural language understanding (NLU) tasks. Previously, for zero-shot cross-lingual evaluation, pre-trained models are only fine-tuned on English data and tested on a variety of target languages. In this paper, we do cross-lingual evaluation on various NLU tasks (sentence classification, sequence labeling, question answering) using prompt-t...

Find SimilarView on arXiv

Analyzing and Adapting Large Language Models for Few-Shot Multilingual NLU: Are We There Yet?

March 4, 2024

92% Match
Evgeniia Razumovskaia, Ivan Vulić, Anna Korhonen
Computation and Language

Supervised fine-tuning (SFT), supervised instruction tuning (SIT) and in-context learning (ICL) are three alternative, de facto standard approaches to few-shot learning. ICL has gained popularity recently with the advent of LLMs due to its simplicity and sample efficiency. Prior research has conducted only limited investigation into how these approaches work for multilingual few-shot learning, and the focus so far has been mostly on their performance. In this work, we present...

Find SimilarView on arXiv

Polyglot Prompt: Multilingual Multitask PrompTraining

April 29, 2022

92% Match
Jinlan Fu, See-Kiong Ng, Pengfei Liu
Computation and Language

This paper aims for a potential architectural improvement for multilingual learning and asks: Can different tasks from different languages be modeled in a monolithic framework, i.e. without any task/language-specific module? The benefit of achieving this could open new doors for future multilingual research, including allowing systems trained on low resources to be further assisted by other languages as well as other tasks. We approach this goal by developing a learning frame...

Find SimilarView on arXiv

Multilingual LLMs are Better Cross-lingual In-context Learners with Alignment

May 10, 2023

92% Match
Eshaan Tanwar, Subhabrata Dutta, ... , Chakraborty Tanmoy
Computation and Language

In-context learning (ICL) unfolds as large language models become capable of inferring test labels conditioned on a few labeled samples without any gradient update. ICL-enabled large language models provide a promising step forward toward bypassing recurrent annotation costs in a low-resource setting. Yet, only a handful of past studies have explored ICL in a cross-lingual setting, in which the need for transferring label-knowledge from a high-resource language to a low-resou...

Find SimilarView on arXiv

In-Context Learning for Text Classification with Many Labels

September 19, 2023

92% Match
Aristides Milios, Siva Reddy, Dzmitry Bahdanau
Computation and Language
Machine Learning

In-context learning (ICL) using large language models for tasks with many labels is challenging due to the limited context window, which makes it difficult to fit a sufficient number of examples in the prompt. In this paper, we use a pre-trained dense retrieval model to bypass this limitation, giving the model only a partial view of the full label space for each inference call. Testing with recent open-source LLMs (OPT, LLaMA), we set new state of the art performance in few-s...

Find SimilarView on arXiv