Alexatm 20b: Few-shot learning using a large-scale multilingual seq2seq model

S Soltan, S Ananthakrishnan, J FitzGerald… - arXiv preprint arXiv …, 2022 - arxiv.org
… -based seq2seq models can be used as few-shot learners, outperforming … multilingual 20
billion parameter seq2seq model, which we will refer to as Alexa Teacher Model (AlexaTM 20B

GeMQuAD: Generating multilingual question answering datasets from large language models using few shot learning

A Namboori, S Mangale, A Rosenbaum… - arXiv preprint arXiv …, 2024 - arxiv.org
… ICL with just one example in the target language using AlexaTM 20B Seq2Seq LLM. …
data using 1-shot ICL on AlexaTM 20B (teacher) model; 2) Apply our semi-supervised learning

CLASP: Few-shot cross-lingual data augmentation for semantic parsing

A Rosenbaum, S Soltan, W Hamza, A Saffari… - arXiv preprint arXiv …, 2022 - arxiv.org
models: we generate synthetic data from AlexaTM 20B to augment the training set for a
modeltraining data for multi-lingual Semantic Parsing by prompting a frozen Large Language …

LINGUIST: Language model instruction tuning to generate annotated utterances for intent classification and slot tagging

A Rosenbaum, S Soltan, W Hamza, Y Versley… - arXiv preprint arXiv …, 2022 - arxiv.org
… ify our results on an internal large-scale multilingual dataset for … to new intents and slots in
challenging few-shot, zero-shot, and … We use AlexaTM-Large 500M, which is trained using the …

Bidirectional language models are also few-shot learners

A Patel, B Li, MS Rasooli, N Constant, C Raffel… - arXiv preprint arXiv …, 2022 - arxiv.org
… mT5 model enables few-shot and zero-shot machine translation and zero-shot multilingual
question … Abstractive text summarization using sequence-to-sequence RNNs and beyond. In …

A review of current trends, techniques, and challenges in large language models (llms)

R Patil, V Gudivada - Applied Sciences, 2024 - mdpi.com
AlexaTM 20B [51] is the first seq2seq model trained using … of those entities derived from
large-scale knowledge bases. … in finetuning and zero/few-shot learning settings. Additionally, …

Several categories of large language models (llms): A short survey

S Pahune, M Chandrasekharan - arXiv preprint arXiv:2307.10188, 2023 - arxiv.org
… issues are covered in this study, including model features, datasets, transformer models, and
LLM … Alexatm 20b: Few-shot learning using a large-scale multilingual seq2seq model. arXiv …

Codet5+: Open code large language models for code understanding and generation

Y Wang, H Le, AD Gotmare, NDQ Bui, J Li… - arXiv preprint arXiv …, 2023 - arxiv.org
… We denote this variant as a sequence-to-sequence (Seq2Seq) … CodeT5+ first on the large-scale
unimodal dataset and then on … in the few-shot evaluation setting. We did observe that our …

In-context examples selection for machine translation

S Agrawal, C Zhou, M Lewis, L Zettlemoyer… - arXiv preprint arXiv …, 2022 - arxiv.org
… from the standard sequence-to-sequence models, where the … -level knowledge from the
large-scale language model using a … Choice of Few-shot Examples We show the distribution of …

DecoMT: Decomposed Prompting for Machine Translation Between Related Languages using Large Language Models

R Puduppully, A Kunchukuttan, R Dabre… - Proceedings of the …, 2023 - aclanthology.org
… We introduce DecoMT, a novel approach of fewshot … We conducted evaluations on multiple
languages using the … If a bucket contains fewer than 20 instances, we merge it with its …