From post

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed.

 

Другие публикации лиц с тем же именем

What Language Model Architecture and Pretraining Objective Works Best for Zero-Shot Generalization?, , , , , , , и . ICML, том 162 из Proceedings of Machine Learning Research, стр. 22964-22984. PMLR, (2022)Mistral 7B., , , , , , , , , и 8 other автор(ы). CoRR, (2023)What Language Model to Train if You Have One Million GPU Hours?, , , , , , , , , и 8 other автор(ы). EMNLP (Findings), стр. 765-782. Association for Computational Linguistics, (2022)The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset., , , , , , , , , и 44 other автор(ы). NeurIPS, (2022)Crosslingual Generalization through Multitask Finetuning., , , , , , , , , и 9 other автор(ы). ACL (1), стр. 15991-16111. Association for Computational Linguistics, (2023)The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset., , , , , , , , , и 44 other автор(ы). CoRR, (2023)Scaling Data-Constrained Language Models., , , , , , , , и . CoRR, (2023)How many data points is a prompt worth?, и . NAACL-HLT, стр. 2627-2636. Association for Computational Linguistics, (2021)Transformers: State-of-the-Art Natural Language Processing., , , , , , , , , и 12 other автор(ы). EMNLP (Demos), стр. 38-45. Association for Computational Linguistics, (2020)Multitask Prompted Training Enables Zero-Shot Task Generalization., , , , , , , , , и 30 other автор(ы). ICLR, OpenReview.net, (2022)