From post

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed.

 

Другие публикации лиц с тем же именем

Discovering Representation Sprachbund For Multilingual Pre-Training., , , , , , и . EMNLP (Findings), стр. 881-894. Association for Computational Linguistics, (2021)Resource Central: Understanding and Predicting Workloads for Improved Resource Management in Large Cloud Platforms., , , , , и . SOSP, стр. 153-167. ACM, (2017)Scalable and Efficient MoE Training for Multitask Multilingual Models., , , , , , , , и . CoRR, (2021)DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders., , , , , , , , и . CoRR, (2021)Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers., , , и . ICML, том 162 из Proceedings of Machine Learning Research, стр. 13782-13792. PMLR, (2022)Toward ML-centric cloud platforms., , , , , , , , , и . Commun. ACM, 63 (2): 50-59 (2020)Improving Multilingual Translation by Representation and Gradient Regularization., , , , , и . EMNLP (1), стр. 7266-7279. Association for Computational Linguistics, (2021)XLM-T: Scaling up Multilingual Machine Translation with Pretrained Cross-lingual Transformer Encoders., , , , , , , , , и 3 other автор(ы). CoRR, (2020)Multilingual Machine Translation Systems from Microsoft for WMT21 Shared Task., , , , , , , , , и 1 other автор(ы). WMT@EMNLP, стр. 446-455. Association for Computational Linguistics, (2021)SEER-MoE: Sparse Expert Efficiency through Regularization for Mixture-of-Experts., , и . CoRR, (2024)