Inproceedings,

A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models.

, , , , , and .
ACL (Findings), page 11239-11246. Association for Computational Linguistics, (2023)

Meta data

Tags

Users

  • @dblp

Comments and Reviews