Advanced search
1 file | 629.88 KB Add to list

When the student becomes the master : learning better and smaller monolingual models from mBERT

Pranaydeep Singh (UGent) and Els Lefever (UGent)
Author
Organization
Abstract
In this research, we present pilot experiments to distil monolingual models from a jointly trained model for 102 languages (mBERT). We demonstrate that it is possible for the target language to outperform the original model, even with a basic distillation setup. We evaluate our methodology for 6 languages with varying amounts of resources and belonging to different language families.
Keywords
LT3

Downloads

  • 2022.coling-1.391.pdf
    • full text (Published version)
    • |
    • open access
    • |
    • PDF
    • |
    • 629.88 KB

Citation

Please use this url to cite or link to this publication:

MLA
Singh, Pranaydeep, and Els Lefever. “When the Student Becomes the Master : Learning Better and Smaller Monolingual Models from MBERT.” Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022), International Committee on Computational Linguistics, 2022, pp. 4434–41.
APA
Singh, P., & Lefever, E. (2022). When the student becomes the master : learning better and smaller monolingual models from mBERT. Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022), 4434–4441. International Committee on Computational Linguistics.
Chicago author-date
Singh, Pranaydeep, and Els Lefever. 2022. “When the Student Becomes the Master : Learning Better and Smaller Monolingual Models from MBERT.” In Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022), 4434–41. International Committee on Computational Linguistics.
Chicago author-date (all authors)
Singh, Pranaydeep, and Els Lefever. 2022. “When the Student Becomes the Master : Learning Better and Smaller Monolingual Models from MBERT.” In Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022), 4434–4441. International Committee on Computational Linguistics.
Vancouver
1.
Singh P, Lefever E. When the student becomes the master : learning better and smaller monolingual models from mBERT. In: Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022). International Committee on Computational Linguistics; 2022. p. 4434–41.
IEEE
[1]
P. Singh and E. Lefever, “When the student becomes the master : learning better and smaller monolingual models from mBERT,” in Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022), Gyeongju, Republic of Korea, 2022, pp. 4434–4441.
@inproceedings{8769631,
  abstract     = {{In this research, we present pilot experiments to distil monolingual models from a jointly trained model for 102 languages (mBERT). We demonstrate that it is possible for the target language to outperform the original model, even with a basic distillation setup. We evaluate our methodology for 6 languages with varying amounts of resources and belonging to different language families.}},
  author       = {{Singh, Pranaydeep and Lefever, Els}},
  booktitle    = {{Proceedings of the 29th International Conference on Computational Linguistic (COLING 2022)}},
  issn         = {{2951-2093}},
  keywords     = {{LT3}},
  language     = {{eng}},
  location     = {{Gyeongju, Republic of Korea}},
  pages        = {{4434--4441}},
  publisher    = {{International Committee on Computational Linguistics}},
  title        = {{When the student becomes the master : learning better and smaller monolingual models from mBERT}},
  url          = {{https://aclanthology.org/2022.coling-1.391}},
  year         = {{2022}},
}