Paper Details

Fast Vocabulary Transfer for Language Model Compression

Authors: Leonidas Gee, Andrea Zugarini, Leonardo Rigutini, Paolo Torroni
Layman Summary

Error: Could not generate summary.

Abstract

Real-world business applications require a trade-off between language model performance and size. We propose a new method for model compression that relies on vocabulary transfer. We evaluate the method on various vertical domains and downstream tasks. Our results indicate that vocabulary transfer can be effectively used in combination with other compression techniques, yielding a significant reduction in model size and inference time while marginally compromising on performance.

Additional Information
  • DOI: 10.18653/v1/2022.emnlp-industry.41
  • Journal Reference: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP 2022): Industry Track