@inproceedings{tan-etal-2023-gkd, title = "{GKD}: A General Knowledge Distillation Framework for Large-scale Pre-trained Language Model", author = "Tan, Shicheng and Tam, Weng Lam and Wang, Yuanchun and Gong, Wenwen and Zhao, Shu and Zhang, Peng and Tang, Jie", editor = "Sitaram, Sunayana and Beigman Klebanov, Beata and Williams, Jason D", booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 5: Industry Track)", month = jul, year = "2023", address = "Toronto, Canada", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2023.acl-industry.15/", doi = "10.18653/v1/2023.acl-industry.15", pages = "134--148" }