Distilling Task-Specific Knowledge from Teacher Model into BiLSTM
nlp pytorch transformer korean ele knowledge-distillation bert korean-nlp electra distillation bilstm roberta gpt-2 huggingface t5 xlnet distilbert kobert bert-multilingua
-
Updated
Jan 4, 2023 - Python