Tweeted By @PyTorch
Check out this tutorial from the Catalyst team on how to distillate BERT models. Distilling BERT models can minimize accuracy loss, reduce model sizes, and speed up inferences.https://t.co/uBvba5NObZ
— PyTorch (@PyTorch) June 24, 2020