Tweeted By @AlecRad
What I've been working on for the past year! https://t.co/CAQMYS1rR7
— Alec Radford (@AlecRad) June 11, 2018
Inspired by CoVE, ELMo, and ULMFiT we show that a single transformer language model can be finetuned to a wide variety of NLP tasks and performs very well with little tuning/tweaking.