Tweeted By @Thom_Wolf
A very nice notebook/blog-post from @PatrickPlaten on how to create and train (to really nice results) an Encoder-Decoder model using pre-trained Bert/Roberta/GPT2/etc models as encoders and as decoders 👇 https://t.co/bZAUTaDv5g
— Thomas Wolf (@Thom_Wolf) November 9, 2020