Tweeted By @thtrieu_
We show large language models trained on massive text corpora (LM1b, CommonCrawl, Gutenberg) can be used for commonsense reasoning and obtain SOTA on Winograd Schema Challenge. Paper at https://t.co/aRndlByWfj, results reproducible at https://t.co/jFOmUYf03O pic.twitter.com/s3uyrksAQz
— Trieu H. Trinh (@thtrieu_) June 11, 2018