We now have a sentencepiece tokenizer in fastai, so you can train NLP models in agglutinative languages like Turkish and non space delimited languages like Chinese.
— Jeremy Howard (@jeremyphoward) June 11, 2019
I just tried it on Turkish sentiment analysis with ULMFiT and easily beat the SoTA! :)https://t.co/6agBdk2tjW pic.twitter.com/PURrIsn7dD