Homepage
Close
Menu

Site Navigation

  • Home
  • Archive(TODO)
    • By Day
    • By Month
  • About(TODO)
  • Stats
Close
by yvespeirsman on 2019-08-27 (UTC).

Transfer Learning works great for Natural Language Processing, but sometimes its large models can be hard to handle. At NLP Town we used model distillation to train @spacy_io text classifiers that rival BERT for sentiment analysishttps://t.co/mLvJnYou0R #deeplearning #NLProc pic.twitter.com/lcl3bQ9S8F

— Yves Peirsman (@yvespeirsman) August 27, 2019
nlpresearch
by chipro on 2019-08-28 (UTC).

Now that we know it's possible to achieve comparable results to BERT using only 66M parameters, can someone find a way to train a 66M param model from scratch instead of distilling? https://t.co/ycJjMwSwsr

— Chip Huyen (@chipro) August 28, 2019
nlpresearchw_code

Tags

learning tutorial misc nlp rstats gan ethics research dataviz survey python tool security kaggle video thought bayesian humour tensorflow w_code bias dataset pytorch cv tip application javascript forecast swift golang rl jax julia gnn causal surey diffusion
© Copyright Philosophy 2018 Site Template by Colorlib