Homepage
Close
Menu

Site Navigation

  • Home
  • Archive(TODO)
    • By Day
    • By Month
  • About(TODO)
  • Stats
Close
by Thom_Wolf on 2018-11-12 (UTC).

I've added FP16 training to our PyTorch BERT repo to easily fine-tune BERT-large on GPU.

The repo has become a showcase of all the tools you can use to train huge NNs 🙂

Got >91 F1 on SQuAD training BERT-large a few hours on 4-GPUs.
Should take less than a day on 1-(recent)-GPU pic.twitter.com/bMhII36gT0

— Thomas Wolf (@Thom_Wolf) November 12, 2018
nlp
by Thom_Wolf on 2018-11-17 (UTC).

Our PyTorch BERT is on pip!
I took extra care to make it both easy to use and modular.
Uses @ai2_allennlp file caching technique to download/cache/load Google's pretrained models
Includes 6 PyTorch models with various architectures, tokenizer & optimizer
👉https://t.co/dChmNPGPKO pic.twitter.com/jETFVQT5RM

— Thomas Wolf (@Thom_Wolf) November 17, 2018
nlpw_code
by PyTorch on 2018-12-14 (UTC).

PyTorch BERT models are now 4x faster, thanks to @nvidia https://t.co/enSldGFAAC

— PyTorch (@PyTorch) December 14, 2018
pytorchnlptool
by ch402 on 2019-06-07 (UTC).

Very cool visualizations of different word senses being represented in later layers of BERT, by @_coenen, Emily Reif, Ann Yuan and collaborators.https://t.co/RiMjXseXxW pic.twitter.com/o6sKdPprAJ

— Chris Olah (@ch402) June 7, 2019
nlpresearch
  • Prev
  • 1
  • 2

Tags

learning tutorial misc nlp rstats gan ethics research dataviz survey python tool security kaggle video thought bayesian humour tensorflow w_code bias dataset pytorch cv tip application javascript forecast swift golang rl jax julia gnn causal surey diffusion
© Copyright Philosophy 2018 Site Template by Colorlib