Homepage
Close
Menu

Site Navigation

  • Home
  • Archive(TODO)
    • By Day
    • By Month
  • About(TODO)
  • Stats
Close
by ml_review on 2019-09-27 (UTC).

ALBERT: A Lite BERT for Self-supervised Learning of Language Representations

Parameter reduction techniques:
(i) Vocabulary embedding matrix factorization. Separates hidden layers size from vocabulary embedding
(ii) cross-layer parameter sharinghttps://t.co/dUamPpHLt1 pic.twitter.com/ehcwdYN6mE

— ML Review (@ml_review) September 27, 2019
nlpresearch
by Miles_Brundage on 2019-09-28 (UTC).

P.S. See also the ALBERT paper, which shows stronger results on these metrics for a similarly sized model, using a different approach (using parameters better to begin with vs. compressing a big model later): https://t.co/4VY2gbuFQu

— Miles Brundage (@Miles_Brundage) September 28, 2019
researchnlp
by GoogleAI on 2019-12-20 (UTC).

ALBERT is a new, open-source architecture for natural language processing that achieves state-of-the-art performance on multiple benchmarks with ~30% fewer parameters than #BERT. Learn all about it below: https://t.co/oYriFcwrQo

— Google AI (@GoogleAI) December 20, 2019
nlpresearch

Tags

learning tutorial misc nlp rstats gan ethics research dataviz survey python tool security kaggle video thought bayesian humour tensorflow w_code bias dataset pytorch cv tip application javascript forecast swift golang rl jax julia gnn causal surey diffusion
© Copyright Philosophy 2018 Site Template by Colorlib