Some people requested the blog post version so here it is. Hopefully this is easier to read ๐https://t.co/pFe4HxNiFf
โ Chip Huyen (@chipro) February 8, 2022
Some people requested the blog post version so here it is. Hopefully this is easier to read ๐https://t.co/pFe4HxNiFf
โ Chip Huyen (@chipro) February 8, 2022
The last couple of weeks, I took a deep dive into @PyTorchLightnin and am positively surprised how flexible it is for research. Just created a tutorial implementing our recent CORN method for ordinal regression: https://t.co/SMNGlY3FJa
โ Sebastian Raschka (@rasbt) January 28, 2022
I just published "Stan Algorithms: Where to Start?" @mcmc_stan https://t.co/2Ui3kuBP1Q
โ Daniel Lee (@djsyclik) January 19, 2022
A nice thread that explains intuitive why self-attention mechanism is needed in order to properly represent the relationships of the inputsโsomething that MLPโs struggle with. Looking forward to the next thread! https://t.co/s5Zrl6cjbH
โ hardmaru (@hardmaru) January 18, 2022
A good summary of various mental models useful for thinking about complex problems in many different areas, from science, systems, economics, to warfare.https://t.co/UHvjAWG7YE
โ hardmaru (@hardmaru) January 17, 2022
Good discussion about ConvNets vs Transformers https://t.co/MmPInthvFH
โ hardmaru (@hardmaru) January 13, 2022
The Annotated S4 (https://t.co/Qyylpwo6J6 /w @siddkaramcheti)
โ Sasha Rush (@srush_nlp) January 12, 2022
A step-by-step guide for building your own 16,000-gram language model... pic.twitter.com/tcVac4pSgd
After my post on real-time machine learning last year, many people asked me how to do it.
โ Chip Huyen (@chipro) January 3, 2022
This post discusses the challenges + solutions for online prediction, online evaluation, and continual learning, with use cases and examples.
Feedback appreciated!https://t.co/lywDVpYnUD
A beautiful music transformer visualization of the final attention heads from @ashVaswani 's talk on "attention is all you need" at RAAIS 2019 https://t.co/XwSIKr4nm3
โ Eric Jang ๐บ๐ธ๐น๐ผ (@ericjang11) December 26, 2021
The model learns to attend to periodic tokens when doing things like tremelos pic.twitter.com/znWCv5WWsU
an alternative way to imagine PCA: pic.twitter.com/VXXGBEDXic
โ Karl Rohe (@karlrohe) December 20, 2021
Here is the sequel to "Just ask for Generalization" - in this blog post I argue that Generalization *is* Language, and suggest how we might be able to re-use Language Models as "generalization modules" for non-NLP domains. Check it out!https://t.co/TI5FrbAU2R
โ Eric Jang ๐บ๐ธ๐น๐ผ (@ericjang11) December 17, 2021
Just came across this eminently readable and beginner friendly introduction to Transformers. Probably the best such text I've seen.
โ Bojan Tunguz (@tunguz) December 7, 2021
"Transformers from Scratch"https://t.co/qwI75Ezu9O#AI #ML #DS #NLP #artificialintelligence #machinelearning #datascience 1/2 pic.twitter.com/vvhFuzRdC9