Imagic: Text-Based Real Image Editing with Diffusion Models
— AK (@_akhaliq) October 18, 2022
abs: https://t.co/xRW6F6w2ZG pic.twitter.com/wGifY74i4w
Imagic: Text-Based Real Image Editing with Diffusion Models
— AK (@_akhaliq) October 18, 2022
abs: https://t.co/xRW6F6w2ZG pic.twitter.com/wGifY74i4w
Robust Preference Learning for Storytelling via Contrastive Reinforcement Learning
— AK (@_akhaliq) October 17, 2022
abs: https://t.co/tL0kvRqEpn pic.twitter.com/pMyAyPFUNk
Prompt-to-Prompt: Latent Diffusion and Stable Diffusion implementation with @huggingface diffusers is out
— AK (@_akhaliq) October 16, 2022
github: https://t.co/B4YcBt7vgo pic.twitter.com/QoIsax3xB1
Mass Editing Memory in a Transformer
— AK (@_akhaliq) October 14, 2022
abs: https://t.co/AYwoYmmmaw
project page: https://t.co/Xsnrh3EtJ8
github: https://t.co/ufsbkROY2d pic.twitter.com/ZNJyW1xOn4
Scalable Neural Video Representations with Learnable Positional Features
— AK (@_akhaliq) October 14, 2022
abs: https://t.co/r8ofrXivQ0
project page: https://t.co/nHTBO27iOG pic.twitter.com/4I7YRnIzGT
Unifying Diffusion Models' Latent Space, with Applications to CycleDiffusion and Guidance
— AK (@_akhaliq) October 12, 2022
abs: https://t.co/d3NRBZTL4a pic.twitter.com/sP0PLwfHHk
This looks like the Vision Transformers architecture we have been waiting for: MaxViT https://t.co/WbzgJ50PjB
— Martin Görner (@martin_gorner) October 11, 2022
1/ State of the Art accuracy on ImageNet (no pre-training on huge datasets)
2/ Linear complexity wrt. image size (thanks to a clever attention design) pic.twitter.com/5bW0N7n3s5
All the related work section should be like the one of the recent "Attention Beats Concatenation for Conditioning Neural Fields" of Rebain et al. https://t.co/MBQfH7sE6S
— Thomas Wolf (@Thom_Wolf) October 11, 2022
Look at this cool git-like branching graph putting all the cited work in perspective! pic.twitter.com/HXLSyXe4xE
Understanding HTML with Large Language Models
— AK (@_akhaliq) October 11, 2022
abs: https://t.co/nKCCunfLxr
project page: https://t.co/cktfbuul4R pic.twitter.com/RzOaKOvGqP
MaxViT : combines ConvNet modules and 2 types of self attention (local n'y block, and on a subsampled grid).
— Yann LeCun (@ylecun) October 10, 2022
Since DETR (hi @alcinos26 !), I've become convinced that combining Conv and attention/dynamic routing was the Right Thing. https://t.co/DNOBsqL54Z
GLM-130B reaches INT4 quantization w/ no perf degradation, allowing effective inference on 4*3090 or 8*2080 Ti GPUs, the most ever affordable GPUs required for using 100B-scale models?
— Tsinghua KEG (@thukeg) October 10, 2022
Paper: https://t.co/f2bj1N8JTN
Model weights & code & demo & lessons: https://t.co/aKZNGEDmks pic.twitter.com/kVRV0b8Y56
Content-Based Search for Deep Generative Models
— AK (@_akhaliq) October 7, 2022
abs: https://t.co/6yAYV5XNqO
project page: https://t.co/fTF1qDsYyh pic.twitter.com/jxEDXIagrJ