Opens profile photo
Follow
Hila Chefer
@hila_chefer
PhD candidate @ Tel Aviv University, student researcher , interested in Deep Learning, Computer Vision, NLP, explainable AI @Hila_Chefer@sigmoid.social
hila-chefer.github.ioJoined December 2020

Hila Chefer’s Tweets

Excited to share that we will be presenting our work in person at #NeurIPS2022 ! Interested in leveraging explainability to improve accuracy and robustness? Come check out our poster and chat 🥳 Code: github.com/hila-chefer/Ro demo: huggingface.co/spaces/Hila/Ro
Quote Tweet
[1/n] Can explainability improve model accuracy? Our latest work shows the answer is yes! arxiv.org/pdf/2206.01161 github.com/hila-chefer/Ro We noticed that ViTs suffer from salient issues- their output is often based on supportive signals (background) rather than the actual object
Show this thread
Image
4
130
Show this thread
info on the papers:
Quote Tweet
[1/n] @eccvconf #ECCV2022 paper thread! 1. Image-based CLIP-Guided Essence Transfer (TargetCLIP)- we extract the essence of a target while preserving realism and source identity. 2. No Token Left Behind- we use explainability to stabilize the unreliable CLIP similarity scores. twitter.com/_akhaliq/statu…
Show this thread
2
Show this thread
[9/n] No Token Left Behind- Using this additional explainability loss, we demonstrate that downstream tasks such as image classification and image generation and editing that use CLIP guidance can be significantly improved.
Image
Image
1
4
Show this thread
[7/n] No Token Left Behind- RT : twitter.com/_akhaliq/statu We find that CLIP similarity scores can be unreliable since they rely on a small subset of the text tokens. E.g., the noisy image scores higher due to the influence of the words "image of" on the similarity score.
Image
Quote Tweet
Image
No Token Left Behind: Explainability-Aided Image Classification and Generation abs: arxiv.org/abs/2204.04908
1
14
Show this thread
[5/n] TargetCLIP - As an alternative to long optimization, we show that one can also fine-tune an inversion encoder to output the essence vector of a target, allowing for instant extraction of the essence for each target!
Image
1
3
Show this thread
[3/n] TargetCLIP - We demonstrate that using CLIP guidance and the powerful StyleGAN, we can extract an essence vector- a vector of semantic properties that correspond to the “signature characteristics” that are also identified by humans as related to the specific target.
Image
2
6
Show this thread
[1/n] #ECCV2022 paper thread! 1. Image-based CLIP-Guided Essence Transfer (TargetCLIP)- we extract the essence of a target while preserving realism and source identity. 2. No Token Left Behind- we use explainability to stabilize the unreliable CLIP similarity scores.
Quote Tweet
Image-Based CLIP-Guided Essence Transfer abs: arxiv.org/abs/2110.12427 github: github.com/hila-chefer/Ta new method creates a blending operator that is optimized to be simultaneously additive in both latent spaces
Image
2
127
Show this thread
Check out our new demo (huggingface.co/spaces/Hila/Ro)- Even for out of domain inputs such as images generated by DALL-E 2 or animations, our method corrects the original model to produce a plausible prediction! (code: github.com/hila-chefer/Ro)
Image
Image
Quote Tweet
Image
Optimizing Relevance Maps of Vision Transformers Improves Robustness abs: arxiv.org/abs/2206.01161 github: github.com/hila-chefer/Ro show that by finetuning the explainability maps of ViTs, a significant increase in robustness can be achieved
35
The most popular Arxiv link yesterday:
Quote Tweet
[1/n] Can explainability improve model accuracy? Our latest work shows the answer is yes! arxiv.org/pdf/2206.01161 github.com/hila-chefer/Ro We noticed that ViTs suffer from salient issues- their output is often based on supportive signals (background) rather than the actual object
Show this thread
Image
8
Thanks for sharing our work ! Check out the thread with more details on the paper 👇 demo coming soon!
Quote Tweet
[1/n] Can explainability improve model accuracy? Our latest work shows the answer is yes! arxiv.org/pdf/2206.01161 github.com/hila-chefer/Ro We noticed that ViTs suffer from salient issues- their output is often based on supportive signals (background) rather than the actual object
Show this thread
Image
35
Cool, a clever usage of explainabiliy for fine-tuning ViT 🧠
Quote Tweet
[1/n] Can explainability improve model accuracy? Our latest work shows the answer is yes! arxiv.org/pdf/2206.01161 github.com/hila-chefer/Ro We noticed that ViTs suffer from salient issues- their output is often based on supportive signals (background) rather than the actual object
Show this thread
Image
4
In this work, we show that one can *directly optimize the explanations*- i.e. use a loss on the explainability signal to ensure that the classification is based on the *right reasons*- the foreground and not the background.
Image
1
12
Show this thread