Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About Visualization Tools #2

Open
KaKa-101 opened this issue Dec 19, 2024 · 4 comments
Open

About Visualization Tools #2

KaKa-101 opened this issue Dec 19, 2024 · 4 comments

Comments

@KaKa-101
Copy link

Thanks for your great work, which is significant for this area.
Could you please provide the visualization tool for Figure 3 ? I am really interested in the visual results of retained image patches and want to try it on my own.
Thanks a lot.

@ChimpOnCloud
Copy link
Collaborator

lol you mean fig.2? Currently we just picked random samples and tracked attention maps to get these data. You can simply get attention distribution in llava_llama.py and get [CLS] attention distribution of visual encoder in clip_encoder.py. Later we will release the complete visualization tool

@KaKa-101
Copy link
Author

KaKa-101 commented Dec 23, 2024

Sry, actually I mean fig.4

@ChimpOnCloud
Copy link
Collaborator

ChimpOnCloud commented Dec 23, 2024

For fig.4, currently we just filtered those patches with top [CLS] attention scores, and manually marked each object with different color for paper readers to see the effectiveness of pruning with [CLS] attention. We consider introducing some automatic tools like SAM to mark different objects in the near future.

@KaKa-101
Copy link
Author

Thanks for your kind reply.
Looking forward to the release of your complete visualization tools~

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants