Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

VLMs: major clean up 🧼 #34502

Open
wants to merge 20 commits into
base: main
Choose a base branch
from
Open

Conversation

zucchini-nlp
Copy link
Member

@zucchini-nlp zucchini-nlp commented Oct 30, 2024

What does this PR do?

We have updated all the configs for VLMs on the hub so this PR removes legacy path for models, as it has been there for already 3 releases from v4.44. Also it fixes some stuff that broke on the way, like generating from only text input in LLaVA models

For Video-LLaVA the hub configs cannot be updated as the hub owner has been silent for several mmonths already. And since there is only one model with such architecture, we can hardcode the default values for patch_num and also remove the legacy path

fixes #34824, fixes #35169

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Collaborator

@ArthurZucker ArthurZucker left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I don't think we need this, we deprecated the legacy path, we can just remove it now no?
I don't remember what we said for for 4.46 but better to go with non legacy now if we can!

@zucchini-nlp
Copy link
Member Author

zucchini-nlp commented Oct 30, 2024

We can remove it after updating the files on the hub and that mean we also need to change warning to error so users have chance to see what is the reason for failure.

I think the earliest we can remove is next release, because the blocking PR will prob be merged next week. After that I will take time to update all hub configs. Maybe then we'll wait for the blocking PR and remove all deprecation warnings?

@zucchini-nlp zucchini-nlp changed the title Fix llava tests VLMs: major clean up 🧼 Nov 24, 2024
@ArthurZucker
Copy link
Collaborator

Sounds good, let's wait a bit!

@zucchini-nlp
Copy link
Member Author

@ArthurZucker i think this can be review now :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

LlavaForConditionalGeneration._merge_input_ids_with_image_features throws error Flash attention 2 broke when batch inference
3 participants