Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inability to apply qwen2-vl or llama-3.2-vision #1100

Open
Yoo-Youngjae opened this issue Jan 24, 2025 · 0 comments
Open

Inability to apply qwen2-vl or llama-3.2-vision #1100

Yoo-Youngjae opened this issue Jan 24, 2025 · 0 comments

Comments

@Yoo-Youngjae
Copy link

Yoo-Youngjae commented Jan 24, 2025

Is your feature request related to a problem? Please describe.

I would like to leverage guidance on vlm models such as qwen2-vl or llama3.2-vision.

However, these are difficult to use because they use the latest transformer's MllamaForConditionalGeneration and Qwen2VLForConditionalGeneration, which causes an error in the current guidance.

Is it possible to update these to support them more widely?

Additional context

Or is it because I am not good at using it?

If anyone has solved it, please let me know the solution.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant