Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does it support multi-modal LLM like LLaVa? #1751

Closed
Privilger opened this issue Nov 22, 2023 · 6 comments
Closed

Does it support multi-modal LLM like LLaVa? #1751

Privilger opened this issue Nov 22, 2023 · 6 comments

Comments

@Privilger
Copy link

The given example shows how to use for LLM like llama.
But how can I use it to accelerate Visual Language Model like LLaVa?

@simon-mo
Copy link
Collaborator

simon-mo commented Nov 22, 2023

LLaVa needs to be added with our lightweight model addition process. Contribution welcomed! https://docs.vllm.ai/en/latest/models/adding_model.html

@l4b4r4b4b4
Copy link

LLaVa needs to be added with our lightweight model addition process. Contribution welcomed! https://docs.vllm.ai/en/latest/models/adding_model.html

You dont generally see an issue in using the lighweight model adidtion for the composed llava model with CLIP, vision encoder and LLM?

@simon-mo
Copy link
Collaborator

It does turns out to be more complex. A working PR is here #3042

@ywang96
Copy link
Member

ywang96 commented Apr 18, 2024

Closing as #3042 adds support for LLaVA1.5

@ywang96 ywang96 closed this as completed Apr 18, 2024
@iamsaurabhgupt
Copy link

is LLaVA-Next supported?
couldn't find a working example

@ywang96
Copy link
Member

ywang96 commented Jun 26, 2024

@iamsaurabhgupt Feel free to take a look at the example here! #4194 (comment). One limitation we have now is that we don't support dynamic shape of input image, so the results might be slightly different from huggingface implementation, but we're on track to support this very soon!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants