Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] support qwen2 vl #1352

Closed
2 tasks done
zhyncs opened this issue Sep 8, 2024 · 5 comments
Closed
2 tasks done

[Feature] support qwen2 vl #1352

zhyncs opened this issue Sep 8, 2024 · 5 comments
Assignees

Comments

@zhyncs
Copy link
Member

zhyncs commented Sep 8, 2024

Checklist

Motivation

https://huggingface.co/collections/Qwen/qwen2-vl-66cee7455501d7126940800d

Related resources

No response

@merrymercy
Copy link
Contributor

cc @JustinLin610

@HiDostoyevskiy
Copy link

I eagerly look forward to the support for the qwen2-vl series models by sglang as soon as possible. Thanks!

@zhyncs
Copy link
Member Author

zhyncs commented Sep 19, 2024

I eagerly look forward to the support for the qwen2-vl series models by sglang as soon as possible. Thanks!

@ispobock's partner is currently working on it.

@bash99
Copy link

bash99 commented Sep 21, 2024

I would like to know will sglang support qwen2 vl gptq-int4 release with more than one GPU?

https://qwen.readthedocs.io/en/latest/quantization/gptq.html said vllm don't work and need padding and quantizated again.

@zhyncs
Copy link
Member Author

zhyncs commented Oct 7, 2024

ref #1546

@zhyncs zhyncs closed this as completed Oct 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants