Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[New Model]: Cogagent #4575

Open
leoozy opened this issue May 3, 2024 · 2 comments
Open

[New Model]: Cogagent #4575

leoozy opened this issue May 3, 2024 · 2 comments
Labels
new model Requests to new models

Comments

@leoozy
Copy link

leoozy commented May 3, 2024

The model to consider.

https://huggingface.co/THUDM/CogAgent

The closest model vllm already supports.

No response

What's your difficulty of supporting the model you want?

Vision models

@leoozy leoozy added the new model Requests to new models label May 3, 2024
@JBurtn
Copy link

JBurtn commented Jun 22, 2024

Need to Wait for / help with #4888 and #4942 before this can be implemented. Maybe even some more stuff.

@afeldman-nm
Copy link
Contributor

Need to Wait for / help with #4888 and #4942 before this can be implemented. Maybe even some more stuff.

Quick update:

#4888 is landed, enabling the xFormers backend to support encoder attention, decoder self-attention, and decoder cross-attention. #4837 and #4888 (both of which have been landed) were prerequisites for #4942 . #4942
completes end-to-end support for encoder/decoder models with the xFormers backend & also introduces the BART model into vLLM. #4942 is still WIP but hoping to complete it soon.

@robertgshaw2-neuralmagic

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
new model Requests to new models
Projects
None yet
Development

No branches or pull requests

3 participants