Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[New Model]: Support Nemotron-4-340B #5722

Closed
dskhudia opened this issue Jun 20, 2024 · 3 comments · Fixed by #6611
Closed

[New Model]: Support Nemotron-4-340B #5722

dskhudia opened this issue Jun 20, 2024 · 3 comments · Fixed by #6611
Labels
new model Requests to new models

Comments

@dskhudia
Copy link
Contributor

🚀 The feature, motivation and pitch

Benchmarks on the Nvidia's latest nemotron model look great. Is there plan or already going work to support it?

Alternatives

No response

Additional context

https://huggingface.co/nvidia/Nemotron-4-340B-Instruct

@mgoin
Copy link
Collaborator

mgoin commented Jun 20, 2024

Once there is support in HF transformers, then it should be relatively straightforward to port into vLLM. It seems there aren't any efforts from searching the transformers issues/PRs

@WoosukKwon WoosukKwon added new model Requests to new models and removed feature request labels Jun 21, 2024
@riverind
Copy link

attention

@DarkLight1337 DarkLight1337 changed the title [Feature]: Support Nemotron-4-340B [Model]: Support Nemotron-4-340B Jun 27, 2024
@DarkLight1337 DarkLight1337 changed the title [Model]: Support Nemotron-4-340B [New Model]: Support Nemotron-4-340B Jun 27, 2024
@natolambert
Copy link

I started a paid bounty to close these issues. Already over $200 of support.
https://x.com/natolambert/status/1814735390877884823

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
new model Requests to new models
Projects
None yet
Development

Successfully merging a pull request may close this issue.

5 participants