Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add conversion support for CodeLlama 34B #25737

Closed
wants to merge 1 commit into from

Conversation

AlpinDale
Copy link
Contributor

There hasn't been any architectural changes, and I was able to convert the 34B by simply adding in the intermediate size for the model. I have converted the 34B model and it appears to work well.

What does this PR do?

This PR should make it possible to convert the CodeLlama models, recently released by Meta AI.

There hasn't been any architectural changes, and I was able to convert the 34B by simply adding in the intermediate size for the model.
@AlpinDale
Copy link
Contributor Author

Doesn't seem to work for 7B and 13B variants.

@AlpinDale
Copy link
Contributor Author

Conversion doesn't account for the modified tokenizer and the new rope theta value. Closing the PR until someone else implements the new changes.

@AlpinDale AlpinDale closed this Aug 24, 2023
@ArthurZucker
Copy link
Collaborator

See #25740

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants