You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I noticed that Llama-3.1 has a rope_scaling parameter set in config.json, and it has also changed the field name for type to rope_type. Moreover, the rope_scaling parameter in the previous Llama-3 was null.
This causes an error in the conversion process of "ctranslate2\converters\transformers.py" whenever it reads rope_scaling["type"]. However, this can be easily fixed by writing the code as follows:
But this is not entirely sufficient because the RoPE scaling for Llama-3 has not been implemented. Thus, during the conversion, the following error occurs:
NotImplementedError: RoPE scaling type 'llama3' is not yet implemented. The following RoPE scaling types are currently supported: linear, su
Hello again, if you plan on supporting LLama3.1 please note that it requires a new category of ROPE scaling. Thanks!
https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct/blob/main/config.json
The text was updated successfully, but these errors were encountered: