-
Notifications
You must be signed in to change notification settings - Fork 4.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Load Huggingface Transformers model using TFAutoModel #6307
Comments
Exalate commented: sara-tagger commented: Thanks for submitting this feature request |
Exalate commented: ftarlaci commented: Hi, Is there any update on this enhancement? I would also like to use some of HF models that are currently not available in the HFTransformersNLP Component. Thank you! |
Exalate commented: dakshvar22 commented: The reason for not being able to use |
Exalate commented: patricebechard commented: Hi @dakshvar22, We were interested in looking into french models (FlauBERT, CamemBERT) for our project. Pretrained weights for these models are not available through the normal |
Exalate commented: dakshvar22 commented: Thanks for the info @patricebechard . Are you looking at them for intent classification and entity recognition or some other tasks? We recently changed the default model weights that |
Exalate commented: Ghostvv commented: @dakshvar22 if we cannot use auto model, shall we close the issue? |
Exalate commented: dakshvar22 commented: The plan is to review this once we update the transformers dependency. Once we are sure that it is not possible to use auto model with the latest dependencies as well, we should close this issue. |
Exalate commented: Ghostvv commented: shall we move it to blocked then? |
Exalate commented: mleimeister commented: @TyDunn Just to raise it to your awareness, I've started a side project to explore this. The current state is here: #10624 If I get to a point where it looks viable, I can check with Enable how this would play together with the scheduled TF/transformers upgrade. |
➤ Maxime Verger commented: 💡 Heads up! We're moving issues to Jira: https://rasa-open-source.atlassian.net/browse/OSS. From now on, this Jira board is the place where you can browse (without an account) and create issues (you'll need a free Jira account for that). This GitHub issue has already been migrated to Jira and will be closed on January 9th, 2023. Do not forget to subscribe to the corresponding Jira issue! ➡️ More information in the forum: https://forum.rasa.com/t/migration-of-rasa-oss-issues-to-jira/56569. |
Description of Problem: I want to use models provided by the HuggingFace transformers library that are not available when using the HFTransformersNLP Component.
Overview of the Solution: Right now, models are loaded using a dictionary of architectures in
rasa/nlu/utils/hugging_face/registry.py
. Using theAutoModel
andAutoTokenizer
features from thetransformers
library, we can get rid of these limitations, as well as not having to define amodel_name
in the config file. We can only use themodel_weights
info to select the proper model and weights of the wanted tokenizers and featurizers.Examples (if relevant): Docs for how to use the "auto" features is available here
Blockers (if relevant):
Definition of Done:
rasa/rasa/nlu/utils/hugging_face/
The text was updated successfully, but these errors were encountered: