Replies: 2 comments 6 replies
-
Quoting myself from #147 for clarity:
As for your question:
This does not make much sense. What you need is some data to finetune on. With (most) LLMs, instead of having an encoder encode the source data, and a decoder decode the target data, you only have a decoder which "reads" the prompt, and continues generation. So you need to show your model how it's supposed to perform such generation. Hence the prompt basically saying [Task][SRC][TGT]. Then you finetune the LLM like any other LLM (see llama2 finetuning for an example, can easily be extended to other models). |
Beta Was this translation helpful? Give feedback.
-
Read this: https://huggingface.co/Unbabel/TowerInstruct-Mistral-7B-v0.2/discussions/3 |
Beta Was this translation helpful? Give feedback.
-
Hi Iʻm keen to test fine-tuning Llama 2 or 3.1 with my bilingual datasets. In the recipe provided for wmt22_with_TowerInstruct-llama2 I donʻt see a yaml config file to train from the Llama model.
Can you explain briefly how I can go about fine-tuning an LLM with my bilingual datasets?
Also, does the promptize_llama2.py script acts like a prompt to start the fine-tuning training?
Beta Was this translation helpful? Give feedback.
All reactions