You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I was trying to run the the IKE editing method on the ZsRE dataset through running the run_knowedit_llama2.py. I tracked the code to check how the metrics to check how the metrics are being computed and I have some questions.
To compute the edit success, the edit_evaluation function calls the compute_icl_edit_quality in the evaluate.py file. Depending on the bool value of pre_edit, the function either passes New Fact: {prompt} {target_new}\nPrompt: {prompt} or the prompt. Why do you change the prompt based on the value of the pre_edit and what difference does it make ?
To compute the edit success, The run_knowedit_llama2.py script does not pass the ground truth values of the edit prompts to the .edit method so when computing the metric the ground truth is <|endoftext|>. Why aren't the groundtruth values passed and what does the success of the edit computed against ?
When computing the locality or portability, you prepend the edited fact to the locality/protability prompts. To edit an instruct tuned model, should one account for the adding the chat template ?
Thanks a lot for your efforts and I am looking forward for your reply.
The text was updated successfully, but these errors were encountered:
pre_edit is the performance of the model before we edit, so we do not pass the `New Fact'.
Well since edit donot require the ground_truth', we just set it endoftext', we only require target_new'. You can change ground truth' as you need.
Yes, we recommend using the chat_template, but in previous work we just used the original prompt. In our new paper CKnowedit, we have applied chat_template and you can select different settings via different settings: editor.generate_edit
Hi
Thanks for the amazing efforts on this toolkit.
I was trying to run the the IKE editing method on the ZsRE dataset through running the
run_knowedit_llama2.py
. I tracked the code to check how the metrics to check how the metrics are being computed and I have some questions.To compute the edit success, the
edit_evaluation
function calls thecompute_icl_edit_quality
in theevaluate.py
file. Depending on the bool value of pre_edit, the function either passesNew Fact: {prompt} {target_new}\nPrompt: {prompt}
or the prompt. Why do you change the prompt based on the value of the pre_edit and what difference does it make ?To compute the edit success, The
run_knowedit_llama2.py
script does not pass the ground truth values of the edit prompts to the.edit
method so when computing the metric the ground truth is<|endoftext|>
. Why aren't the groundtruth values passed and what does the success of the edit computed against ?When computing the locality or portability, you prepend the edited fact to the locality/protability prompts. To edit an instruct tuned model, should one account for the adding the chat template ?
Thanks a lot for your efforts and I am looking forward for your reply.
The text was updated successfully, but these errors were encountered: