-
Notifications
You must be signed in to change notification settings - Fork 30
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ways to compute ROC-AUC and the label #262
Comments
Hi, thank you for your question.
|
Here is an example for launching the benchmark scripts:
|
When I'm labeling, I find that sometimes it outputs' nan '. When calculating ROC-AUC, is it filtered along with the corresponding uncertainty score? |
Hi, thank you for your question! Yes, when the uncertainty estimator outputs 'NaN' for certain claims, those claims are excluded from the ROC-AUC calculations. Additionally, some claims cannot be determined as either 'True' or 'False' (e.g., when GPT-4 outputs 'Not known' instead of a definitive 'True' or 'False'). These claims are labeled as 'NaN' in OpenAIFactCheck and are similarly excluded from the computations. |
Thank you again for providing this repository. I would like to know how to use this repository to evaluate the obtained uncertainty after obtaining it, and how the labels used in the evaluation process are obtained, such as the ROC-AUC value of the CCP method based on biographical claims. Could you please provide an example?
The text was updated successfully, but these errors were encountered: