You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The f1 metric from the package seems to be wrongly implemented. The first two lines are already an indication (unique), making the rest of the code also wrong.
This is wrong as one would expect (for binary classification) with 1 as the positive label, an F1 score of 0.86. If the positive label were 0 the F1 score would be 0.8.
The text was updated successfully, but these errors were encountered:
Hi! Thank you for raising this issue. I agree that this function is confusing. In the documentation for the function, I explain that this is actually the f1 score in the context of information retrieval problems. It is not the f1 score in the context of binary classification problems. I recognize that the name f1 score is more commonly used in binary classification, which is why this function is confusing. I tried to make that clear in the documentation. What do you think about the documentation?
Thanks for the quick reply. I have to admit that I didn't know exactly what "information retrieval context" meant. I was actually using this package indirectly from the iml package which uses it as a dependency.
Thanks for pointing that out and I also notice that you implemented an fbeta_score which works for my case for beta = 1.
As a suggestion, additional thing that would be great to see from the package would be an f1 score for multi-class cases (i.e., micro and macro cases).
The f1 metric from the package seems to be wrongly implemented. The first two lines are already an indication
(unique)
, making the rest of the code also wrong.Reproducible example:
This is wrong as one would expect (for binary classification) with
1
as the positive label, an F1 score of0.86
. If the positive label were0
the F1 score would be0.8
.The text was updated successfully, but these errors were encountered: