Skip to content

yg211/explainable-metrics

Repository files navigation

Explainable NLG Metrics

This project aims at explaining state-of-the-art NLG metrics, including

We provide explanations by breaking down the score to show the contribution of each word. The break down scores are computed using the SHAP method.

sts-example

The above example uses BertScore to measure the semantic similarity between sentences. It shows that the contribution of word hates is negative, suggesting that its appearance harms the similarity score.

xmover-example In the example above, the quality of a translation is measured by XMoverScore, by comparing the semantic similarity between the source and the translation (without using any references). The score breakdown suggests that word dislikes harms the score.

More monolingual examples can be found at here, and crosslingual examples can be found at here

Contact person: Yang Gao@Royal Holloway, Unversity of London. Don't hesitate to drop me an e-mail if something is broken or if you have any questions.

License

Apache License Version 2.0

Releases

No releases published

Packages

No packages published