Evaluation metric using BERT sentence representations
BERTScore is a metric for automatic evaluation of machine translation that calculates the similarity between a machine translation output and a reference translation using sentence representation.
BERTScore addresses two common pitfalls in n-gram-based metrics.
[…] First, such methods often fail to robustly match paraphrases.
[…] Second, n-gram models fail to capture distant dependencies and penalize semantically-critical ordering changes.
For example, given a small window of size two, BLEU will only mildly penalize swapping of cause and effect clauses (e.g. A because B instead of B because A), especially when the arguments A and B are long phrases.
In contrast, contextualized embeddings are trained to effectively capture distant dependencies and ordering.
The BERTScore metric uses sentence representations from BERT, a deep learning model.
BERTScore computes precision, recall, and F1 measure.