[Adapt] What Can We Learn from Collective Human Opinions on Natural Language Inference Data?

Li Zitong AutSky_JadeK at outlook.com
Wed Sep 15 02:00:43 CST 2021

Hi Adapters,

This week, I will give you the talk about a paper from EMNLP2020 : "What Can We Learn from Collective Human Opinions on Natural Language Inference Data?".

Despite the subjective nature of many NLP tasks, most NLU evaluations have focused on using the majority label with presumably high agreement as the ground truth. Less attention has been paid to the distribution of human opinions. The authors collect ChaosNLI, a dataset with a total of 464,500 annotations to study Collective HumAn OpinionS in oft-used NLI evaluation sets. This dataset is created by collecting 100 annotations per example for 3,113 examples in SNLI and MNLI and 1,532 examples in αNLI. Analysis reveals that: (1) high human disagreement exists in a noticeable amount of examples in these datasets; (2) the state-of-the-art models lack the ability to recover the distribution over human labels; (3) models achieve near-perfect accuracy on the subset of data with a high level of human agreement, whereas they can barely beat a random guess on the data with low levels of human agreement, which compose most of the common errors made by state-of-the-art models on the evaluation sets. This questions the validity of improving model performance on old metrics for the low-agreement part of evaluation datasets. Hence, the authors argue for a detailed examination of human agreement in future data collection efforts, and evaluating model outputs against the distribution over collective human opinions.

I hope my talk makes you feel interesting and helpful.

Time: Wed 4:00pm
Venue: SEIEE 3-414
Best wishes.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://cs.sjtu.edu.cn/pipermail/adapt/attachments/20210914/0c2c9f92/attachment.htm>

More information about the Adapt mailing list