| | --- |
| | task_categories: |
| | - text-classification |
| | - multiple-choice |
| | language: |
| | - en |
| | tags: |
| | - explanation |
| | --- |
| | https://github.com/wangcunxiang/Sen-Making-and-Explanation |
| | ``` |
| | @inproceedings{wang-etal-2019-make, |
| | title = "Does it Make Sense? And Why? A Pilot Study for Sense Making and Explanation", |
| | author = "Wang, Cunxiang and |
| | Liang, Shuailong and |
| | Zhang, Yue and |
| | Li, Xiaonan and |
| | Gao, Tian", |
| | booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", |
| | month = jul, |
| | year = "2019", |
| | address = "Florence, Italy", |
| | publisher = "Association for Computational Linguistics", |
| | url = "https://www.aclweb.org/anthology/P19-1393", |
| | pages = "4020--4026", |
| | abstract = "Introducing common sense to natural language understanding systems has received increasing research attention. It remains a fundamental question on how to evaluate whether a system has the sense-making capability. Existing benchmarks measure common sense knowledge indirectly or without reasoning. In this paper, we release a benchmark to directly test whether a system can differentiate natural language statements that make sense from those that do not make sense. In addition, a system is asked to identify the most crucial reason why a statement does not make sense. We evaluate models trained over large-scale language modeling tasks as well as human performance, showing that there are different challenges for system sense-making.", |
| | } |
| | ``` |