Markchom, T., Dhruva, B., Pravin, C. and Liang, H. (2020) UoR at SemEval-2020 task 4: pre-trained sentence transformer models for commonsense validation and explanation. In: International Workshop on Semantic Evaluation 2020, December 12-13, 2020, Barcelona, Spain, pp. 430-436.
Abstract/Summary
SemEval Task 4 Commonsense Validation and Explanation Challenge is to validate whether a system can differentiate natural language statements that make sense from those that do not make sense. Two subtasks, A and B, are focused in this work, i.e., detecting against-common-sense statements and selecting explanations of why they are false from the given options. Intuitively, commonsense validation requires additional knowledge beyond the given statements. Therefore, we propose a system utilising pre-trained sentence transformer models based on BERT, RoBERTa and DistillBERT architectures to embed the statements before classification. According to the results, these embeddings can improve the performance of the typical MLP and LSTM classifiers as downstream models of both subtasks compared to regular tokenised statements. These embedded statements are shown to comprise additional information from external resources which help validate common sense in natural language.
Item Type | Conference or Workshop Item (Paper) |
URI | https://reading-clone.eprints-hosting.org/id/eprint/92759 |
Item Type | Conference or Workshop Item |
Refereed | Yes |
Divisions | Science > School of Mathematical, Physical and Computational Sciences > Department of Computer Science |
Download/View statistics | View download statistics for this item |
Downloads
Downloads per month over past year
University Staff: Request a correction | Centaur Editors: Update this record