TY - GEN
T1 - Synthetic Dataset for Evaluating Complex Compositional Knowledge for Natural Language Inference
AU - Akoju, Sushma Anand
AU - Vacareanu, Robert
AU - Riaz, Haris
AU - Surdeanu, Mihai
AU - Blanco, Eduardo
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - We introduce a synthetic dataset called Sentences Involving Complex Compositional Knowledge (SICCK) and a novel analysis that investigates the performance of Natural Language Inference (NLI) models to understand compositionality in logic. We produce 1,304 sentence pairs by modifying 15 examples from the SICK dataset (Marelli et al., 2014). To this end, we modify the original texts using a set of phrases - modifiers that correspond to universal quantifiers, existential quantifiers, negation, and other concept modifiers in Natural Logic (NL) (MacCartney, 2009). We use these phrases to modify the subject, verb, and object parts of the premise and hypothesis. Lastly, we annotate these modified texts with the corresponding entailment labels following NL rules. We conduct a preliminary verification of how well the change in the structural and semantic composition is captured by neural NLI models, in both zero-shot and fine-tuned scenarios. We found that the performance of NLI models under the zero-shot setting is poor, especially for modified sentences with negation and existential quantifiers. After fine-tuning this dataset, we observe that models continue to perform poorly over negation, existential and universal modifiers.
AB - We introduce a synthetic dataset called Sentences Involving Complex Compositional Knowledge (SICCK) and a novel analysis that investigates the performance of Natural Language Inference (NLI) models to understand compositionality in logic. We produce 1,304 sentence pairs by modifying 15 examples from the SICK dataset (Marelli et al., 2014). To this end, we modify the original texts using a set of phrases - modifiers that correspond to universal quantifiers, existential quantifiers, negation, and other concept modifiers in Natural Logic (NL) (MacCartney, 2009). We use these phrases to modify the subject, verb, and object parts of the premise and hypothesis. Lastly, we annotate these modified texts with the corresponding entailment labels following NL rules. We conduct a preliminary verification of how well the change in the structural and semantic composition is captured by neural NLI models, in both zero-shot and fine-tuned scenarios. We found that the performance of NLI models under the zero-shot setting is poor, especially for modified sentences with negation and existential quantifiers. After fine-tuning this dataset, we observe that models continue to perform poorly over negation, existential and universal modifiers.
UR - http://www.scopus.com/inward/record.url?scp=85175397948&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85175397948&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85175397948
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 157
EP - 168
BT - 1st Workshop on Natural Language Reasoning and Structured Explanations, NLRSE 2023 @ACL 2023 - Proceedings of the Workshop
A2 - Mishra, Bhavana Dalvi
A2 - Durrett, Greg
A2 - Jansen, Peter
A2 - Ribeiro, Danilo Neves
A2 - Wei, Jason
PB - Association for Computational Linguistics (ACL)
T2 - 1st Workshop on Natural Language Reasoning and Structured Explanations, NLRSE 2023, co-located with ACL 2023
Y2 - 13 June 2023
ER -