TY - GEN
T1 - Semi-supervised teacher-student architecture for relation extraction
AU - Luo, Fan
AU - Nagesh, Ajay
AU - Sharp, Rebecca
AU - Surdeanu, Mihai
N1 - Publisher Copyright:
© 2019 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - Generating a large amount of training data for information extraction (IE) is either costly (if annotations are created manually), or runs the risk of introducing noisy instances (if distant supervision is used). On the other hand, semi-supervised learning (SSL) is a cost-efficient solution to combat lack of training data. In this paper, we adapt Mean Teacher (Tarvainen and Valpola, 2017), a denoising SSL framework to extract semantic relations between pairs of entities. We explore the sweet spot of amount of supervision required for good performance on this binary relation extraction task. Additionally, different syntax representations are incorporated into our models to enhance the learned representation of sentences. We evaluate our approach on the Google-IISc Distant Supervision (GDS) dataset, which removes test data noise present in all previous distance supervision datasets, which makes it a reliable evaluation benchmark (Jat et al., 2017). Our results show that the SSL Mean Teacher approach nears the performance of fully-supervised approaches even with only 10% of the labeled corpus. Further, the syntax-aware model outperforms other syntax-free approaches across all levels of supervision.
AB - Generating a large amount of training data for information extraction (IE) is either costly (if annotations are created manually), or runs the risk of introducing noisy instances (if distant supervision is used). On the other hand, semi-supervised learning (SSL) is a cost-efficient solution to combat lack of training data. In this paper, we adapt Mean Teacher (Tarvainen and Valpola, 2017), a denoising SSL framework to extract semantic relations between pairs of entities. We explore the sweet spot of amount of supervision required for good performance on this binary relation extraction task. Additionally, different syntax representations are incorporated into our models to enhance the learned representation of sentences. We evaluate our approach on the Google-IISc Distant Supervision (GDS) dataset, which removes test data noise present in all previous distance supervision datasets, which makes it a reliable evaluation benchmark (Jat et al., 2017). Our results show that the SSL Mean Teacher approach nears the performance of fully-supervised approaches even with only 10% of the labeled corpus. Further, the syntax-aware model outperforms other syntax-free approaches across all levels of supervision.
UR - https://www.scopus.com/pages/publications/85092915434
UR - https://www.scopus.com/pages/publications/85092915434#tab=citedBy
M3 - Conference contribution
AN - SCOPUS:85092915434
T3 - NLP@NAACL-HLT 2019 - 3rd Workshop on Structured Prediction for NLP, Proceedings
SP - 29
EP - 37
BT - NLP@NAACL-HLT 2019 - 3rd Workshop on Structured Prediction for NLP, Proceedings
PB - Association for Computational Linguistics (ACL)
T2 - 3rd Workshop on Structured Prediction for NLP, NLP@NAACL-HLT 2019
Y2 - 7 June 2019
ER -