TY - GEN
T1 - Semi-supervised teacher-student architecture for relation extraction
AU - Luo, Fan
AU - Nagesh, Ajay
AU - Sharp, Rebecca
AU - Surdeanu, Mihai
N1 - Funding Information:
This work was supported by the Defense Advanced Research Projects Agency (DARPA) under the World Modelers program, grant number W911NF1810014, and by the Bill and Melinda Gates Foundation HBGDki Initiative. Mihai Sur-deanu declares a financial interest in lum.ai. This interest has been properly disclosed to the University of Arizona Institutional Review Committee and is managed in accordance with its conflict of interest policies.
Publisher Copyright:
© 2019 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - Generating a large amount of training data for information extraction (IE) is either costly (if annotations are created manually), or runs the risk of introducing noisy instances (if distant supervision is used). On the other hand, semi-supervised learning (SSL) is a cost-efficient solution to combat lack of training data. In this paper, we adapt Mean Teacher (Tarvainen and Valpola, 2017), a denoising SSL framework to extract semantic relations between pairs of entities. We explore the sweet spot of amount of supervision required for good performance on this binary relation extraction task. Additionally, different syntax representations are incorporated into our models to enhance the learned representation of sentences. We evaluate our approach on the Google-IISc Distant Supervision (GDS) dataset, which removes test data noise present in all previous distance supervision datasets, which makes it a reliable evaluation benchmark (Jat et al., 2017). Our results show that the SSL Mean Teacher approach nears the performance of fully-supervised approaches even with only 10% of the labeled corpus. Further, the syntax-aware model outperforms other syntax-free approaches across all levels of supervision.
AB - Generating a large amount of training data for information extraction (IE) is either costly (if annotations are created manually), or runs the risk of introducing noisy instances (if distant supervision is used). On the other hand, semi-supervised learning (SSL) is a cost-efficient solution to combat lack of training data. In this paper, we adapt Mean Teacher (Tarvainen and Valpola, 2017), a denoising SSL framework to extract semantic relations between pairs of entities. We explore the sweet spot of amount of supervision required for good performance on this binary relation extraction task. Additionally, different syntax representations are incorporated into our models to enhance the learned representation of sentences. We evaluate our approach on the Google-IISc Distant Supervision (GDS) dataset, which removes test data noise present in all previous distance supervision datasets, which makes it a reliable evaluation benchmark (Jat et al., 2017). Our results show that the SSL Mean Teacher approach nears the performance of fully-supervised approaches even with only 10% of the labeled corpus. Further, the syntax-aware model outperforms other syntax-free approaches across all levels of supervision.
UR - http://www.scopus.com/inward/record.url?scp=85092915434&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85092915434&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85092915434
T3 - NLP@NAACL-HLT 2019 - 3rd Workshop on Structured Prediction for NLP, Proceedings
SP - 29
EP - 37
BT - NLP@NAACL-HLT 2019 - 3rd Workshop on Structured Prediction for NLP, Proceedings
PB - Association for Computational Linguistics (ACL)
T2 - 3rd Workshop on Structured Prediction for NLP, NLP@NAACL-HLT 2019
Y2 - 7 June 2019
ER -