TY - GEN
T1 - Exploration of noise strategies in semi-supervised named entity classification
AU - Narayan, Pooja Lakshmi
AU - Nagesh, Ajay
AU - Surdeanu, Mihai
N1 - Funding Information:
★ This work has been supported by the Spanish R+D National Plan, project TIN2007-67407-C03-03, “BRAVO: Búsqueda de Respuestas Avanzada Multimodal y Multil-ingüe - Recuperación de Información” and by Madrid R+D Regional Plan, MAVIR Project, S-0505/TIC/000267 http://www.mavir.net
Publisher Copyright:
© 2019 Association for Computational Linguistics
PY - 2019
Y1 - 2019
N2 - Noise is inherent in real world datasets and modeling noise is critical during training as it is effective in regularization. Recently, novel semi-supervised deep learning techniques have demonstrated tremendous potential when learning with very limited labeled training data in image processing tasks. A critical aspect of these semi-supervised learning techniques is augmenting the input or the network with noise to be able to learn robust models. While modeling noise is relatively straightforward in continuous domains such as image classification, it is not immediately apparent how noise can be modeled in discrete domains such as language. Our work aims to address this gap by exploring different noise strategies for the semi-supervised named entity classification task, including statistical methods such as adding Gaussian noise to input embeddings, and linguistically-inspired ones such as dropping words and replacing words with their synonyms. We compare their performance on two benchmark datasets (OntoNotes and CoNLL) for named entity classification. Our results indicate that noise strategies that are linguistically informed perform at least as well as statistical approaches, while being simpler and requiring minimal tuning.
AB - Noise is inherent in real world datasets and modeling noise is critical during training as it is effective in regularization. Recently, novel semi-supervised deep learning techniques have demonstrated tremendous potential when learning with very limited labeled training data in image processing tasks. A critical aspect of these semi-supervised learning techniques is augmenting the input or the network with noise to be able to learn robust models. While modeling noise is relatively straightforward in continuous domains such as image classification, it is not immediately apparent how noise can be modeled in discrete domains such as language. Our work aims to address this gap by exploring different noise strategies for the semi-supervised named entity classification task, including statistical methods such as adding Gaussian noise to input embeddings, and linguistically-inspired ones such as dropping words and replacing words with their synonyms. We compare their performance on two benchmark datasets (OntoNotes and CoNLL) for named entity classification. Our results indicate that noise strategies that are linguistically informed perform at least as well as statistical approaches, while being simpler and requiring minimal tuning.
UR - http://www.scopus.com/inward/record.url?scp=85094530393&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85094530393&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85094530393
T3 - *SEM@NAACL-HLT 2019 - 8th Joint Conference on Lexical and Computational Semantics
SP - 186
EP - 191
BT - *SEM@NAACL-HLT 2019 - 8th Joint Conference on Lexical and Computational Semantics
PB - Association for Computational Linguistics (ACL)
T2 - 8th Joint Conference on Lexical and Computational Semantics, *SEM@NAACL-HLT 2019
Y2 - 6 June 2019 through 7 June 2019
ER -