TY - GEN
T1 - This before that
T2 - 15th Workshop on Biomedical Natural Language Processing, BioNLP 2016
AU - Hahn-Powell, Gus
AU - Bell, Dane
AU - Valenzuela-Escárcega, Marco A.
AU - Surdeanu, Mihai
N1 - Funding Information:
This work was funded by the Defense Advanced Research Projects Agency (DARPA) Big Mechanism program under ARO contract W911NF-14-1-0395.
Publisher Copyright:
© BioNLP 2016. All rights reserved.
PY - 2016
Y1 - 2016
N2 - Causal precedence between biochemical interactions is crucial in the biomedical domain, because it transforms collections of individual interactions, e.g., bindings and phosphorylations, into the causal mechanisms needed to inform meaningful search and inference. Here, we analyze causal precedence in the biomedical domain as distinct from open-domain, temporal precedence. First, we describe a novel, hand-annotated text corpus of causal precedence in the biomedical domain. Second, we use this corpus to investigate a battery of models of precedence, covering rule-based, feature-based, and latent representation models. The highestperforming individual model achieved a micro F1 of 43 points, approaching the best performers on the simpler temporalonly precedence tasks. Feature-based and latent representation models each outperform the rule-based models, but their performance is complementary to one another. We apply a sieve-based architecture to capitalize on this lack of overlap, achieving a micro F1 score of 46 points.
AB - Causal precedence between biochemical interactions is crucial in the biomedical domain, because it transforms collections of individual interactions, e.g., bindings and phosphorylations, into the causal mechanisms needed to inform meaningful search and inference. Here, we analyze causal precedence in the biomedical domain as distinct from open-domain, temporal precedence. First, we describe a novel, hand-annotated text corpus of causal precedence in the biomedical domain. Second, we use this corpus to investigate a battery of models of precedence, covering rule-based, feature-based, and latent representation models. The highestperforming individual model achieved a micro F1 of 43 points, approaching the best performers on the simpler temporalonly precedence tasks. Feature-based and latent representation models each outperform the rule-based models, but their performance is complementary to one another. We apply a sieve-based architecture to capitalize on this lack of overlap, achieving a micro F1 score of 46 points.
UR - http://www.scopus.com/inward/record.url?scp=85120077674&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85120077674&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85120077674
T3 - BioNLP 2016 - Proceedings of the 15th Workshop on Biomedical Natural Language Processing
SP - 146
EP - 155
BT - BioNLP 2016 - Proceedings of the 15th Workshop on Biomedical Natural Language Processing
A2 - Cohen, Kevin Bretonnel
A2 - Demner-Fushman, Dina
A2 - Ananiadou, Sophia
A2 - Tsujii, Jun-ichi
PB - Association for Computational Linguistics (ACL)
Y2 - 12 August 2016
ER -