TY - GEN
T1 - Transferring Legal Natural Language Inference Model from a US State to Another
T2 - 5th Natural Legal Language Processing Workshop, NLLP 2023
AU - Kwak, Alice Saebom
AU - Forte, Gaetano Vincent
AU - Bambauer, Derek E.
AU - Surdeanu, Mihai
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - This study investigates whether a legal natural language inference (NLI) model trained on the data from one US state can be transferred to another state. We fine-tuned a pre-trained model on the task of evaluating the validity of legal will statements, once with the dataset containing the Tennessee wills and once with the dataset containing the Idaho wills. Each model's performance on the in-domain setting and the out-of-domain setting are compared to see if the models can across the states. We found that the model trained on one US state can be mostly transferred to another state. However, it is clear that the model's performance drops in the out-of-domain setting. The F1 scores of the Tennessee model and the Idaho model are 96.41 and 92.03 when predicting the data from the same state, but they drop to 66.32 and 81.60 when predicting the data from another state. Subsequent error analysis revealed that there are two major sources of errors. First, the model fails to recognize equivalent laws across states when there are stylistic differences between laws. Second, difference in statutory section numbering system between the states makes it difficult for the model to locate laws relevant to the cases being predicted on. This analysis provides insights on how the future NLI system can be improved. Also, our findings offer empirical support to legal experts advocating the standardization of legal documents.
AB - This study investigates whether a legal natural language inference (NLI) model trained on the data from one US state can be transferred to another state. We fine-tuned a pre-trained model on the task of evaluating the validity of legal will statements, once with the dataset containing the Tennessee wills and once with the dataset containing the Idaho wills. Each model's performance on the in-domain setting and the out-of-domain setting are compared to see if the models can across the states. We found that the model trained on one US state can be mostly transferred to another state. However, it is clear that the model's performance drops in the out-of-domain setting. The F1 scores of the Tennessee model and the Idaho model are 96.41 and 92.03 when predicting the data from the same state, but they drop to 66.32 and 81.60 when predicting the data from another state. Subsequent error analysis revealed that there are two major sources of errors. First, the model fails to recognize equivalent laws across states when there are stylistic differences between laws. Second, difference in statutory section numbering system between the states makes it difficult for the model to locate laws relevant to the cases being predicted on. This analysis provides insights on how the future NLI system can be improved. Also, our findings offer empirical support to legal experts advocating the standardization of legal documents.
UR - http://www.scopus.com/inward/record.url?scp=85185007433&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85185007433&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85185007433
T3 - NLLP 2023 - Natural Legal Language Processing Workshop 2023, Proceedings of the Workshop
SP - 215
EP - 222
BT - NLLP 2023 - Natural Legal Language Processing Workshop 2023, Proceedings of the Workshop
A2 - Preotiuc-Pietro, Daniel
A2 - Goanta, Catalina
A2 - Chalkidis, Ilias
A2 - Barrett, Leslie
A2 - Spanakis, Gerasimos
A2 - Aletras, Nikolaos
PB - Association for Computational Linguistics (ACL)
Y2 - 7 December 2023
ER -