TY - GEN
T1 - Classify First, and Then Extract
T2 - 6th Natural Legal Language Processing Workshop 2024, NLLP 2024, co-located with the 2024 Conference on Empirical Methods in Natural Language Processing
AU - Kwak, Alice Saebom
AU - Morrison, Clayton T.
AU - Bambauer, Derek E.
AU - Surdeanu, Mihai
N1 - Publisher Copyright:
©2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - This work presents a new task-aware prompt design and example retrieval approach for information extraction (IE) using a prompt chaining technique. Our approach divides IE tasks into two steps: (1) text classification to understand what information (e.g., entity or event types) is contained in the underlying text and (2) information extraction for the identified types. Initially, we use a large language model (LLM) in a few-shot setting to classify the contained information. The classification output is used to select the relevant prompt and retrieve the examples relevant to the input text. Finally, we ask a LLM to do the information extraction with the generated prompt. By evaluating our approach on legal IE tasks with two different LLMs, we demonstrate that the prompt chaining technique improves the LLM's overall performance in a few-shot setting when compared to the baseline in which examples from all possible classes are included in the prompt. Our approach can be used in a low-resource setting as it does not require a large amount of training data. Also, it can be easily adapted to many different IE tasks by simply adjusting the prompts. Lastly, it provides a cost benefit by reducing the number of tokens in the prompt.
AB - This work presents a new task-aware prompt design and example retrieval approach for information extraction (IE) using a prompt chaining technique. Our approach divides IE tasks into two steps: (1) text classification to understand what information (e.g., entity or event types) is contained in the underlying text and (2) information extraction for the identified types. Initially, we use a large language model (LLM) in a few-shot setting to classify the contained information. The classification output is used to select the relevant prompt and retrieve the examples relevant to the input text. Finally, we ask a LLM to do the information extraction with the generated prompt. By evaluating our approach on legal IE tasks with two different LLMs, we demonstrate that the prompt chaining technique improves the LLM's overall performance in a few-shot setting when compared to the baseline in which examples from all possible classes are included in the prompt. Our approach can be used in a low-resource setting as it does not require a large amount of training data. Also, it can be easily adapted to many different IE tasks by simply adjusting the prompts. Lastly, it provides a cost benefit by reducing the number of tokens in the prompt.
UR - http://www.scopus.com/inward/record.url?scp=85212793257&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85212793257&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85212793257
T3 - NLLP 2024 - Natural Legal Language Processing Workshop 2024, Proceedings of the Workshop
SP - 303
EP - 317
BT - NLLP 2024 - Natural Legal Language Processing Workshop 2024, Proceedings of the Workshop
A2 - Aletras, Nikolaos
A2 - Chalkidis, Ilias
A2 - Barrett, Leslie
A2 - Goanta, Catalina
A2 - Preotiuc-Pietro, Daniel
A2 - Spanakis, Gerasimos
PB - Association for Computational Linguistics (ACL)
Y2 - 16 November 2024
ER -