TY - GEN
T1 - PDDLEGO
T2 - 13th Joint Conference on Lexical and Computational Semantics, StarSEM 2024
AU - Zhang, Li
AU - Jansen, Peter
AU - Zhang, Tianyi
AU - Clark, Peter
AU - Callison-Burch, Chris
AU - Tandon, Niket
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - Planning in textual environments have been shown to be a long-standing challenge even for current models. A recent, promising line of work uses LLMs to generate a formal representation of the environment that can be solved by a symbolic planner. However, existing methods rely on a fully-observed environment where all entity states are initially known, so a one-off representation can be constructed, leading to a complete plan. In contrast, we tackle partially-observed environments where there is initially no sufficient information to plan for the end-goal. We propose PDDLEGO that iteratively construct a planning representation that can lead to a partial plan for a given sub-goal. By accomplishing the sub-goal, more information is acquired to augment the representation, eventually achieving the end-goal. We show that plans produced by few-shot PDDLEGO are 43% more efficient than generating plans end-to-end on the Coin Collector simulation, with strong performance (98%) on the more complex Cooking World simulation where end-to-end LLMs fail to generate coherent plans (4%).
AB - Planning in textual environments have been shown to be a long-standing challenge even for current models. A recent, promising line of work uses LLMs to generate a formal representation of the environment that can be solved by a symbolic planner. However, existing methods rely on a fully-observed environment where all entity states are initially known, so a one-off representation can be constructed, leading to a complete plan. In contrast, we tackle partially-observed environments where there is initially no sufficient information to plan for the end-goal. We propose PDDLEGO that iteratively construct a planning representation that can lead to a partial plan for a given sub-goal. By accomplishing the sub-goal, more information is acquired to augment the representation, eventually achieving the end-goal. We show that plans produced by few-shot PDDLEGO are 43% more efficient than generating plans end-to-end on the Coin Collector simulation, with strong performance (98%) on the more complex Cooking World simulation where end-to-end LLMs fail to generate coherent plans (4%).
UR - http://www.scopus.com/inward/record.url?scp=85207852365&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85207852365&partnerID=8YFLogxK
U2 - 10.18653/v1/2024.starsem-1.17
DO - 10.18653/v1/2024.starsem-1.17
M3 - Conference contribution
AN - SCOPUS:85207852365
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 212
EP - 221
BT - StarSEM 2024 - 13th Joint Conference on Lexical and Computational Semantics, Proceedings of the Conference
A2 - Bollegala, Danushka
A2 - Bollegala, Danushka
A2 - Shwartz, Vered
PB - Association for Computational Linguistics (ACL)
Y2 - 20 June 2024 through 21 June 2024
ER -