Learning to reason over scene graphs: a case study of finetuning GPT-2 into a robot language model for grounded task planning.
grounding
language models (LMs)
pretrained models
robot learning
scene graphs
task planning
Journal
Frontiers in robotics and AI
ISSN: 2296-9144
Titre abrégé: Front Robot AI
Pays: Switzerland
ID NLM: 101749350
Informations de publication
Date de publication:
2023
2023
Historique:
received:
12
05
2023
accepted:
03
07
2023
medline:
31
8
2023
pubmed:
31
8
2023
entrez:
31
8
2023
Statut:
epublish
Résumé
Long-horizon task planning is essential for the development of intelligent assistive and service robots. In this work, we investigate the applicability of a smaller class of large language models (LLMs), specifically GPT-2, in robotic task planning by learning to decompose tasks into subgoal specifications for a planner to execute sequentially. Our method grounds the input of the LLM on the domain that is represented as a scene graph, enabling it to translate human requests into executable robot plans, thereby learning to reason over long-horizon tasks, as encountered in the ALFRED benchmark. We compare our approach with classical planning and baseline methods to examine the applicability and generalizability of LLM-based planners. Our findings suggest that the knowledge stored in an LLM can be effectively grounded to perform long-horizon task planning, demonstrating the promising potential for the future application of neuro-symbolic planning methods in robotics.
Identifiants
pubmed: 37649810
doi: 10.3389/frobt.2023.1221739
pii: 1221739
pmc: PMC10464606
doi:
Types de publication
Journal Article
Langues
eng
Pagination
1221739Informations de copyright
Copyright © 2023 Chalvatzaki, Younes, Nandha, Le, Ribeiro and Gurevych.
Déclaration de conflit d'intérêts
Author LR was employed by Amazon Alexa. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Références
Nature. 2015 May 28;521(7553):436-44
pubmed: 26017442
IEEE Robot Autom Lett. 2019 Apr;4(2):1255-1262
pubmed: 31058229