Recently, many works show promising results by using Large Language Models (LLMs) as translators to convert natural language into intermediate expressions that can be used by existing algorithms, which makes directly using LLMs to solve problems less promising. Although the method of directly using LLMs to solve problems does not perform well at the current stage, we argue that it is too early to conclude that we should favor the hybrid method. We propose that we should dive deeper and analyze what impedes the performance of LLMs and find ways to tackle these issues. In this work, we investigate the potential of using solely LLMs to solve Vehicle Routing Problems (VRP), which are widely used to model robot task planning, by directly generating code from natural language task descriptions. We begin by constructing a dataset comprising 80 problem instances, including 8 types of single- and multi-vehicle routing problems, to evaluate the performance of LLMs. We then design different frameworks and varying types of context. After that, we perform evaluations based on different frameworks and different contexts. We conduct an extensive study on what impedes LLMs from performing well and propose several potential directions for future work to improve the performance of LLMs in solving VRP.
Citation
@article{huang2024words,
title={From Words to Routes: Applying Large Language Models to Vehicle Routing},
author={Huang, Zhehui and Shi, Guangyao and Sukhatme, Gaurav S},
journal={arXiv preprint arXiv:2403.10795},
year={2024}
}