具有随机道路通行能力的多停车场车辆路径问题及其推广算法的简化两阶段随机整数线性规划模型

作者: 时间:2021-07-28 点击数:

Wadi Khalid Anuar, Lai Soon Lee, Hsin-Vonn Seow, Stefan Pickl



    A matheuristic approach based on a reduced two-stage Stochastic Integer Linear Programming (SILP) model is presented. The proposed approach is suitable for obtaining a policy constructed dynamically on the go during the rollout algorithm. The rollout algorithm is part of the Approximate Dynamic Programming (ADP) lookahead solution approach for a Markov Decision Processes (MDP) framed Multi-Depot Dynamic Vehicle Routing Problem with Stochastic Road Capacity (MDDVRPSRC). First, a Deterministic Multi-Depot VRP with Road Capacity (D-MDVRPRC) is presented. Then an extension, MDVRPSRC-2S, is presented as an offline two-stage SILP model of the MDDVRPSRC. These models are validated using small simulated instances with CPLEX. Next, two reduced versions of the MDVRPSRC-2S model (MDVRPSRC-2S1 and MDVRPSRC-2S2) are derived. They have a specific task in routing: replenishment and delivering supplies. These reduced models are to be utilised interchangeably depending on the capacity of the vehicle, and repeatedly during the execution of rollout in reinforcement learning. As a result, it is shown that a base policy consisting of an exact optimal decision at each decision epoch can be obtained constructively through these reduced two-stage stochastic integer linear programming models. The results obtained from the resulting rollout policy with CPLEX execution during rollout are also presented to validate the reduced model and the matheuristic algorithm. This approach is proposed as a simple implementation when performing rollout for the lookahead approach in ADP.

Key words:reinforcement learning; approximate dynamic programming; rollout algorithm; matheuristic; two-stage stochastic programming; vehicle routing problem

DOI:https://doi.org/10.3390/math9131572

Date:2021-7-4


Copyright© 2019 广西中国-东盟综合交通国际联合重点实验室  地址:广西南宁市龙亭路8号广西中国-东盟综合交通国际联合重点实验室大楼  电话:0771-5900869 邮编:530200  桂ICP 备11008250号