Please use this identifier to cite or link to this item: http://localhost:8081/jspui/handle/123456789/18873
Full metadata record
DC FieldValueLanguage
dc.contributor.authorKumar, Shivam-
dc.date.accessioned2026-02-05T11:38:26Z-
dc.date.available2026-02-05T11:38:26Z-
dc.date.issued2024-06-
dc.identifier.urihttp://localhost:8081/jspui/handle/123456789/18873-
dc.guideGupta, Manu Kumaren_US
dc.description.abstractThe Vehicle Routing Problem (VRP) is a well-known optimization challenge with practical applications in various areas, including delivery route optimization, transportation logistics, and mobile resource allocation. This work specifically focuses on the capacitated vehicle routing problem (CVRP), where each customer has a demand that must be met by vehicles with capacity constraints. The goal is to minimize the total cost or distance traveled by vehicles while ensuring that each customer’s demand is satisfied and vehicles stay within their capacity limits. Traditional methods for solving CVRP typically use heuristics and mathematical programming techniques. However, these approaches can struggle with large-scale instances and dynamic environments. In recent years, reinforcement learning (RL) has emerged as a promising solution for complex optimization problems. In this study, we use RL techniques to address the CVRP. Our approach harnesses the power of deep RL algorithms, specifically combining deep neural networks and actor-critic methods, to learn effective policies for route planning and optimization. By framing the CVRP as a Markov Decision Process, we develop an RL agent that learns to make sequential decisions regarding vehicle movements and load allocations. We evaluate our proposed RL framework on a set of benchmark VRP instances, comparing its performance against the state-of-the-art solver Gurobi. The experimental results demonstrate that our approach achieves competitive solution quality and computational efficiency, especially in larger problems. Additionally, we explore the robustness and generalization capability of the learned policies by assessing their performance on unseen problem instances and assessing the performance of RL(trained for larger problems) on smaller problem instances. Overall, our work underscores the potential of reinforcement learning as a promising methodology for solving the Vehicle Routing Problem. Through the integration of deep RL algorithms and problem-specific insights, we illustrate that RL can provide efficient and effective solutions to this challenging optimization problem, opening avenues for further advancements in the field of transportation logistics and route planning.en_US
dc.language.isoenen_US
dc.publisherIIT, Roorkeeen_US
dc.titleMINIMIZING ENERGY USAGE FOR CAPACITATED VEHICLE ROUTING PROBLEM USING POLICY GRADIENT (RL) AND COMPARING SOLUTIONS WITH GUROBIen_US
dc.typeDissertationsen_US
Appears in Collections:MASTERS' THESES (MFSDS & AI)

Files in This Item:
File Description SizeFormat 
22566020_SHIVAM KUMAR.pdf1.7 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.