Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Improved CP-Based Lagrangian Relaxation Approach with an Application to the TSP

Authors: Raphaël Boudreault, Claude-Guy Quimper

IJCAI 2021 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental The experimental results on TSP instances show that our algorithms allow significant gains on the resolution time and the size of the search space when compared to the state-of-the-art implementation.
Researcher Affiliation Academia Rapha el Boudreault and Claude-Guy Quimper Universit e Laval, Qu ebec, Canada EMAIL, EMAIL
Pseudocode Yes Algorithm 1: SIMPLE(i, j) if {i, j} T then {k, l} GETREPLACEMENTEDGE(i, j) 1 U (ZLR(λ) + w(k, l) w(i, j)) if i / {1, k, l} deg T (i) 2 then // Decrease λi 2 v ( 1) MAXDECREASE(i, j) (v (deg T (i) 2) v) if j / {1, k, l} deg T (j) 2 then // Decrease λj 3 v ( 1) MAXDECREASE(j, i) (v (deg T (j) 2) v) if < 0 then dom(x{i,j}) dom(x{i,j}) \ {0} else {k, l} GETSUPPORTEDGE(i, j) 4 U (ZLR(λ) + w(i, j) w(k, l)) if i / {1, k, l} deg T (i) 2 then // Increase λi 5 v MAXINCREASE(i, j) (v (deg T (i) 2) + v) if j / {1, k, l} deg T (j) 2 then // Increase λj 6 v MAXINCREASE(j, i) (v (deg T (j) 2) + v) if < 0 then dom(x{i,j}) dom(x{i,j}) \ {1} Function MAXDECREASE(i, j) D1 α min{GETREDUCEDCOST(i, k): {i, k} E \ T} { } if j = 1 then r GETREPLACEMENTEDGE(i, j) D2 β min{ w(i, k) w(r): {i, k} R{i,j}} { } α min{α, β} return α Function MAXINCREASE(i, j) I1 α min{GETREPLACEMENTCOST(i, k): {i, k} T \ M} { } if j = 1 then s GETSUPPORTEDGE(i, j) I2 β min{ w(s) w(i, k): {i, k} C{i,j}\M} { } α min{α, β} return α
Open Source Code Yes The code is available at http://www2.ift.ulaval.ca/ quimper/publications.php.
Open Datasets Yes We chose the symmetric TSP instances from the TSPLIB library [Reinelt, 1991] between 96 and 500 nodes that could be solved by Choco under 8 hours and with at least 100 search nodes.
Dataset Splits No The paper does not provide specific train/validation/test dataset splits. TSP instances are typically solved as a whole for optimization, not split for training/validation/testing in the conventional machine learning sense.
Hardware Specification Yes The experiments were performed on a Cent OS Linux 7 machine using an Intel Xeon Silver 4110 CPU at 2.10 GHz and 32 GB of RAM.
Software Dependencies Yes The algorithms were implemented in Java 14 using the solver Choco 4.0.6 [Jussien et al., 2008] and its extension Choco Graph 4.2.3 [Fages, 2014].
Experiment Setup Yes The TSP was modeled using the WEIGHTEDCIRCUIT constraint already implemented in Choco Graph using the state-of-the-art algorithms [Benchimol et al., 2012]. It uses a subgradient descent algorithm to optimize the Lagrangian multipliers. We kept the default parameters, but we did not restart the algorithm when the lower bound of the 1-tree was increased. As an initial upper bound on the objective variable, we gave the bound provided by the LKH-2.0.9 heuristic [Helsgaun, 2000]. The search strategy was fixed to max Cost with the LCFirst policy [Fages et al., 2016]. ... for HYBRID, since α-SETS is slowed down by the number of constraints, it was only called when |E| 2|V | with a limit of 2 on the cardinality of the α-sets and a maximum of 10 iterations when trying to reach the fixed point.