DocumentCode :
3723200
Title :
RLBS: An Adaptive Backtracking Strategy Based on Reinforcement Learning for Combinatorial Optimization
Author :
Ilyess Bachiri;Jonathan Gaudreault;Claude-Guy Quimper;Brahim Chaib-draa
Author_Institution :
FORAC Res. Consortium, Univ. Laval, Quebec City, QC, Canada
fYear :
2015
Firstpage :
936
Lastpage :
942
Abstract :
Combinatorial optimization problems are often very difficult to solve and the choice of a search strategy has a tremendous influence over the solver´s performance. A search strategy is said to be adaptive when it dynamically adapts to the structure of the problem instance and identifies the areas of the search space that contain good solutions. We introduce an algorithm (RLBS) that learns to efficiently backtrack when searching non-binary trees. Branching can be carried on using any usual variable/value selection strategy. However, when backtracking is needed, the selection of the node to target involves reinforcement learning. As the trees are non-binary, we have the opportunity to backtrack many times to each node during the search, which allows learning which nodes generally lead to the best rewards (that is, to the most interesting leaves). RLBS is evaluated for a scheduling problem using real industrial data. It outperforms classic (nonadaptive) backtracking strategies (DFS, LDS) as well as an adaptive branching strategy (IBS).
Keywords :
"Search problems","Learning (artificial intelligence)","Heuristic algorithms","Cost function","Input variables","Job shop scheduling"
Publisher :
ieee
Conference_Titel :
Tools with Artificial Intelligence (ICTAI), 2015 IEEE 27th International Conference on
ISSN :
1082-3409
Type :
conf
DOI :
10.1109/ICTAI.2015.135
Filename :
7372232
Link To Document :
بازگشت