Learning combinatorial optimization algorithms over graphs
NIPS 2017 9-22-2020
Last updated
Was this helpful?
NIPS 2017 9-22-2020
Last updated
Was this helpful?
The current methods for combinatorial optimization problem cannot learn from experience.
Can we learn from experience?
Given a graph optimization problem and a distribution of problem instances, can we learn a better greedy heuristics that generalize to unseen instances from ?
Given weighted graph: , : edge weight function, is the weight of edge
Given a graph , find a subset of nodes , s.t. every edge is covered.
is minimized
Given a graph , find a subset of nodes , s.t. the weight of cut-set is maximized.
cut-set: , the set of edges with one end in and the other end in
basic idea
original graph with initial state
input the graph to an embedding model (structure2vec) for T steps
each node has a corresponding score based on structure2vec
pick up the node with the highest score (after one iteration)
go back to step 2 and repeat until termination
Reinforcement Learning
MVC (Minimum Vertex Cover)
score we earned at current step
current screen
current selected nodes
move your board left/right
select a node
Action value function
predicted future total rewards
structure2vec embedding
How to choose the action
Reward
State
Action
Policy