2 ZIANG CHEN, JIALIN LIU, XINSHANG WANG, JIANFENG LU, AND WOTAO YIN
key strategies in MILP solvers, and these mappings/strategies usually take an MILP instance
as input and output its key peroperties.
The graph neural network (GNN), due to its nice properties, say permutation invariance, is
considered a suitable model to represent such mappings/strategies for MILP. More specifically,
permutations on variables or constraints of an MILP do not essentially change the problem
itself, reliable ML models such as GNNs should satisfy such properties, otherwise the model may
overfit to the variable/constraint orders in the training data. [17] proposed that an MILP can be
encoded into a bipartite graph on which one can use a GNN to approximate Strong Branching.
[13] proposed to represent MILP with a tripartite graph. Since then, GNNs have been adopted
to represent mappings/strategies for MILP, for example, approximating Strong Branching [20,
21, 31, 40], approximating optimal solution [25, 31], parameterizing cutting strategies [32], and
parameterizing branching strategies [33, 38].
However, theoretical foundations in this direction still remain unclear. A key problem is the
ability of GNN to approximate important mappings related with MILP. We ask the following
questions:
Is GNN able to predict whether an MILP is feasible?(Q1)
Is GNN able to approximate the optimal objective value of an MILP?(Q2)
Is GNN able to approximate an optimal solution of an MILP?(Q3)
To answer questions (Q1 - Q3), one needs theories of separation power and representation
power of GNN. The separation power of GNN is measured by whether it can distinguish
two non-isomorphic graphs. Given two graphs G1, G2, we say a mapping F(e.g. a GNN) has
strong separation power if F(G1)̸=F(G2) as long as G1, G2that are not the isomorphic. In
our settings, since MILPs are represented by graphs, the separation power actually indicates
the ability of GNN to distinguish two different MILP instances. The representation power of
GNN refers to how well it can approximate mappings with permutation invariant properties. In
our settings, we study whether GNN can map an MILP to its feasiblity, optimal objective value
and an optimal solution. The separation power and representation power of GNNs are closely
related to the Weisfeiler-Lehman (WL) test [43], a classical algorithm to identify whether two
graphs are isomorphic or not. It has been shown that GNN has the same separation power
with the WL test [44], and, based on this result, GNNs can universally approximate continuous
graph-input mappings with separation power no stronger than the WL test [5,18].
Our contributions. With the above tools in hand, one still cannot directly answer questions
(Q1 - Q3) since the relationship between characteristics of general MILPs and properties of
graphs are not clear yet. Although there are some works studying the representation power
of GNN on some graph-related optimization problems [30, 36] and linear programming [12],
representing general MILPs with GNNs are still not theoretically studied, to the best of our
knowledge. Our contributions are listed below:
•(Limitation of GNNs for MILP) We show with an example that GNNs do not have strong
enough separation power to distinguish any two different MILP instances. There exist two
MILPs such that one of them is feasible while the other one is not, but, unfortunately, all