
Knowledge Distillation based Contextual Relevance Matching
for E-commerce Product Search
Ziyang Liu§,Chaokun Wang§*, Hao Feng§, Lingfei Wu†, Liqun Yang‡
§Tsinghua University,†JD.com,‡CNAEIT
§liu-zy21@mails.tsinghua.edu.cn,chaokun@tsinghua.edu.cn
†lwu@email.wm.edu, ‡yanglq@cnaeit.com
Abstract
Online relevance matching is an essential task
of e-commerce product search to boost the
utility of search engines and ensure a smooth
user experience. Previous work adopts ei-
ther classical relevance matching models or
Transformer-style models to address it. How-
ever, they ignore the inherent bipartite graph
structures that are ubiquitous in e-commerce
product search logs and are too inefficient to
deploy online. In this paper, we design an
efficient knowledge distillation framework for
e-commerce relevance matching to integrate
the respective advantages of Transformer-style
models and classical relevance matching mod-
els. Especially for the core student model of
the framework, we propose a novel method us-
ing k-order relevance modeling. The experi-
mental results on large-scale real-world data
(the size is 6∼174 million) show that the pro-
posed method significantly improves the pre-
diction accuracy in terms of human relevance
judgment. We deploy our method to the anony-
mous online search platform. The A/B test-
ing results show that our method significantly
improves 5.7% of UV-value under price sort
mode.
1 Introduction
Relevance matching
(Guo et al.,2016;Rao et al.,
2019;Wang et al.,2020) is an important task in the
field of ad-hoc information retrieval (Zhai and Laf-
ferty,2017), which aims to return a sequence of in-
formation resources related to a user query (Huang
et al.,2020;Chang et al.,2021a;Sun and Duh,
2020). Generally, texts are the dominant form of
the user query and returned information resources.
Given two sentences, the target of relevance match-
ing is to estimate their relevance score and then
judge whether they are relevant or not. However,
text similarity does not mean semantic similarity.
For example, while “mac pro 1.7GHz” and “mac
*Chaokun Wang is the corresponding author.
lipstick 1.7ml” look alike, they describe two differ-
ent and irrelevant products. Therefore, relevance
matching is important, especially for long-term
user satisfaction of e-commerce search (Cheng
et al.,2022;Niu et al.,2020;Xu et al.,2021;Zhu
et al.,2020).
Related work.
With the rapid development
of deep learning, the current research on rel-
evance matching can be grouped into two
camps (see Appendix Afor further details):
1.
Classical Relevance Matching Models.
For the
given query and item, the classical relevance match-
ing models either learn their individual embeddings
or learn an overall embedding based on the calcula-
tion from word-level interaction to sentence-level
interaction. The representative methods include
ESIM (Chen et al.,2017) and BERT2DNN (Jiang
et al.,2020). 2.
Transformer-style Models.
These
models adopt the multi-layer Transformer net-
work structure (Vaswani et al.,2017a). They have
achieved breakthroughs on many NLP tasks and
even reached human-level accuracy. The represen-
tative methods include BERT (Devlin et al.,2019)
and ERNIE (Sun et al.,2019b).
Although Transformer-style models show satis-
factory performance on relevance matching, they
are hard to deploy to the online environment due
to their high time complexity. Moreover, the above
methods cannot deal with the abundant context
information (i.e., the neighbor features in a query-
item bipartite graph) in e-commerce product search.
Last but not least, when applied to real-world sce-
narios, existing classical relevance matching mod-
els directly use user behaviors as labeling informa-
tion (Fig. 1). However, this solution is not directly
suitable for relevance matching because user be-
haviors are often noisy and deviate from relevance
signals (Mao et al.,2019;Liu and Mao,2020).
In this paper, we propose to incorporate bipartite
graph embedding into the knowledge distillation
framework (Li et al.,2021a;Dong et al.,2021;
arXiv:2210.01701v1 [cs.IR] 4 Oct 2022