rate, whether debit note exists, whether borrowers refuse to
repay the debt, and so on). They can help the model understand
legal cases’ judgments well [6]. They also play a pivotal role
in affecting experts’ judgment of legal relevance in manual
search. An example of legal element is shown in Table I. In
general IR, the two cases are considered as relevant texts due to
many shared words and phrases. In legal case retrieval, they are
irrelevant and have totally different judgments because Case B
has a surprisingly high rate of interest and becomes usury. We
can observe that legal elements have a significant impact on
judgments and legal relevance. In practice, even though they
are essential for the definition of legal relevance, it is difficult
to model legal elements in current neural networks [1].
To tackle the above challenges, we propose an interaction-
focused network with a Multi-View Contrastive Learning
objective. The whole model is entitled MVCL. The proposed
model builds word-level interactions based on pre-trained lan-
guage models (PLMs) and then aggregates these interactions
to get the matching representation between a query and a
candidate case. We also introduce a multi-view contrastive
learning objective, including case-view and element-view, to
tackle the challenges mentioned earlier. The case-view con-
trastive learning can help the model understand long texts
more thoroughly (Challenge 1). The element-view contrastive
learning builds positive and negative instances by changing
legal elements of cases. In this way, legal elements are
integrated into our neural network. We also introduce a legal
element knowledge-aware indicator to detect elements in legal
cases. Through legal element-oriented modeling (i.e., element-
view contrastive learning), the model can better understand
the query and candidate cases, decide whether two cases’
judgments can support each other, and compute legal relevance
(Challenge 2). Therefore, the MVCL model can learn legal
relevance instead of the general definition.
To evaluate our model, we conduct experiments with bench-
marks of legal case retrieval in CAIL2019-SCM. We achieved
state-of-the-art results in terms of automatic evaluation. In
summary, our main contributions include:
•We propose a multi-view contrastive learning objective
for legal case retrieval, which can effectively use the cases
and elements to align representations of legal cases.
•We first introduce a legal element knowledge-aware indi-
cator as an external knowledge module of legal elements
to improve the model’s understanding of legal relevance.
•We empirically verify the effectiveness of MVCL on the
CAIL2019-SCM dataset and surpass 711 teams in the
China AI&Law Challenge1.
II. RELATED WORK
A. Traditional Ad-hoc Retrival
Ad-hoc retrieval has received a lot of attention in the past
few decades. Traditional models are built on the bag-of-words
model and measure the relevance between texts by comparing
1http://cail.cipsc.org.cn:2019/
tokens in texts, including BM25 [14] and VSM [15]. In-
evitably, they face the problem of sparse and high-dimensional
representation.
The recent development of deep learning has also inspired
applications of neural models in IR. Typically, most ap-
proaches can be categorized into two types: representation-
focused and interaction-focused [16]. To measure the relevance
for documents, models that are representation-focused (i.e.,
Siamese models) learn the representations of a query and a
document separately and then compare two representations by
computing distance with functions like cosine, dot product,
bilinear, or Euclidean distance. The advantage of these ap-
proaches is that shared parameters make Siamese architecture
smaller and far more efficient. Nevertheless, representing an
entire text using a single vector is not sufficient to capture
all crucial information. On the other hand, interaction-focused
approaches first compute a query-document word-by-word
similarity matrix to build local interactions and then aggregate
these word-level matching results to generate the final rele-
vance score. Well-known methods include MatchPyramid [17],
Match-SRNN [18], DRMM [8], and HCRN [19]. These mod-
els get better performance in traditional ad-hoc retrieval tasks
due to the detailed comparison of queries and documents. Ad-
ditionally, there are neural models that combine representation-
focused and interaction-focused approaches [10, 20].
Regarding the length of query and document, those works
mentioned above mainly concentrate on matching between
short texts, or between short and long texts. When adopting to
matching between long texts, they don’t work well and lead
to huge memory cost. Matching between long texts was less
explored. In this line of work, SMASH RNN [11] is proposed
to learn representation using a hierarchical architecture of
RNN to model multiple abstraction levels of the document
structure, including words, sentences, and paragraphs. Yang
et al. [12] proposed a Transformer-based hierarchical archi-
tecture to better model document structure and to increase
maximum input text length of general Transformer-based
models.
B. Legal Case Retrieval
Legal case retrieval is still a challenging task, due to its con-
cept of relevance [13], professional terms and expressions, text
length [6] and logical structure behind natural languages [21].
Many works tried to build a retrieval engine using links or
citations between case documents and legal statutes [2, 3, 4].
Recently, Shao et al. [22] and Liu et al. [23] studied user
behaviors and practical search interactions between users and
systems and observed significant differences between legal
case retrieval and general web search.
To attain the purpose of computing semantic content,
deep learning methods has been applied to Legal Case Re-
trieval. Tran et al. [24] builds an CNN-based model, which also
utilizes summarization information and lexical features. Zhong
et al. [1] adopted a few deep learning models to get a better
view of the current progress of Legal Case Retrieval, like
BiDAF [25], SMASH RNN [11] and so on. Shao et al. [6]