of hypergraphs and their clique-expansion graphs [
26
,
27
], which is computationally expensive as
multiple neural networks of different modalities (hypergraphs and variants of expanded graphs) need
to be optimized. More importantly, contrasting on clique expansion has the risk of losing higher-order
information via pulling representations of hypergraphs and graphs close.
Contributions.
Motivated by [
12
,
14
] that appropriate data augmentations suffice for the effective
contrastive views, and intuitively they are more capable of preserving higher-order relations in
hypergraphs compared to clique expansion, we explore on the question in this paper, how to design
augmented views of hypergraphs in contrastive learning (
HyperGCL
). Our answers are in two folds.
We first assay whether
fabricated
augmentations guided by domain knowledge are suited for Hy-
perGCL. Since hypergraphs are composed of hyperedges and vertices, to augment hyperedges, we
propose two strategies that (i) directly perturb on hyperedges, and (ii) perturb on the “edges” between
hyperedges and vertices in the converted bipartite graph; To augment vertices, we adopt three schemes
of vertex dropping, attribute masking and subgraph from graph-structured data [
14
]. Our finding is
that, different from the fact that vertex augmentations benefit more on graphs, hypergraphs mostly ben-
efit from hyperedge augmentations (up to 9% improvement), revealing that higher-order information
encoded in hyperedges is usually more downstream-relevant (than information in vertices).
Furthermore, in search of even better augmented views but in a data-driven manner, we study
whether/how augmentations of hypergraphs could be learned during contrastive learning. To this
end, for the first time, we propose a novel variational hypergraph auto-encoder architecture, as a
hypergraph
generative
model, to parameterize a certain augmentation space of hypergraphs. In addi-
tion, we propose an end-to-end differentiable pipeline utilizing Gumbel-Softmax [
28
], to jointly learn
hypergraph augmentations and model parameters. Our observation is that generative augmentations
can better capture the higher-order information and achieve state-of-the-art performance on most of
the benchmark data sets (up to 20% improvement).
The aforementioned empirical evidences (for generalizability) are drawn from comprehensive experi-
ments on 13 datasets. Moreover, we introduce the robustness and fairness evaluation for hypergraphs,
and show that HyperGCL in addition boosts robustness against adversarial attacks and imposes
fairness with regard to sensitive attributes.
The rest of the paper is organized as follows. We discuss the related work in Section 2, introduce
HyperGCL in Section 3, present the experimental results in Section 4, and conclude in Section 5.
2 Related Work
Hypergraph neural networks.
Hypergraphs, which are able to encode higher-order relationships,
have attracted significant attentions in recent years. In the machine learning community, hypergraph
neural networks are developed for effective hypergraph representations. HGNN [
1
] adopt the clique
expansion technique and designs the weighted hypergraph Laplacian for message passing. HyperGCN
[
2
] proposes the generalized hypergraph Laplacian and explores adding the hyperedge information
through mediators. The attention mechanism [
29
,
30
] is also designed to learn the importance within
hypergraphs. However, the expanded graph will inevitably cause distortion and lead to unsatisfactory
performance. There is also another line of works such as UniGNN [
31
] and HyperSAGE [
32
] which
try to perform message passing directly on the hypergraph to avoid the information loss. A recent
work [
3
] provides an AllSet framework to unify the existing studies with high expressive power and
achieves state-of-the-art performance on comprehensive benchmarks. The work utilizes deep multiset
functions [33] to identify the propagation and aggregation rules in a data-driven manner.
Contrastive self-supervised learning.
Contrastive self-supervision [
12
,
34
,
35
] has achieved un-
precedented success in computer vision. The core idea is to learn an embedding space where samples
from the same instance are pulled closer and samples from different instances are pushed apart. Re-
cent works start to cross-pollinate between contrastive learning and graph neural networks to for more
generalizable graph representations. Typically, they design some fabricated augmentations guided by
domain knowledge, such as edge perturbation, feature masking or vertex dropping, etc. Nevertheless,
contrastive learning on hypergraphs remains largely unexplored. Most existing works [
6
,
36
,
26
,
37
]
design pretext tasks for hypergraphs and mainly focus on recommender systems [
38
,
39
,
40
,
41
], via
contrasting between graphs and hypergraphs which might lose important higher-order information.
In this work, we explore on the structure of hypergraph itself to construct contrastive views.
2