Learning in RKHM: a C∗-Algebraic Twist for Kernel Machines
Yuka Hashimoto1,2Masahiro Ikeda2,3Hachem Kadri4
1. NTT Network Service Systems Laboratories, NTT Corporation, Tokyo, Japan
2. Center for Advanced Intelligence Project, RIKEN, Tokyo, Japan
3. Keio University, Yokohama, Japan
4. Aix-Marseille University, CNRS, LIS, Marseille, France
Abstract
Supervised learning in reproducing kernel Hilbert space (RKHS) and vector-valued RKHS (vvRKHS)
has been investigated for more than 30 years. In this paper, we provide a new twist to this rich literature by
generalizing supervised learning in RKHS and vvRKHS to reproducing kernel Hilbert C∗-module (RKHM),
and show how to construct effective positive-definite kernels by considering the perspective of C∗-algebra.
Unlike the cases of RKHS and vvRKHS, we can use C∗-algebras to enlarge representation spaces. This
enables us to construct RKHMs whose representation power goes beyond RKHSs, vvRKHSs, and existing
methods such as convolutional neural networks. Our framework is suitable, for example, for effectively
analyzing image data by allowing the interaction of Fourier components.
1 INTRODUCTION
Supervised learning in reproducing kernel Hilbert space (RKHS) has been actively investigated since the early
1990s (Murphy, 2012; Christmann & Steinwart, 2008; Shawe-Taylor & Cristianini, 2004; Sch¨olkopf & Smola,
2002; Boser et al., 1992). The notion of reproducing kernels as dot products in Hilbert spaces was first brought to
the field of machine learning by Aizerman et al. (1964), while the theoretical foundation of reproducing kernels
and their Hilbert spaces dates back to at least Aronszajn (1950). By virtue of the representer theorem (Sch¨olkopf
et al., 2001), we can compute the solution of an infinite-dimensional minimization problem in RKHS with given
finite samples. In addition to the standard RKHSs, applying vector-valued RKHSs (vvRKHSs) to supervised
learning has also been proposed and used in analyzing vector-valued data (Micchelli & Pontil, 2005; ´
Alvarez
et al., 2012; Kadri et al., 2016; Minh et al., 2016; Brouard et al., 2016; Laforgue et al., 2020; Huusari & Kadri,
2021). Generalization bounds of the supervised problems in RKHS and vvRKHS are also derived (Mohri et al.,
2018; Caponnetto & De Vito, 2007; Audiffren & Kadri, 2013; Huusari & Kadri, 2021).
Reproducing kernel Hilbert C∗-module (RKHM) is a generalization of RKHS and vvRKHS by means of
C∗-algebra. C∗-algebra is a generalization of the space of complex values. It has a product and an involution
structures. Important examples are the C∗-algebra of bounded linear operators on a Hilbert space and the
C∗-algebra of continuous functions on a compact space. RKHMs have been originally studied for pure operator
algebraic and mathematical physics problems (Manuilov & Troitsky, 2000; Heo, 2008; Moslehian, 2022). Re-
cently, applying RKHMs to data analysis has been proposed by Hashimoto et al. (2021). They generalized the
representer theorem in RKHS to RKHM, which allows us to analyze structured data such as functional data
with C∗-algebras.
In this paper, we investigate supervised learning in RKHM. This provides a new twist to the state-of-the-art
kernel-based learning algorithms and the development of a novel kind of reproducing kernels. An advantage of
RKHM over RKHS and vvRKHS is that we can enlarge the C∗-algebra characterizing the RKHM to construct
a representation space. This allows us to represent more functions than the case of RKHS and make use of the
product structure in the C∗-algebra. Our main contributions are:
1
arXiv:2210.11855v3 [stat.ML] 26 Jun 2024