2
target model. Within the definition scope, we require that the
server should simultaneously assure the correctness of the
target model, the unlearning process and the newly predictive
model, such that the data owner (as a honest verifier) can be
convinced of the fulfillment of his/her unlearning problem.
PoUL is different from a recent excellent art [45], proof-of-
learning (PoL), which facilitates proofs for a learning process
based on the idea of re-execution. Specifically, PoL offers a
verifier a document on learning trajectory for convincing him
of the truth that the learning trajectory is used to generate
a particular model with overwhelming probability. While
the PoL is easy to understand and implement, it cannot be
extended to implement the PoUL, due to different problem
statements and threat models, figured out by Thudi et al. [73].
Recent attack examples [87] can efficiently forge learning
trajectory to generate a fake but valid PoL proof, which
further demonstrates the difficulty of implementing the PoUL.
Lastly, our PoUL especially emphasizes that subsequent
prediction services should be offered by an unlearned model
as expected, which is not considered by the PoL.
To pursue practice-oriented solutions, we subsequently
highlight practicality requirements needed for our PoUL: (a)
Enabling generic models. The unlearning problem in MLaaS
can happen in various models, which requires proof tech-
niques to support generic computations involved by various
model architectures. (b) Limited invasiveness. Proof techniques
should be maximally compatible with the underlying learning
algorithms, so as to maintain the original ML services quality.
(c) Minimal overhead. Standing on the already workload-
intensive ML pipeline, additional proving workloads should be
as affordable as possible. Besides the server-side requirements,
PoUL should also meet the following verifier-side require-
ments, such that the data owner can enjoy predictive services
nearly as usual beyond verifying unlearning. (d) Concise proof.
Generated proofs should be short enough to ease the storage
cost of the data owner. (e) Efficient verification. Proofs should
also be cheaply verified considering a usually thin verifier.
A. Solution Overview
We now present our solution roadmap to implement PoUL,
towards complying with the deletion obligation. We ob-
serve that PoUL relates to proof-based verifiable computation
(VC) [80], especially a line of “authenticate first and prove
later” proof systems. Simply speaking, a data owner can
authorize her data to the server, along with a data digest
generated by authenticated data structures (ADS) [71], and
then keep track of particular zero-knowledge proofs issued
by the server, with respect to the correctness of a target
model, an unlearning process and a newly predictive model,
which are not transparent to the data owner. Yet, aware of the
mentioned practicality requirements, the line of pure crypto-
assisted proof systems [80] are not the preferred choices (see
more in Appendix A).
From another perspective, trusted execution environments
(TEEs) are quickly becoming trusted computing services
offered by dominant service providers, e.g., Alibaba Cloud [8],
IBM Cloud [43] and Microsoft Azure [2], and many
application efforts [47], [26], [30] demonstrate that TEEs can
assist the providers in performing with obligation compliance,
e.g., enforcing data usage with GDPR compliance [68], [61].
Therefore, we are naturally encouraged to instantiate PoUL
with the TEEs-backed offerings, and thereby enable native
implementation and practical deployment on the server side.
Also, we admit that recent TEEs-empowered ML works [75],
[48], [38], [74], [54], [41], [42], [86], [88], [53], [39], [11]
may help mitigate a certain issue within our PoUL, but we
need new and holistic designs tailored for proving end-to-end
RTBF compliance in MLaaS.
Our unlearning setting. We begin with Bourtoule et al.’s
efficient retraining-based unlearning framework [14], towards
enabling complete deletion [35]. Specifically, the server
trains multiple chunked submodels in isolation over disjoint
shards of data, and meanwhile, each chunked submodel is
incrementally learned with non-overlapping slices of data
points (named sliced data hereafter); when one data point
is deleted, only particular submodels exactly trained on or
impacted by the data point need to be retrained, which is
therefore faster than retraining a complete model from scratch.
TEEs-capable PoUL implementation. Standing on the
unlearning setting, we incorporate ADS with Intel Software
Guard Extensions (SGX) [3], [24] on the server side, to issue
proofs in the Setup and Deletion phases previously described.
Our incorporation implies (i)an authentication layer for
tracing which sliced data are learned or unlearned, and which
chunked submodels are retrained or used to offer predictions;
and (ii)a proving layer for attesting the execution correctness
of the learning or prediction processes that exactly use
particular authenticated data or submodels.
However, it is challenging to implement the two-layer
incorporation, due to the incompatible features between our
unlearning setting and Intel SGX.
Challenge (i): Memory-Efficient Authentication. In the
authentication layer, we require tracking which sliced
data impact which chunked submodels, and supporting
authenticated updates on the impacted submodels when a data
point is deleted. But both data and submodels are often large-
size and the unlearning problem might cause a large number
of submodels to be updated, which inevitably incurs large
memory footprints far beyond the original memory limitation
of SGX1. Hence, we provide a memory-efficient authentication
design tailored for the traceability from the sliced data to the
updatable submodels while protecting the integrity.
Challenge (ii): Proof of Stateful Computation and Fast
Verification. In the proving layer, the computations involving
a chain of incremental retraining processes to update affected
submodels, and subsequent prediction processes using the
newly retrained model should be attested, so that a verifier
is convinced of the fulfillment of data deletion. Yet, the
computations are stateful, since the incremental retraining and
prediction processes require previously generated states, e.g.,
submodels, which is in conflict with the SGX design of pri-
marily protecting stateless computations. We extend the trust
outside the SGX’s protected memory to securely save previous
1It is about 94 MB available to applications in a widely adopted version,
and a recent release [6] supports 188 MB.