
Federated Boosted Decision Trees with Dierential Privacy∗
Samuel Maddock†
University of Warwick
Graham Cormode
Meta AI
Tianhao Wang‡
University of Virginia
Carsten Maple
University of Warwick
Somesh Jha‡
University of Wisconsin-Madison
ABSTRACT
There is great demand for scalable, secure, and ecient privacy-
preserving machine learning models that can be trained over dis-
tributed data. While deep learning models typically achieve the
best results in a centralized non-secure setting, dierent models can
excel when privacy and communication constraints are imposed. In-
stead, tree-based approaches such as XGBoost have attracted much
attention for their high performance and ease of use; in particular,
they often achieve state-of-the-art results on tabular data. Conse-
quently, several recent works have focused on translating Gradient
Boosted Decision Tree (GBDT) models like XGBoost into feder-
ated settings, via cryptographic mechanisms such as Homomorphic
Encryption (HE) and Secure Multi-Party Computation (MPC). How-
ever, these do not always provide formal privacy guarantees, or
consider the full range of hyperparameters and implementation
settings. In this work, we implement the GBDT model under Dier-
ential Privacy (DP). We propose a general framework that captures
and extends existing approaches for dierentially private decision
trees. Our framework of methods is tailored to the federated setting,
and we show that with a careful choice of techniques it is possible to
achieve very high utility while maintaining strong levels of privacy.
KEYWORDS
Gradient Boosting, Dierential Privacy, Federated Learning
1 INTRODUCTION
It is well known that machine learning models can leak private in-
formation about individuals in the training set [
14
,
58
]. Dierential
privacy (DP) [
21
] is a popular denition that has been developed to
mitigate such privacy risks and has become the dominant notion
of privacy in recent years. Much of the current research on private
machine learning is focused on training deep learning models with
dierential privacy [
1
,
35
,
41
,
60
]. DP is often combined with feder-
ated learning, where data resides on client devices, and only small
information about model updates is collected from clients, in order
to further minimize the privacy risk [36].
While deep learning models are powerful for a range of real-
world tasks in a centralized setting, they are sometimes beaten by
“simpler” models on tabular datasets. One such competitor is Gradi-
ent Boosted Decision Trees (GBDTs) [
22
,
31
,
59
] [
32
]. GBDT meth-
ods build an ensemble of weak decision trees that incrementally
correct for past mistakes in training to improve predictions. Many
GBDT frameworks such as XGBoost [
16
], LightGBM [
38
], and Cat-
Boost [
19
] have seen widespread industry adoption [
6
,
11
,
40
,
44
].
∗Full version of a paper to appear at ACM CCS’22
†Author correspondence to s.maddock@warwick.ac.uk
‡Work was done while part-time at Meta AI
GBDT methods are an attractive alternative to deep learning due
to their speed, scalability, ease of use, and impressive performance
on tabular datasets.
Recent works have studied GBDT implementations such as XG-
Boost under secure training in the federated setting [
17
,
18
,
24
,
48
].
These methods typically rely on cryptographic techniques such
as Homomorphic Encryption (HE) or Secure Multi-Party Compu-
tation (MPC). While this allows secure joint training of a GBDT
model without any participant directly releasing their data, the
end model may not necessarily be private and will not guarantee
formal dierential privacy (DP) [
23
]. For instance, in the case of
decision trees, split decisions in a tree can directly reveal sensitive
information regarding the training set. Moreover, such reliance on
heavyweight cryptographic techniques such as HE or MPC often
makes methods computationally intensive or require a large num-
ber of communication rounds, making them impractical to scale
beyond more than a few participants [17, 42].
In parallel, many works have studied decision tree models under
the central model of DP [
25
,
52
,
65
]. Most studies focus on training
random forest (RF) models and there has been little research to
explore trade-os between gradient boosting and DP; those that do
often use central DP mechanisms that cannot easily be extended to
federated settings [
33
]. It therefore remains an open problem to im-
plement GBDTs in the federated DP setting, and show how to obtain
utility comparable to their centralized non-private counterparts.
Our focus is on DP-GBDT methods that operate within the fed-
erated setting via lightweight MPC methods such as secure aggre-
gation [
7
,
10
]. This setting has recently risen to prominence, as it
promises an attractive trade-o between the computational e-
ciency of central DP techniques and the security of cryptographic
methods. Recent federated works that consider GBDTs have pro-
posed methods under the local model of DP, but due to the use of
local noise, incur a signicant loss in utility [43, 61, 62].
In this paper, we bring together existing methods under a unied
framework where we propose techniques to satisfy DP that are
well suited to the federated setting. We nd that by dissecting the
GBDT algorithm into its constituent parts and carefully considering
the options for each component of the algorithm we can identify
specic combinations that achieve the best balance of privacy and
utility. We also emphasise variants that can train such private GBDT
models in only a small number of communication rounds, which is
of particular importance to the federated setting.
Our high-level nding is that it is possible to achieve high perfor-
mance with GBDT models, even comparable to that of non-private
methods. In order to do so, one must allocate privacy budget to
the quantities that are most important for the learning process. For
example, we show that spending such budget on computing split
decisions of trees is not as important as spending it on the leaf
arXiv:2210.02910v1 [cs.CR] 6 Oct 2022