
2
to the regression setting, based on Efron [14]’s redistribution-of-mass construction. From a
different perspective, [45] employed a martingale-based approach for fitting CQR, and the
resulting method has been shown to be closely related to [48]’s method [43,46]. Both [48]’s
and [45]’s methods, along with their variants, involve solving a series of quantile regression
problems that can be reformulated as linear programs, solvable by the simplex or interior
point method [3,49,38]. Statistical properties of the aforementioned methods have been well
studied, assuming that the number of covariates, p, is fixed [43,45,50,46]. To this date, the
impact of dimensionality in the increasing-pregime, in which pis allowed to increase with
the number of observations, has remained unclear in the presence of censored outcomes.
In the high-dimensional setting in which p>n, convex and nonconvex penalty functions
are often employed to perform variable selection and to achieve a trade-off between statistical
bias and model complexity. While penalized Cox proportional hazards and AFT models have
been well studied [16,29,7,5], existing work on penalized CQR under the framework of
[48] and [45] in the high-dimensional setting is still lagging. Large-sample properties of
penalized CQR estimators were first derived under the fixed-psetting (p<n), mainly due
to the technical challenges introduced by the sequential nature of the procedure [55,62,65].
More recently, [70] studied a penalized CQR estimator, extending the method of [45] to the
high-dimensional setting (p > n). They showed that the estimation error (under `2-norm) of
the `1-penalized CQR estimator is upper bounded by Oexp(Cs)pslog(p)/nwith high
probability, where C > 0is a dimension-free constant. Compared to the `1-penalized QR
for uncensored data [4], whose convergence rate is of order Opslog(p)/n, there is a
substantial gap in terms of the impact of the sparsity parameter s.
In addition to the above theoretical issues, our study is motivated by the computational
hardness of CQR under the framework of [48] and [45] for problems with large dimension.
Recall that this framework involves fitting a series of quantile regressions sequentially over
a dense grid of quantile indexes, each of which is solvable by the Frisch-Newton algorithm
with computational complexity that grows as a cubic function of p[49]. Moreover, under the
regime in which p < n, the asymptotic covariance matrix of the estimator is rather compli-
cated and thus resampling methods are often used to perform statistical inference [48,45].
A sample-based inference procedure (without resampling) for Peng-Huang’s estimator [45]
is available by adapting the plug-in covariance estimation method from [56]. In the high-
dimensional setting (p > n), computation of the `1-penalized QR is based on either refor-
mulation as linear programs [39] or alternating direction method of multiplier algorithms
[68,22]. These algorithms are generic and applicable to a broad spectrum of problems but
lack scalability. Since the `1-penalized CQR not only requires the estimation of the whole
quantile regression process, but also relies on cross-validation to select the sequence of
(mostly different) penalty levels, the state-of-the-art methods [70,17] can be highly inef-
ficient when applied to large-pproblems.
To illustrate the computational challenge for CQR, we compare the `1-penalized CQR pro-
posed by Zheng et al. [70] and our proposed method by analyzing a gene expression dataset
studied in [53]. In this study, 22,283 genes from 442 lung adenocarcinomas are incorporated
to predict the survival time in lung cancer, with 46.6% subjects that are censored. We im-
plement both methods with quantile grid set as {0.1,0.11,...,0.7}, and use a predetermined
sequence of regularization parameters. For Zheng et al. [70], we use the rqPen package to
compute the `1-penalized QR estimator at each quantile level [54]. The computational time
and maximum allocated memory are reported in Table 1. The reference machine for this
experiment is a worker node with 2.5 GHz 32-core processor and 512 GB of memory in a
high-performance computing cluster.
In this paper, we develop a smoothed framework for CQR that is scalable to problems
with large dimension pin both low- and high-dimensional settings. Our proposed method