major components: the cost function, the search method, and the penalty/constraint
(Truong et al.,2020). The choice of the cost function and search method has a crucial
impact on the method’s computational complexity. As increasingly larger data sets are
being collected in modern applications, there is an urgent need to develop more efficient
algorithms to handle such big data sets. Examples include testing the structure breaks
for genetics data and detecting changes in the volatility of big financial data.
One popular way to tackle the change-point detection problem is to cast it into a
model-selection problem by solving a penalized optimization problem over possible num-
bers and locations of change-points. The framework includes several well-established
procedures, such as the penalized likelihood and minimum description length. The cor-
responding optimization can be solved exactly using dynamic programming (Auger and
Lawrence,1989;Jackson et al.,2005) whose computational cost is PT
t=1 Pt
s=1 q(s), where
Tis the number of data points and q(s) denotes the time complexity for calculating the
cost function value based on sdata points. Killick et al. (2012) introduced the pruned
exact linear time (PELT) algorithm with a pruning step in dynamic programming. PELT
reduces the computational cost without affecting the exactness of the resulting segmenta-
tion. Rigaill (2010) proposed an alternative pruned dynamic programming algorithm with
the aim of reducing the computational effort. However, in the worst case scenario, the
computational cost of dynamic programming coupled with the above pruning strategies
remains the order of O(PT
t=1 Pt
s=1 q(s)).
Unlike the pruning strategy, this paper aims to improve the computational efficiency
of dynamic programming from a different perspective. We focus on the class of problems
where the cost function involves solving a non-trivial optimization problem without a
closed-form solution. Dynamic programming requires repeatedly solving the optimization
over different data sequence segments, which can be very time-consuming for big data.
This paper makes the following contributions to address the issue.
1. A new sequential updating method (SE) that can be coupled with the gradient
descent (SeGD) and quasi-Newton’s method (SeN) is proposed to update the pa-
rameter estimate and the cost value in dynamic programming. The new strategy
avoids repeatedly optimizing the objective function based on each data segment.
It thus significantly improves the computational efficiency of the vanilla PELT, es-
pecially when the cost function involves solving a non-trivial optimization problem
without a closed-form solution. Though our algorithm is no longer exact, numerical
studies suggest that the new method achieves almost the same estimation accuracy
as PELT does.
2. SeGD is related to the stochastic gradient descent (SGD) without-replacement sam-
pling (Shamir,2016;Nagaraj et al.,2019;Rajput et al.,2020). The main difference
is that our update is along the time order of the data points, and hence no sam-
pling or additional randomness is introduced. Using some techniques from SGD
and transductive learning theory, we obtain the convergence rate of the approxi-
mate cost value derived from the algorithm to the true cost value.
3. The proposed method applies to a broad class of statistical models, such as para-
metric likelihood models, generalized linear models, nonparametric models, and
penalized regression.
Finally, we mention two other routes to reduce the computational complexity in
change-point analysis. The first one is to relax the l0penalty on the number of parameters
2