larger than the sample size n, such as Ollila et al. (2021). Third, some methods propose a
choice of hyperparameter(s) through cross-validation, such as Yu et al. (2017); Yi and Tyler
(2021), which can be computationally expensive. In this paper, we address these problems
by developing a simple algorithm based on nonlinear shrinkage (Ledoit and Wolf (2012,
2015,2020,2022b)), inspired by the above robust approaches and the work of Hediger and
N¨af (2022). In essence, the algorithm applies the quadratic inverse shrinkage (QIS) method
of Ledoit and Wolf (2022b) to appropriately standardized data, thereby greatly increasing
its finite-sample performance in heavy-tailed models. Thus, we refer to the new method
as “Robust Nonlinear Shrinkage” (R-NL); in particular, we extend the proposal of Hediger
and N¨af (2022) from a parametric model to general elliptical distributions. This approach
includes an iteration over the space of orthogonal matrices, which we prove converges to a
stationary point. We motivate our approach using properties of elliptical distributions along
the lines of Chen et al. (2011); Zhang and Wiesel (2016); Ashurbekova et al. (2021) and
demonstrate the favorable performance of our method in a wide range of settings. Notably,
our approach (i) greatly improves the performance of (standard) nonlinear shrinkage in
heavy-tailed settings; does not deteriorate when moving from heavy to Gaussian tails; (iii)
can handle the case pąn; and (iv) does not require the choice of a tuning parameter.
The remainder of the article is organized as follows. Section 1.1 lists our contribu-
tions. Section 2presents an example to motivate our methodology. Section 3describes
the proposed new methodology and provides results concerning the convergence of the new
algorithm. Section 4showcases the performance of our method in a simulation study using
various settings for both pănand pąn. Section 5applies our method to financial data,
illustrating the performance of the method on real data.
1.1 Contributions
To the best of our knowledge, no paper has so far attempted to combine nonlinear shrinkage
of Ledoit and Wolf (2012,2015,2020,2022b) with Tyler’s method. As such, our approach
differs markedly from previous ones. It is partly based on an M-estimator interpretation,
but also adds the nonparametric nonlinear shrinkage approach. A downside of this approach
is that theoretical convergence results are harder to come by. Nonetheless, we are able to
show that the iterative part of our algorithm converges to a stationary point, a crucial
result for the practical usefulness of the algorithm.
Maybe the closest paper to our method is Breloy et al. (2019), where the eigenvalues of
Tyler’s estimator are iteratively shrunken towards predetermined target eigenvalues, with a
parameter αdetermining the shrinkage strength. Through different objectives, they arrive
at an algorithm from which the iterative part of our Algorithm 2can be recovered when
setting α“ 8. Additionally, using the eigenvalues from nonlinear shrinkage as the target
eigenvalues, their method presents an alternative way of combining Tyler’s estimator with
nonlinear shrinkage. Though they did not originally propose this, this was suggested by
an anonymous reviewer. However, while there is an overlap in the two algorithms for the
corner case of α“ 8, they arrive at their Algorithm 1 from a different angle than we do.
Consequently, their theoretical results cannot be applied in our analysis. Moreover, they
do not suggest how to choose the tuning parameter α. In Appendix A, simulations indicate
that when the target eigenvalues are obtained from nonlinear shrinkage, setting α“ 8, and
thus maximally shrinking towards the nonlinear shrinkage eigenvalues, is usually beneficial.
3