Abstract

We attempt to recover an n-dimensional vector observed in white noise, where n is large and the vector is known to be sparse, but the degree of sparsity is unknown. We consider three different ways of defining sparsity of a vector: using the fraction of nonzero terms; imposing power-law decay bounds on the ordered entries; and controlling the ℓ<sub>p</sub> norm for p small. We obtain a procedure which is asymptotically minimax for ℓ<sup>r</sup> loss, simultaneously throughout a range of such sparsity classes.\n¶ The optimal procedure is a data-adaptive thresholding scheme, driven by control of the false discovery rate (FDR). FDR control is a relatively recent innovation in simultaneous testing, ensuring that at most a certain expected fraction of the rejected null hypotheses will correspond to false rejections.\n¶ In our treatment, the FDR control parameter q<sub>n</sub> also plays a determining role in asymptotic minimaxity. If q=lim q<sub>n</sub>∈[0,1/2] and also q<sub>n</sub>>γ/log(n), we get sharp asymptotic minimaxity, simultaneously, over a wide range of sparse parameter spaces and loss functions. On the other hand, q=lim q<sub>n</sub>∈(1/2,1] forces the risk to exceed the minimax risk by a factor growing with q.\n¶ To our knowledge, this relation between ideas in simultaneous inference and asymptotic decision theory is new.\n¶ Our work provides a new perspective on a class of model selection rules which has been introduced recently by several authors. These new rules impose complexity penalization of the form 2⋅log(potential model size/actual model sizes). We exhibit a close connection with FDR-controlling procedures under stringent control of the false discovery rate.

Keywords

MinimaxMathematicsFalse discovery rateThresholdingFraction (chemistry)Range (aeronautics)InferenceMultiple comparisons problemNorm (philosophy)Applied mathematicsCombinatoricsMathematical optimizationDiscrete mathematicsStatisticsComputer scienceArtificial intelligence

Affiliated Institutions

Related Publications

A Direct Approach to False Discovery Rates

Summary Multiple-hypothesis testing involves guarding against much more complicated errors than single-hypothesis testing. Whereas we typically control the type I error rate for...

2002 Journal of the Royal Statistical Soci... 5607 citations

Publication Info

Year
2006
Type
article
Volume
34
Issue
2
Citations
458
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

458
OpenAlex

Cite This

Felix Abramovich, Yoav Benjamini, David L. Donoho et al. (2006). Adapting to unknown sparsity by controlling the false discovery rate. The Annals of Statistics , 34 (2) . https://doi.org/10.1214/009053606000000074

Identifiers

DOI
10.1214/009053606000000074