Refine
Document Type
- Article (19)
- Preprint (4)
- Monograph/Edited Volume (1)
- Other (1)
Is part of the Bibliography
- yes (25)
Keywords
- adaptive estimation (2)
- conjugate gradient (2)
- discrepancy principle (2)
- early stopping (2)
- minimax convergence rates (2)
- partial least squares (2)
- reproducing kernel Hilbert space (2)
- Banach-valued process (1)
- Bernstein inequality (1)
- Classification (1)
- Distributed Learning (1)
- False discovery rate (1)
- Gaussian sequence model (1)
- Inference post model-selection (1)
- Inverse problem (1)
- Kernel regression (1)
- Landweber iteration (1)
- Linear inverse problems (1)
- Minimax Optimality (1)
- Minimax convergence rates (1)
- Minimax hypothesis testing (1)
- Nonparametric regression (1)
- PoSI constants (1)
- Reproducing kernel Hilbert space (1)
- Spectral Regularization (1)
- Spectral regularization (1)
- Stability selection (1)
- Statistical learning (1)
- Subsampling (1)
- Variable selection (1)
- classification with partial labels (1)
- concentration (1)
- confidence intervals (1)
- consistency (1)
- continuous testing (1)
- eigenvalue decay (1)
- false discovery rate (1)
- generalization bounds (1)
- high-dimensional inference (1)
- kernel method (1)
- label noise (1)
- learning kernels (1)
- least favorable configuration (1)
- linear inverse problems (1)
- linear regression (1)
- local Rademacher complexity (1)
- minimax optimality (1)
- minimax rate (1)
- mixed membership models (1)
- mixture proportion estimation (1)
- multiclass classification with label noise (1)
- multiple kernel learning (1)
- multiple testing (1)
- multiple testing; (1)
- mutual contamination models (1)
- nonasymptotic minimax separation rate (1)
- nonparametric regression (1)
- oracle inequalities (1)
- oracle inequality (1)
- positive correlation (1)
- restricted isometry property (1)
- sparsity (1)
- spectral cut-off (1)
- spectral regularization (1)
- statistical inverse problem (1)
- step-up (1)
- stochastic process (1)
- surrogate loss (1)
- topic modeling (1)
- truncated SVD (1)
- weak dependence (1)
Institute
We consider composite-composite testing problems for the expectation in the Gaussian sequence model where the null hypothesis corresponds to a closed convex subset C of R-d. We adopt a minimax point of view and our primary objective is to describe the smallest Euclidean distance between the null and alternative hypotheses such that there is a test with small total error probability. In particular, we focus on the dependence of this distance on the dimension d and variance 1/n giving rise to the minimax separation rate. In this paper we discuss lower and upper bounds on this rate for different smooth and non-smooth choices for C.
Uniformly valid confidence intervals post model selection in regression can be constructed based on Post-Selection Inference (PoSI) constants. PoSI constants are minimal for orthogonal design matrices, and can be upper bounded in function of the sparsity of the set of models under consideration, for generic design matrices. In order to improve on these generic sparse upper bounds, we consider design matrices satisfying a Restricted Isometry Property (RIP) condition. We provide a new upper bound on the PoSI constant in this setting. This upper bound is an explicit function of the RIP constant of the design matrix, thereby giving an interpolation between the orthogonal setting and the generic sparse setting. We show that this upper bound is asymptotically optimal in many settings by constructing a matching lower bound.
We obtain a Bernstein-type inequality for sums of Banach-valued random variables satisfying a weak dependence assumption of general type and under certain smoothness assumptions of the underlying Banach norm. We use this inequality in order to investigate in the asymptotical regime the error upper bounds for the broad family of spectral regularization methods for reproducing kernel decision rules, when trained on a sample coming from a tau-mixing process.
Many machine learning problems can be characterized by mutual contamination models. In these problems, one observes several random samples from different convex combinations of a set of unknown base distributions and the goal is to infer these base distributions. This paper considers the general setting where the base distributions are defined on arbitrary probability spaces. We examine three popular machine learning problems that arise in this general setting: multiclass classification with label noise, demixing of mixed membership models, and classification with partial labels. In each case, we give sufficient conditions for identifiability and present algorithms for the infinite and finite sample settings, with associated performance guarantees.
We investigate if kernel regularization methods can achieve minimax convergence rates over a source condition regularity assumption for the target function. These questions have been considered in past literature, but only under specific assumptions about the decay, typically polynomial, of the spectrum of the the kernel mapping covariance operator. In the perspective of distribution-free results, we investigate this issue under much weaker assumption on the eigenvalue decay, allowing for more complex behavior that can reflect different structure of the data at different scales.