Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
It is of soaring demand to develop statistical analysis tools that are robust against contamination as well as preserving individual data owners' privacy. In spite of the fact that both topics host a rich body of literature, to the best of our knowledge, we are the first to systematically study the connections between the optimality under Huber's contamination model and the local differential privacy (LDP) constraints. We start with a general minimax lower bound result, which disentangles the costs of being robust against Huber's contamination and preserving LDP. We further study four concrete examples: a two-point testing problem, a potentially-diverging mean estimation problem, a nonparametric density estimation problem and a univariate median estimation problem. For each problem, we demonstrate procedures that are optimal in the presence of both contamination and LDP constraints, comment on the connections with the state-of-the-art methods that are only studied under either contamination or privacy constraints, and unveil the connections between robustness and LDP via partially answering whether LDP procedures are robust and whether robust procedures can be efficiently privatised. Overall, our work showcases a promising prospect of joint study for robustness and local differential privacy.
This is joint work with Mengchu Li and Yi Yu.
[-]
It is of soaring demand to develop statistical analysis tools that are robust against contamination as well as preserving individual data owners' privacy. In spite of the fact that both topics host a rich body of literature, to the best of our knowledge, we are the first to systematically study the connections between the optimality under Huber's contamination model and the local differential privacy (LDP) constraints. We start with a general ...
[+]
62C20 ; 62G35 ; 62G10
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
2 y
In the fist part of the talk, we will look to some statistical inverse problems for which the natural framework is no more an Euclidian one.
In the second part we will try to give the initial construction of (not orthogonal) wavelets -of the 80 - by Frazier, Jawerth,Weiss, before the Yves Meyer ORTHOGONAL wavelets theory.
In the third part we will propose a construction of a geometric wavelet theory. In the Euclidian case, Fourier transform plays a fundamental role. In the geometric situation this role is given to some "Laplacian operator" with some properties.
In the last part we will show that the previous theory could help to revisit the topic of regularity of Gaussian processes, and to give a criterium only based on the regularity of the covariance operator.
[-]
In the fist part of the talk, we will look to some statistical inverse problems for which the natural framework is no more an Euclidian one.
In the second part we will try to give the initial construction of (not orthogonal) wavelets -of the 80 - by Frazier, Jawerth,Weiss, before the Yves Meyer ORTHOGONAL wavelets theory.
In the third part we will propose a construction of a geometric wavelet theory. In the Euclidian case, Fourier transform ...
[+]
42C15 ; 43A85 ; 46E35 ; 58J35 ; 43A80 ; 62G05 ; 62G10 ; 62G20
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
When performing multiple testing, adjusting the distribution of the null hypotheses is ubiquitous in applications. However, the effect of such an operation remains largely unknown, especially in terms of false discovery proportion (FDP) and true discovery proportion (TDP). In this talk, we explore this issue in the most classical case where the null distributions are Gaussian with an unknown rescaling parameters (mean and variance) and where the Benjamini-Hochberg (BH) procedure is applied after a datarescaling step.
[-]
When performing multiple testing, adjusting the distribution of the null hypotheses is ubiquitous in applications. However, the effect of such an operation remains largely unknown, especially in terms of false discovery proportion (FDP) and true discovery proportion (TDP). In this talk, we explore this issue in the most classical case where the null distributions are Gaussian with an unknown rescaling parameters (mean and variance) and where the ...
[+]
62G10 ; 62C20 ; 62G30
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
In this paper we study asymptotic properties of random forests within the framework of nonlinear time series modeling. While random forests have been successfully applied in various fields, the theoretical justification has not been considered for their use in a time series setting. Under mild conditions, we prove a uniform concentration inequality for regression trees built on nonlinear autoregressive processes and, subsequently, use this result to prove consistency for a large class of random forests. The results are supported by various simulations. (This is joint work with Mikkel Slot Nielsen.)
[-]
In this paper we study asymptotic properties of random forests within the framework of nonlinear time series modeling. While random forests have been successfully applied in various fields, the theoretical justification has not been considered for their use in a time series setting. Under mild conditions, we prove a uniform concentration inequality for regression trees built on nonlinear autoregressive processes and, subsequently, use this ...
[+]
62G10 ; 60G10 ; 60J05 ; 62M05 ; 62M10
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
We propose a general method for constructing confidence sets and hypothesis tests that have finite-sample guarantees without regularity conditions. We refer to such procedures as “universal.” The method is very simple and is based on a modified version of the usual likelihood ratio statistic, that we call “the split likelihood ratio test” (split LRT) statistic. The (limiting) null distribution of the classical likelihood ratio statistic is often intractable when used to test composite null hypotheses in irregular statistical models. Our method is especially appealing for statistical inference in these complex setups. The method we suggest works for any parametric model and also for some nonparametric models, as long as computing a maximum likelihood estimator (MLE) is feasible under the null. Canonical examples arise in mixture modeling and shape-constrained inference, for which constructing tests and confidence sets has been notoriously difficult. We also develop various extensions of our basic methods. We show that in settings when computing the MLE is hard, for the purpose of constructing valid tests and intervals, it is sufficient to upper bound the maximum likelihood. We investigate some conditions under which our methods yield valid inferences under model-misspecification. Further, the split LRT can be used with profile likelihoods to deal with nuisance parameters, and it can also be run sequentially to yield anytime-valid p-values and confidence sequences. Finally, when combined with the method of sieves, it can be used to perform model selection with nested model classes.
[-]
We propose a general method for constructing confidence sets and hypothesis tests that have finite-sample guarantees without regularity conditions. We refer to such procedures as “universal.” The method is very simple and is based on a modified version of the usual likelihood ratio statistic, that we call “the split likelihood ratio test” (split LRT) statistic. The (limiting) null distribution of the classical likelihood ratio statistic is often ...
[+]
62C05 ; 62F03 ; 62G10 ; 62L12
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
Déposez votre fichier ici pour le déplacer vers cet enregistrement.