DACO seminar

×

Modal title

Modal content

Bitte abonnieren Sie hier, wenn Sie über diese Veranstaltungen per E-Mail benachrichtigt werden möchten. Ausserdem können Sie auch den iCal/ics-Kalender abonnieren.

Herbstsemester 2024

Datum / Zeit Referent:in Titel Ort
17. September 2024
14:15-15:15
Daniil Dmitriev
ETH Zurich
Details

DACO Seminar

Titel Robust Mixture Learning when Outliers Overwhelm Small Groups
Referent:in, Affiliation Daniil Dmitriev, ETH Zurich
Datum, Zeit 17. September 2024, 14:15-15:15
Ort HG G 19.1
Abstract We study the problem of estimating the means of well-separated mixtures when an adversary may add arbitrary outliers. While strong guarantees are available when the outlier fraction is significantly smaller than the minimum mixing weight, much less is known when outliers may crowd out low-weight clusters - a setting we refer to as list-decodable mixture learning (LD-ML). In this case, adversarial outliers can simulate additional spurious mixture components. Hence, if all means of the mixture must be recovered up to a small error in the output list, the list size needs to be larger than the number of (true) components. We propose an algorithm that obtains order-optimal error guarantees for each mixture mean with a minimal list-size overhead, significantly improving upon list-decodable mean estimation, the only existing method that is applicable for LD-ML. Although improvements are observed even when the mixture is non-separated, our algorithm achieves particularly strong guarantees when the mixture is separated: it can leverage the mixture structure to partially cluster the samples before carefully iterating a base learner for list-decodable mean estimation at different scales.
Robust Mixture Learning when Outliers Overwhelm Small Groupsread_more
HG G 19.1
24. September 2024
14:15-15:15
Vanessa Piccolo
ENS Lyon, FR
Details

DACO Seminar

Titel Dynamics of optimization in high dimensions for multi-spiked tensor PCA
Referent:in, Affiliation Vanessa Piccolo, ENS Lyon, FR
Datum, Zeit 24. September 2024, 14:15-15:15
Ort HG G 19.1
Abstract We will study the high-dimensional statistical problem of multi-spiked tensor PCA, where the goal is to infer a finite number of unknown, orthogonal signal vectors (or spikes) from noisy observations of a p-tensor. I will present our recent results on the sample complexity required for stochastic gradient descent to efficiently recover the signal vectors from natural initializations. In particular, we will show that it is possible to recover a permutation of all spikes provided a number of sample scaling as N^{p-2}, aligning with the computational threshold identified in the rank-one tensor PCA problem. The recovery process is governed by a sequential elimination phenomenon. As one correlation exceeds an explicit critical threshold, all correlations that share a row or column index become sufficiently small to be negligible, allowing the subsequent correlation to grow and become macroscopic. The order in which correlations become macroscopic is determined by their initial values and the associated signal-to-noise ratios. Based on recent joint work with Gérard Ben Arous (NYU) and Cédric Gerbelot (NYU).
Dynamics of optimization in high dimensions for multi-spiked tensor PCAread_more
HG G 19.1
JavaScript wurde auf Ihrem Browser deaktiviert