Universal Robust Regression via Maximum Mean Discrepancy

نویسندگان

چکیده

Abstract Many modern datasets are collected automatically and thus easily contaminated by outliers. This has led to a renewed interest in robust estimation, including new notions of robustness such as adversarial contamination the data. However, most estimation methods designed for specific model. Notably, many were proposed recently obtain estimators linear models, or generalized few developed very settings, example beta regression sample selection models. In this paper we develop approach arbitrary based on maximum mean discrepancy minimization. We build two which both proven be Huber-type contamination. non-asymptotic error bound them show that it is also contamination, but estimator computationally more expensive use practice than other one. As by-product our theoretical analysis derive results kernel conditional embedding distributions independent interest.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Optimized Maximum Mean Discrepancy

We propose a method to optimize the representation and distinguishability of samples from two probability distributions, by maximizing the estimated power of a statistical test based on the maximum mean discrepancy (MMD). This optimized MMD is applied to the setting of unsupervised learning by generative adversarial networks (GAN), in which a model attempts to generate realistic samples, and a ...

متن کامل

Maximum Mean Discrepancy Imitation Learning

Imitation learning is an efficient method for many robots to acquire complex skills. Some recent approaches to imitation learning provide strong theoretical performance guarantees. However, there remain crucial practical issues, especially during the training phase, where the training strategy may require execution of control policies that are possibly harmful to the robot or its environment. M...

متن کامل

Generative Models and Model Criticism via Optimized Maximum Mean Discrepancy

We propose a method to optimize the representation and distinguishability of samples from two probability distributions, by maximizing the estimated power of a statistical test based on the maximum mean discrepancy (MMD). This optimized MMD is applied to the setting of unsupervised learning by generative adversarial networks (GAN), in which a model attempts to generate realistic samples, and a ...

متن کامل

Training generative neural networks via Maximum Mean Discrepancy optimization

We consider training a deep neural network to generate samples from an unknown distribution given i.i.d. data. We frame learning as an optimization minimizing a two-sample test statistic—informally speaking, a good generator network produces samples that cause a twosample test to fail to reject the null hypothesis. As our two-sample test statistic, we use an unbiased estimate of the maximum mea...

متن کامل

Testing Hypotheses by Regularized Maximum Mean Discrepancy

Do two data samples come from different distributions? Recent studies of this fundamental problem focused on embedding probability distributions into sufficiently rich characteristic Reproducing Kernel Hilbert Spaces (RKHSs), to compare distributions by the distance between their embeddings. We show that Regularized Maximum Mean Discrepancy (RMMD), our novel measure for kernel-based hypothesis ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Biometrika

سال: 2023

ISSN: ['0006-3444', '1464-3510']

DOI: https://doi.org/10.1093/biomet/asad031