Expectation-Maximization via Pretext-Invariant Representations
نویسندگان
چکیده
Contrastive learning methods have been widely adopted in numerous unsupervised and self-supervised visual representation methods. Such algorithms aim to maximize the cosine similarity between two positive samples while minimizing that of negative samples. Recently, Grill et al. propose an algorithm, BYOL [1], utilize only samples, completely giving up on ones, by introducing a Siamese-like asymmetric architecture. Although many recent state-of-the-art (SOTA) adopt architecture, most them simply introduce additional neural network, predictor, without much exploration asymmetrical In contrast, He SimSiam [2], simple Siamese architecture relying stop-gradient operation instead momentum encoder describe framework from perspective Expectation-Maximization. We argue BYOL-like attain suboptimal performance due inconsistency during training. this work, we explain novel objective, Expectation-Maximization via Pretext-Invariant Representations (EMPIR), which enhances Expectation-Maximization-based optimization enforcing augmentation invariance within local region k nearest neighbors, resulting consistent learning. other words, as core task architectures. show it consistently outperforms decent margin. also demonstrate its transfer capabilities downstream image recognition tasks.
منابع مشابه
Expectation Maximization
The Expectation Maximization (EM) algorithm [1, 2] is one of the most widely used algorithms in statistics. Suppose we are given some observed data X and a model family parametrized by θ, and would like to find the θ which maximizes p(X |θ), i.e. the maximum likelihood estimator. The basic idea of EM is actually quite simple: when direct maximization of p(X |θ) is complicated we can augment the...
متن کاملWeek 9: Expectation Maximization
Last week, we saw how we could represent clustering with a probabilistic model. In this model, called a Gaussian mixture model, we model each datapoint x i as originating from some cluster, with a corresponding cluster label y i distributed according to p(y), and the corresponding distribution for that cluster given by a multivariate Gaussian: p(x|y = k) =
متن کاملRelational Neural Expectation Maximization
Common-sense physical reasoning is an essential ingredient for any intelligent agent operating in the real-world. For example, it can be used to simulate the environment, or to infer the state of parts of the world that are currently unobserved. In order to match real-world conditions this causal knowledge must be learned without access to supervised data. To solve this problem, we present a no...
متن کاملNeural Expectation Maximization
We introduce a novel framework for clustering that combines generalized EM with neural networks and can be implemented as an end-to-end differentiable recurrent neural network. It learns its statistical model directly from the data and can represent complex non-linear dependencies between inputs. We apply our framework to a perceptual grouping task and empirically verify that it yields the inte...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2023
ISSN: ['2169-3536']
DOI: https://doi.org/10.1109/access.2023.3289589