(self-supervised learning)Event Camera Data Pre-training

发布时间:2023年12月30日

Publisher:?ICCV 2023?

MOTIVATION OF READING: 自监督学习、稀疏事件 = NILM

link:?https://arxiv.org/pdf/2301.01928.pdf

Code:?GitHub - Yan98/Event-Camera-Data-Pre-training


1. Overview

Contributions are summarized as follows:

1.?A self-supervised framework for event camera data pre-training. The pre-trained model can be transferred to diverse downstream tasks;

2.?A family of event data augmentations, generating meaningful event images;

3.?A conditional masking strategy, sampling informative event patches for network training;

4.?An embedding projection loss, using paired RGB embeddings to regularize event embeddings to avoid model collapse;

5.?A probability distribution alignment loss for aligning embeddings from the paired event and RGB images.

6.?We achieve state-of-the-art performance in standard event benchmark datasets.

2. Related work

The SSL frameworks can be generally divided into two categories: contrastive learning and masked modeling.

2.1?Contrastive learning

This approach generally assumes augmentation invariance of images.?one notable drawback
of contrastive learning is suffering from model collapse and training instability.

2.2 Masked modeling

Reconstructing masked inputs from the (i. e., unmasked) visible ones is a popular selfsupervised
learning objective motivated by the idea of autoencoding. (Bert, GPT)

3. Methodology

For pre-training, our method takes event data E and its paired natural RGB image I as inputs, and outputs a pre-trained network fe.

Firstly, consecutively perform data augmentations, event image generation, and conditional masking to obtain two patch sets (xq, xk).

Secondly,?fe extracts features from event patch set xq,?and he_img?and he_evt?separately project features from fe to latent embeddings q_img and q_evt.

fm and hm_evt?are the momentum of fe and he_evt, and are updated by the exponential moving average (EMA). (momentum的含义可以参考MOCO论文)

The momentum network takes patch set xk as input and generates an embedding k_evt.

At the same time, the natural RGB image I is embeded into y = f1(h1(I)).

Finally, we perform event discrimination, and event and natural RGB image discrimination to train our model. 这里不用INFONCE直接对q_evt和k_evt进行相似度计算是因为这么做会导致embedding collapse使得embedding过于相似。原因是事件图像是稀疏离散。因此使用RGB图像的映射。

L_evt is an event embedding projection loss aiming to pull together paired event embeddings qevt and kevt, for event discrimination.

L_RGB aims to pull together paired event and RGB embeddings q_evt and y,?for event and natural RGB image discrimination.

L_k1?aims to drive fe learning discriminative event embeddings, towards well-structured embedding space of natural RGB images.


InfoNCE loss?Contrastive learning aims to pull together embeddings q and k+, and pushes away embeddings q and {k?}.

Event embedding projection loss

ζ(v1, v2) is the projection function.

Event and RGB image discrimination

Considering the sparsity of the event image, a single event image is less informative than an RGB image, possessing difficulty for self-supervised event network training.

We pull together embeddings of paired event and RGB images, xq and I.

we first compute the pairwise embedding similarity and then fit an exponential kernel to the similarities to compute probability scores.?The probability score of the (i, j)-th pair is given by,

Our probability distribution alignment loss is given by,

Total Loss

where λ1 is a hyper-parameter for balancing the losses.

4. Experiment

We evaluate our method?on three downstream tasks: object recognition, optical flow estimation, and semantic segmentation.

文章来源:https://blog.csdn.net/shadowismine/article/details/135275039
本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。