SELF-SUPERVISED REPRESENTATION LEARNING USING BOOTSTRAPPED LATENT REPRESENTATIONS
Applicants
GDM Holding LLC
Inventors
GRILL, Jean-Bastien François Laurent, STRUB, Florian, ALTCHÉ, Florent, TALLEC, Corentin, RICHEMOND, Pierre, PIRES, Bernardo Avila, GUO, Zhaohan, AZAR, Mohammad Gheshlaghi, PIOT, Bilal, MUNOS, Remi, VALKO, Michal
Abstract
A computer-implemented method of training a neural network. The method comprises processing a first transformed view of a training data item, e.g. an image, with a target neural network to generate a target output, processing a second transformed view of the training data item, e.g. image, with an online neural network to generate a prediction of the target output, updating parameters of the online neural network to minimize an error between the prediction of the target output and the target output, and updating parameters of the target neural network based on the parameters of the online neural network. The method can effectively train an encoder neural network without using labelled training data items, and without using a contrastive loss, i.e. without needing "negative examples" which comprise transformed views of different data items.
IPC Classifications
Designated States
AL, AT, BE, BG, CH, CY, CZ, DE, DK, EE, ES, FI, FR, GB, GR, HR, HU, IE, IS, IT, LI, LT, LU, LV, MC, MK, MT, NL, NO, PL, PT, RO, RS, SE, SI, SK, SM, TR