Cross-modal Knowledge Transfer - Improving the Word Embedding of Apple by Looking at Oranges (Both et al. 2017)

Overview

Key Results

Constructing a tri-modal concept space

E

Transferring multi-modal knowledge to single modalities

E

  1. Feature Mask:

    • For common concepts in two embedding spaces, calculate correlation(differences in each feature, similarity between two vectors)

    • Normalize this correlation vector by a sigmoid

    • Element-wise product between the single embedding vector and the feature mask

  2. Feature Rebuilding

    • Train a mapping function (in this case, NN) between the single embedding vector for concept A --> feature n of concept A in the fused embedding space

    • Train a mapping function for every dimension of the fused embedding space

    • Normalize this artificial vector and concatenate it with the originam single embedding vector

Comments