Multi-Modal Joint Embedding Space Learning for Cross-Modalit Retrieval

  • Matthias Dorfer (Speaker)

Activity: Talk or presentationInvited talkscience-to-science

Description

Cross-modality retrieval encompasses retrieval tasks where the fetched items are of a different type than the search query, e.g., retrieving pictures relevant to a given text query. The state-of-the-art approach to cross-modality retrieval relies on learning a joint embedding space of the two modalities, where items from either modality are retrieved using nearest-neighbor search. In my talk I will review two different learning paradigms -- Deep Canonical Correlation Analysis and Pairwise Ranking Losses — which both yield embedding spaces exhibiting properties beneficial for retrieval. I will present potential application scenarios as well as experimental retrieval results on two different modality pairs, namely text and images as well as audio and sheet-music.
Period16 Oct 2018
Event titleunbekannt/unknown
Event typeOther
LocationAustriaShow on map

Fields of science

  • 202002 Audiovisual media
  • 102 Computer Sciences
  • 102001 Artificial intelligence
  • 102015 Information systems
  • 102003 Image processing

JKU Focus areas

  • Computation in Informatics and Mathematics
  • Engineering and Natural Sciences (in general)