Picture for Thomas Hummel

Thomas Hummel

EgoCVR: An Egocentric Benchmark for Fine-Grained Composed Video Retrieval

Add code
Jul 23, 2024
Viaarxiv icon

Video-adverb retrieval with compositional adverb-action embeddings

Add code
Sep 26, 2023
Viaarxiv icon

Text-to-feature diffusion for audio-visual few-shot learning

Add code
Sep 07, 2023
Figure 1 for Text-to-feature diffusion for audio-visual few-shot learning
Figure 2 for Text-to-feature diffusion for audio-visual few-shot learning
Figure 3 for Text-to-feature diffusion for audio-visual few-shot learning
Figure 4 for Text-to-feature diffusion for audio-visual few-shot learning
Viaarxiv icon

Semantic Image Synthesis with Semantically Coupled VQ-Model

Add code
Sep 06, 2022
Figure 1 for Semantic Image Synthesis with Semantically Coupled VQ-Model
Figure 2 for Semantic Image Synthesis with Semantically Coupled VQ-Model
Figure 3 for Semantic Image Synthesis with Semantically Coupled VQ-Model
Figure 4 for Semantic Image Synthesis with Semantically Coupled VQ-Model
Viaarxiv icon

Temporal and cross-modal attention for audio-visual zero-shot learning

Add code
Jul 20, 2022
Figure 1 for Temporal and cross-modal attention for audio-visual zero-shot learning
Figure 2 for Temporal and cross-modal attention for audio-visual zero-shot learning
Figure 3 for Temporal and cross-modal attention for audio-visual zero-shot learning
Figure 4 for Temporal and cross-modal attention for audio-visual zero-shot learning
Viaarxiv icon

Where and When: Space-Time Attention for Audio-Visual Explanations

Add code
May 04, 2021
Figure 1 for Where and When: Space-Time Attention for Audio-Visual Explanations
Figure 2 for Where and When: Space-Time Attention for Audio-Visual Explanations
Figure 3 for Where and When: Space-Time Attention for Audio-Visual Explanations
Figure 4 for Where and When: Space-Time Attention for Audio-Visual Explanations
Viaarxiv icon

Crossmodal Language Grounding in an Embodied Neurocognitive Model

Add code
Jun 24, 2020
Figure 1 for Crossmodal Language Grounding in an Embodied Neurocognitive Model
Figure 2 for Crossmodal Language Grounding in an Embodied Neurocognitive Model
Figure 3 for Crossmodal Language Grounding in an Embodied Neurocognitive Model
Figure 4 for Crossmodal Language Grounding in an Embodied Neurocognitive Model
Viaarxiv icon