Picture for Dong-Jin Kim

Dong-Jin Kim

djnjusa@kaist.ac.kr

Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality

Add code
Oct 07, 2024
Figure 1 for Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality
Figure 2 for Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality
Figure 3 for Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality
Figure 4 for Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality
Viaarxiv icon

IFCap: Image-like Retrieval and Frequency-based Entity Filtering for Zero-shot Captioning

Add code
Sep 26, 2024
Viaarxiv icon

Self-Sufficient Framework for Continuous Sign Language Recognition

Add code
Mar 21, 2023
Viaarxiv icon

Semi-Supervised Image Captioning by Adversarially Propagating Labeled Data

Add code
Jan 26, 2023
Viaarxiv icon

Signing Outside the Studio: Benchmarking Background Robustness for Continuous Sign Language Recognition

Add code
Nov 01, 2022
Figure 1 for Signing Outside the Studio: Benchmarking Background Robustness for Continuous Sign Language Recognition
Figure 2 for Signing Outside the Studio: Benchmarking Background Robustness for Continuous Sign Language Recognition
Figure 3 for Signing Outside the Studio: Benchmarking Background Robustness for Continuous Sign Language Recognition
Figure 4 for Signing Outside the Studio: Benchmarking Background Robustness for Continuous Sign Language Recognition
Viaarxiv icon

Single-Modal Entropy based Active Learning for Visual Question Answering

Add code
Nov 18, 2021
Figure 1 for Single-Modal Entropy based Active Learning for Visual Question Answering
Figure 2 for Single-Modal Entropy based Active Learning for Visual Question Answering
Figure 3 for Single-Modal Entropy based Active Learning for Visual Question Answering
Figure 4 for Single-Modal Entropy based Active Learning for Visual Question Answering
Viaarxiv icon

ACP++: Action Co-occurrence Priors for Human-Object Interaction Detection

Add code
Sep 09, 2021
Figure 1 for ACP++: Action Co-occurrence Priors for Human-Object Interaction Detection
Figure 2 for ACP++: Action Co-occurrence Priors for Human-Object Interaction Detection
Figure 3 for ACP++: Action Co-occurrence Priors for Human-Object Interaction Detection
Figure 4 for ACP++: Action Co-occurrence Priors for Human-Object Interaction Detection
Viaarxiv icon

MCDAL: Maximum Classifier Discrepancy for Active Learning

Add code
Jul 23, 2021
Figure 1 for MCDAL: Maximum Classifier Discrepancy for Active Learning
Figure 2 for MCDAL: Maximum Classifier Discrepancy for Active Learning
Figure 3 for MCDAL: Maximum Classifier Discrepancy for Active Learning
Figure 4 for MCDAL: Maximum Classifier Discrepancy for Active Learning
Viaarxiv icon

Distribution-Aware Semantics-Oriented Pseudo-label for Imbalanced Semi-Supervised Learning

Add code
Jun 10, 2021
Figure 1 for Distribution-Aware Semantics-Oriented Pseudo-label for Imbalanced Semi-Supervised Learning
Figure 2 for Distribution-Aware Semantics-Oriented Pseudo-label for Imbalanced Semi-Supervised Learning
Figure 3 for Distribution-Aware Semantics-Oriented Pseudo-label for Imbalanced Semi-Supervised Learning
Figure 4 for Distribution-Aware Semantics-Oriented Pseudo-label for Imbalanced Semi-Supervised Learning
Viaarxiv icon

Dealing with Missing Modalities in the Visual Question Answer-Difference Prediction Task through Knowledge Distillation

Add code
Apr 13, 2021
Figure 1 for Dealing with Missing Modalities in the Visual Question Answer-Difference Prediction Task through Knowledge Distillation
Figure 2 for Dealing with Missing Modalities in the Visual Question Answer-Difference Prediction Task through Knowledge Distillation
Figure 3 for Dealing with Missing Modalities in the Visual Question Answer-Difference Prediction Task through Knowledge Distillation
Figure 4 for Dealing with Missing Modalities in the Visual Question Answer-Difference Prediction Task through Knowledge Distillation
Viaarxiv icon