Picture for Yuki M. Asano

Yuki M. Asano

Learning to Ground VLMs without Forgetting

Add code
Oct 14, 2024
Viaarxiv icon

TULIP: Token-length Upgraded CLIP

Add code
Oct 13, 2024
Viaarxiv icon

TVBench: Redesigning Video-Language Evaluation

Add code
Oct 10, 2024
Viaarxiv icon

Do better language models have crisper vision?

Add code
Oct 09, 2024
Figure 1 for Do better language models have crisper vision?
Figure 2 for Do better language models have crisper vision?
Figure 3 for Do better language models have crisper vision?
Figure 4 for Do better language models have crisper vision?
Viaarxiv icon

Self-Masking Networks for Unsupervised Adaptation

Add code
Sep 11, 2024
Figure 1 for Self-Masking Networks for Unsupervised Adaptation
Figure 2 for Self-Masking Networks for Unsupervised Adaptation
Figure 3 for Self-Masking Networks for Unsupervised Adaptation
Figure 4 for Self-Masking Networks for Unsupervised Adaptation
Viaarxiv icon

Foundation Model or Finetune? Evaluation of few-shot semantic segmentation for river pollution

Add code
Sep 05, 2024
Viaarxiv icon

Rethinking Image Super-Resolution from Training Data Perspectives

Add code
Sep 01, 2024
Viaarxiv icon

SelEx: Self-Expertise in Fine-Grained Generalized Category Discovery

Add code
Aug 26, 2024
Viaarxiv icon

NeCo: Improving DINOv2's spatial representations in 19 GPU hours with Patch Neighbor Consistency

Add code
Aug 20, 2024
Figure 1 for NeCo: Improving DINOv2's spatial representations in 19 GPU hours with Patch Neighbor Consistency
Figure 2 for NeCo: Improving DINOv2's spatial representations in 19 GPU hours with Patch Neighbor Consistency
Figure 3 for NeCo: Improving DINOv2's spatial representations in 19 GPU hours with Patch Neighbor Consistency
Figure 4 for NeCo: Improving DINOv2's spatial representations in 19 GPU hours with Patch Neighbor Consistency
Viaarxiv icon

Scaling Backwards: Minimal Synthetic Pre-training?

Add code
Aug 03, 2024
Figure 1 for Scaling Backwards: Minimal Synthetic Pre-training?
Figure 2 for Scaling Backwards: Minimal Synthetic Pre-training?
Figure 3 for Scaling Backwards: Minimal Synthetic Pre-training?
Figure 4 for Scaling Backwards: Minimal Synthetic Pre-training?
Viaarxiv icon