Picture for Xiaoqin Zhang

Xiaoqin Zhang

COMPrompter: reconceptualized segment anything model with multiprompt network for camouflaged object detection

Add code
Nov 28, 2024
Viaarxiv icon

Historical Test-time Prompt Tuning for Vision Foundation Models

Add code
Oct 27, 2024
Figure 1 for Historical Test-time Prompt Tuning for Vision Foundation Models
Figure 2 for Historical Test-time Prompt Tuning for Vision Foundation Models
Figure 3 for Historical Test-time Prompt Tuning for Vision Foundation Models
Figure 4 for Historical Test-time Prompt Tuning for Vision Foundation Models
Viaarxiv icon

LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language Models

Add code
Oct 15, 2024
Figure 1 for LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language Models
Figure 2 for LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language Models
Figure 3 for LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language Models
Figure 4 for LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language Models
Viaarxiv icon

Exploring Deeper! Segment Anything Model with Depth Perception for Camouflaged Object Detection

Add code
Jul 17, 2024
Figure 1 for Exploring Deeper! Segment Anything Model with Depth Perception for Camouflaged Object Detection
Figure 2 for Exploring Deeper! Segment Anything Model with Depth Perception for Camouflaged Object Detection
Figure 3 for Exploring Deeper! Segment Anything Model with Depth Perception for Camouflaged Object Detection
Figure 4 for Exploring Deeper! Segment Anything Model with Depth Perception for Camouflaged Object Detection
Viaarxiv icon

CorrMAE: Pre-training Correspondence Transformers with Masked Autoencoder

Add code
Jun 09, 2024
Figure 1 for CorrMAE: Pre-training Correspondence Transformers with Masked Autoencoder
Figure 2 for CorrMAE: Pre-training Correspondence Transformers with Masked Autoencoder
Figure 3 for CorrMAE: Pre-training Correspondence Transformers with Masked Autoencoder
Figure 4 for CorrMAE: Pre-training Correspondence Transformers with Masked Autoencoder
Viaarxiv icon

One-shot Training for Video Object Segmentation

Add code
May 22, 2024
Viaarxiv icon

MonoMAE: Enhancing Monocular 3D Detection through Depth-Aware Masked Autoencoders

Add code
May 13, 2024
Viaarxiv icon

Observation, Analysis, and Solution: Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training

Add code
Apr 18, 2024
Figure 1 for Observation, Analysis, and Solution: Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training
Figure 2 for Observation, Analysis, and Solution: Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training
Figure 3 for Observation, Analysis, and Solution: Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training
Figure 4 for Observation, Analysis, and Solution: Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-Training
Viaarxiv icon

Masked AutoDecoder is Effective Multi-Task Vision Generalist

Add code
Mar 14, 2024
Figure 1 for Masked AutoDecoder is Effective Multi-Task Vision Generalist
Figure 2 for Masked AutoDecoder is Effective Multi-Task Vision Generalist
Figure 3 for Masked AutoDecoder is Effective Multi-Task Vision Generalist
Figure 4 for Masked AutoDecoder is Effective Multi-Task Vision Generalist
Viaarxiv icon

Weakly Supervised Monocular 3D Detection with a Single-View Image

Add code
Feb 29, 2024
Figure 1 for Weakly Supervised Monocular 3D Detection with a Single-View Image
Figure 2 for Weakly Supervised Monocular 3D Detection with a Single-View Image
Figure 3 for Weakly Supervised Monocular 3D Detection with a Single-View Image
Figure 4 for Weakly Supervised Monocular 3D Detection with a Single-View Image
Viaarxiv icon