Picture for Tao Ji

Tao Ji

Mitigating Object Hallucinations in MLLMs via Multi-Frequency Perturbations

Add code
Mar 19, 2025
Viaarxiv icon

The Role of Visual Modality in Multimodal Mathematical Reasoning: Challenges and Insights

Add code
Mar 06, 2025
Viaarxiv icon

Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs

Add code
Feb 20, 2025
Viaarxiv icon

EvoLlama: Enhancing LLMs' Understanding of Proteins via Multimodal Structure and Sequence Representations

Add code
Dec 16, 2024
Figure 1 for EvoLlama: Enhancing LLMs' Understanding of Proteins via Multimodal Structure and Sequence Representations
Figure 2 for EvoLlama: Enhancing LLMs' Understanding of Proteins via Multimodal Structure and Sequence Representations
Figure 3 for EvoLlama: Enhancing LLMs' Understanding of Proteins via Multimodal Structure and Sequence Representations
Figure 4 for EvoLlama: Enhancing LLMs' Understanding of Proteins via Multimodal Structure and Sequence Representations
Viaarxiv icon

AntLM: Bridging Causal and Masked Language Models

Add code
Dec 04, 2024
Viaarxiv icon

Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision

Add code
Nov 25, 2024
Figure 1 for Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision
Figure 2 for Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision
Figure 3 for Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision
Figure 4 for Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision
Viaarxiv icon

Multi-Programming Language Sandbox for LLMs

Add code
Oct 30, 2024
Figure 1 for Multi-Programming Language Sandbox for LLMs
Figure 2 for Multi-Programming Language Sandbox for LLMs
Figure 3 for Multi-Programming Language Sandbox for LLMs
Figure 4 for Multi-Programming Language Sandbox for LLMs
Viaarxiv icon

Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs

Add code
Oct 15, 2024
Figure 1 for Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs
Figure 2 for Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs
Figure 3 for Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs
Figure 4 for Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs
Viaarxiv icon

Generation with Dynamic Vocabulary

Add code
Oct 11, 2024
Figure 1 for Generation with Dynamic Vocabulary
Figure 2 for Generation with Dynamic Vocabulary
Figure 3 for Generation with Dynamic Vocabulary
Figure 4 for Generation with Dynamic Vocabulary
Viaarxiv icon

Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) Models

Add code
Oct 04, 2024
Figure 1 for Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) Models
Figure 2 for Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) Models
Figure 3 for Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) Models
Figure 4 for Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) Models
Viaarxiv icon