Picture for Ajay Divakaran

Ajay Divakaran

Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification

Add code
Jul 02, 2024
Figure 1 for Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification
Figure 2 for Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification
Figure 3 for Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification
Figure 4 for Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification
Viaarxiv icon

Empowering Interdisciplinary Insights with Dynamic Graph Embedding Trajectories

Add code
Jun 25, 2024
Viaarxiv icon

BloomVQA: Assessing Hierarchical Multi-modal Comprehension

Add code
Dec 20, 2023
Viaarxiv icon

A Video is Worth 10,000 Words: Training and Benchmarking with Diverse Captions for Better Long Video Retrieval

Add code
Nov 30, 2023
Figure 1 for A Video is Worth 10,000 Words: Training and Benchmarking with Diverse Captions for Better Long Video Retrieval
Figure 2 for A Video is Worth 10,000 Words: Training and Benchmarking with Diverse Captions for Better Long Video Retrieval
Figure 3 for A Video is Worth 10,000 Words: Training and Benchmarking with Diverse Captions for Better Long Video Retrieval
Figure 4 for A Video is Worth 10,000 Words: Training and Benchmarking with Diverse Captions for Better Long Video Retrieval
Viaarxiv icon

DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback

Add code
Nov 16, 2023
Figure 1 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Figure 2 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Figure 3 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Figure 4 for DRESS: Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Language Feedback
Viaarxiv icon

Demonstrations Are All You Need: Advancing Offensive Content Paraphrasing using In-Context Learning

Add code
Oct 16, 2023
Figure 1 for Demonstrations Are All You Need: Advancing Offensive Content Paraphrasing using In-Context Learning
Figure 2 for Demonstrations Are All You Need: Advancing Offensive Content Paraphrasing using In-Context Learning
Figure 3 for Demonstrations Are All You Need: Advancing Offensive Content Paraphrasing using In-Context Learning
Figure 4 for Demonstrations Are All You Need: Advancing Offensive Content Paraphrasing using In-Context Learning
Viaarxiv icon

Confidence Calibration for Systems with Cascaded Predictive Modules

Add code
Sep 21, 2023
Viaarxiv icon

Measuring and Improving Chain-of-Thought Reasoning in Vision-Language Models

Add code
Sep 08, 2023
Viaarxiv icon

TIJO: Trigger Inversion with Joint Optimization for Defending Multimodal Backdoored Models

Add code
Aug 07, 2023
Viaarxiv icon

Probing Conceptual Understanding of Large Visual-Language Models

Add code
Apr 07, 2023
Figure 1 for Probing Conceptual Understanding of Large Visual-Language Models
Figure 2 for Probing Conceptual Understanding of Large Visual-Language Models
Figure 3 for Probing Conceptual Understanding of Large Visual-Language Models
Figure 4 for Probing Conceptual Understanding of Large Visual-Language Models
Viaarxiv icon