Picture for Chitta Baral

Chitta Baral

Shammie

TripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives

Add code
Nov 04, 2024
Figure 1 for TripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives
Figure 2 for TripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives
Figure 3 for TripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives
Figure 4 for TripletCLIP: Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives
Viaarxiv icon

ToW: Thoughts of Words Improve Reasoning in Large Language Models

Add code
Oct 21, 2024
Figure 1 for ToW: Thoughts of Words Improve Reasoning in Large Language Models
Figure 2 for ToW: Thoughts of Words Improve Reasoning in Large Language Models
Figure 3 for ToW: Thoughts of Words Improve Reasoning in Large Language Models
Figure 4 for ToW: Thoughts of Words Improve Reasoning in Large Language Models
Viaarxiv icon

VL-GLUE: A Suite of Fundamental yet Challenging Visuo-Linguistic Reasoning Tasks

Add code
Oct 17, 2024
Figure 1 for VL-GLUE: A Suite of Fundamental yet Challenging Visuo-Linguistic Reasoning Tasks
Figure 2 for VL-GLUE: A Suite of Fundamental yet Challenging Visuo-Linguistic Reasoning Tasks
Figure 3 for VL-GLUE: A Suite of Fundamental yet Challenging Visuo-Linguistic Reasoning Tasks
Figure 4 for VL-GLUE: A Suite of Fundamental yet Challenging Visuo-Linguistic Reasoning Tasks
Viaarxiv icon

ActionCOMET: A Zero-shot Approach to Learn Image-specific Commonsense Concepts about Actions

Add code
Oct 17, 2024
Figure 1 for ActionCOMET: A Zero-shot Approach to Learn Image-specific Commonsense Concepts about Actions
Figure 2 for ActionCOMET: A Zero-shot Approach to Learn Image-specific Commonsense Concepts about Actions
Figure 3 for ActionCOMET: A Zero-shot Approach to Learn Image-specific Commonsense Concepts about Actions
Figure 4 for ActionCOMET: A Zero-shot Approach to Learn Image-specific Commonsense Concepts about Actions
Viaarxiv icon

Help Me Identify: Is an LLM+VQA System All We Need to Identify Visual Concepts?

Add code
Oct 17, 2024
Figure 1 for Help Me Identify: Is an LLM+VQA System All We Need to Identify Visual Concepts?
Figure 2 for Help Me Identify: Is an LLM+VQA System All We Need to Identify Visual Concepts?
Figure 3 for Help Me Identify: Is an LLM+VQA System All We Need to Identify Visual Concepts?
Figure 4 for Help Me Identify: Is an LLM+VQA System All We Need to Identify Visual Concepts?
Viaarxiv icon

REVISION: Rendering Tools Enable Spatial Fidelity in Vision-Language Models

Add code
Aug 05, 2024
Viaarxiv icon

Step-by-Step Reasoning to Solve Grid Puzzles: Where do LLMs Falter?

Add code
Jul 20, 2024
Viaarxiv icon

UnSeenTimeQA: Time-Sensitive Question-Answering Beyond LLMs' Memorization

Add code
Jul 03, 2024
Viaarxiv icon

Multi-LogiEval: Towards Evaluating Multi-Step Logical Reasoning Ability of Large Language Models

Add code
Jun 24, 2024
Figure 1 for Multi-LogiEval: Towards Evaluating Multi-Step Logical Reasoning Ability of Large Language Models
Figure 2 for Multi-LogiEval: Towards Evaluating Multi-Step Logical Reasoning Ability of Large Language Models
Figure 3 for Multi-LogiEval: Towards Evaluating Multi-Step Logical Reasoning Ability of Large Language Models
Figure 4 for Multi-LogiEval: Towards Evaluating Multi-Step Logical Reasoning Ability of Large Language Models
Viaarxiv icon

Investigating and Addressing Hallucinations of LLMs in Tasks Involving Negation

Add code
Jun 08, 2024
Viaarxiv icon