Picture for Chirag Agarwal

Chirag Agarwal

Analyzing Memorization in Large Language Models through the Lens of Model Attribution

Add code
Jan 09, 2025
Figure 1 for Analyzing Memorization in Large Language Models through the Lens of Model Attribution
Figure 2 for Analyzing Memorization in Large Language Models through the Lens of Model Attribution
Figure 3 for Analyzing Memorization in Large Language Models through the Lens of Model Attribution
Figure 4 for Analyzing Memorization in Large Language Models through the Lens of Model Attribution
Viaarxiv icon

HALLUCINOGEN: A Benchmark for Evaluating Object Hallucination in Large Visual-Language Models

Add code
Dec 29, 2024
Viaarxiv icon

On the Impact of Fine-Tuning on Chain-of-Thought Reasoning

Add code
Nov 22, 2024
Viaarxiv icon

Towards Operationalizing Right to Data Protection

Add code
Nov 16, 2024
Viaarxiv icon

On the Hardness of Faithful Chain-of-Thought Reasoning in Large Language Models

Add code
Jun 15, 2024
Viaarxiv icon

Towards Safe and Aligned Large Language Models for Medicine

Add code
Mar 06, 2024
Viaarxiv icon

Understanding the Effects of Iterative Prompting on Truthfulness

Add code
Feb 09, 2024
Viaarxiv icon

Faithfulness vs. Plausibility: On the (Un)Reliability of Explanations from Large Language Models

Add code
Feb 08, 2024
Viaarxiv icon

Quantifying Uncertainty in Natural Language Explanations of Large Language Models

Add code
Nov 06, 2023
Viaarxiv icon

Are Large Language Models Post Hoc Explainers?

Add code
Oct 10, 2023
Viaarxiv icon