Picture for Roee Aharoni

Roee Aharoni

Keep Guessing? When Considering Inference Scaling, Mind the Baselines

Add code
Oct 20, 2024
Figure 1 for Keep Guessing? When Considering Inference Scaling, Mind the Baselines
Figure 2 for Keep Guessing? When Considering Inference Scaling, Mind the Baselines
Figure 3 for Keep Guessing? When Considering Inference Scaling, Mind the Baselines
Figure 4 for Keep Guessing? When Considering Inference Scaling, Mind the Baselines
Viaarxiv icon

Localizing Factual Inconsistencies in Attributable Text Generation

Add code
Oct 09, 2024
Figure 1 for Localizing Factual Inconsistencies in Attributable Text Generation
Figure 2 for Localizing Factual Inconsistencies in Attributable Text Generation
Figure 3 for Localizing Factual Inconsistencies in Attributable Text Generation
Figure 4 for Localizing Factual Inconsistencies in Attributable Text Generation
Viaarxiv icon

Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs

Add code
Aug 20, 2024
Figure 1 for Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs
Figure 2 for Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs
Figure 3 for Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs
Figure 4 for Beneath the Surface of Consistency: Exploring Cross-lingual Knowledge Representation Sharing in LLMs
Viaarxiv icon

Can Few-shot Work in Long-Context? Recycling the Context to Generate Demonstrations

Add code
Jun 19, 2024
Viaarxiv icon

Can Large Language Models Faithfully Express Their Intrinsic Uncertainty in Words?

Add code
May 27, 2024
Viaarxiv icon

Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?

Add code
May 09, 2024
Figure 1 for Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Figure 2 for Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Figure 3 for Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Figure 4 for Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Viaarxiv icon

MiMiC: Minimally Modified Counterfactuals in the Representation Space

Add code
Feb 16, 2024
Viaarxiv icon

A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains

Add code
Feb 02, 2024
Figure 1 for A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains
Figure 2 for A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains
Figure 3 for A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains
Figure 4 for A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains
Viaarxiv icon

Narrowing the Knowledge Evaluation Gap: Open-Domain Question Answering with Multi-Granularity Answers

Add code
Jan 09, 2024
Viaarxiv icon

Multilingual Instruction Tuning With Just a Pinch of Multilinguality

Add code
Jan 08, 2024
Viaarxiv icon