Picture for Orna Raz

Orna Raz

Uncovering Code Insights: Leveraging GitHub Artifacts for Deeper Code Understanding

Add code
Nov 05, 2025
Viaarxiv icon

Vintage Code, Modern Judges: Meta-Validation in Low Data Regimes

Add code
Oct 31, 2025
Viaarxiv icon

Statistical multi-metric evaluation and visualization of LLM system predictive performance

Add code
Jan 30, 2025
Figure 1 for Statistical multi-metric evaluation and visualization of LLM system predictive performance
Figure 2 for Statistical multi-metric evaluation and visualization of LLM system predictive performance
Figure 3 for Statistical multi-metric evaluation and visualization of LLM system predictive performance
Figure 4 for Statistical multi-metric evaluation and visualization of LLM system predictive performance
Viaarxiv icon

Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity

Add code
Aug 22, 2024
Figure 1 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Figure 2 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Figure 3 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Figure 4 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Viaarxiv icon

Generating Unseen Code Tests In Infinitum

Add code
Jul 29, 2024
Figure 1 for Generating Unseen Code Tests In Infinitum
Figure 2 for Generating Unseen Code Tests In Infinitum
Figure 3 for Generating Unseen Code Tests In Infinitum
Figure 4 for Generating Unseen Code Tests In Infinitum
Viaarxiv icon

Using Combinatorial Optimization to Design a High quality LLM Solution

Add code
May 15, 2024
Viaarxiv icon

Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations

Add code
Mar 09, 2024
Figure 1 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 2 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 3 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 4 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Viaarxiv icon

Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations

Add code
Mar 08, 2024
Viaarxiv icon

Unveiling Safety Vulnerabilities of Large Language Models

Add code
Nov 07, 2023
Figure 1 for Unveiling Safety Vulnerabilities of Large Language Models
Figure 2 for Unveiling Safety Vulnerabilities of Large Language Models
Figure 3 for Unveiling Safety Vulnerabilities of Large Language Models
Figure 4 for Unveiling Safety Vulnerabilities of Large Language Models
Viaarxiv icon

Predicting Question-Answering Performance of Large Language Models through Semantic Consistency

Add code
Nov 02, 2023
Figure 1 for Predicting Question-Answering Performance of Large Language Models through Semantic Consistency
Figure 2 for Predicting Question-Answering Performance of Large Language Models through Semantic Consistency
Figure 3 for Predicting Question-Answering Performance of Large Language Models through Semantic Consistency
Figure 4 for Predicting Question-Answering Performance of Large Language Models through Semantic Consistency
Viaarxiv icon