Picture for Eitan Farchi

Eitan Farchi

Statistical multi-metric evaluation and visualization of LLM system predictive performance

Add code
Jan 30, 2025
Viaarxiv icon

Exploring Straightforward Conversational Red-Teaming

Add code
Sep 07, 2024
Figure 1 for Exploring Straightforward Conversational Red-Teaming
Figure 2 for Exploring Straightforward Conversational Red-Teaming
Figure 3 for Exploring Straightforward Conversational Red-Teaming
Figure 4 for Exploring Straightforward Conversational Red-Teaming
Viaarxiv icon

Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity

Add code
Aug 22, 2024
Figure 1 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Figure 2 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Figure 3 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Figure 4 for Can You Trust Your Metric? Automatic Concatenation-Based Tests for Metric Validity
Viaarxiv icon

A Novel Metric for Measuring the Robustness of Large Language Models in Non-adversarial Scenarios

Add code
Aug 04, 2024
Viaarxiv icon

Generating Unseen Code Tests In Infinitum

Add code
Jul 29, 2024
Figure 1 for Generating Unseen Code Tests In Infinitum
Figure 2 for Generating Unseen Code Tests In Infinitum
Figure 3 for Generating Unseen Code Tests In Infinitum
Figure 4 for Generating Unseen Code Tests In Infinitum
Viaarxiv icon

Using Combinatorial Optimization to Design a High quality LLM Solution

Add code
May 15, 2024
Viaarxiv icon

Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations

Add code
Mar 09, 2024
Figure 1 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 2 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 3 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Figure 4 for Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations
Viaarxiv icon

Alignment Studio: Aligning Large Language Models to Particular Contextual Regulations

Add code
Mar 08, 2024
Viaarxiv icon

Unveiling Safety Vulnerabilities of Large Language Models

Add code
Nov 07, 2023
Figure 1 for Unveiling Safety Vulnerabilities of Large Language Models
Figure 2 for Unveiling Safety Vulnerabilities of Large Language Models
Figure 3 for Unveiling Safety Vulnerabilities of Large Language Models
Figure 4 for Unveiling Safety Vulnerabilities of Large Language Models
Viaarxiv icon

Predicting Question-Answering Performance of Large Language Models through Semantic Consistency

Add code
Nov 02, 2023
Viaarxiv icon