Picture for Golnoosh Farnadi

Golnoosh Farnadi

Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML

Add code
Nov 19, 2024
Figure 1 for Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML
Figure 2 for Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML
Figure 3 for Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML
Figure 4 for Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML
Viaarxiv icon

Beyond the Safety Bundle: Auditing the Helpful and Harmless Dataset

Add code
Nov 12, 2024
Viaarxiv icon

Multilingual Hallucination Gaps in Large Language Models

Add code
Oct 23, 2024
Viaarxiv icon

FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation

Add code
Oct 22, 2024
Figure 1 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Figure 2 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Figure 3 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Figure 4 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Viaarxiv icon

Hallucination Detox: Sensitive Neuron Dropout (SeND) for Large Language Model Training

Add code
Oct 20, 2024
Viaarxiv icon

On the Implicit Relation Between Low-Rank Adaptation and Differential Privacy

Add code
Sep 26, 2024
Viaarxiv icon

Understanding the Local Geometry of Generative Model Manifolds

Add code
Aug 15, 2024
Viaarxiv icon

Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild

Add code
Jul 16, 2024
Figure 1 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Figure 2 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Figure 3 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Figure 4 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Viaarxiv icon

Towards More Realistic Extraction Attacks: An Adversarial Perspective

Add code
Jul 02, 2024
Viaarxiv icon

Position: Cracking the Code of Cascading Disparity Towards Marginalized Communities

Add code
Jun 03, 2024
Viaarxiv icon