Picture for Golnoosh Farnadi

Golnoosh Farnadi

Beyond the Safety Bundle: Auditing the Helpful and Harmless Dataset

Add code
Nov 12, 2024
Viaarxiv icon

Multilingual Hallucination Gaps in Large Language Models

Add code
Oct 23, 2024
Viaarxiv icon

FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation

Add code
Oct 22, 2024
Figure 1 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Figure 2 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Figure 3 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Figure 4 for FairLoRA: Unpacking Bias Mitigation in Vision Models with Fairness-Driven Low-Rank Adaptation
Viaarxiv icon

Hallucination Detox: Sensitive Neuron Dropout (SeND) for Large Language Model Training

Add code
Oct 20, 2024
Viaarxiv icon

On the Implicit Relation Between Low-Rank Adaptation and Differential Privacy

Add code
Sep 26, 2024
Viaarxiv icon

Understanding the Local Geometry of Generative Model Manifolds

Add code
Aug 15, 2024
Viaarxiv icon

Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild

Add code
Jul 16, 2024
Figure 1 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Figure 2 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Figure 3 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Figure 4 for Trust No Bot: Discovering Personal Disclosures in Human-LLM Conversations in the Wild
Viaarxiv icon

Towards More Realistic Extraction Attacks: An Adversarial Perspective

Add code
Jul 02, 2024
Viaarxiv icon

Position: Cracking the Code of Cascading Disparity Towards Marginalized Communities

Add code
Jun 03, 2024
Viaarxiv icon

Mitigating Disparate Impact of Differential Privacy in Federated Learning through Robust Clustering

Add code
May 29, 2024
Viaarxiv icon