Picture for Yuri Nakao

Yuri Nakao

Accountability of Generative AI: Exploring a Precautionary Approach for "Artificially Created Nature"

Add code
May 12, 2025
Viaarxiv icon

EARN Fairness: Explaining, Asking, Reviewing and Negotiating Artificial Intelligence Fairness Metrics Among Stakeholders

Add code
Jul 16, 2024
Figure 1 for EARN Fairness: Explaining, Asking, Reviewing and Negotiating Artificial Intelligence Fairness Metrics Among Stakeholders
Figure 2 for EARN Fairness: Explaining, Asking, Reviewing and Negotiating Artificial Intelligence Fairness Metrics Among Stakeholders
Figure 3 for EARN Fairness: Explaining, Asking, Reviewing and Negotiating Artificial Intelligence Fairness Metrics Among Stakeholders
Figure 4 for EARN Fairness: Explaining, Asking, Reviewing and Negotiating Artificial Intelligence Fairness Metrics Among Stakeholders
Viaarxiv icon

Exploring the Impact of Lay User Feedback for Improving AI Fairness

Add code
Dec 18, 2023
Figure 1 for Exploring the Impact of Lay User Feedback for Improving AI Fairness
Figure 2 for Exploring the Impact of Lay User Feedback for Improving AI Fairness
Figure 3 for Exploring the Impact of Lay User Feedback for Improving AI Fairness
Figure 4 for Exploring the Impact of Lay User Feedback for Improving AI Fairness
Viaarxiv icon

Towards Responsible AI: A Design Space Exploration of Human-Centered Artificial Intelligence User Interfaces to Investigate Fairness

Add code
Jun 01, 2022
Figure 1 for Towards Responsible AI: A Design Space Exploration of Human-Centered Artificial Intelligence User Interfaces to Investigate Fairness
Figure 2 for Towards Responsible AI: A Design Space Exploration of Human-Centered Artificial Intelligence User Interfaces to Investigate Fairness
Figure 3 for Towards Responsible AI: A Design Space Exploration of Human-Centered Artificial Intelligence User Interfaces to Investigate Fairness
Figure 4 for Towards Responsible AI: A Design Space Exploration of Human-Centered Artificial Intelligence User Interfaces to Investigate Fairness
Viaarxiv icon

Towards Involving End-users in Interactive Human-in-the-loop AI Fairness

Add code
Apr 22, 2022
Figure 1 for Towards Involving End-users in Interactive Human-in-the-loop AI Fairness
Figure 2 for Towards Involving End-users in Interactive Human-in-the-loop AI Fairness
Figure 3 for Towards Involving End-users in Interactive Human-in-the-loop AI Fairness
Figure 4 for Towards Involving End-users in Interactive Human-in-the-loop AI Fairness
Viaarxiv icon

One-vs.-One Mitigation of Intersectional Bias: A General Method to Extend Fairness-Aware Binary Classification

Add code
Oct 26, 2020
Figure 1 for One-vs.-One Mitigation of Intersectional Bias: A General Method to Extend Fairness-Aware Binary Classification
Figure 2 for One-vs.-One Mitigation of Intersectional Bias: A General Method to Extend Fairness-Aware Binary Classification
Figure 3 for One-vs.-One Mitigation of Intersectional Bias: A General Method to Extend Fairness-Aware Binary Classification
Figure 4 for One-vs.-One Mitigation of Intersectional Bias: A General Method to Extend Fairness-Aware Binary Classification
Viaarxiv icon