paint-brush
What We’ve Learned About BiasPainter’s Accuracy and Limitationsby@mediabias

What We’ve Learned About BiasPainter’s Accuracy and Limitations

tldt arrow

Too Long; Didn't Read

Potential threats to BiasPainter’s validity include imperfections in AI techniques, limited input data diversity, and a narrow range of evaluated models. Mitigations include using advanced APIs, diverse and comprehensive input data, and high accuracy human annotations. Future work aims to broaden testing across more models and systems.
featured image - What We’ve Learned About BiasPainter’s Accuracy and Limitations
Tech Media Bias [Research Publication] HackerNoon profile picture

Authors:

(1) Wenxuan Wang, The Chinese University of Hong Kong, Hong Kong, China;

(2) Haonan Bai, The Chinese University of Hong Kong, Hong Kong, China

(3) Jen-tse Huang, The Chinese University of Hong Kong, Hong Kong, China;

(4) Yuxuan Wan, The Chinese University of Hong Kong, Hong Kong, China;

(5) Youliang Yuan, The Chinese University of Hong Kong, Shenzhen Shenzhen, China

(6) Haoyi Qiu University of California, Los Angeles, Los Angeles, USA;

(7) Nanyun Peng, University of California, Los Angeles, Los Angeles, USA

(8) Michael Lyu, The Chinese University of Hong Kong, Hong Kong, China.

Abstract

1 Introduction

2 Background

3 Approach and Implementation

3.1 Seed Image Collection and 3.2 Neutral Prompt List Collection

3.3 Image Generation and 3.4 Properties Assessment

3.5 Bias Evaluation

4 Evaluation

4.1 Experimental Setup

4.2 RQ1: Effectiveness of BiasPainter

4.3 RQ2 - Validity of Identified Biases

4.4 RQ3 - Bias Mitigation

5 Threats to Validity

6 Related Work

7 Conclusion, Data Availability, and References

5 THREATS TO VALIDITY

The validity of this work may be subject to some threats. The first threat lies in the AI techniques adopted by BiasPainter for bias identification. Due to the imperfect nature of AI techniques, the biases identified by BiasPainter may be false positives, or BiasPainter may miss some biased generation, leading to false negatives. To relieve this threat, BiasPainter calls commercial-level APIs and deploys complicated pipelines to analyze the race, gender, and age properties, aiming to ensure the soundness. In addition, we also


Table 5: Bias Mitigation Results on Profession


conducted human annotation to show that BiasPainter can achieve high accuracy (i.e., 90.8%) in detecting bias.


The second threat is that the input data of BiasPainter are predefined, both seed images and prompt lists, which may hurt the comprehensiveness of the testing results. To mitigate this threat, we collected diverse and comprehensive seed images and prompt words, all of which are from the real world on the Internet and manually annotated by researchers. Also, we want to highlight that what we provide is a workflow: select seed images, design prompt lists, generate test cases, and find social bias. If users are willing to evaluate other properties on other images, they could follow this workflow to add more images and prompts.


The third threat lies in the image generation models under test in the evaluation. We do not evaluate the performance of BiasPainter on other systems. To mitigate this threat, we chose to test the most widely used commercial image generation software and SOTA academic models provided by famous organizations. In the future, we could test more commercial software and research models to further mitigate this threat.


This paper is available on arxiv under CC0 1.0 DEED license.