Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Runtime Monitoring and Enforcement of Conditional Fairness in Generative AIs

Created by
  • Haebom

Author

Chih-Hong Cheng, Changshun Wu, Xingyu Zhao, Saddek Bensalem, Harald Ruess

Outline

This paper presents novel characterization and application techniques specific to GenAI to address fairness issues arising when deploying generative AI (GenAI) models. Unlike conventional AI that performs specific tasks, GenAI's broad functionality requires conditional fairness tailored to the context in which it is generated (e.g., demographic fairness in generating images of poor and successful business people). We define two levels of fairness: the first evaluates the fairness of generated outputs independent of prompts and models, and the second evaluates intrinsic fairness using neutral prompts. Given the complexity of GenAI and the difficulty of specifying fairness, we focus on limiting the worst-case scenario by deeming the GenAI system unfair if the distance between the appearances of a specific group exceeds a predefined threshold. We also explore combinatorial testing to assess the relative completeness of cross-sectional fairness. By limiting the worst-case scenario, we develop a prompt injection method that applies conditional fairness with minimal intervention within an agent-based framework, and validate it on a state-of-the-art GenAI system.

Takeaways, Limitations

Takeaways:
A Novel Approach to GenAI's Conditional Fairness Problem
Fairness can be assessed and applied in a way that limits the worst case scenario.
Enhancing efficient fairness through agent-based prompt injection.
Using Combinatorial Testing for Cross-Functional Fairness Assessment
Limitations:
Further research is needed to determine the appropriateness and generalizability of the established thresholds.
The need to verify generalizability across various GenAI models and application areas.
Analysis of the safety and potential exploitability of prompt injection methods is needed.
Focusing on the worst case scenario may not accurately reflect the level of fairness in real-world situations.
👍