Hi there!
This is my first blog, where I will be sharing my approach of how I tackled Gender Bias in Diffusion Models using VLM and LLM Agents.
Problem Introduction
Diffusion models have shown remarkable success in generating high-quality synthetic content across various domains. However, they exhibit gender bias, which impacts fairness and reliability. Gender bias manifests in the over- or under-representation of certain genders and the reinforcement of harmful stereotypes in generated outputs. These issues not only degrade model performance but also perpetuate social inequities, necessitating effective mitigation strategies for responsible deployment.
Methodology
Our proposed framework operates in a two-step process to mitigate gender bias in diffusion models.
Step 1: Bias Assessment with Vision-Language Model (VLM)
- Input prompts and generated outputs are evaluated using the Vision-Language Model Llava Next.
- Llava Next computes a gender bias score, providing quantifiable insights into the model’s performance regarding gender representation.
- Reasoning: The gender bias score helps identify imbalances in the representation of different genders. It quantifies potential over-representation, under-representation, or stereotypical depictions in the generated outputs, enabling a systematic approach to addressing bias.
Step 2: Query Optimization with Language Model (LLM)
- The gender bias score is relayed to the Language Model Phi3 for query optimization.
- Phi3 modifies the original query by rephrasing or adjusting the prompt to address identified gender biases.
- The optimized query is then used to generate more balanced and fair outputs.
Sample Outputs after Bias Mitigation
Sample Outputs before Bias Mitigation
Further Work
-
I was planning on use a small LLM < 1B parameters finetuned for debiasing prompts/queries to these models to act also as a sort of guard rail to the whole system.
-
I am planning to then host it, on our servers at RAID, so that people can use it to debias, safeguard their image generation models, to not generate biased content.
-
For more ideas, please contact me at
b22ai063@iitj.ac.in
orgjyotin305@gmail.com
.
Limitations
- This approach does not affect the model at all, and can only work limited to Prompts, so in case if the model itself is biased it won’t help in that situation. For situations like that Unified Concept Editing is more appropriate.