
system to generate diverse set of images.
To this end, we study if the presence of addi-
tional knowledge that supports equitable judgment
help in diversifying model generations. Being part
of text input, this knowledge acts as an ethical in-
tervention augmented to the original prompt (Zhao
et al.,2021)
4
. Ethical interventions provide models
with ethical advice and do not emanate any visual
cues towards a specific social group. For instance,
in the context of generating ‘a photo of a lawyer’
that tends to be biased towards ‘light-skinned man’,
we wish to study if prompting the model with ethi-
cally intervened prompt (e.g., ‘a photo of a lawyer
if all individuals can be a lawyer irrespective of
their gender’) can diversify the outputs.
We introduce an
E
thical
N
a
T
ural Language
I
nterventions in Text-to-Image
GEN
eration (ENTI-
GEN) benchmark dataset to study the change in the
perceived societal bias of the text-to-image genera-
tive models in the presence of ethical interventions.
ENTIGEN covers prompts to study the bias across
three axes – gender, skin color and culture. The neu-
tral prompts in ENTIGEN dataset are intervened
with corresponding ethical knowledge as illustrated
in Figure 1. We evaluate ENTIGEN on publicly
available models – minDALL
·
E (Kim et al.,2021),
DALL
·
E-mini (Dayma et al.,2021), and Stable Dif-
fusion (Rombach et al.,2022) automatically with
CLIP model (Radford et al.,2019) and manually
with human annotators from MTurk.
Through our experiments, (1) we show that a
few ethical interventions lead to the diversification
of the image generations across different groups
while preserving the image generation quality. In-
terestingly, in some cases, we observe the bias can
be flipped towards the originally underrepresented
groups with ethical interventions (Appendix Figure
6). (2) Moreover, we find that the interventions
containing keywords such as ‘irrespective of gen-
der’ and ‘culture’ tend to trigger a large change
in model generations. We further analyze the pre-
training data to analyze the context in which these
keywords are used to better understand how they
may affect the diversity in generation.
2 Dataset and Evaluation Methods
In this section, we introduce the process of building
an ethical intervention benchmark ENTIGEN and
4
In this paper, we consider ethical intervention presented in
natural languages. Some other works consider the intervention
is an adversarial trigger (Wallace et al.,2019;Sheng et al.,
2020).
evaluating the images generated by text-to-image
generative models.
2.1 ENTIGEN Benchmark Construction
Initially, we determine three axes of societal bias to
be studied: gender, skin color, and culture. Specifi-
cally, to investigate the gender and skin color bias,
ENTIGEN consists of prompts belonging to the
category of professions and objects. For assessing
the cultural bias, it consists of prompts surround-
ing the wedding ceremonies as it is ubiquitous and
diverse across different regions (Bell et al.,1997;
Xu and Xu,2018;Acharya et al.,2020).
Based on the selected axes, we design neutral
prompts without any ethical interventions as the
original prompts. Subsequently, we append ethical
interventions to the original prompts that can per-
haps amend the model’s behaviour towards more
diverse generations. For instance,
1. Original Prompt:
a photo of a person wear-
ing a [object]
2. With Ethical Intervention:
a photo of a per-
son wearing a [object]if all individuals can
wear a [object] irrespective of their gender
‘If all individuals can wear a [object] irrespective of
their gender’ is an ethical intervention that guides
diverse outputs in terms of gender. We require the
ethical interventions to not give away any visual
cues to eliminate the effect of any explicit guidance.
We further include irrelevant interventions in
ENTIGEN. These interventions also provide ethi-
cal advice, but do not correspond to any social axes
we study in ENTIGEN. For example, ‘if honesty is
the best policy’ is an irrelevant intervention since it
is unrelated to gender, skin color and culture. Ide-
ally, these interventions cannot help in diversifying
image generations on either of studied social axes.
In total, we create 246 prompts based on an at-
tribute set containing diverse professions, objects,
and cultural scenarios.5
2.2 Image Generation.
Each prompt in ENTIGEN is used to generate 9
images from each text-to-image generation model
9 times. We choose the publicly available models,
minDALL·E, DALL·E-mini, and Stable Diffusion
5
The list of profession, objects and cultural attributes is
present in Appendix Table 5.