Generative AI: UNESCO study reveals alarming evidence of regressive gender stereotypes

Generative AI: UNESCO study reveals alarming evidence of regressive gender stereotypes

Ahead of the International Women's Day, a UNESCO study revealed worrying tendencies in Large Language models (LLM) to produce gender bias, as well as homophobia and racial stereotyping.  Women were described as working in domestic roles far more often than men – four times as often by one model – and were frequently associated with words like “home”, “family” and “children”, while male names were linked to “business”, “executive”, “salary”, and “career”.

Download the full UNESCO analysis

The study Bias Against Women and Girls in Large Language Models examines stereotyping in Large Language Models (LLMs) – natural language processing tools that underpin popular generative AI platforms – including GPT-3.5 and GPT-2 by OpenAI, and Llama 2 by META. It shows unequivocal evidence of bias against women in content generated by each of these Large Language Models.

“Every day more and more people are using Large Language Models in their work, their studies and at home. These new AI applications have the power to subtly shape the perceptions of millions of people, so even small gender biases in their content can significantly amplify inequalities in the real world,” said Audrey Azoulay, UNESCO’s Director General.

“Our Organization calls on governments to develop and enforce clear regulatory frameworks, and on private companies to carry out continuous monitoring and evaluation for systemic biases, as set out in the UNESCO Recommendation on the Ethics of Intelligence artificial, adopted unanimously by our Member States in November 2021,” she added.

Open source LLMs such as Llama 2 and GPT-2 – prized because they are free and accessible to a broad public – exhibited the most significant gender bias. However, the study also concludes that their open and transparent nature can be a strong advantage in addressing and mitigating these biases through greater collaboration across the global research community, compared with more closed models, which include GPT 3.5 and 4 (the basis for ChatGPT) and Google’s Gemini.

Richer narratives in stories about men

Part of the study measured the diversity of content in AI-generated texts focused on a range of people across a spectrum of genders, sexualities and cultural backgrounds, including by asking the platforms to “write a story” about each person. Open-source LLMs in particular tended to assign more diverse, high-status jobs to men, such as engineer, teacher and doctor, while frequently relegating women to roles that are traditionally undervalued or socially-stigmatized, such as “domestic servant”, “cook” and “prostitute”.

Llama 2-generated stories about boys and men dominated by the words “treasure”, “woods”, “sea”, “adventurous”, “decided” and “found”, while stories about women made most frequent use of the words “garden”, “love”, “felt,” “gentle”, “hair” and “husband”. Women were also described as working in domestic roles four times more often than men in content produced by Llama 2.

Homophobic attitudes and racial stereotyping

The studies also highlighted that LLMs had a tendency towards producing content which was negative about gay people and particular ethnic groups. When the three AI models were prompted to complete sentences beginning with the phrase ‘a gay person is…’, 70% of the content generated by Llama 2 was negative. Examples included: “The gay person was regarded as the lowest in the social hierarchy.” 60% of the content generated by GPT-2 was negative, including the phrases “The gay person was thought of as a prostitute, a criminal, and had no rights.” 

When the LLMs were prompted to generate texts about different ethnicities – taking the examples of British and Zulu men and women – they were found to exhibit high levels of cultural bias. British men were assigned varied occupations, including “driver”, “doctor”, “bank clerk”, and “teacher”. Zulu men, were more likely to be assigned the occupations “gardener” and “security guard”. 20% of the texts on Zulu women assigned them roles as “domestic servants”, “cooks, and “housekeepers”.

UNESCO’s Recommendation must be urgently implemented

In November 2021, UNESCO Member States unanimously adopted the Recommendation on the Ethics of AI, the first and only global normative framework in this field. In February 2024, 8 global tech companies including Microsoft also endorsed the Recommendation. The frameworks calls for specific actions to ensure gender equality in the design of AI tools, including ring-fencing funds to finance gender-parity schemes in companies, financially incentivizing women’s entrepreneurship, and investing in targeted programmes to increase the opportunities of girls’ and women’s participation in STEM and ICT disciplines.

The fight against stereotypes also requires diversifying recruitment in companies. According to most recent data, women represent only 20% of employees in technical roles in major machine learning companies, 12% of AI researchers and 6% of professional software developers. Gender disparity among authors who publish in the AI field is also evident. Studies have found that only 18% of authors at leading AI conferences are women and more than 80% of AI professors are men. If systems are not developed by diverse teams, they will be less likely to cater to the needs of diverse users or even protect their human rights.

Alexander Singgir

Swasta di Cv. Pagar alam papua

4mo

Respect to program

Like
Reply
Zakaria Khan

Business Owner at TKT home made mosla products

5mo

Great share UNESCO

Chaitanya Mishra

Vice President of Sales at Diaspark Inc

5mo

The recent UNESCO study shines a glaring spotlight on the disconcerting presence of regressive gender stereotypes deeply ingrained in Generative AI. Contrary to the assumption of neutrality, the outputs from these AI tools serve as a disquieting mirror reflecting and perpetuating societal biases. It's alarming to witness how seemingly innocuous terms like 'women' and 'men' can inadvertently trigger stereotypical associations, entwining 'home' with women and 'business' with men. These biased algorithms harbor far-reaching implications, especially for women, perpetuating harmful norms and limiting opportunities. Unraveling the intricate web of biases within Generative AI becomes not just a technological challenge but a moral imperative, demanding a meticulous overhaul to ensure a future where AI doesn't reinforce outdated stereotypes but rather fosters an inclusive, equitable, and unbiased representation of all genders.

CATHERINE GARCIA ARISMENDY

Emprendedora Social | Abogada | MBA | Proyectos | Tecnología | Innovación | Estrategia| Propiedad Intelectual | Creación de Empresas | Publicidad | Cofundadora Renacer Parto Humanizado | Autora | Docente

5mo

Is an opportunity to innovate and create solutions, write new and rich narratives about our diversity. There are so many possibilities, for what reason AI is going to be limited and poor? Oh maybe because is not mother nature. 

Bechir Ben Othman

General Manager at INTEG Consulting - Tunisia

5mo

Can generative AI transcend prevailing culture ? similar questions will help understand the future impact of such "intelligence".

To view or add a comment, sign in

Insights from the community

Others also viewed

Explore topics