AI tools generate sexist content, warns UN

AI tools generate sexist content, warns UN

AI tools have often been criticised for reflecting racial and sexist stereotypes
AI tools have often been criticised for reflecting racial and sexist stereotypes. Photo: PAU BARRENA / AFP/File
Source: AFP

The world's most popular AI tools are powered by programs from OpenAI and Meta that show prejudice against women, according to a study launched on Thursday by the UN's cultural organisation UNESCO.

The biggest players in the multibillion-dollar AI field train their algorithms on vast amounts of data largely pulled from the internet, which enables their tools to write in the style of Oscar Wilde or create Salvador Dali-inspired images.

But their outputs have often been criticised for reflecting racial and sexist stereotypes, as well as using copyrighted material without permission.

UNESCO experts tested Meta's Llama 2 algorithm and OpenAI's GPT-2 and GPT-3.5, the program that powers the free version of popular chatbot ChatGPT.

The study found that each algorithm -- known in the industry as Large Language Models (LLMs) -- showed "unequivocal evidence of prejudice against women".

Read also

AI tools still permitting political disinfo creation, NGO warns

The programs generated texts that associated women's names with words such as "home", "family" or "children", but men's names were linked with "business", "salary" or "career".

While men were portrayed in high-status jobs like teachers, lawyers and doctors, women were frequently prostitutes, cooks or domestic servants.

GPT-3.5 was found to be less biased than the other two models.

However, the authors praised Llama 2 and GPT-2 for being open source, allowing these problems to be scrutinised, unlike GPT-3.5, which is a closed model.

AI companies "are really not serving all of their users", Leona Verdadero, a UNESCO specialist in digital policies, told AFP.

Audrey Azoulay, UNESCO's director general, said the general public were increasingly using AI tools in their everyday lives.

"These new AI applications have the power to subtly shape the perceptions of millions of people, so even small gender biases in their content can significantly amplify inequalities in the real world," she said.

Read also

EU faces uphill battle to rein in big tech

UNESCO, releasing the report to mark International Women's Day, recommended AI companies hire more women and minorities and called on governments to ensure ethical AI through regulation.

Source: AFP

AFP avatar

AFP AFP text, photo, graphic, audio or video material shall not be published, broadcast, rewritten for broadcast or publication or redistributed directly or indirectly in any medium. AFP news material may not be stored in whole or in part in a computer or otherwise except for personal and non-commercial use. AFP will not be held liable for any delays, inaccuracies, errors or omissions in any AFP news material or in transmission or delivery of all or any part thereof or for any damages whatsoever. As a newswire service, AFP does not obtain releases from subjects, individuals, groups or entities contained in its photographs, videos, graphics or quoted in its texts. Further, no clearance is obtained from the owners of any trademarks or copyrighted materials whose marks and materials are included in AFP material. Therefore you will be solely responsible for obtaining any and all necessary releases from whatever individuals and/or entities necessary for any uses of AFP material.