XCLUB-COOL STUFF AROUND YOU

 找回密码
 Register
查看: 75|回复: 3
打印 上一主题 下一主题

AI voice assistants reinforce harmful gender stereotypes, new UN report says

[复制链接]

3712

主题

3万

帖子

4万

积分

Security Supervisors

Rank: 8Rank: 8

Happy children's dayHot PartyDiamondKiss Day MedalTeddy Day MedalACTIVE STAR4.0xclubXCLUB TOP FAN photographers2019post star1post star2post star3post star4sign star1sign star2sign star3sign star4post star5X'Club badge exclusive for Indiasign star5

跳转到指定楼层
#1
发表于 2019-05-22 02:30:37 来自手机 | 只看该作者 |只看大图 回帖奖励 |倒序浏览 |阅读模式
[br] [br][br]Artificial intelligence-powered voice assistants, many of which default to female-sounding voices, are reinforcing harmful gender stereotypes, according to a new study published by the United Nations.[br][br]Titled “I’d blush if I could,” after a response Siri utters when receiving certain sexually explicit commands, the paper explores the effects of bias in AI research and product development and the potential long-term negative implications of conditioning society, particularly children, to treat these digital voice assistants as unquestioning helpers who exist only to serve owners unconditionally. It was authored by the United Nations Educational, Scientific, and Cultural Organization, otherwise known as UNESCO.[br][br]The paper argues that by naming voice assistants with traditionally female names, like Alexa and Siri, and rendering the voices as female-sounding by default, tech companies have already preconditioned users to fall back upon antiquated and harmful perceptions of women. Going further, the paper argues that tech companies have failed to build in proper safeguards against hostile, abusive, and gendered language. Instead, most assistants, as Siri does, tend to deflect aggression or chime in with a sly joke. For instance, ask Siri to make you a sandwich, and the voice assistant will respond with, “I can’t. I don’t have any condiments.”[br][br]Tech companies are perpetuating harmful stereotypes about women through AI[br][br]“Companies like Apple and Amazon, staffed by overwhelmingly male engineering teams, have built AI systems that cause their feminized digital assistants to greet verbal abuse with catch-me-if-you-can flirtation,” the report states. “Because the speech of most voice assistants is female, it sends a signal that women are ... docile and eager-to-please helpers, available at the touch of a button or with a blunt voice command like ‘hey’ or ‘OK’. The assistant holds no power of agency beyond what the commander asks of it. It honours commands and responds to queries regardless of their tone or hostility.”[br][br]Much has been written about the pitfalls of tech companies having built their entire consumer-facing AI platforms in the image of traditional, Hollywood-influenced ideas of subservient intelligences. In the future, it’s likely voice assistants will be the primary mode of interaction with hardware and software with the rise of so-called ambient computing, when all manner of internet-connected gadgets exist all around us at all times. (Think Spike Jonze’s Her, which seems like the most accurate depiction of the near-future in film you can find today.) How we interact with the increasingly sophisticated intelligences powering these platforms could have profound cultural and sociological effects on how we interact with other human beings, with service workers, and with humanoid robots that take on more substantial roles in daily life and the labor force.[br][br]However, as Business Insider reported last September, Amazon chose a female-sounding voice because market research indicated it would be received as more “sympathetic” and therefore more helpful. Microsoft, on the other hand, named its assistant Cortana to bank on the existing recognition of the very much female-identifying AI character in its Halo video game franchise; you can’t change Cortana’s voice to a male one, and the company hasn’t said when it plans to let users do so. Siri, for what it’s worth, is a Scandinavian name traditionally for females that means “beautiful victory” in Old Norse. In other words, these decisions about gender with regard to AI assistants were made on purpose, and after what sounds like extensive feedback.[br][br]The AI research field is predominantly white and male[br][br]Tech companies have made an effort to move away from these early design decisions steeped in stereotypes. Google now refers to its various Assistant voice options, which now include different accents with male and female options for each, represented by colors. You can no longer select a “male” or “female” version; each color is randomly assigned to one of eight voice options for each user. The company also rolled out an initiative called Pretty Please that rewards young children when they use phrases like “please” and “thank you” while interacting with Google Assistant. Amazon released something similar last year to encourage polite behavior when talking to Alexa.[br][br]Yet as the report says, these features and gender voice options don’t go far enough; the problem may be baked into the AI and tech industries themselves. The field of AI research is predominantly white and male, a new report from last month found. Eighty percent of AI academics are men, and just 15 percent of AI researchers at Facebook and just 10 percent at Google are women.[br][br]UNESCO says solutions to this issue would be to create as close to gender-neutral assistant voices as possible and to create systems to discourage gender-based insults. Additionally, the report says tech companies should stray away from conditioning users to treat AI as they would a lesser, subservient human being, and that the only way to avoid perpetuating harmful stereotypes like these is to remake voice assistants as purposefully non-human entities.[br][br][br][br]source : THE VERGE
THE STAR OF XFANS INDIA {:xb_4:}{:xb_4:}{:16_24:}dr.gautam{:16_24:}{:xb_4:}{:xb_4:}
回复

使用道具 举报

1万

主题

8万

帖子

9万

积分

Sapphire Diamond LV48

Crazy TechnologyDaily Check-inTech Fans2020PL KINGHappy children's dayEid PartyHot PartyDiamondHappy Easter Medal.pngGood Wallpaper DesignerChocolate Day MedalHug Day MedalKiss Day MedalPromise Day MedalPropose Day MedalTeddy Day Medalsingel medalACTIVE STAR2020 Wish Medal20M4.0xclub2019post star1post star2post star3sign star1sign star2sign star3sign star4post star4post star5X'Club badge exclusive for Indiasign star5

#2
发表于 2019-05-22 09:53:48 来自手机 | 只看该作者
Good share
回复

使用道具 举报

2万

主题

18万

帖子

20万

积分

Super CP

couple medalStay HomeHappy Mother's DaySmileHappy children's dayEid PartyBest ReviewsX'Club badge exclusive for NigeriaXclub Kolpost star1sign star1post star2sign star2PL KINGTech Fans2020Weekly Tech Star2020Daily Check-inCrazy Technology

#3
发表于 2019-05-22 10:48:26 来自手机 | 只看该作者
Nice share brother
回复

使用道具 举报

294

主题

3979

帖子

4411

积分

Jade Star LV10

Chocolate Day MedalHug Day MedalKiss Day MedalPromise Day MedalPropose Day MedalTeddy Day MedalValentines Lovesign star1sign star2sign star3post star1post star2sign star4X'Club badge exclusive for India

#4
发表于 2019-05-22 12:12:20 来自手机 | 只看该作者
good info share.
回复

使用道具 举报

高级模式
B Color Link Quote Code Smilies |上传

本版积分规则

Infinix Official Website|Infinix official mall|infinix Note 4|XCLUB-COOL STUFF AROUND YOU

GMT+8, 2025-06-29 04:50 , Processed in 0.037662 second(s), 22 queries .

Powered by Discuz! X3.4

© 2001-2017 Comsenz Inc.

快速回复 返回顶部 返回列表