Notes_Man2Programmer@Woman2Homemaker

探讨了英语词汇嵌入中固有的性别偏见问题,并提出了一种去除这些偏见的方法。研究发现,预先训练的词嵌入如word2vec中存在明显的性别偏见,例如将程序员更多地与男性联系起来,而护士则更多地与女性联系。为了解决这个问题,研究人员通过识别并移除性别中立词汇中的性别偏差维度来修正这种偏见。

Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings

introduction
Bias especially gender stereotypes in word embeddings:

e.g. Man - woman = programmer - homemaker

Pretrained embeddings: word2vec / 300dimensions / Google News

Quantify bias:

Compare a word vector to the vectors of a pair of gender-specific words. for example, nurse close to woman is not bias itself, because nurse close to humans, but closer than man suggest bias.

consider the distinction between gender specific words that are associated with a gender by definition (e.g. brother / sister), which close to a specfic gender is not bias, and the remaining gender neutral words (e.g. programmer / nurse).

We will use the gender specific words to learn a gender subspace ( Surprisingly, there exists a low dimensional subspace in the embedding that captures much of the gender bias.) in the embedding. Removes the bias only from the gender neutral words while respecting gender specific words.

Gender biases in English

Implicit Association Tests have uncovered gender-word biases that people do not self-report and may not even be aware of. Biases are shown in morphology as well as while there are more words referring to males, there are many more words that sexualize females than males.

Biases in algorithms

A number of online systems have been shown to exhibit various biases.Schmidt identified the bias present in word embeddings and proposed debiasing by entirely removing multiple gender dimensions. His approach is entirely remove gender from embeddings. At the same time, the difficulty of evaluating embedding quality (as compared to supervised learning) parallels the difficulty of defining bias in an embedding.

word embeddings

Embeddings form: wϵRd,||w||=1. Assume F-M pair PϵRdRd, gender neutral word NϵW, similiarity is cosine similarity:

cos(u,v)=uv|u||v|

so similarity between embeddings is
cos(w1,w2)=w1w2(2)
Crowd experiments
Geometry of Gender and Bias in Word Embeddings

understand biases present in embeddings(i.e which words more close to he/she etc.) and to which extent biases agree with human notion of stereotypes.

Occupational stereotypes

Ask the crowdworkers to evaluate whether an occupation is con-sidered female-stereotypic, male-stereotypic, or neutral. Spearman r=.51(strongly correlated):

the geometric biases of embedding vectors is aligned with crowd judgment.

Analogies exhibiting stereotypes

(To Be Continued…)

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值