Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
7 views62 pages

Ethics

The document discusses the ethical implications of Natural Language Processing (NLP) systems, particularly how they can affect individuals' lives through biased data and decision-making processes. It highlights various sources of bias in machine learning, including biased training data, objectives, and amplification of biases in models. The document emphasizes the importance of addressing these biases to ensure fair and equitable outcomes in NLP applications.

Uploaded by

codersento
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views62 pages

Ethics

The document discusses the ethical implications of Natural Language Processing (NLP) systems, particularly how they can affect individuals' lives through biased data and decision-making processes. It highlights various sources of bias in machine learning, including biased training data, objectives, and amplification of biases in models. The document emphasizes the importance of addressing these biases to ensure fair and equitable outcomes in NLP applications.

Uploaded by

codersento
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 62

ethics in NLP

CS 685, Spring 2022


Introduction to Natural Language Processing
http://people.cs.umass.edu/~miyyer/cs685/

Mohit Iyyer
College of Information and Computer Sciences
University of Massachusetts Amherst

many slides from Yulia Tsvetkov & Mark Yatskar


OpenAI PALMS: https://
openai.com/blog/improving-
language-model-behavior/

Fine-tune LMs on values-


targeted datasets
Fine-tune on small set of QA pairs

Solaiman & Dennison, 2021


And change the behavior of the model!

Solaiman & Dennison, 2021


Solaiman & Dennison, 2021
Demo: https://delphi.allenai.org/
what are we talking about today?
• many NLP systems affect actual people
• systems that interact with people (conversational agents)
• perform some reasoning over people (e.g.,
recommendation systems, targeted ads)
• make decisions about people’s lives (e.g., parole
decisions, employment, immigration)

• questions of ethics arise in all of these applications!


why are we talking about it?
• the explosion of data, in particular user-generated
data (e.g., social media)

• machine learning models that leverage huge amounts


of this data to solve certain tasks
Learn to Assess AI Systems Adversarially

● Who could benefit from such a technology?


● Who can be harmed by such a technology?

● Representativeness of training data


● Could sharing this data have major effect on people’s lives?

● What are confounding variables and corner cases to control for?


● Does the system optimize for the “right” objective?
● Could prediction errors have major effect on people’s lives?
let’s start with the data…
BIASED

A
I
Online data is riddled with SOCIAL STEREOTYPES
Racial Stereotypes

● June 2016: web search query “three black teenagers”


Gender/Race/Age Stereotypes

● June 2017: image search query “Doctor”


Gender/Race/Age Stereotypes
● June 2017: image search query “Nurse”
Gender/Race/Age Stereotypes

● June 2017: image search query “Homemaker”


Gender/Race/Age Stereotypes
● June 2017: image search query “CEO”
BIASED

A
I
Consequence: models are biased
Gender Biases on the Web

● The dominant class is often portrayed and perceived as relatively more


professional (Kay, Matuszek, and Munson 2015)
● Males are over-represented in the reporting of web-based news articles
(Jia, Lansdall-Welfare, and Cristianini 2015)
● Males are over-represented in twitter conversations (Garcia, Weber, and
Garimella 2014)
● Biographical articles about women on Wikipedia disproportionately discuss
romantic relationships or family-related issues (Wagner et al. 2015)
● IMDB reviews written by women are perceived as less useful (Otterbacher
2013)
Biased NLP Technologies

● Bias in word embeddings (Bolukbasi et al. 2017; Caliskan et al.


2017; Garg et al. 2018)
● Bias in Language ID (Blodgett & O'Connor. 2017; Jurgens et al.
2017)
● Bias in Visual Semantic Role Labeling (Zhao et al. 2017)
● Bias in Natural Language Inference (Rudinger et al. 2017)
● Bias in Coreference Resolution (At NAACL: Rudinger et al. 2018;
Zhao et al. 2018 )
● Bias in Automated Essay Scoring (At NAACL: Amorim et al. 2018)
Zhao et al., NAACL 2018
Sources of Human Biases in Machine Learning

● Bias in data and sampling

● Optimizing towards a biased objective

● Inductive bias

● Bias amplification in learned models


Sources of Human Biases in Machine Learning

● Bias in data and sampling

● Optimizing towards a biased objective

● Inductive bias

● Bias amplification in learned models


Types of Sampling Bias in Naturalistic Data
● Self-Selection Bias
○ Who decides to post reviews on Yelp and why?
Who posts on Twitter and why?
● Reporting Bias
○ People do not necessarily talk about things in the world in
proportion to their empirical distributions
(Gordon and Van Durme 2013)

● Proprietary System Bias


○ What results does Twitter return for a particular
query of interest and why? Is it possible to know?

● Community / Dialect / Socioeconomic Biases


○ What linguistic communities are over- or under-represented?
leads to community-specific model performance (Jorgensen et al. 2015)
credit: Brendan O’Connor
Example: Bias in Language Identification

● Most applications employ off-the-shelf LID systems which


are highly accurate

*Slides on LID by David Jurgens


(Jurgens et al. ACL’17)
McNamee, P., “Language identification: a solved problem suitable
for undergraduate instruction” Journal of Computing Sciences in
Colleges 20(3) 2005.

“This paper describes […]


how even the most simple of
these methods using data
obtained from the World
Wide Web achieve accuracy
approaching 100% on a test
suite comprised of ten
European languages”
● Language identification degrades significantly on African American
Vernacular English
(Blodgett et al. 2016) Su-Lin Blodgett just got her PhD from UMass!
LID Usage Example: Health Monitoring
LID Usage Example: Health Monitoring
Socioeconomic Bias in Language Identification

● Off-the-shelf LID systems under-represent populations in


less-developed countries

Jurgens et al. ACL’17


Better Social Representation through
Network-based Sampling

● Re-sampling from strategically-diverse corpora


Topical Geographic

Socia
Multilingual
l

Jurgens et al. ACL’17


Estimated accuracy for
English tweets

Human Development Index of


text’s origin country

Jurgens et al. ACL’17


Sources of Human Biases in Machine Learning

● Bias in data and sampling

● Optimizing towards a biased objective

● Inductive bias

● Bias amplification in learned models


Optimizing Towards a Biased Objective

● Northpointe vs ProPublica
Optimizing Towards a Biased Objective

“what is the probability that this person will commit a serious


crime in the future, as a function of the sentence you give
them now?”
Optimizing Towards a Biased Objective

“what is the probability that this person will commit a serious crime
in the future, as a function of the sentence you give them now?”

● COMPAS system
○ balanced training data about people of all races
○ race was not one of the input features
● Objective function
○ labels for “who will commit a crime” are unobtainable
○ a proxy for the real, unobtainable data: “who is more likely to be
convicted”

what are some issues with


this proxy objective?
Predicting prison sentences
given case descriptions

Chen et al., EMNLP 2019, “Charge-based prison term prediction…”


Is this sufficient consideration of ethical
issues of this work? Should the work
have been done at all?

Chen et al., EMNLP 2019, “Charge-based prison term prediction…”


Sources of Human Biases in Machine Learning

● Bias in data and sampling

● Optimizing towards a biased objective

● Inductive bias

● Bias amplification in learned models


what is inductive bias?
• the assumptions used by our model. examples:

• recurrent neural networks for NLP assume that the


sequential ordering of words is meaningful

• features in discriminative models are assumed to be


useful to map inputs to outputs
Bias in Word Embeddings

1. Caliskan, A., Bryson, J. J. and Narayanan, A. (2017) Semantics derived


automatically from language corpora contain human-like biases.
Science

2. Bolukbasi T., Chang K.-W., Zou J., Saligrama V., Kalai A. (2016) Man is to
Computer Programmer as Woman is to Homemaker? Debiasing Word
Embeddings. NIPS

3. Nikhil Garg, Londa Schiebinger, Dan Jurafsky, James Zou. (2018) Word
embeddings quantify 100 years of gender and ethnic stereotypes.
PNAS.
Biases in Embeddings: Another Take
Towards Debiasing

1. Identify gender subspace: B


Gender Subspace

The top PC captures the gender


subspace
Towards Debiasing

1. Identify gender subspace: B


2. Identify gender-definitional (S) and gender-neutral
words (N)
Gender-definitional vs. Gender-neutral Words
Towards Debiasing

1. Identify gender subspace: B


2. Identify gender-definitional (S) and gender-neutral words
(N)
3. Apply transform matrix (T) to the embedding matrix (W)
such that
a. Project away the gender subspace B from the gender-neutral words N
b. But, ensure the transformation doesn’t change the embeddings too much

Don’t modify Minimize gender


embeddings too component
much

T - the desired debiasing transformation B - biased space


W - embedding matrix
N - embedding matrix of gender neutral words
Sources of Human Biases in Machine Learning

● Bias in data and sampling

● Optimizing towards a biased objective

● Inductive bias

● Bias amplification in learned models


Bias Amplification

Zhao, J., Wang, T., Yatskar, M., Ordonez, V and Chang, M.-
W. (2017) Men Also Like Shopping: Reducing Gender
Bias Amplification using Corpus-level Constraint.
EMNLP
imSitu Visual Semantic Role Labeling (vSRL)

Slides by Mark Yatskar https://homes.cs.washington.edu/~my89/talks/ZWYOC17_slide.pdf


imSitu Visual Semantic Role Labeling (vSRL)
Dataset Gender Bias
Model Bias After Training
Algorithmic Bias
Quantifying Dataset Bias
Quantifying Dataset Bias: Dev Set
Model Bias Amplification
Reducing Bias Amplification (RBA)
Discussion

● Applications that are built from online data, generated by


people, learn also real-world stereotypes
● Should our ML models represent the “real world”?
● Or should we artificially skew data distribution?
● If we modify our data, what are guiding principles on what
our models should or shouldn't learn?

You might also like