• Aucun résultat trouvé

Recognizing and Reducing Bias in NLP Applications

N/A
N/A
Protected

Academic year: 2022

Partager "Recognizing and Reducing Bias in NLP Applications"

Copied!
1
0
0

Texte intégral

(1)

Recognizing and Reducing Bias in NLP Applications

Dirk Hovy Universit` a Bocconi, Italia dirk.hovy@unibocconi.it

Abstract

As NLP technology becomes used in ever more settings, it has ever more impact on the lives of people all around the world. As NLP practitioners, we have become increasingly aware that we have the responsibility to evaluate the effects of our research and prevent or at least mitigate harmful outcomes.

This is true for academic researchers, government labs, and industry devel- opers. However, without experience of how to recognize and engage with the many ethical conundrums in NLP, it is easy to become overwhelmed and remain inactive. One of the most central ethical issues in NLP is the impact of hidden biases that affect performance unevenly, and thereby disadvantage certain user groups.

This tutorial aims to empower NLP practitioners with the tools spot these biases, and a number of other common ethical pitfalls of our practice. We will cover both high-level strategies, as well as go through specific case sample exercises. This is a highly interactive workshop with room for debate and questions from the attendees. The workshop will cover the following broad topics:

• Biases: Understanding the different ways in which biases affect NLP data, models, and input representations, including including strategies to test for and reduce bias in all of them.

• Dual Use: Learning to anticipate how a system could be repurposed for harmful or negative purposes, rather than its intended goal.

• Privacy: Protecting the privacy of users both in corpus construction and model building.

Références

Documents relatifs

Unlike other machine learning methods, this topological data analysis method im- poses less burden in the sense that the human-in- the-loop does not need to select the right metric

The Tint pipeline is based on Stanford CoreNLP (Manning et al., 2014), an open-source framework written in Java, that provides most of the com- mon Natural Language Processing

The OHSU Knight Cancer Institute has begun a program to create a natural language processing (NLP) capability to extract, store, and link data from free text sources at the

Stereotypes, therefore, represent biases toward and against others. In this pa- per, we formalise stereotypical bias within trust evaluations. Then, using the iterated

The SIRO model for minimal information breaks down the protocol in key ele- ments that we have found to be common to our corpus of experimental protocols: i) Sample/ Specimen (S),

Well-structured and accurately described protocols (procesable by humans and machines) should facilitate experimental reproducibility. In this poster we present the

on evidence. This argumentation structure makes very clear that S is not supported by any evidence. Thus the analyst should lower her confidence in the

य़JT IZQPUIFTJT IBT CFFO QVU UP UIF UFTU CZ UFBDIJOH EJ੖FSFOU QIPOPMPHJDBM QB॒FSOT UP MFBSOFST UP TFF JG DFSUBJO QB॒FSOT BSF MFBSOU CF॒FS PS NPSF RVJDLMZ UIBO PUIFST 'PS FYBNQMF