Dynamically Disentangling Social Bias from Task-Oriented Representations with Adversarial Attack

Abstract

Representation learning is widely used in NLP for a vast range of tasks. However, representations derived from text corpora often reflect social biases. This phenomenon is pervasive and consistent across different neural models, causing serious concern. Previous methods mostly rely on a pre-specified, user-provided direction or suffer from unstable training. In this paper, we propose an adversarial disentangled debiasing model to dynamically decouple social bias attributes from the intermediate representations trained on the main task. We aim to denoise bias information while training on the downstream task, rather than completely remove social bias and pursue static unbiased representations. Experiments show the effectiveness of our method, both on the effect of debiasing and the main task performance.

Publication
NAACL 2021
Liwen Wang
Liwen Wang
Postgraduate Student

Spoken Language Understading and related applications

Yuanmeng Yan
Yuanmeng Yan
Postgraduate Student

Spoken Language Understanding, Pre-training Language Model

Keqing He
Postgraduate Student

Dialogue System, Summarization, Pre-training Language Model

Yanan Wu
Yanan Wu
Postgraduate Student

Spoken Language Understading

Weiran Xu
Weiran Xu
Associate Professor, Master Supervisor, Ph.D Supervisor

Information Retrieval, Pattern Recognition, Machine Learning