Filters








6,561 Hits in 4.2 sec

Adversarial Removal of Demographic Attributes from Text Data [article]

Yanai Elazar, Yoav Goldberg
2018 arXiv   pre-print
The implication is that decisions of classifiers trained on textual data are not agnostic to -- and likely condition on -- demographic attributes.  ...  We show that demographic information of authors is encoded in -- and can be recovered from -- the intermediate representations learned by text-based neural classifiers.  ...  Recent work on creating private representation in the text domain (Li et al., 2018) share our motivation of removing unintended demographic attributes from the learned representation using adversarial  ... 
arXiv:1808.06640v2 fatcat:j6vrcvpvajhdxpzzgwjlaal75q

Adversarial Removal of Demographic Attributes from Text Data

Yanai Elazar, Yoav Goldberg
2018 Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing  
The implication is that decisions of classifiers trained on textual data are not agnostic to-and likely condition on-demographic attributes.  ...  We show that demographic information of authors is encoded in-and can be recovered from-the intermediate representations learned by text-based neural classifiers.  ...  Recent work on creating private representation in the text domain (Li et al., 2018) share our motivation of removing unintended demographic attributes from the learned representation using adversarial  ... 
doi:10.18653/v1/d18-1002 dblp:conf/emnlp/ElazarG18 fatcat:6j7hkybrq5aftcicdcwe7ivovm

Adversarial Scrubbing of Demographic Information for Text Classification [article]

Somnath Basu Roy Chowdhury, Sayan Ghosh, Yiyuan Li, Junier B. Oliva, Shashank Srivastava, Snigdha Chaturvedi
2021 arXiv   pre-print
Contextual representations learned by language models can often encode undesirable attributes, like demographic associations of the users, while being trained for an unrelated target task.  ...  In this paper, we present an adversarial learning framework "Adversarial Scrubber" (ADS), to debias contextual representations.  ...  In this work, we focus on removing demographic attributes encoded in data representations during training text classification systems.  ... 
arXiv:2109.08613v1 fatcat:evjk4v3ubzaljnl5p65ypcjjda

Demoting Racial Bias in Hate Speech Detection [article]

Mengzhou Xia, Anjalie Field, Yulia Tsvetkov
2020 arXiv   pre-print
This bias in annotated training data and the tendency of machine learning models to amplify it cause AAE text to often be mislabeled as abusive/offensive/hate speech with a high false positive rate by  ...  In this paper, we use adversarial training to mitigate this bias, introducing a hate speech classifier that learns to detect toxic sentences while demoting confounds corresponding to AAE texts.  ...  Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the NSF.  ... 
arXiv:2005.12246v1 fatcat:klt4rbqn3fbtrnlpyjkzuw5nyu

Disentangling Document Topic and Author Gender in Multiple Languages: Lessons for Adversarial Debiasing

Erenay Dayanik, Sebastian Padó
2021 Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis  
In this paper, we experiment with text classification of the correlated attributes of document topic and author gender, using a novel multilingual parallel corpus of TED talk transcripts.  ...  In social media analysis, this problem surfaces for demographic user classes such as language, topic, or gender, which influence how an author writes a text to a substantial extent.  ...  Elazar and Goldberg (2018) apply the idea to the removal of demographic bias; McHardy et al. (2019) remove publication source as a bias variable from a satire detection model. reported that adversarial  ... 
dblp:conf/wassa/DayanikP21 fatcat:2kjbzfu3ejes3ghdxplrdmf2n4

Neural User Factor Adaptation for Text Classification: Learning to Generalize Across Author Demographics

Xiaolei Huang, Michael J. Paul
2019 Proceedings of the Eighth Joint Conference on Lexical and Computational Semantics (*  
In this study, we examine empirically how text data can vary across four demographic factors: gender, age, country, and region.  ...  We propose a multitask neural model to account for demographic variations via adversarial training.  ...  The authors thank Zijiao Yang for helping evaluate inference accuracy of the Microsoft Face API. This work was supported in part by the National Science Foundation under award number IIS-1657338.  ... 
doi:10.18653/v1/s19-1015 dblp:conf/starsem/HuangP19 fatcat:pl2vtoywsnbzzgyrrmsibcxrfq

Privacy-Aware Text Rewriting

Qiongkai Xu, Lizhen Qu, Chenchen Xu, Ran Cui
2019 Proceedings of the 12th International Conference on Natural Language Generation  
Instead of relying on unknown decision systems or human decisionmakers, we argue that a better way to protect data providers is to remove the trails of sensitive information before publishing the data.  ...  Recent work from the NLP community focuses on building systems that make fair decisions based on text.  ...  Related Work Achieving fairness or preserving privacy through removing sensitive information from text has been explored by adversarial training (Li et al., 2018; Figure 2: Sample of original text,  ... 
doi:10.18653/v1/w19-8633 dblp:conf/inlg/XuQXC19 fatcat:oxjgrkqctza7xdivajsx2f4iuy

Towards Robust and Privacy-preserving Text Representations [article]

Yitong Li, Timothy Baldwin, Trevor Cohn
2018 arXiv   pre-print
Written text often provides sufficient clues to identify the author, their gender, age, and other important attributes.  ...  In this paper, we propose an approach to explicitly obscure important author characteristics at training time, such that representations learned are invariant to these attributes.  ...  expect a larger impact from bias removal using ADV training.  ... 
arXiv:1805.06093v1 fatcat:pyc3hmzl7vaxlcpwfgpwjkykay

Towards Robust and Privacy-preserving Text Representations

Yitong Li, Timothy Baldwin, Trevor Cohn
2018 Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)  
Written text often provides sufficient clues to identify the author, their gender, age, and other important attributes.  ...  In this paper, we propose an approach to explicitly obscure important author characteristics at training time, such that representations learned are invariant to these attributes.  ...  expect a larger impact from bias removal using ADV training.  ... 
doi:10.18653/v1/p18-2005 dblp:conf/acl/LiBC18 fatcat:3hr32nulofcb5hrxm4pqd2mlwi

Balancing out Bias: Achieving Fairness Through Balanced Training [article]

Xudong Han, Timothy Baldwin, Trevor Cohn
2022 arXiv   pre-print
We extend the method in the form of a gated model, which incorporates protected attributes as input, and show that it is effective at reducing bias in predictions through demographic input perturbation  ...  Group bias in natural language processing tasks manifests as disparities in system error rates across texts authorized by different demographic groups, typically disadvantaging minority groups.  ...  The baseline models are: STANDARD, which is a naively-trained MLP classifier; INLP (Ravfogel et al., 2020) , which removes demographic information from text representations through iterative nullspace  ... 
arXiv:2109.08253v2 fatcat:vdtzadpf3jhifmeljk62xtkpyi

Fairness in Deep Learning: A Computational Perspective [article]

Mengnan Du, Fan Yang, Na Zou, Xia Hu
2020 arXiv   pre-print
We provide a review covering recent progresses to tackle algorithmic fairness problems of deep learning from the computational perspective.  ...  We also discuss fairness mitigation approaches categorized according to three stages of deep learning life-cycle, aiming to push forward the area of fairness in deep learning and build genuinely fair and  ...  In-processing Adversarial Learning: From model training perspective, adversarial training [45] is a representative solution to remove information about sensitive attributes from intermediate representation  ... 
arXiv:1908.08843v2 fatcat:kaaevm64fbctpjfdycv5uz3dhi

A Prompt Array Keeps the Bias Away: Debiasing Vision-Language Models with Adversarial Learning [article]

Hugo Berg, Siobhan Mackenzie Hall, Yash Bhalgat, Wonsuk Yang, Hannah Rose Kirk, Aleksandar Shtedritski, Max Bain
2022 arXiv   pre-print
We address both of these challenges in this paper: First, we evaluate different bias measures and propose the use of retrieval metrics to image-text representations via a bias measuring framework.  ...  Prior proposed bias measurements lack robustness and feature degradation occurs when mitigating bias without access to pretraining data.  ...  unbiased, defined as outputting similar distributions of scores across attributes for a given text query which should be unrelated to demographic affiliation.  ... 
arXiv:2203.11933v2 fatcat:d5ahgqbrmbhlndyr6ln5tjvbwa

How to keep text private? A systematic review of deep learning methods for privacy-preserving natural language processing [article]

Samuel Sousa, Roman Kern
2022 arXiv   pre-print
Further, we discuss open challenges in privacy-preserving NLP regarding data traceability, computation overhead, dataset size, the prevalence of human biases in embeddings, and the privacy-utility tradeoff  ...  Data protection laws, such as the European Union's General Data Protection Regulation (GDPR), thereby enforce the need for privacy.  ...  Reducing or removing human biases from embedding models mainly relies on adversarial training.  ... 
arXiv:2205.10095v1 fatcat:rksy7oxxlbde5bol3ay44yycru

Obfuscating Gender in Social Media Writing

Sravana Reddy, Kevin Knight
2016 Proceedings of the First Workshop on NLP and Computational Social Science  
The vast availability of textual data on social media has led to an interest in algorithms to predict user attributes such as gender based on the user's writing.  ...  away their demographic identities.  ...  Paraphrasing algorithms benefit from parallel data: texts expressing the same message written by users from different demographic groups.  ... 
doi:10.18653/v1/w16-5603 dblp:conf/acl-nlpcss/ReddyK16 fatcat:2fmco776fvacnemldd77vxmkle

FairRank: Fairness-aware Single-tower Ranking Framework for News Recommendation [article]

Chuhan Wu, Fangzhao Wu, Tao Qi, Yongfeng Huang
2022 arXiv   pre-print
We apply adversarial learning to both of them to reduce the biases brought by sensitive user attributes.  ...  However, these models can easily inherit the biases related to users' sensitive attributes (e.g., demographics) encoded in training click data, and may generate recommendation results that are unfair to  ...  In fact, click behavior data inherently encodes biases related to users' sensitive attributes such as demographics [1, 5] .  ... 
arXiv:2204.00541v1 fatcat:ludwigtdeffdbfjzxfjl55rbwa
« Previous Showing results 1 — 15 out of 6,561 results