DADD: Discovering and Attesting Digital Discrimination

Lead Research Organisation: King's College London
Department Name: Informatics

Abstract

In digital discrimination, users are treated unfairly, unethically or just differently based on their personal data. Examples include low-income neighborhoods targeted with high-interest loans; women being undervalued by 21% in online marketing; and online ads suggestive of arrest records appearing more often with searches of black-sounding names than white-sounding names. Digital discrimination very often reproduces existing instances of discrimination in the offline world by either inheriting the biases of prior decision makers, or simply reflecting widespread prejudices in society. Digital discrimination may also have an even more perverse result, it may exacerbate existing inequalities by causing less favourable treatment for historically disadvantaged groups, suggesting they actually deserve that treatment. As more and more tasks are delegated to computers, mobile devices, and autonomous systems, digital discrimination is becoming a huge problem.

Digital discrimination can be the result of algorithmic biases, i.e., the way in which a particular algorithm has been designed creates discriminatory outcomes, but it also occurs using non-biased algorithms when they are fed or trained with biased data. Research has been conducted on so-called fair algorithms, tackling biased input data, demonstrating learned biases, and measuring relative influence of data attributes, which can quantify and limit the extent of bias introduced by an algorithm or dataset. But, how much bias is too much? That is, what is legal, ethical and/or socially-acceptable? And even more importantly, how do we translate those legal, ethical, or social expectations into automated methods that attest digital discrimination in datasets and algorithms?

DADD (Discovering and Attesting Digital Discrimination) is a *novel cross-disciplinary collaboration* to address these open research questions following a continuously-running co-creation process with academic (Computer Science, Digital Humanities, Law and Ethics) and non-academic partners (Google, AI Club), and the general public, including technical and non-technical users. DADD will design ground-breaking methods to certify whether or not datasets and algorithms discriminate by automatically verifying computational non-discrimination norms, which will in turn be formalised based on socio-economic, cultural, legal, and ethical dimensions, creating the new *transdisciplinary field of digital discrimination certification*.

Planned Impact

DADD will run a programme of activities with a view to maximising:

*Long-term cultural change* will be pursued by: i) actively engaging with the four research themes in King's Research Strategy and Vision 2029 to foster and encourage cross-disciplinary research in the College, and in particular, DADD is part of and will work closely with the Social Justice theme (one of the four themes); ii) building project team's skills (PDRAs and investigators), leveraging the co-creation and the way of working together in the project, and KCL's Centre for Research Staff Development and KCL's Engagement Services Team; and iii) training the next generation of digital discrimination researchers and professionals (PDRAs, PhD and MSc students), leveraging the PhD studentship committed by KCL, the London Interdisciplinary Social Science Doctoral Training Partnership, and KCL MSc programmes like the MSc in Data Science.

*Influencing other ICT researchers to foster adoption of cross-disciplinarity and co-creation* will be delivered by organising workshops at academic conferences, at King's and in conjunction with the Science and Engineering South (SES), and at other venues in collaboration with other projects funded in this call, together with publications (magazines like CACM) on cross-disciplinary and co-creation methodologies, their benefits, and know-how gained throughout the project.

*Sustainable economic and public sector impact* will be delivered by working directly with companies and the public sector (e.g. e-government) using a two-staged approach. In the first stage, a business and public-sector-focused workshop will run to engage with potentially interested businesses and professionals, seeking engagement with the Digital Catapult. We have already secured support and collaboration from Google and AI Club who will also be invited to these events. In the second stage, activities with already secured partners and other selected partners from stage one will run, e.g., Google confirmed that they will collaborate with DADD (advisory board and project activities). DADD will utilise internal (IAA) and external (KTP, Google schemes) grant mechanisms for follow-up collaborations beyond its life, and it will engage with KCL's IP & Licensing Team to explore commercialization options for the Toolkit.

*Influencing policy and regulation on digital discrimination* will leverage the extensive expertise at KCL through its Policy Institute to reach out to and engage with policy makers, and campaign for and raise awareness about digital discrimination, to update them with project findings, and explore possible areas of impact in current and prospective regulations. In particular, relevant regulation bodies like the Information Commissioner's Office (ICO) and Prudential Regulation Authority (PRA) will be contacted and invited to engagement workshops. Also, the PI will leverage his policy contacts as member of the Policy Fellows Network at Centre for Science and Policy, University of Cambridge (a network of academics and policy professionals).

*Raising public awareness of digital discrimination* is embedded in the participatory research methods used in DADD in which the general public will take part (e.g. techno-cultural workshops). Public Awareness will also be delivered by means of a highly active Social Media and Press Releases strategy.

Publications

10 25 50
publication icon
Criado N (2019) Algorithmic Regulation

publication icon
Criado N (2019) Algorithmic Regulation

publication icon
Criado N (2021) Attesting Digital Discrimination Using Norms in International Journal of Interactive Multimedia and Artificial Intelligence

publication icon
Ferrer X (2021) Bias and Discrimination in AI: A Cross-Disciplinary Perspective in IEEE Technology and Society Magazine

publication icon
Ferrer-Aran X (2023) Discovering and Interpreting Biased Concepts in Online Communities in IEEE Transactions on Knowledge and Data Engineering

 
Description We have found from that bias is sometimes confused with discrimination, and that the two are different. We have also found that we can spot biases and discrimination in an machine learning model trained with human language from online communities. We also found that the transparency needed to know whether AI models discriminate varies depending on the stakeholders, and different types and degrees of transparency are needed for engineers, users, and legal practitioners.

We created a method to automatically find biases in Machine-learning-based natural language processing models. We have applied this to study the language people use online in social media to discover the biases they have. We have also created a method to automatically check non-discrimination norms in general-purpose machine learning systems.
Exploitation Route We anticipate some of the results will be of a lot of interest to the social sciences as an automated method to study biases in language.
Sectors Communities and Social Services/Policy,Creative Economy,Education,Financial Services, and Management Consultancy,Government, Democracy and Justice

URL https://dadd-project.github.io/
 
Description Part of the tools we have developed are currently being used to teach digital humanities courses at King's College London and UC Berkeley (USA). We engaged with Science Gallery London to create a strand in which visitors could interact with the language biases people have, particularly when talking about gender, in the SGL season "Genders: shaping and breaking the Binary". Project DADD led to a keynote at the European Central Bank (ECB) in Frankfurt, a keynote at the CHIST-ERA conference in Estonia, an invited talk at PwC headquarters in London, an invited talk at Twitter headquarters in London, an invited talk at the Gender Summit in London, an invited talk at the Financial Conduct Authority (FCA) in London,.
Sector Education,Government, Democracy and Justice
Impact Types Cultural,Societal

 
Title Collection of comments of two subreddits from Reddit.com (r/theredpill and r/atheism) 
Description >r/theredpill dataset: The Red Pill defines itself as a forum for the discussion of sexual strategy in a culture increasingly lacking a positive identity for men'.' Swallowing the pill', in the community parlance, denotes the acceptance of the belief that men, not women, have been socially disenfranchised in the west. Critical and feminist scholars have argued that the community of TRP propagates values like chauvinism and traditional male-female role definitions. Reddit placed the subreddit, which had more than 300.000 subscribers at the time, in 'quarantine' in September 2018. As a highly controversial discourse community, our question is to what extent word embeddings are able to trace bias surrounding gender. >r/atheism dataset: Created from the subreddit r/atheism, a large community (with about 2.5 million members), calling itself 'the web's largest atheist forum', on which[a]ll topics related to atheism, agnosticism and secular living are welcome'. Religious biases between Islam and Christian are explored in this dataset (see submitted paper for a detailed list wrt attribute concept word sets). 
Type Of Material Database/Collection of data 
Year Produced 2020 
Provided To Others? Yes  
Impact We discovered word and conceptual biases between different concepts related to gender, religion and nationality in various Reddit communities. The discoveries show strong gender and religion biases in the two communities explored. 
URL https://osf.io/qmf62/?view_only=6be755746530433da0a5d985ffa69579
 
Title Unfair treatment by automated computational systems 
Description This dataset describes the results from a prescreened survey of 663 participants describing their experiences with unfair treatment caused by automated computational systems. After cleaning, the dataset contains a list of 620 participant quotes and their demographics in an Excel spreadsheet. The data describes experiences by users who are faced with automated decisions, strategies for harm reduction, and perceptions of fairness and discrimination. The data also includes questions on participants' self-perceived technical literacy, and several demographic questions. Participants have been anonymised. Participants were recruited through research recruitment platform Prolific, and oversampled for "at-risk characteristics" (see paper). The data excludes 9 participants who failed at least one attention check, and 24 participants who did not finish the survey. The DOI of the accompanying research paper is https://doi.org/10.1145/3555546. The dataset can be shared on request for 12 months after the end of the study (30 June 2022) in accordance with participant consent and EPSRC guidelines. 
Type Of Material Database/Collection of data 
Year Produced 2022 
Provided To Others? Yes  
URL https://kcl.figshare.com/articles/dataset/Unfair_treatment_by_automated_computational_systems/204992...
 
Title Unfair treatment by automated computational systems 
Description This dataset describes the results from a prescreened survey of 663 participants describing their experiences with unfair treatment caused by automated computational systems. After cleaning, the dataset contains a list of 620 participant quotes and their demographics in an Excel spreadsheet. The data describes experiences by users who are faced with automated decisions, strategies for harm reduction, and perceptions of fairness and discrimination. The data also includes questions on participants' self-perceived technical literacy, and several demographic questions. Participants have been anonymised. Participants were recruited through research recruitment platform Prolific, and oversampled for "at-risk characteristics" (see paper). The data excludes 9 participants who failed at least one attention check, and 24 participants who did not finish the survey. The DOI of the accompanying research paper is https://doi.org/10.1145/3555546. The dataset can be shared on request for 12 months after the end of the study (30 June 2022) in accordance with participant consent and EPSRC guidelines. 
Type Of Material Database/Collection of data 
Year Produced 2022 
Provided To Others? Yes  
URL https://kcl.figshare.com/articles/dataset/Unfair_treatment_by_automated_computational_systems/204992...
 
Description Collaboration with Science Gallery London 
Organisation King's College London
Department Science Gallery London
Country United Kingdom 
Sector Academic/University 
PI Contribution We contributed one strand and two workshops to run during the "Genders: Breaking the binary" exhibition period at Science Gallery London. This used our developments in the project to check for biases in language to have a tool to show visitors how different words are more or less associated to genders. Also, we will be running two techno-cultural workshops during the exhibition process, which will provide very valuable data for the project.
Collaborator Contribution They gave us the opportunity to run a workshop that will contribute to gather data for our project.
Impact Still ongoing exhibition and the workshops are to run, we will have more of this next year.
Start Year 2019
 
Title A Normative approach to Attest Digital Discrimination Source Code 
Description This repository contains the source code of the original paper 'A Normative approach to Attest Digital Discrimination' accepted at AI4EQ Workshop of ECAI 2020, and part of the project Discovering and Attesting Digital Discrimination (DADD). Digital discrimination is a form of discrimination whereby users are automatically treated unfairly, unethically or just differently based on their personal data by a machine learning (ML) system. Examples of digital discrimination include low-income neighborhood's targeted with high-interest loans or low credit scores, and women being undervalued by 21% in online marketing. Recently, different techniques and tools have been proposed to detect biases that may lead to digital discrimination. These tools often require technical expertise to be executed and for their results to be interpreted. To allow non-technical users to benefit from ML, simpler notions and concepts to represent and reason about digital discrimination are needed. In this paper, we use norms as an abstraction to represent different situations that may lead to digital discrimination. In particular, we formalise non-discrimination norms in the context of ML systems and propose an algorithm to check whether ML systems violate these norms. 
Type Of Technology Software 
Year Produced 2020 
Open Source License? Yes  
Impact This is the source code of the original paper 'A Normative approach to Attest Digital Discrimination' accepted at AI4EQ Workshop of ECAI 2020, and part of the project Discovering and Attesting Digital Discrimination (DADD). 
URL https://github.com/xfold/NormativeApproachToDiscrimination
 
Title Discovering And Interpreting Conceptual Biases Source Code 
Description This repository contains the source code of the original paper Discovering and Interpreting Conceptual Biases in Online Communities. This work is part of the project Discovering and Attesting Digital Discrimination (DADD). Related to this work, we created the Language Bias Visualiser, an interactive web-based platform that helps to explore gender biases found in various Reddit datasets. 
Type Of Technology Software 
Year Produced 2020 
Open Source License? Yes  
Impact This software allowed the analysis of conceptual biases in online communities that led to the publication with the same name. 
URL https://github.com/xfold/DiscoveringAndInterpretingConceptualBiases
 
Title Discovering and Interpreting Conceptual Biases in Online Communities 
Description Language carries implicit biases, functioning both as a reflection and a perpetuation of stereotypes that people carry with them. Recently, advances in Artificial Intelligence have made it possible to use machine learning techniques to trace linguistic biases. One of the most promising approaches in this field involves word embeddings, which transform text into high-dimensional vectors and capture semantic relations between words, and which has been successfully used to quantify human biases in large textual datasets. The Language Bias Visualiser helps to explore word and contextual biases found in data collected from different Reddit communities in an interactive and visually pleasing way. 
Type Of Technology Webtool/Application 
Year Produced 2020 
Open Source License? Yes  
Impact The tool is already being used in two top universities in the USA and the UK to teach biases and discrimination in online communities in digital humanities and social sciences courses. 
URL https://xfold.github.io/Web-DiscoveringAndInterpretingConceptualBiases/
 
Title Gender Bias Visualiser in Reddit 
Description The online tool helps to explore gender biases found in four different Reddit communities in an interactive and visually pleasing way. 
Type Of Technology Webtool/Application 
Year Produced 2020 
Open Source License? Yes  
Impact The tool is already being used in two top universities in the USA and the UK to teach gender biases in language, and discrimination in online communities in digital humanities and social sciences courses. 
URL https://xfold.github.io/WE-GenderBiasVisualisationWeb/
 
Title Language Biases In Reddit Source Code 
Description This repository contains the source code of the original paper "Discovering and Categorising Language Biases in Reddit" accepted at the International Conference on Web and Social Media (ICWSM 2021). This work is part of the project Discovering and Attesting Digital Discrimination (DADD). Related to this work, we created the Language Bias Visualiser, an interactive web-based platform that helps exploring gender biases found in various Reddit datasets. In this work we present a data-driven approach using word embeddings to discover and categorise language biases on the discussion platform Reddit. As spaces for isolated user communities, platforms such as Reddit are increasingly connected to issues of racism, sexism and other forms of discrimination. Hence, there is a need to monitor the language of these groups. One of the most promising AI approaches to trace linguistic biases in large textual datasets involves word embeddings, which transform text into high-dimensional dense vectors and capture semantic relations between words. Yet, previous studies require predefined sets of potential biases to study, e.g., whether gender is more or less associated with particular types of jobs. This makes these approaches unfit to deal with smaller and community-centric datasets such as those on Reddit, which contain smaller vocabularies and slang, as well as biases that may be particular to that community. This paper proposes a data-driven approach to automatically discover language biases encoded in the vocabulary of online discourse communities on Reddit. In our approach, protected attributes are connected to evaluative words found in the data, which are then categorised through a semantic analysis system. We verify the effectiveness of our method by comparing the biases we discover in the Google News dataset with those found in previous literature. We then successfully discover gender bias, religion bias, and ethnic bias in different Reddit communities. We conclude by discussing potential application scenarios and limitations of this data-driven bias discovery method. 
Type Of Technology Software 
Year Produced 2020 
Open Source License? Yes  
Impact This is the source code of the original paper "Discovering and Categorising Language Biases in Reddit" accepted at the International Conference on Web and Social Media (ICWSM 2021). This work is part of the project Discovering and Attesting Digital Discrimination (DADD). 
URL https://github.com/xfold/LanguageBiasesInReddit
 
Description Inivited talk Twitter 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Industry/Business
Results and Impact One of the PostDocs of the project delivered a talk at Twitter (London premises) with employees from Twitter joining via teleconference from around the world. The talk detailed some of the advances we are making in the project and how that could be applied to online platforms like Twitter.
Year(s) Of Engagement Activity 2019
 
Description Invited Talk - PwC 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach National
Primary Audience Industry/Business
Results and Impact One of the CoIs of the project delivered an invited talk about the project at PwC's headquarters in London.
Year(s) Of Engagement Activity 2019
 
Description Keynote Chist-Era 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Other audiences
Results and Impact The PI was invited to deliver the Keynote of the CHIST-ERA Conference, which included funders and academics across Europe to delineate a call for projects.
Year(s) Of Engagement Activity 2019
 
Description Language Bias Visualizer 
Form Of Engagement Activity Engagement focused website, blog or social media channel
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Public/other audiences
Results and Impact The DADD Language Bias Visualiser users Word Embeddings to connect target concepts such as `male' or `female' to evaluative attributes found in online data, which are then categorised through clustering algorithms and labelled through a semantic analysis system into more general (conceptual) biases. Categorising biases allows us to give a broad picture of the biases present in discourse communities, such as those on Reddit.
Year(s) Of Engagement Activity 2020
URL https://xfold.github.io/WE-GenderBiasVisualisationWeb/
 
Description SOCINFO2020 TUTORIAL - DISCOVERING GENDER BIAS AND DISCRIMINATION IN LANGUAGE 
Form Of Engagement Activity Participation in an activity, workshop or similar
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Other audiences
Results and Impact The tutorial focuses on the issue of digital discrimination, particularly towards gender. Its main goal is to help participants improve their digital literacy by understanding the social issues at stake in digital (gender) discrimination, and learning about technical applications and solutions. The tutorial is divided in four parts: it basically iterates twice through the social and technical dimensions. We make use of our own research in language modelling and Word Embeddings in order to clarify how human gender biases may be incorporated into AI/ML models. We first offer a short introduction to digital discrimination and (gender) bias. We give examples of gender discrimination in the field of AI/ML, and discuss the clear gender binary (M/F) that is presupposed when dealing with a computational bias towards gender. We then move to a technical perspective, introducing the DADD Language Bias Visualiser which allows us to discover and analyze gender bias using Word Embeddings. Finally, we show how computational models of bias and discrimination are built on implicit binaries, and discuss with participants the difficulties pertaining to these assumptions in times of post-binary gender attribution.
Year(s) Of Engagement Activity 2020
URL http://dadd-project.org/2020/10/06/socinfo2020-tutorial-discovering-gender-bias-and-discrimination-i...
 
Description Science Gallerey London App 
Form Of Engagement Activity Engagement focused website, blog or social media channel
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Public/other audiences
Results and Impact The wonderful app that Steve Brown built for the exhibit with Science Gallery London, which uses Word Embeddings models built by the project. Users play a small game to explore language bias in a Google News dataset, and The Red Pill, a notorious community on Reddit.
Year(s) Of Engagement Activity 2020
URL http://sgl.stevebrown.co/dadd
 
Description Talk to the FCA 
Form Of Engagement Activity A talk or presentation
Part Of Official Scheme? No
Geographic Reach National
Primary Audience Policymakers/politicians
Results and Impact We were invited to deliver a talk about the project by the Financial Conduct Authority (FCA) on their premises.
Year(s) Of Engagement Activity 2019
 
Description Video Lectures 
Form Of Engagement Activity A broadcast e.g. TV/radio/film/podcast (other than news/press)
Part Of Official Scheme? No
Geographic Reach International
Primary Audience Public/other audiences
Results and Impact This was a series of video lectures for digital humanities students that we made available through youtube and the project website. They are very introductory and hence serve for the general public too.

There are several URLs for this:
http://dadd-project.org/2020/03/26/bias-and-discrimination-web-lecture/
http://dadd-project.org/2020/03/21/dadd-video-lectures-word-embeddings-and-language-bias/
Year(s) Of Engagement Activity 2020