Join Our Discord (940+ Members)

Think About the Stakeholders First! Towards an Algorithmic Transparency Playbook for Regulatory Compliance

Content License: cc-by

Think About the Stakeholders First! Towards an Algorithmic Transparency Playbook for Regulatory Compliance

Papers is Alpha. This content is part of an effort to make research more accessible, and (most likely) has lost some details from the original. You can find the original paper here.

Introduction

In the past decade, there has been widespread proliferation of artificial intelligence (AI) systems into the private and public sectors. These systems have been implemented in a broad range of contexts, including employment, healthcare, lending, criminal justice, and more. The rapid development and implementation of AI technologies has greatly outpaced public oversight, creating a “wild-west”-style regulatory environment. As policy makers struggle to catch up, the issues of unregulated AI have become glaringly obvious, especially for underprivileged and marginalized communities. Famously, ProPublica revealed that the AI-driven system COMPAS used to assess the likelihood of a prisoner recidivating was highly discriminatory against black individuals. In another example, Amazonbuilt and implemented an automated resume screening and hiring AI system–only to later find out that the system was biased against hiring women. In an effort to address these issues, countries around the world have begun regulating the use of AI systems. Over 50 nations and intergovernmental organizations have published AI strategies, actions plans, policy papers or directives. A survey of existing and proposed regulation around AI transparency is given in Section sec-laws.

Unfortunately, most strategies, directives and laws to date lack specificity on how AI regulation should be carried out in practice by technologists. Where there is specificity, there is a lack of mechanisms for enforcing laws and holding institutions using AI accountable.Documents on AI governance have focused on what to do (or what not to do) with respect to AI, but leave the brunt of the work to practitioners to figure out how things should be done. This tension plays out heavily in regulations governing the transparency of AI systems (called “explainability” by AI practitioners). The most prominent example of this is the “right to explanation” of data use that is included in the EU’s General Data Protection Regulation (GDPR). Despite being passed into law in 2016, the meaning and scope of the right is still being debated by legal scholars, with little of the discussion resulting in concrete benefits for citizens.

While regulation can help weigh the benefits of new technology against the risks, developingeffective regulation is difficult, as is establishing effective mechanisms to comply with existing regulation. This paper aims to fill a gap in the existing literature by writing to technologists and AI practitioners about the existing AI regulatory landscape, and speaks to their role in designing complaint systems. We make a case for why AI practitioners should be leading efforts to ensure the transparency of AI systems, and to this end, we propose a novel framework for implementing regulatory-compliant explanations for stakeholders. We also consider an instantiation of our stakeholder-first approach in the context of a real-world example using work done by a national employment agency.

We make the following three contributions: (1) provide a survey of existing and proposed regulations on the transparency and explainability of AI systems; (2) propose a novel framework for a stakeholder-first approach to designing transparent AI systems; and (3) present a case-study that illustrates how this stakeholder-first approach could be used in practice.

Existing and Emerging Regulations

In recent years, countries around the world have increasingly been drafting strategies, action plans, and policy directives to govern the use of AI systems. To some extent, regulatory approaches vary by country and region. For example, policy strategies in the US and the EU reflect their respective strengths: free-market ideas for the former, and citizen voice for the latter. Yet, despite country-level variation, many AI policies contain similar themes and ideas. A meta-analysis of over 80 AI ethics guidelines and soft-laws found that 87% mention transparency, and include an effort to increase the explainability of AI systems. Unfortunately, all documents to date have one major limitation: they are filled with uncertainty on how transparency and explainability should actually be implemented in a way that is compliant with the evolving regulatory landscape. This limitation has 3 main causes: (1) it is difficult to design transparency regulations that can easily be standardized across different fields of AI, such as self-driving cars, robotics, and predictive modeling; (2) when it comes to transparency, there is a strong information asymmetry between technologists and policymakers, and, ultimately, the individuals who are impacted by AI systems; (3) there is no normative consensus around AI transparency, and most policy debates are focused on the risks of AI rather than the opportunities. For the purposes of scope, we will focus on regulations in the United States and Europe. However, its important noting that there is meaningful AI regulation emerging in Latin and South America, Asia, Africa, and beyond, and summarizing those regulations is an avenue for future work. For example, in 2021, Chile presented it’s first national action plan on AI policy [https://www.gob.cl/en/news/chile-presents-first-national-policy-artificial-intelligence/ ].

United States

In 2019 the US took two major steps in the direction of AI regulation. First, Executive Order 13859 was issued with the purpose of establishing federal principles for AI systems, and to promote AI research, economic competitiveness, and national security. Importantly, the order mandates that AI algorithms implemented for use by public bodies must be “understandable”, “transparent”, “responsible”, and “accountable.” Second, the Algorithmic Accountability Act of 2019 was introduced to the House of Representatives, and more recently reintroduced under the name Algorithmic Accountability Act of 2022. If passed into law, the Algorithmic Accountability Act would be a landmark legalisation for AI regulation in the US. The purpose of the bill is to create transparency and prevent disparate outcomes for AI systems, and it would require companies to assess the impacts of the AI systems they use and sell. The bill describes the impact assessment in detail — which must be submitted to an oversight committee— and states that the assessment must address “the transparency and explainability of [an AI system] and the degree to which a consumer may contest, correct, or appeal a decision or opt out of such system or process”, which speaks directly to what AI practitioners refer to as “recourse”, or the ability of an individual to understand the outcome of an AI system and what they could do to change that outcome.

In 2019 the OPEN Government Data Act was passed into law, requiring that federal agencies maintain and publish their information online as open data. The data also must be cataloged on Data.gov, a public data repository created by the the US government. While this law only applies to public data, it demonstrates how policy can address transparency within the whole pipeline of an AI system, from the data to the algorithm to the system outcome.

There are also some industry-specific standards for transparency that could act as a model for future cross-industry regulations. Under the Equal Credit Opportunity Act, creditors who deny loan applicants must provide a specific reason for the denial. This includes denials made by AI systems. The explanations for a denial come from a standardized list of numeric reason codes, such as: “U4: Too many recently opened accounts with balances[https://www.fico.com/en/latest-thinking/solution-sheet/us-fico-score-reason-codes ].”

European Union

In 2019 the EU published a white paper titled “Ethics Guidelines for Trustworthy AI,” containing a legal framework that outlines ethical principles and legal obligations for EU member states to follow when deploying AI[https://ec.europa.eu/digital-single-market/en/news/ethics-guidelines-trustworthy-ai ]. While the white paper is non-binding, it lays out expectations on how member-states should regulate the transparency of AI systems: “… data, system and AI business models should be transparent. Traceability mechanisms can help achieving this. Moreover, AI systems and their decisions should be explained in a manner adapted to the stakeholder concerned. Humans need to be aware that they are interacting with an AI system, and must be informed of the system’s capabilities and limitations.”

Currently, the European Commission is reviewing the Artificial Intelligence Act[https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A52021PC0206 ], which would create a common legal framework for governing all types of AI used in all non-military sectors in Europe. The directive takes the position that AI systems pose a significant risk to the health, safety and fundamental rights of persons, and governs from that perspective. With respect to transparency, the directive delineates between non-high-risk and high-risk AI systems (neither of which are rigorously defined at this time). It states that for “non-high-risk AI systems, only very limited transparency obligations are imposed, for example in terms of the provision of information to flag the use of an AI system when interacting with humans.” Yet, for high-risk systems, “the requirements of high quality data, documentation and traceability, transparency, human oversight, accuracy and robustness, are strictly necessary to mitigate the risks to fundamental rights and safety posed by AI and that are not covered by other existing legal frameworks.” Notably, as in the Algorithmic Accountability Act in the United States, the document contains explicit text mentioning recourse (referred to as “redress”) for persons affected by AI systems.

The EU has also passed Regulation (EU) 2019/1150 that sets guidelines for the transparency of rankings for online search.[https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32019R1150 ] In practice, this means that online stores and search engines should be required to disclose the algorithmic parameters used to rank goods and services on their site. The regulation also states that explanations about rankings should contain redress mechanisms for individuals and businesses affected by the rankings.

Right to Explanation.

The Right to Explanation is a proposed fundamental human right that would guarantee individuals access to an explanation for any AI system decision that affects them. The Right to Explanation was written into the EU’s 2016 GDPR regulations, and reads as follows: “[the data subject should have] the right … to obtain an explanation of the decision reached.”[https://www.privacy-regulation.eu/en/r71.htm ] The legal meaning and obligation of the text has been debated heavily by legal scholars, who are unsure under which circumstances it applies, what constitutes an explanation, and how the right is applicable to different AI systems. The Right to Explanation is an example of how emerging AI technologies may “reveal” additional rights that need to be considered by lawmakers and legal experts.

The EU’s recently proposed Artificial Intelligence Act simultaneously reinforces the idea that explanations about AI systems are a human right, while slightly rolling back the Right to Explanation by acknowledging that there are both non-high-risk and high-risk AI systems. Discussions about the Right are likely to continue, and will be a central part of debates on regulating AI transparency. In fact, some local governing bodies have already taken steps to adopt the Right to Explanation. France passed the Digital Republic Act in 2016, which gives the Right to Explanation for individuals affected by an AI system in the public sector. Hungary also has a similar law.

Local

There has been significant movement on the regulation of specific forms of AI systems at local levels of government. In response to the well-documented biases of facial recognition software when identifying people of different races and ethnicities, Washington State signed Senate Bill 6820 into law in 2020, which prohibits the use of facial recognition software in surveillance and limits its use in criminal investigation.[<https://app.leg.wa.gov/billsummary?BillNumber=6280 Initiative=false Year=2019>] Detroit has also reacted to concerns about facial recognition, and its City Council approved legislation that mandates transparency and accountability for the procurement process of video and camera surveillance contracts used in the city.[https://www.detroitnews.com/story/news/local/detroit-city/2021/05/25/detroit-council-approves-ordinance-boost-transparency-surveillance-camera-contracts/7433185002/ ] The New York City Council recently regulated the use of AI systems in relation to employment decisions (Local Law 144 of 2021).[<https://legistar.council.nyc.gov/LegislationDetail.aspx?ID=4344524 GUID=B051915D-A9AC-451E-81F8-6596032FA3F9 Options=Advanced Search>] The bill requires that AI tools for hiring employees be subject to yearly bias audits. An additional requirement is to notify job seekers that they were screened by a tool, and to disclose to them what “qualifications or characteristics” were used by the tool as basis of decisions. Finally, in the Netherlands, the municipality of Rotterdam has created a Data-Driven Working program which has been critical of transparency surrounding the algorithms used for fraud detection.[https://nos.nl/artikel/2376810-rekenkamer-rotterdam-risico-op-vooringenomen-uitkomsten-door-gebruik-algoritmes ]

The Role of Technologists

The continuously evolving regulatory landscape of AI, combined with the limitations of existing regulation in providing clarity on how transparency should be implemented into AI systems, has left open questions concerning responsibilities for AI design and implementation. We argue that (1) practitioners should bear the bulk of the responsibility for designing and implementing compliant, transparent AI systems (2) it is in the best interest of practitioners to bear this responsibility. Researchers have also shown that there may be risks of only partially complying with AI regulations, and that fusll compliance is the best way forward. Technologists include AI practitioners, researchers, designers, programmers, and developers.

Practitioners have the right technical expertise. Transparency has been a central topic of AI research for the past decade, and is motivated beyond just regulatory compliance by ideas like making systems more efficient, debugging systems, and giving decision making agency to the data subjects (i.e., those affected by AI-assisted decisions) or to the users of AI systems (i.e., those making decisions with the help of AI). New technologies in transparent AI are being created at a fast pace, and there is no indication that the rapid innovation of explainable AI will slow any time soon, meaning that of all the stakeholders involved in the socio-technical environment of AI systems, technologists are the most likely to be aware of available tools for creating transparent AI systems. Furthermore, there are currently no objective measures for the quality of transparency in AI systems, and so technologists are necessary to discern the difference between a “good explanation” and a “bad explanation” about a system.

Practitioners are the least-cost avoiders. This idea is based on the principle of the least-cost avoider, which states that obligations and liabilities should be allocated entirely to the party with the lowest cost of care. AI practitioners are the least-cost avoiders because they are already equipped with the technical know-how for building and implementing transparency tools into AI systems, especially when compared to policymakers and the individuals affected by the outcome of the system. Notably, given the wide range of existing transparency tools, implementing the “bare minimum” is trivially easy for most technologists.

One argument practitioners give against building transparent systems is that they may be less accurate than highly complex, black-box systems. However, there has been a growing amount of evidence suggesting that building transparent systems actually results into little to no trade-off in the accuracy of AI systems. In other words: building transparent systems is not a Pareto-reducing constraint for practitioners.

Practitioners already bear the responsibility for implementing transparency into AI systems. A study interviewing AI practitioners found that using AI responsibly in their work is viewed as the practitioner’s burden, not the institutions for which they work. Practitioners noted that existing structures within institutions are often antithetical to the goals of responsible AI, and that it is up to them to push for structural change within that institution. Section sec-laws shows that AI regulation is converging on requiring transparent AI systems that offer meaningful explanations to stakeholders. Therefore, it is in the best interest of practitioners to continue the bottom-up approach of building transparent AI systems in the face of looming regulations.

A Stakeholder-First Approach to Designing Transparent ADS

Definitions

Technologists and AI researchers have not agreed on a definition of transparency for AI systems. Instead, a number of terms have been used, including explainability, interpretability, intelligibility, understandability, and comprehensibility. There is no consensus on the meaning of these terms and they are often defined differently by different authors or used interchangeably. Furthermore, transparency and its related terms cannot trivially be quantified or measured, and transparency for one stakeholder does not automatically imply the same for different stakeholders.

While having multiple definitions of transparency has been useful for distinguishing nuance in a research setting, it also poses a challenge for policy making. In contrast to technologists, policymakers favor definitions of transparency that are about human thought and behavior such as accountability or legibility. Table terms outlines terms related to transparency commonly used by policymakers versus those used by technologists.

Transparency. For the purposes of this paper, we choose to use only the term “transparency,” in the broadest possible sense, so that it encompasses all the definitions above. This is most similar to the way “explainability” is used by technologists.Here we use the definition adapted from work by Christoph Molnar and define transparency as “the degree to which a human can understand an AI system.”

Explanation. We use the term “explanation” to refer to an instantiation of transparency. For example, to ensure transparency for a system, a technologist may create an explanation about the data it uses.

Automated Decision Systems. The approach described in this paper applies to all Automated Decision Systems (ADS), which is any system that processes data to make decisions about people. This means that AI systems are a subset of ADS, but there are two key distinctions: (1) an ADS is underpinned by any algorithm and not just AI or machine learning, and (2) an ADS implies a context of use and some kind of impact. For a formal definition of ADS, see. Henceforth, we will use the term ADS.

Notably, while many regulations are written to specifically mention “AI systems”, all the ideas they contain about transparency could be applied to all ADS. It is likely that future regulations will focus broadly on ADS, as seen in NYC Local Law 144 of 2021 and France’s Digital Republic Act.

Running Example: Predicting Unemployment in Portugal

To make the discussion concrete, we use a running example of an ADS implemented in Portugal to try and prevent long-term unemployment (being unemployed for 12 months or more). The long-term unemployed are particularly vulnerable persons, and tend to earn less once they find new jobs, have poorer health and have children with worse academic performance as compared to those who had continuous employment. The Portuguese national employment agency, the Institute for Employment and Vocational Training (IEFP), uses an ADS to allocate unemployment resources to at-risk unemployed persons. The system is based on demographic data about the individual, including their age, unemployment length, and profession, along with other data on macroeconomic trends in Portugal.

The ADS is used by job counselors who work at the IEFP unemployment centers spread across Portugal. This interaction model, where an ML system makes a prediction and a human ultimately makes a final determination informed by the system’s predictions, is referred to as having a “human-in-the-loop” (HITL). Having a HITL is an increasingly common practice for implementing ADS. The ADS assigns unemployed persons as low, medium, or high risk for remaining unemployed, and then job counselors have the responsibility of assigning them to interventions such as re-skilling, resume building, or job search training.

This is a useful case study for three reasons: (1) people’s access to economic opportunity is at stake, and as a result, systems for predicting long-term unemployment are used widely around the world; (2) the ADS exists in a dynamic setting which includes several stakeholders, like unemployed persons, job counselors who act as the human-in-the-loop, policymakers who oversee the implementation of the tool, and the technologists who developed the tool; (3) lessons from this case about designing stakeholder-first transparent systems generalize well to other real-world uses cases of ADS.

The Approach

There are many purposes, goals, use-cases and methods for the transparency of ADS, which have been categorized in a number of taxonomies and frameworks. Theapproach we propose here has three subtle — yet important — differences from much of the existing work in this area: (1) our approach is stakeholder-first, capturing an emerging trend among researchers in this field to reject existing method or use-case driven approaches; (2) our approach is focused on improving the design of transparent ADS, rather than attempting to categorize the entire field of transparency; (3) our approach is aimed at designing ADS that comply with transparency regulations.

Our approach can be seen in Figure fig-taxonomy and is made up of the following components: stakeholders, goals, purpose, and methods. We describe each component in the remainder of this section, and explain how they apply to the running example.

A stakeholder-first approach for creating transparent ADS. The framework is made up of four components: stakeholders, goals, purpose, and methods. We recommend that transparency be thought of first by stakeholders, second by goals, before thirdly defining the purpose, and lastly choosing an appropriate method to serve said purpose. Using the framework is simple: starting at the top, one should consider each bubble in a component before moving onto the next component.fig-taxonomy

A stakeholder-first approach for creating transparent ADS. The framework is made up of four components: stakeholders, goals, purpose, and methods. We recommend that transparency be thought of first by stakeholders, second by goals, before thirdly defining the purpose, and lastly choosing an appropriate method to serve said purpose. Using the framework is simple: starting at the top, one should consider each bubble in a component before moving onto the next component.

Definitions and examples of stakeholder goals for the 6 categories of ADS transparency goals.

Table Label: tab-goals

Download PDF to view table

Stakeholders

Much of ADS transparency research is focused on creating novel and innovative transparency methods for algorithms, and then later trying to understand how these methods can be used to meet stakeholders needs. Counter to this rationale, we propose a starting point that focuses on ADS stakeholders: assuming algorithmic transparency is intended to improve the understanding of a human stakeholder, technologists designing transparent ADS must first consider the stakeholders of the system, before thinking about the system’s goals or the technical methods for creating transparency.

The existing literature and taxonomies on ADS transparency have identified a number of important stakeholders, which include technologists, policymakers, auditors, regulators, humans-in-the-loop, and those individuals affected by the output of the ADS.While there is some overlap in how these stakeholders may think about transparency, in general, there is no single approach to designing transparent systems for these disparate stakeholder groups, and each of them has their own goals and purposes for wanting to understand an ADS.In fact, even within a stakeholder group there may be variations on how they define meaningful transparency. Designers of ADS may also want to weight the needs of separate stakeholders differently. For example, it may be more meaningful to meet the transparency needs of affected individuals over AI managers or auditors.

Importantly, by staking transparency on the needs of stakeholders, technologists will be compliant with citizen-aware regulations like the Right to Explanation, and those that require audits of ADS.

Running example. In the ADS used by IEFP in Portugal, there are four main stakeholders: the technologists who developed the ADS, the policymakers who reviewed the ADS and passed laws for its implementation, the job counselors who use the system, and the affected individuals who are assessed for long-term unemployment.In the development of the AI, explanations were created to meet the varying goals of many of these stakeholders including practitioners, policymakers, and the job counselors.Unfortunately, and significantly, affected individuals were not considered.Had the practitioners adopted a robust stakeholder-first approach to designing transparent systems they could have better considered how to meet the goals of this key stakeholder group. For example, a person may want to appeal being predicted low risk because they feel they are high risk for long-term unemployment and need access to better interventions.

Goals.

There has been little consensus in the literature on how ADS goals should be classified. Some researchers have focused broadly, classifying the goals of ADS as evaluating, justifying, managing, improving, or learning about the outcome of an ADS.Others have defined goals more closely to what can be accomplished by known transparency methods, including building trust, establishing causality, and achieving reliability, fairness, and privacy.Amarasinghe et al. identified five main goals (designated as use-cases) of transparency specifically in a policy setting:model-debugging, trust and adoption, whether or not to intervene, improving intervention assignments, and for recourse. In this context, the term intervention refers to a policy action associated with the outcome of an ADS.

Notably, the goals of transparency are distinct from the purpose. The purpose addresses a context-specific aim of the ADS. For example, if an explanation is created for an ADS with the purpose of explaining to an individual why their loan was rejected, the goal may be to offer individual recourse against the rejection. This distinction is made clear in purpose.

For our stakeholder-first approach we make two changes to the existing body of research work. First, we require that the goal of transparent design must start with a stakeholder. Since all transparency elements of an ADS are intended for a human audience, defining a stakeholder is implicit in defining goals. Second, we have established 6 goal categories, which encompass those found in literature.These categories are validity, trust, learning and support, recourse, fairness and privacy, and are defined in Table tab-goals alongside concrete examples of how these goals may be implemented.

  • Validity.Validity refers to making sure that an AI is constructed correctly and is reasonable.It encompasses ideas like making sure the AI is reliable and robust.
  • Trust. In the context of an AI, trust refers knowing “how often a model is right” and “for which examples it is right”.Importantly, trust is related to the adoption of an ADS.
  • Learning and Support.Learning and support refers to when the goal of transparency in an ADS is to satisfy human curiosity, or increase understanding about how an AI is supporting a real-world recommendation.
  • Recourse.Recourse refers to allowing a stakeholder to take some action against the outcome of an AI.
  • Fairness.Fairness refers to making sure that an AI is fair based on some metric, and transparency can be used to find bias within a model.
  • Privacy.Privacy refers to making sure that an AI respects the data privacy of an individual.Transparency may be used to understand if privacy is being respected in an AI.

An important discussion surrounding goals are the justifications for pursuing them.For example, fairness and privacy goals may be justified for humanitarian reasons (they are perceived by the stakeholders as the “right thing to do”).Other justifications may be to prevent harm, like offering recourse to stakeholders against an outcome of an ADS, or for a reward, like an explanation that supports a doctor’s correct diagnosis.For reasons of scope we will not delve into the issue of goal justification in this paper.

Running example. In our case study, transparency is built into the ADS with the goal of offering learning and support to job counselors. The ADS generates explanations about what factors contribute to an individual being classified as low, medium, or high risk for long-term unemployment, which job counselors use to help make better treatment decision.Furthermore, the job counselor may also use the explanation to offer recommendations for recourse against a high risk score.

Purpose

Miller proposed that the purpose of transparency is to answer a “why” question, and gives the following example: In the context where a system is predicting if a credit loan is accepted or rejected, one may ask, “why was a particular loan rejected?” Liao et al. expanded on this significantly by creating a “question bank” which is a mapping from a taxonomy of technical transparency methodology to different types of user questions.Instead of just answering why questions, the works shows that transparency can be used to answer 10 categories of questions:questions about the input, output, and performance of the system, how, why, why not, what if, how to be that, how to still be this, and others. These questions have two important characteristics. First, they are context-specific and should address a direct transparency goal of the stakeholder. Second, and importantly for technologists, these questions can be mapped onto known methods for creating explanations, meaning that a well-defined purpose for transparency acts a bridge between the goals and methods.

Thoughtfully defining the goals and purpose of transparency in ADS is critical for technologists to be compliant with regulators. It is not sufficient to try and apply general, one-size-fits-all design like simply showing the features that were used by an ADS. For instance, both the proposed Algorithmic Accountability Act in the United States and the Artificial Intelligence Act in the European Union specifically mention that ADS should have transparency mechanisms that allow individuals to have recourse against a system outcome. Researchers have noted that feature-highlighting transparency lacks utility when there is a disconnect between the explanation and real-world actions. For instance, if someone is rejected for a loan and the reason for that decision is the person’s age, there is no action that they can effectively take for recourse against that decision.

Running example. In the long-term unemployment use case, there were two main purposes of transparency: to understand why an individual was assigned to a particular risk category, and to understand what could be done to help high risk individuals lower their chances of remaining long-term unemployed.

Methods.

Once the stakeholders, goals, and purposes for algorithmic transparency have been established, it is time for the technologist to pick the appropriate transparency method (somtimes called explainablity method). Over the past decade there has been significant work in transparent ADS research (sometimes called “explainable AI” research or XAI) on developing new methods for understanding opaque ADS.There are several existing taxonomies of these methods, which show that explanations can be classified on a number of attributes like the scope (local or global), intrinsic or post-hoc, data or model, model-agnostic or model-specific, surrogate or model behavior, and static or interactive. Furthermore, researchers have created a number of different tools to accomplish transparency in ADS.

In contrast to the complex classification of transparency methods by technologists, regulations have focused on two elements of ADS: (1) what aspect of the ADS pipeline is being explained (the data, algorithm, or outcome)?, and (2) what is the scope of the explanation (for one individual or the entire system)? Table tab-laws shows how different regulations speak to different combinations of pipeline and scope. In our stakeholder first-approach to transparency, we focus on these two main attributes. We will not discuss specific methods in detail, but for the convenience of technologists we have underlined them throughout this discussion.

How different laws regulate the aspects the ADS pipeline (the data, algorithm or outcome), and within what scope (local or global).

Table Label: tab-laws

Download PDF to view table

Data, algorithm, or outcome. Transparency methods have focused on generating explanations for three different “points in time” in an ADS pipeline:the data (pre-processing), the model/algorithm (in-processing, intrinsic), or the outcome (post-processing, post-hoc).Importantly, transparency is relevant for each part of the machine learning pipeline because issues likes bias can arise within each component.

Transparency techniques that focus on the pre-processing component of the pipeline, that is, on the data used to create an ADS, typically include descriptive statistics or data visualizations.

Data visualizations have proved useful for informing users and making complex information more accessible and digestible, and have even been found to have a powerful persuasive effect. Therefore, it is advisable to use data visualization if it can easilyaddress the purpose of an explanation. However, visualizations should be deployed thoughtfully, as they have the ability to be abused and can successfully misrepresent a message through techniques like exaggeration or understatement.

Techniques for creating in-processing or post-processing explanations call into question the important consideration of using explainable versus black-box algorithms when designing AI. The machine learning community accepts two classifications of models:those that are intrinsically transparent by their nature (sometimes called directly interpretable or white-box models), and those that are not (called black box models). Interpretable models, like linear regression, decision trees, or rules-based models, have intrinsic transparency mechanisms that offer algorithmic transparency, like the linear formula, the tree diagram, and the set of rules, respectively. There are also methods like select-regress-round that simplify black-box models into interpretable models that use a similar set of features.

As an important design consideration for technologists, researchers have studied the effect of the complexity of a model and how it impacts its ability to be understood by a stakeholder. A user study found that the understanding of a machine learning model is negatively correlated with it’s complexity, and found decision trees to be among the model types most understood by users. An additional, lower-level design consideration is that model complexity is not fixed to a particular model type, but rather to the way that the model is constructed. For example, a decision tree with 1,000 nodes will be understood far less well than a tree with only 3 or 5 nodes.

In contrast to in-process transparency, which is intrinsically built into a model or algorithm, post-hoc transparency aims to answer questions about a model or algorithm after is has already been created.Some of the most popular post-hoc methods are LIME, SHAP, SAGE, and QII. These methods are considered model-agnostic because they can be used to create explanations for any model, from linear models to random forests to neural networks. Some methods create a transparent surrogate model that mimics the behavior of a black-box model. For example, LIME creates a linear regression to approximate an underlying black-box model. More work needs to be done in this direction, but one promising study has shown that post-hoc explanations can actually improve the perceived trust in the outcome of an algorithm. However, post-hoc transparency methods have been shown to have two weaknesses that technologists should be aware of: (1) in many cases, these methods are at-best approximations of the black-box they are trying to explain, and (2) these methods may be vulnerable to adversarial attacks and exploitation.Some researchers have also called into question the utility of black-box models and post-hoc explanation methods altogether, and have cautioned against their use in real-world contexts like clinical settings.

Scope

. There are two levels at which a transparent explanation about an ADS can operate:it either explains its underlying algorithm fully, called a “global” explanation; or it explains how the algorithm operates on one specific instance, called a “local” explanation. Molnar further subdivides each of these levels into two sub-levels: global explanations can either be holistic (applying to an entire algorithm, which includes all of its features, and in the case of an ensemble algorithm, all of the component algorithms) or modular, meaning they explain on part of the holistic explanation and local explanations can either be applied to a single individual, or aggregated to provide local explanations for an entire group.

The scope of an explanation is highly relevant to the stakeholder and goals of an explanation, and is related to whether the stakeholder operates at a system or individual level. Researchers found that the scope of explanation can influence whether or not an individual thinks a model is fair. Policymakers and ADS compliance officers are more apt to be concerned with system level goals, like ensuring that the ADS is fair, respects privacy, and is valid overall, while humans-in-the-loop and those individuals affected by the outcome of an ADS are likely more interested in seeing local explanations to pertain to their specific cases. Technologists should consider both.

Naturally, there is considerable overlap between stakeholders’ scope needs (for example, an auditor may want to inspect a model globally and look at local cases), but generally, it is important which scope an explanation has.Therefore designers of ADS explanations should be thoughtful of how they select the scope of an explanation based on a stakeholder and their goals.

Running-example. In the IEFP use case, SHAP factors were given to job counselors to show the top factors influencing the score of a candidate both positively and negatively. The transparency provided by SHAP provided a local explanation about the outcome of the model. A bias audit was also conducted on the entire algorithm, and presented to policy officials within IEFP.

Overall, researchers found that the explanations improved the confidence of the decisions, but counter-intuitively, had a somewhat negative effect on the quality of those decisions.

Putting the Approach into Practice

The stakeholder-first approach describe in Section sec-approach is meant to act as a guide for technologists creating regulatory-compliant ADS. Putting this approach into practice is simple: starting at the first component in Figure fig-taxonomy (stakeholders), one should consider each bubble, before moving onto the next component and again considering each bubble. By the time one has finished worked their way through the figure, they should have considered all the possible stakeholders, goals, purposes, and methods of an ADS. An instantiation of the approach can be found throughout Section sec-approach in the running example of building an ADS that predicts the risk of long-term unemployment in Portugal.

It’s important to note that our proposed stakeholder-first approach is only a high-level tool for thinking about ADS transparency through the perspective of stakeholders and their needs. Beyond this approach, there are meaningful low-level steps that can be taken by technologists when it comes to actually implement transparency into ADS. One such step is the use of participatory design, where stakeholders are included directly in design conversations. In one promising study researchers used participatory design to successfully create better algorithmic explanations for users in the field of communal energy accounting.

Concluding Remarks

If there is to be a positive, ethical future for the use of AI systems, there needs to be stakeholder-driven design for creating transparency algorithms — and who better to lead this effort than technologists. Here we proposed a stakeholder-first approach that technologists can use to guide their design of transparent AI systems that are compliant with existing and proposed AI regulations. While there is still significant research that needs to be done in understanding how the transparency of AI systems can be most useful for stakeholders, and in the policy design of AI regulation, this paper aims to be a step in the right direction.

There are several important research steps that could be taken to extend this work. First, the stakeholder-first approach described here lays the foundation for creating a complete playbook to designing transparent systems. This playbook would be useful to a number of audiences including technologists, humans-in-the-loop, and policymakers. Second, a repository of examples and use cases of regulatory-compliant systems derived from this approach could be created, to act as a reference to technologists.

Bibliography

   1@article{rakova2020responsible,
   2  year = {2020},
   3  journal = {arXiv preprint arXiv:2006.12358},
   4  author = {Rakova, Bogdana and Yang, Jingying and Cramer, Henriette and Chowdhury, Rumman},
   5  title = {Where Responsible AI meets Reality: Practitioner Perspectives on Enablers for shifting Organizational Practices},
   6}
   7
   8@article{DBLP:journals/corr/abs-1906-11668,
   9  bibsource = {dblp computer science bibliography, https://dblp.org},
  10  biburl = {https://dblp.org/rec/journals/corr/abs-1906-11668.bib},
  11  timestamp = {Mon, 01 Jul 2019 13:00:07 +0200},
  12  eprint = {1906.11668},
  13  archiveprefix = {arXiv},
  14  url = {http://arxiv.org/abs/1906.11668},
  15  year = {2019},
  16  volume = {abs/1906.11668},
  17  journal = {CoRR},
  18  title = {Artificial Intelligence: the global landscape of ethics guidelines},
  19  author = {Anna Jobin and
  20Marcello Ienca and
  21Effy Vayena},
  22}
  23
  24@inproceedings{DBLP:conf/fat/BuolamwiniG18,
  25  bibsource = {dblp computer science bibliography, https://dblp.org},
  26  biburl = {https://dblp.org/rec/conf/fat/BuolamwiniG18.bib},
  27  timestamp = {Wed, 03 Apr 2019 18:17:20 +0200},
  28  url = {http://proceedings.mlr.press/v81/buolamwini18a.html},
  29  year = {2018},
  30  publisher = {{PMLR}},
  31  pages = {77--91},
  32  volume = {81},
  33  series = {Proceedings of Machine Learning Research},
  34  booktitle = {Conference on Fairness, Accountability and Transparency, {FAT} 2018,
  3523-24 February 2018, New York, NY, {USA}},
  36  title = {Gender Shades: Intersectional Accuracy Disparities in Commercial Gender
  37Classification},
  38  editor = {Sorelle A. Friedler and
  39Christo Wilson},
  40  author = {Joy Buolamwini and
  41Timnit Gebru},
  42}
  43
  44@article{angwin2016machine,
  45  year = {2016},
  46  journal = {See https://www. propublica. org/article/machine-bias-risk-assessments-in-criminal-sentencing},
  47  author = {Angwin, Julia and Larson, Jeff and Mattu, Surya and Kirchner, Lauren},
  48  title = {Machine bias. ProPublica},
  49}
  50
  51@article{KUZIEMSKI2020101976,
  52  abstract = {The rush to understand new socio-economic contexts created by the wide adoption of AI is justified by its far-ranging consequences, spanning almost every walk of life. Yet, the public sector's predicament is a tragic double bind: its obligations to protect citizens from potential algorithmic harms are at odds with the temptation to increase its own efficiency - or in other words - to govern algorithms, while governing by algorithms. Whether such dual role is even possible, has been a matter of debate, the challenge stemming from algorithms' intrinsic properties, that make them distinct from other digital solutions, long embraced by the governments, create externalities that rule-based programming lacks. As the pressures to deploy automated decision making systems in the public sector become prevalent, this paper aims to examine how the use of AI in the public sector in relation to existing data governance regimes and national regulatory practices can be intensifying existing power asymmetries. To this end, investigating the legal and policy instruments associated with the use of AI for strenghtening the immigration process control system in Canada; “optimising” the employment services” in Poland, and personalising the digital service experience in Finland, the paper advocates for the need of a common framework to evaluate the potential impact of the use of AI in the public sector. In this regard, it discusses the specific effects of automated decision support systems on public services and the growing expectations for governments to play a more prevalent role in the digital society and to ensure that the potential of technology is harnessed, while negative effects are controlled and possibly avoided. This is of particular importance in light of the current COVID-19 emergency crisis where AI and the underpinning regulatory framework of data ecosystems, have become crucial policy issues as more and more innovations are based on large scale data collections from digital devices, and the real-time accessibility of information and services, contact and relationships between institutions and citizens could strengthen – or undermine - trust in governance systems and democracy.},
  53  keywords = {Artificial intelligence, Public sector innovation, Automated decision making, Algorithmic accountability},
  54  author = {Maciej Kuziemski and Gianluca Misuraca},
  55  url = {http://www.sciencedirect.com/science/article/pii/S0308596120300689},
  56  doi = {https://doi.org/10.1016/j.telpol.2020.101976},
  57  issn = {0308-5961},
  58  note = {Artificial intelligence, economy and society},
  59  year = {2020},
  60  pages = {101976},
  61  number = {6},
  62  volume = {44},
  63  journal = {Telecommunications Policy},
  64  title = {AI governance in the public sector: Three tales from the frontiers of automated decision-making in democratic settings},
  65}
  66
  67@article{DBLP:journals/corr/abs-2001-09734,
  68  bibsource = {dblp computer science bibliography, https://dblp.org},
  69  biburl = {https://dblp.org/rec/journals/corr/abs-2001-09734.bib},
  70  timestamp = {Thu, 30 Jan 2020 18:46:36 +0100},
  71  eprint = {2001.09734},
  72  archiveprefix = {arXiv},
  73  url = {https://arxiv.org/abs/2001.09734},
  74  year = {2020},
  75  volume = {abs/2001.09734},
  76  journal = {CoRR},
  77  title = {One Explanation Does Not Fit All: The Promise of Interactive Explanations
  78for Machine Learning Transparency},
  79  author = {Kacper Sokol and
  80Peter A. Flach},
  81}
  82
  83@inproceedings{DBLP:conf/fat/SelbstP18,
  84  bibsource = {dblp computer science bibliography, https://dblp.org},
  85  biburl = {https://dblp.org/rec/conf/fat/SelbstP18.bib},
  86  timestamp = {Wed, 03 Apr 2019 18:17:20 +0200},
  87  url = {http://proceedings.mlr.press/v81/selbst18a.html},
  88  year = {2018},
  89  publisher = {{PMLR}},
  90  pages = {48},
  91  volume = {81},
  92  series = {Proceedings of Machine Learning Research},
  93  booktitle = {Conference on Fairness, Accountability and Transparency, {FAT} 2018,
  9423-24 February 2018, New York, NY, {USA}},
  95  title = {"Meaningful Information" and the Right to Explanation},
  96  editor = {Sorelle A. Friedler and
  97Christo Wilson},
  98  author = {Andrew Selbst and
  99Julia Powles},
 100}
 101
 102@article{reed2018,
 103  url = {http://dx.doi.org/10.1098/rsta.2017.0360},
 104  year = {2018},
 105  volume = {376},
 106  journal = {Trans. R. Soc. A},
 107  title = {How should we regulate artificial intelligence?},
 108  author = {C. Reed},
 109}
 110
 111@article{1894_showreel,
 112  author+an = {1=self},
 113  keywords = {public},
 114  day = {12},
 115  month = {11},
 116  year = {2020},
 117  url = {https://dataresponsibly.github.io/documents/Bill1894Showreel.pdf},
 118  journal = {NYU Center for Responsible AI},
 119  title = {Public Engagement Showreel, Int 1894},
 120  author = {Julia Stoyanovich and Steven Kuyan and Meghan McDermott and Maria Grillo and Mona Sloane},
 121}
 122
 123@article{DBLP:journals/pvldb/StoyanovichHJ20,
 124  addendum = {paper accompanying keynote presentation at the 46th International Conference on Very Large Data Bases, {VLDB}},
 125  author+an = {1=self},
 126  keywords = {invited,selected},
 127  doi = {10.14778/3415478.3415570},
 128  pages = {3474-3489},
 129  number = {12},
 130  volume = {13},
 131  year = {2020},
 132  journal = {PVLDB},
 133  title = {Responsible Data Management},
 134  author = {Julia Stoyanovich and Bill Howe and H.V. Jagadish},
 135}
 136
 137@inproceedings{DBLP:conf/aies/KrafftYKHB20,
 138  bibsource = {dblp computer science bibliography, https://dblp.org},
 139  biburl = {https://dblp.org/rec/conf/aies/KrafftYKHB20.bib},
 140  timestamp = {Fri, 08 Jan 2021 08:52:12 +0100},
 141  doi = {10.1145/3375627.3375835},
 142  url = {https://doi.org/10.1145/3375627.3375835},
 143  year = {2020},
 144  publisher = {{ACM}},
 145  pages = {72--78},
 146  booktitle = {{AIES} '20: {AAAI/ACM} Conference on AI, Ethics, and Society, New
 147York, NY, USA, February 7-8, 2020},
 148  title = {Defining {AI} in Policy versus Practice},
 149  editor = {Annette N. Markham and
 150Julia Powles and
 151Toby Walsh and
 152Anne L. Washington},
 153  author = {P. M. Krafft and
 154Meg Young and
 155Michael A. Katell and
 156Karen Huang and
 157Ghislain Bugingo},
 158}
 159
 160@book{molnar2019,
 161  subtitle = {A Guide for Making Black Box Models Explainable},
 162  year = {2019},
 163  note = {\url{https://christophm.github.io/interpretable-ml-book/}},
 164  author = {Christoph Molnar},
 165  title = {Interpretable Machine Learning},
 166}
 167
 168@article{DBLP:journals/corr/abs-2010-14374,
 169  bibsource = {dblp computer science bibliography, https://dblp.org},
 170  biburl = {https://dblp.org/rec/journals/corr/abs-2010-14374.bib},
 171  timestamp = {Mon, 02 Nov 2020 18:17:09 +0100},
 172  eprint = {2010.14374},
 173  archiveprefix = {arXiv},
 174  url = {https://arxiv.org/abs/2010.14374},
 175  year = {2020},
 176  volume = {abs/2010.14374},
 177  journal = {CoRR},
 178  title = {Explainable Machine Learning for Public Policy: Use Cases, Gaps, and
 179Research Directions},
 180  author = {Kasun Amarasinghe and
 181Kit T. Rodolfa and
 182Hemank Lamba and
 183Rayid Ghani},
 184}
 185
 186@article{meske,
 187  doi = {10.1080/10580530.2020.1849465},
 188  journal = {Information Systems Management},
 189  title = {Explainable Artificial Intelligence: Objectives, Stakeholders and Future Research Opportunities},
 190  pages = {},
 191  month = {12},
 192  year = {2020},
 193  author = {Meske, Christian and Bunde, Enrico and Schneider, Johannes and Gersch, Martin},
 194}
 195
 196@inproceedings{DBLP:conf/chi/LiaoGM20,
 197  bibsource = {dblp computer science bibliography, https://dblp.org},
 198  biburl = {https://dblp.org/rec/conf/chi/LiaoGM20.bib},
 199  timestamp = {Fri, 25 Dec 2020 01:14:19 +0100},
 200  doi = {10.1145/3313831.3376590},
 201  url = {https://doi.org/10.1145/3313831.3376590},
 202  year = {2020},
 203  publisher = {{ACM}},
 204  pages = {1--15},
 205  booktitle = {{CHI} '20: {CHI} Conference on Human Factors in Computing Systems,
 206Honolulu, HI, USA, April 25-30, 2020},
 207  title = {Questioning the {AI:} Informing Design Practices for Explainable {AI}
 208User Experiences},
 209  editor = {Regina Bernhaupt and
 210Florian 'Floyd' Mueller and
 211David Verweij and
 212Josh Andres and
 213Joanna McGrenere and
 214Andy Cockburn and
 215Ignacio Avellino and
 216Alix Goguey and
 217Pernille Bj{\o}n and
 218Shengdong Zhao and
 219Briane Paul Samson and
 220Rafal Kocielnik},
 221  author = {Q. Vera Liao and
 222Daniel M. Gruen and
 223Sarah Miller},
 224}
 225
 226@article{DBLP:journals/corr/Miller17a,
 227  bibsource = {dblp computer science bibliography, https://dblp.org},
 228  biburl = {https://dblp.org/rec/journals/corr/Miller17a.bib},
 229  timestamp = {Mon, 13 Aug 2018 16:47:22 +0200},
 230  eprint = {1706.07269},
 231  archiveprefix = {arXiv},
 232  url = {http://arxiv.org/abs/1706.07269},
 233  year = {2017},
 234  volume = {abs/1706.07269},
 235  journal = {CoRR},
 236  title = {Explanation in Artificial Intelligence: Insights from the Social Sciences},
 237  author = {Tim Miller},
 238}
 239
 240@article{DBLP:journals/jmlr/AryaBCDHHHLLMMP20,
 241  bibsource = {dblp computer science bibliography, https://dblp.org},
 242  biburl = {https://dblp.org/rec/journals/jmlr/AryaBCDHHHLLMMP20.bib},
 243  timestamp = {Wed, 18 Nov 2020 15:58:12 +0100},
 244  url = {http://jmlr.org/papers/v21/19-1035.html},
 245  year = {2020},
 246  pages = {130:1--130:6},
 247  volume = {21},
 248  journal = {J. Mach. Learn. Res.},
 249  title = {{AI} Explainability 360: An Extensible Toolkit for Understanding Data
 250and Machine Learning Models},
 251  author = {Vijay Arya and
 252Rachel K. E. Bellamy and
 253Pin{-}Yu Chen and
 254Amit Dhurandhar and
 255Michael Hind and
 256Samuel C. Hoffman and
 257Stephanie Houde and
 258Q. Vera Liao and
 259Ronny Luss and
 260Aleksandra Mojsilovic and
 261Sami Mourad and
 262Pablo Pedemonte and
 263Ramya Raghavendra and
 264John T. Richards and
 265Prasanna Sattigeri and
 266Karthikeyan Shanmugam and
 267Moninder Singh and
 268Kush R. Varshney and
 269Dennis Wei and
 270Yunfeng Zhang},
 271}
 272
 273@inproceedings{yang2020fairness,
 274  year = {2020},
 275  booktitle = {HILDA workshop at SIGMOD},
 276  author = {Yang, Ke and Huang, Biao and Stoyanovich, Julia and Schelter, Sebastian},
 277  title = {Fairness-aware instrumentation of preprocessing pipelines for machine learning},
 278}
 279
 280@article{zhang2019should,
 281  year = {2019},
 282  journal = {arXiv preprint arXiv:1904.12991},
 283  author = {Zhang, Yujia and Song, Kuangyan and Sun, Yiming and Tan, Sarah and Udell, Madeleine},
 284  title = {" Why Should You Trust My Explanation?" Understanding Uncertainty in LIME Explanations},
 285}
 286
 287@inproceedings{DBLP:conf/aies/SlackHJSL20,
 288  bibsource = {dblp computer science bibliography, https://dblp.org},
 289  biburl = {https://dblp.org/rec/conf/aies/SlackHJSL20.bib},
 290  timestamp = {Mon, 24 Feb 2020 12:40:26 +0100},
 291  doi = {10.1145/3375627.3375830},
 292  url = {https://doi.org/10.1145/3375627.3375830},
 293  year = {2020},
 294  publisher = {{ACM}},
 295  pages = {180--186},
 296  booktitle = {{AIES} '20: {AAAI/ACM} Conference on AI, Ethics, and Society, New
 297York, NY, USA, February 7-8, 2020},
 298  title = {Fooling {LIME} and {SHAP:} Adversarial Attacks on Post hoc Explanation
 299Methods},
 300  editor = {Annette N. Markham and
 301Julia Powles and
 302Toby Walsh and
 303Anne L. Washington},
 304  author = {Dylan Slack and
 305Sophie Hilgard and
 306Emily Jia and
 307Sameer Singh and
 308Himabindu Lakkaraju},
 309}
 310
 311@article{platform2018tackling,
 312  year = {2018},
 313  author = {Anette Scoppetta and Arthur Buckenleib.},
 314  title = {Tackling Long-Term Unemployment through Risk Profiling and Outreach},
 315}
 316
 317@article{loxha2014profiling,
 318  publisher = {World Bank Group, Washington, DC},
 319  year = {2014},
 320  author = {Loxha, Artan and Morgandi, Matteo},
 321  title = {Profiling the Unemployed: a review of OECD experiences and implications for emerging economies},
 322}
 323
 324@inproceedings{riipinen2011risk,
 325  year = {2011},
 326  pages = {8--9},
 327  booktitle = {Power Point presentation at the European Commission’s “PES to PES Dialogue Dissemination Conference,” Brussels, September},
 328  author = {Riipinen, T},
 329  title = {Risk profiling of long-term unemployment in Finland},
 330}
 331
 332@article{caswell2010unemployed,
 333  publisher = {Sage Publications Sage UK: London, England},
 334  year = {2010},
 335  pages = {384--404},
 336  number = {3},
 337  volume = {30},
 338  journal = {Critical Social Policy},
 339  author = {Caswell, Dorte and Marston, Greg and Larsen, J{\o}rgen Elm},
 340  title = {Unemployed citizen or ‘at risk’client? Classification systems and employment services in Denmark and Australia},
 341}
 342
 343@book{matty2013predicting,
 344  publisher = {Corporate Document Services},
 345  year = {2013},
 346  author = {Matty, Simon},
 347  title = {Predicting Likelihood of Long-term Unemployment: The Development of a UK Jobseekers' Classification Instrument},
 348}
 349
 350@article{sztandar2018changing,
 351  year = {2018},
 352  number = {2},
 353  volume = {16},
 354  journal = {Social Work \& Society},
 355  author = {Sztandar-Sztanderska, Karolina and Zielenska, Marianna},
 356  title = {Changing social citizenship through information technology},
 357}
 358
 359@article{raso2017displacement,
 360  publisher = {Cambridge University Press},
 361  year = {2017},
 362  pages = {75--95},
 363  number = {1},
 364  volume = {32},
 365  journal = {Canadian Journal of Law and Society},
 366  author = {Raso, Jennifer},
 367  title = {Displacement as regulation: New regulatory technologies and front-line decision-making in Ontario works},
 368}
 369
 370@article{wagner2019liable,
 371  publisher = {Wiley Online Library},
 372  year = {2019},
 373  pages = {104--122},
 374  number = {1},
 375  volume = {11},
 376  journal = {Policy \& Internet},
 377  author = {Wagner, Ben},
 378  title = {Liable, but not in control? Ensuring meaningful human agency in automated decision-making systems},
 379}
 380
 381@article{gillingham2019can,
 382  publisher = {Wiley Online Library},
 383  year = {2019},
 384  pages = {114--126},
 385  number = {2},
 386  volume = {28},
 387  journal = {Child abuse review},
 388  author = {Gillingham, Philip},
 389  title = {Can predictive algorithms assist decision-making in social work with children and families?},
 390}
 391
 392@article{gill2020policy,
 393  publisher = {Duke Global Working Paper Series},
 394  year = {2020},
 395  author = {Gill, Indermit S},
 396  title = {Policy Approaches to Artificial Intelligence Based Technologies in China, European Union and the United States},
 397}
 398
 399@article{narayanan2018humans,
 400  year = {2018},
 401  journal = {arXiv preprint arXiv:1802.00682},
 402  author = {Narayanan, Menaka and Chen, Emily and He, Jeffrey and Kim, Been and Gershman, Sam and Doshi-Velez, Finale},
 403  title = {How do humans understand explanations from machine learning systems? an evaluation of the human-interpretability of explanation},
 404}
 405
 406@article{unicri,
 407  url = {http://www.unicri.it/index.php/topics/ai_robotics},
 408  year = {2020},
 409  author = {UNICRI},
 410  title = {Towards Responsible Artificial Intelligence Innovation},
 411}
 412
 413@article{DBLP:journals/corr/abs-1909-03567,
 414  bibsource = {dblp computer science bibliography, https://dblp.org},
 415  biburl = {https://dblp.org/rec/journals/corr/abs-1909-03567.bib},
 416  timestamp = {Tue, 17 Sep 2019 11:23:44 +0200},
 417  eprint = {1909.03567},
 418  archiveprefix = {arXiv},
 419  url = {http://arxiv.org/abs/1909.03567},
 420  year = {2019},
 421  volume = {abs/1909.03567},
 422  journal = {CoRR},
 423  title = {What You See Is What You Get? The Impact of Representation Criteria
 424on Human Bias in Hiring},
 425  author = {Andi Peng and
 426Besmira Nushi and
 427Emre Kiciman and
 428Kori Inkpen and
 429Siddharth Suri and
 430Ece Kamar},
 431}
 432
 433@inproceedings{DBLP:conf/softcomp/BekriKH19,
 434  bibsource = {dblp computer science bibliography, https://dblp.org},
 435  biburl = {https://dblp.org/rec/conf/softcomp/BekriKH19.bib},
 436  timestamp = {Tue, 29 Dec 2020 18:31:06 +0100},
 437  doi = {10.1007/978-3-030-20055-8\_4},
 438  url = {https://doi.org/10.1007/978-3-030-20055-8\_4},
 439  year = {2019},
 440  publisher = {Springer},
 441  pages = {35--46},
 442  volume = {950},
 443  series = {Advances in Intelligent Systems and Computing},
 444  booktitle = {14th International Conference on Soft Computing Models in Industrial
 445and Environmental Applications {(SOCO} 2019) - Seville, Spain, May
 44613-15, 2019, Proceedings},
 447  title = {A Study on Trust in Black Box Models and Post-hoc Explanations},
 448  editor = {Francisco Mart{\'{\i}}nez{-}{\'{A}}lvarez and
 449Alicia Troncoso Lora and
 450Jos{\'{e}} Ant{\'{o}}nio S{\'{a}}ez Mu{\~{n}}oz and
 451H{\'{e}}ctor Quinti{\'{a}}n and
 452Emilio Corchado},
 453  author = {Nadia El Bekri and
 454Jasmin Kling and
 455Marco F. Huber},
 456}
 457
 458@article{lipton2018mythos,
 459  publisher = {ACM New York, NY, USA},
 460  year = {2018},
 461  pages = {31--57},
 462  number = {3},
 463  volume = {16},
 464  journal = {Queue},
 465  author = {Lipton, Zachary C},
 466  title = {The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery.},
 467}
 468
 469@article{doshi2017towards,
 470  year = {2017},
 471  journal = {arXiv preprint arXiv:1702.08608},
 472  author = {Doshi-Velez, Finale and Kim, Been},
 473  title = {Towards a rigorous science of interpretable machine learning},
 474}
 475
 476@inproceedings{ventocilla2018towards,
 477  year = {2018},
 478  pages = {151--157},
 479  booktitle = {XAI Workshop on Explainable Artificial Intelligence},
 480  author = {Ventocilla, Elio and Helldin, Tove and Riveiro, Maria and Bae, Juhee and Boeva, Veselka and Falkman, G{\"o}ran and Lavesson, Niklas},
 481  title = {Towards a taxonomy for interpretable and interactive machine learning},
 482}
 483
 484@article{stoyanovich2016revealing,
 485  year = {2016},
 486  journal = {Freedom to Tinker (August 5 2016)},
 487  author = {Stoyanovich, Julia and Goodman, Ellen P},
 488  title = {Revealing algorithmic rankers},
 489}
 490
 491@article{DBLP:journals/internet/GasserA17,
 492  bibsource = {dblp computer science bibliography, https://dblp.org},
 493  biburl = {https://dblp.org/rec/journals/internet/GasserA17.bib},
 494  timestamp = {Mon, 26 Oct 2020 09:03:54 +0100},
 495  doi = {10.1109/MIC.2017.4180835},
 496  url = {https://doi.org/10.1109/MIC.2017.4180835},
 497  year = {2017},
 498  pages = {58--62},
 499  number = {6},
 500  volume = {21},
 501  journal = {{IEEE} Internet Comput.},
 502  title = {A Layered Model for {AI} Governance},
 503  author = {Urs Gasser and
 504Virg{\'{\i}}lio A. F. Almeida},
 505}
 506
 507@article{wachter2017transparent,
 508  year = {2017},
 509  author = {Wachter, Sandra and Mittelstadt, Brent and Floridi, Luciano},
 510  title = {Transparent, explainable, and accountable AI for robotics},
 511}
 512
 513@article{gosiewska2019not,
 514  year = {2019},
 515  journal = {arXiv preprint arXiv:1903.11420},
 516  author = {Gosiewska, Alicja and Biecek, Przemyslaw},
 517  title = {Do not trust additive explanations},
 518}
 519
 520@article{rudin2019stop,
 521  publisher = {Nature Publishing Group},
 522  year = {2019},
 523  pages = {206--215},
 524  number = {5},
 525  volume = {1},
 526  journal = {Nature Machine Intelligence},
 527  author = {Rudin, Cynthia},
 528  title = {Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead},
 529}
 530
 531@article{DBLP:journals/corr/abs-2101-09429,
 532  bibsource = {dblp computer science bibliography, https://dblp.org},
 533  biburl = {https://dblp.org/rec/journals/corr/abs-2101-09429.bib},
 534  timestamp = {Sat, 30 Jan 2021 18:02:51 +0100},
 535  eprint = {2101.09429},
 536  archiveprefix = {arXiv},
 537  url = {https://arxiv.org/abs/2101.09429},
 538  year = {2021},
 539  volume = {abs/2101.09429},
 540  journal = {CoRR},
 541  title = {Explainable Artificial Intelligence Approaches: {A} Survey},
 542  author = {Sheikh Rabiul Islam and
 543William Eberle and
 544Sheikh Khaled Ghafoor and
 545Mohiuddin Ahmed},
 546}
 547
 548@inproceedings{DBLP:conf/scai/AllahyariL11,
 549  bibsource = {dblp computer science bibliography, https://dblp.org},
 550  biburl = {https://dblp.org/rec/conf/scai/AllahyariL11.bib},
 551  timestamp = {Fri, 19 May 2017 01:25:16 +0200},
 552  doi = {10.3233/978-1-60750-754-3-11},
 553  url = {https://doi.org/10.3233/978-1-60750-754-3-11},
 554  year = {2011},
 555  publisher = {{IOS} Press},
 556  pages = {11--19},
 557  volume = {227},
 558  series = {Frontiers in Artificial Intelligence and Applications},
 559  booktitle = {Eleventh Scandinavian Conference on Artificial Intelligence, {SCAI}
 5602011, Trondheim, Norway, May 24th - 26th, 2011},
 561  title = {User-oriented Assessment of Classification Model Understandability},
 562  editor = {Anders Kofod{-}Petersen and
 563Fredrik Heintz and
 564Helge Langseth},
 565  author = {Hiva Allahyari and
 566Niklas Lavesson},
 567}
 568
 569@misc{rodolfa2020machine,
 570  primaryclass = {cs.LG},
 571  archiveprefix = {arXiv},
 572  eprint = {2012.02972},
 573  year = {2020},
 574  author = {Kit T. Rodolfa and Hemank Lamba and Rayid Ghani},
 575  title = {Machine learning for public policy: Do we need to sacrifice accuracy to make models fair?},
 576}
 577
 578@article{DBLP:journals/debu/StoyanovichH19,
 579  bibsource = {dblp computer science bibliography, https://dblp.org},
 580  biburl = {https://dblp.org/rec/journals/debu/StoyanovichH19.bib},
 581  timestamp = {Tue, 10 Mar 2020 16:23:50 +0100},
 582  url = {http://sites.computer.org/debull/A19sept/p13.pdf},
 583  year = {2019},
 584  pages = {13--23},
 585  number = {3},
 586  volume = {42},
 587  journal = {{IEEE} Data Eng. Bull.},
 588  title = {Nutritional Labels for Data and Models},
 589  author = {Julia Stoyanovich and
 590Bill Howe},
 591}
 592
 593@article{campos2011nutrition,
 594  publisher = {Cambridge University Press},
 595  year = {2011},
 596  pages = {1496--1506},
 597  number = {8},
 598  volume = {14},
 599  journal = {Public health nutrition},
 600  author = {Campos, Sarah and Doxey, Juliana and Hammond, David},
 601  title = {Nutrition labels on pre-packaged foods: a systematic review},
 602}
 603
 604@inproceedings{ribeiro2016should,
 605  year = {2016},
 606  pages = {1135--1144},
 607  booktitle = {Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
 608  author = {Ribeiro, Marco Tulio and Singh, Sameer and Guestrin, Carlos},
 609  title = {" Why should i trust you?" Explaining the predictions of any classifier},
 610}
 611
 612@inproceedings{datta2016algorithmic,
 613  organization = {IEEE},
 614  year = {2016},
 615  pages = {598--617},
 616  booktitle = {2016 IEEE symposium on security and privacy (SP)},
 617  author = {Datta, Anupam and Sen, Shayak and Zick, Yair},
 618  title = {Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems},
 619}
 620
 621@inproceedings{DBLP:conf/nips/LundbergL17,
 622  bibsource = {dblp computer science bibliography, https://dblp.org},
 623  biburl = {https://dblp.org/rec/conf/nips/LundbergL17.bib},
 624  timestamp = {Thu, 21 Jan 2021 15:15:21 +0100},
 625  url = {https://proceedings.neurips.cc/paper/2017/hash/8a20a8621978632d76c43dfd28b67767-Abstract.html},
 626  year = {2017},
 627  pages = {4765--4774},
 628  booktitle = {Advances in Neural Information Processing Systems 30: Annual Conference
 629on Neural Information Processing Systems 2017, December 4-9, 2017,
 630Long Beach, CA, {USA}},
 631  title = {A Unified Approach to Interpreting Model Predictions},
 632  editor = {Isabelle Guyon and
 633Ulrike von Luxburg and
 634Samy Bengio and
 635Hanna M. Wallach and
 636Rob Fergus and
 637S. V. N. Vishwanathan and
 638Roman Garnett},
 639  author = {Scott M. Lundberg and
 640Su{-}In Lee},
 641}
 642
 643@inproceedings{DBLP:conf/fat/SokolF20,
 644  bibsource = {dblp computer science bibliography, https://dblp.org},
 645  biburl = {https://dblp.org/rec/conf/fat/SokolF20.bib},
 646  timestamp = {Fri, 24 Jan 2020 19:41:57 +0100},
 647  doi = {10.1145/3351095.3372870},
 648  url = {https://doi.org/10.1145/3351095.3372870},
 649  year = {2020},
 650  publisher = {{ACM}},
 651  pages = {56--67},
 652  booktitle = {FAT* '20: Conference on Fairness, Accountability, and Transparency,
 653Barcelona, Spain, January 27-30, 2020},
 654  title = {Explainability fact sheets: a framework for systematic assessment
 655of explainable approaches},
 656  editor = {Mireille Hildebrandt and
 657Carlos Castillo and
 658Elisa Celis and
 659Salvatore Ruggieri and
 660Linnet Taylor and
 661Gabriela Zanfir{-}Fortuna},
 662  author = {Kacper Sokol and
 663Peter A. Flach},
 664}
 665
 666@inproceedings{DBLP:conf/chi/HohmanHCDD19,
 667  bibsource = {dblp computer science bibliography, https://dblp.org},
 668  biburl = {https://dblp.org/rec/conf/chi/HohmanHCDD19.bib},
 669  timestamp = {Fri, 24 Jan 2020 16:59:38 +0100},
 670  doi = {10.1145/3290605.3300809},
 671  url = {https://doi.org/10.1145/3290605.3300809},
 672  year = {2019},
 673  publisher = {{ACM}},
 674  pages = {579},
 675  booktitle = {Proceedings of the 2019 {CHI} Conference on Human Factors in Computing
 676Systems, {CHI} 2019, Glasgow, Scotland, UK, May 04-09, 2019},
 677  title = {Gamut: {A} Design Probe to Understand How Data Scientists Understand
 678Machine Learning Models},
 679  editor = {Stephen A. Brewster and
 680Geraldine Fitzpatrick and
 681Anna L. Cox and
 682Vassilis Kostakos},
 683  author = {Fred Hohman and
 684Andrew Head and
 685Rich Caruana and
 686Robert DeLine and
 687Steven Mark Drucker},
 688}
 689
 690@article{gunaratne2017using,
 691  publisher = {Wiley Online Library},
 692  year = {2017},
 693  pages = {1836--1849},
 694  number = {8},
 695  volume = {68},
 696  journal = {Journal of the Association for Information Science and Technology},
 697  author = {Gunaratne, Junius and Nov, Oded},
 698  title = {Using interactive “Nutrition labels” for financial products to assist decision making under uncertainty},
 699}
 700
 701@article{zejnilovic2020machine,
 702  year = {2020},
 703  journal = {Available at SSRN 3715529},
 704  author = {Zejnilovic, Leid and Lavado, Susana and Soares, Carlos and Rituerto de Troya, {\'I}{\~n}igo and Bell, Andrew and Ghani, Rayid},
 705  title = {Machine Learning Informed Decision-Making: A Field Intervention in Public Employment Service},
 706}
 707
 708@article{zejnilovic2020algorithmic,
 709  publisher = {University of California Press},
 710  year = {2020},
 711  number = {1},
 712  volume = {1},
 713  journal = {Global Perspectives},
 714  author = {Zejnilovi{\'c}, Leid and Lavado, Susana and Mart{\'\i}nez de Rituerto de Troya, {\'I}{\~n}igo and Sim, Samantha and Bell, Andrew},
 715  title = {Algorithmic Long-Term Unemployment Risk Assessment in Use: Counselors’ Perceptions and Use Practices},
 716}
 717
 718@article{doshi2017accountability,
 719  year = {2017},
 720  journal = {arXiv preprint arXiv:1711.01134},
 721  author = {Doshi-Velez, Finale and Kortz, Mason and Budish, Ryan and Bavitz, Chris and Gershman, Sam and O'Brien, David and Scott, Kate and Schieber, Stuart and Waldo, James and Weinberger, David and others},
 722  title = {Accountability of AI under the law: The role of explanation},
 723}
 724
 725@article{DBLP:journals/corr/abs-2012-01805,
 726  bibsource = {dblp computer science bibliography, https://dblp.org},
 727  biburl = {https://dblp.org/rec/journals/corr/abs-2012-01805.bib},
 728  timestamp = {Fri, 04 Dec 2020 12:07:23 +0100},
 729  eprint = {2012.01805},
 730  archiveprefix = {arXiv},
 731  url = {https://arxiv.org/abs/2012.01805},
 732  year = {2020},
 733  volume = {abs/2012.01805},
 734  journal = {CoRR},
 735  title = {Interpretability and Explainability: {A} Machine Learning Zoo Mini-tour},
 736  author = {Ricards Marcinkevics and
 737Julia E. Vogt},
 738}
 739
 740@article{wilson2021building,
 741  year = {2021},
 742  author = {Wilson, Christo and Ghosh, Avijit and Jiang, Shan and Mislove, Alan and Baker, Lewis and Szary, Janelle and Trindel, Kelly and Polli, Frida},
 743  title = {Building and Auditing Fair Algorithms: A Case Study in Candidate Screening},
 744}
 745
 746@article{meyers2007street,
 747  publisher = {sage Publications London, UK},
 748  year = {2007},
 749  pages = {153--163},
 750  journal = {The handbook of public administration},
 751  author = {Meyers, Marcia K and Vorsanger, Susan and Peters, B Guy and Pierre, Jon},
 752  title = {Street-level bureaucrats and the implementation of public policy},
 753}
 754
 755@article{tal2016blinded,
 756  publisher = {SAGE Publications Sage UK: London, England},
 757  year = {2016},
 758  pages = {117--125},
 759  number = {1},
 760  volume = {25},
 761  journal = {Public Understanding of Science},
 762  author = {Tal, Aner and Wansink, Brian},
 763  title = {Blinded with science: Trivial graphs and formulas increase ad persuasiveness and belief in product efficacy},
 764}
 765
 766@inproceedings{pandey2015deceptive,
 767  year = {2015},
 768  pages = {1469--1478},
 769  booktitle = {Proceedings of the 33rd annual acm conference on human factors in computing systems},
 770  author = {Pandey, Anshul Vikram and Rall, Katharina and Satterthwaite, Margaret L and Nov, Oded and Bertini, Enrico},
 771  title = {How deceptive are deceptive visualizations? An empirical analysis of common distortion techniques},
 772}
 773
 774@article{DBLP:journals/tvcg/PandeyMNSB14,
 775  bibsource = {dblp computer science bibliography, https://dblp.org},
 776  biburl = {https://dblp.org/rec/journals/tvcg/PandeyMNSB14.bib},
 777  timestamp = {Wed, 14 Nov 2018 10:22:06 +0100},
 778  doi = {10.1109/TVCG.2014.2346419},
 779  url = {https://doi.org/10.1109/TVCG.2014.2346419},
 780  year = {2014},
 781  pages = {2211--2220},
 782  number = {12},
 783  volume = {20},
 784  journal = {{IEEE} Trans. Vis. Comput. Graph.},
 785  title = {The Persuasive Power of Data Visualization},
 786  author = {Anshul Vikram Pandey and
 787Anjali Manivannan and
 788Oded Nov and
 789Margaret Satterthwaite and
 790Enrico Bertini},
 791}
 792
 793@article{wachter2017counterfactual,
 794  publisher = {HeinOnline},
 795  year = {2017},
 796  pages = {841},
 797  volume = {31},
 798  journal = {Harv. JL \& Tech.},
 799  author = {Wachter, Sandra and Mittelstadt, Brent and Russell, Chris},
 800  title = {Counterfactual explanations without opening the black box: Automated decisions and the GDPR},
 801}
 802
 803@inproceedings{ustun2019actionable,
 804  year = {2019},
 805  pages = {10--19},
 806  booktitle = {Proceedings of the conference on fairness, accountability, and transparency},
 807  author = {Ustun, Berk and Spangher, Alexander and Liu, Yang},
 808  title = {Actionable recourse in linear classification},
 809}
 810
 811@article{edwards2018enslaving,
 812  publisher = {IEEE},
 813  year = {2018},
 814  pages = {46--54},
 815  number = {3},
 816  volume = {16},
 817  journal = {IEEE Security \& Privacy},
 818  author = {Edwards, Lilian and Veale, Michael},
 819  title = {Enslaving the algorithm: From a “Right to an Explanation” to a “Right to Better Decisions”?},
 820}
 821
 822@article{malgieri2019automated,
 823  publisher = {Elsevier},
 824  year = {2019},
 825  pages = {105327},
 826  number = {5},
 827  volume = {35},
 828  journal = {Computer law \& security review},
 829  author = {Malgieri, Gianclaudio},
 830  title = {Automated decision-making in the EU Member States: The right to explanation and other “suitable safeguards” in the national legislations},
 831}
 832
 833@misc{bhatt2020explainable,
 834  primaryclass = {cs.LG},
 835  archiveprefix = {arXiv},
 836  eprint = {1909.06342},
 837  year = {2020},
 838  author = {Umang Bhatt and Alice Xiang and Shubham Sharma and Adrian Weller and Ankur Taly and Yunhan Jia and Joydeep Ghosh and Ruchir Puri and José M. F. Moura and Peter Eckersley},
 839  title = {Explainable Machine Learning in Deployment},
 840}
 841
 842@misc{slack2020fooling,
 843  primaryclass = {cs.LG},
 844  archiveprefix = {arXiv},
 845  eprint = {1911.02508},
 846  year = {2020},
 847  author = {Dylan Slack and Sophie Hilgard and Emily Jia and Sameer Singh and Himabindu Lakkaraju},
 848  title = {Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods},
 849}
 850
 851@article{guidotti2018survey,
 852  bibsource = {dblp computer science bibliography, https://dblp.org},
 853  biburl = {https://dblp.org/rec/journals/csur/GuidottiMRTGP19.bib},
 854  timestamp = {Sat, 08 Jan 2022 02:23:15 +0100},
 855  doi = {10.1145/3236009},
 856  url = {https://doi.org/10.1145/3236009},
 857  year = {2019},
 858  pages = {93:1--93:42},
 859  number = {5},
 860  volume = {51},
 861  journal = {{ACM} Comput. Surv.},
 862  title = {A Survey of Methods for Explaining Black Box Models},
 863  author = {Riccardo Guidotti and
 864Anna Monreale and
 865Salvatore Ruggieri and
 866Franco Turini and
 867Fosca Giannotti and
 868Dino Pedreschi},
 869}
 870
 871@article{DBLP:journals/corr/abs-2004-00668,
 872  bibsource = {dblp computer science bibliography, https://dblp.org},
 873  biburl = {https://dblp.org/rec/journals/corr/abs-2004-00668.bib},
 874  timestamp = {Fri, 26 Nov 2021 16:33:35 +0100},
 875  eprint = {2004.00668},
 876  eprinttype = {arXiv},
 877  url = {https://arxiv.org/abs/2004.00668},
 878  year = {2020},
 879  volume = {abs/2004.00668},
 880  journal = {CoRR},
 881  title = {DBLP:journals/corr/abs-2004-00668 Feature Contributions Through Additive Importance
 882Measures},
 883  author = {Ian Covert and
 884Scott M. Lundberg and
 885Su{-}In Lee},
 886}
 887
 888@article{Black_2020,
 889  month = {Jan},
 890  year = {2020},
 891  author = {Black, Emily and Yeom, Samuel and Fredrikson, Matt},
 892  publisher = {ACM},
 893  journal = {Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency},
 894  doi = {10.1145/3351095.3372845},
 895  url = {http://dx.doi.org/10.1145/3351095.3372845},
 896  isbn = {9781450369367},
 897  title = {FlipTest},
 898}
 899
 900@article{osti_10182459,
 901  author = {Yang, Ke and Huang, Biao and Stoyanovich, Julia and Schelter, Sebastian},
 902  journal = {Workshop on Human-In-the-Loop Data Analytics (HILDA'20)},
 903  abstractnote = {Surfacing and mitigating bias in ML pipelines is a complex topic, with a dire need to provide system-level support to data scientists. Humans should be empowered to debug these pipelines, in order to control for bias and to improve data quality and representativeness. We propose fairDAGs, an open-source library that extracts directed acyclic graph (DAG) representations of the data flow in preprocessing pipelines for ML. The library subsequently instruments the pipelines with tracing and visualization code to capture changes in data distributions and identify distortions with respect to protected group membership as the data travels through the pipeline. We illustrate the utility of fairDAGs, with experiments on publicly available ML pipelines.},
 904  doi = {10.1145/3398730.3399194},
 905  url = {https://par.nsf.gov/biblio/10182459},
 906  title = {Fairness-Aware Instrumentation of Preprocessing~Pipelines for Machine Learning},
 907  place = {Country unknown/Code not available},
 908}
 909
 910@misc{marcinkevics2020interpretability,
 911  primaryclass = {cs.LG},
 912  archiveprefix = {arXiv},
 913  eprint = {2012.01805},
 914  year = {2020},
 915  author = {Ričards Marcinkevičs and Julia E. Vogt},
 916  title = {Interpretability and Explainability: A Machine Learning Zoo Mini-tour},
 917}
 918
 919@misc{saha2020measuring,
 920  primaryclass = {cs.CY},
 921  archiveprefix = {arXiv},
 922  eprint = {2001.00089},
 923  year = {2020},
 924  author = {Debjani Saha and Candice Schumann and Duncan C. McElfresh and John P. Dickerson and Michelle L. Mazurek and Michael Carl Tschantz},
 925  title = {Measuring Non-Expert Comprehension of Machine Learning Fairness Metrics},
 926}
 927
 928@article{STUMPF2009639,
 929  abstract = {Although machine learning is becoming commonly used in today's software, there has been little research into how end users might interact with machine learning systems, beyond communicating simple “right/wrong” judgments. If the users themselves could work hand-in-hand with machine learning systems, the users’ understanding and trust of the system could improve and the accuracy of learning systems could be improved as well. We conducted three experiments to understand the potential for rich interactions between users and machine learning systems. The first experiment was a think-aloud study that investigated users’ willingness to interact with machine learning reasoning, and what kinds of feedback users might give to machine learning systems. We then investigated the viability of introducing such feedback into machine learning systems, specifically, how to incorporate some of these types of user feedback into machine learning systems, and what their impact was on the accuracy of the system. Taken together, the results of our experiments show that supporting rich interactions between users and machine learning systems is feasible for both user and machine. This shows the potential of rich human–computer collaboration via on-the-spot interactions as a promising direction for machine learning systems and users to collaboratively share intelligence.},
 930  keywords = {Intelligent user interfaces, Rich feedback, Explanations, Machine learning},
 931  author = {Simone Stumpf and Vidya Rajaram and Lida Li and Weng-Keen Wong and Margaret Burnett and Thomas Dietterich and Erin Sullivan and Jonathan Herlocker},
 932  url = {https://www.sciencedirect.com/science/article/pii/S1071581909000457},
 933  doi = {https://doi.org/10.1016/j.ijhcs.2009.03.004},
 934  issn = {1071-5819},
 935  year = {2009},
 936  pages = {639-662},
 937  number = {8},
 938  volume = {67},
 939  journal = {International Journal of Human-Computer Studies},
 940  title = {Interacting meaningfully with machine learning systems: Three experiments},
 941}
 942
 943@article{Obermeyer447,
 944  journal = {Science},
 945  eprint = {https://science.sciencemag.org/content/366/6464/447.full.pdf},
 946  url = {https://science.sciencemag.org/content/366/6464/447},
 947  issn = {0036-8075},
 948  abstract = {The U.S. health care system uses commercial algorithms to guide health decisions. Obermeyer et al. find evidence of racial bias in one widely used algorithm, such that Black patients assigned the same level of risk by the algorithm are sicker than White patients (see the Perspective by Benjamin). The authors estimated that this racial bias reduces the number of Black patients identified for extra care by more than half. Bias occurs because the algorithm uses health costs as a proxy for health needs. Less money is spent on Black patients who have the same level of need, and the algorithm thus falsely concludes that Black patients are healthier than equally sick White patients. Reformulating the algorithm so that it no longer uses costs as a proxy for needs eliminates the racial bias in predicting who needs extra care.Science, this issue p. 447; see also p. 421Health systems rely on commercial prediction algorithms to identify and help patients with complex health needs. We show that a widely used algorithm, typical of this industry-wide approach and affecting millions of patients, exhibits significant racial bias: At a given risk score, Black patients are considerably sicker than White patients, as evidenced by signs of uncontrolled illnesses. Remedying this disparity would increase the percentage of Black patients receiving additional help from 17.7 to 46.5\%. The bias arises because the algorithm predicts health care costs rather than illness, but unequal access to care means that we spend less money caring for Black patients than for White patients. Thus, despite health care cost appearing to be an effective proxy for health by some measures of predictive accuracy, large racial biases arise. We suggest that the choice of convenient, seemingly effective proxies for ground truth can be an important source of algorithmic bias in many contexts.},
 949  publisher = {American Association for the Advancement of Science},
 950  doi = {10.1126/science.aax2342},
 951  year = {2019},
 952  pages = {447--453},
 953  number = {6464},
 954  volume = {366},
 955  title = {Dissecting racial bias in an algorithm used to manage the health of populations},
 956  author = {Obermeyer, Ziad and Powers, Brian and Vogeli, Christine and Mullainathan, Sendhil},
 957}
 958
 959@article{BARTLETT2021,
 960  abstract = {U.S. fair-lending law prohibits lenders from making credit determinations that disparately affect minority borrowers if those determinations are based on characteristics unrelated to creditworthiness. Using an identification under this rule, we show risk-equivalent Latinx/Black borrowers pay significantly higher interest rates on GSE-securitized and FHA-insured loans, particularly in high-minority-share neighborhoods. We estimate these rate differences cost minority borrowers over $450 million yearly. FinTech lenders’ rate disparities were similar to those of non-Fintech lenders for GSE mortgages, but lower for FHA mortgages issued in 2009–2015 and for FHA refi mortgages issued in 2018–2019.},
 961  keywords = {Discrimination, FinTech, GSE mortgages, Credit scoring, Algorithmic underwriting, Big-data lending, Platform loans, Statistical discrimination, Legitimate business necessity},
 962  author = {Robert Bartlett and Adair Morse and Richard Stanton and Nancy Wallace},
 963  url = {https://www.sciencedirect.com/science/article/pii/S0304405X21002403},
 964  doi = {https://doi.org/10.1016/j.jfineco.2021.05.047},
 965  issn = {0304-405X},
 966  year = {2021},
 967  journal = {Journal of Financial Economics},
 968  title = {Consumer-lending discrimination in the FinTech Era},
 969}
 970
 971@misc{holstein2021equity,
 972  primaryclass = {cs.HC},
 973  archiveprefix = {arXiv},
 974  eprint = {2104.12920},
 975  year = {2021},
 976  author = {Kenneth Holstein and Shayan Doroudi},
 977  title = {Equity and Artificial Intelligence in Education: Will "AIEd" Amplify or Alleviate Inequities in Education?},
 978}
 979
 980@misc{baker_hawn_2021,
 981  month = {Mar},
 982  year = {2021},
 983  author = {Baker, Ryan S and Hawn, Aaron},
 984  publisher = {EdArXiv},
 985  doi = {10.35542/osf.io/pbmvz},
 986  url = {edarxiv.org/pbmvz},
 987  title = {Algorithmic Bias in Education},
 988}
 989
 990@misc{hu_2020,
 991  month = {Jun},
 992  year = {2020},
 993  author = {Hu, Qian|Rangwala},
 994  publisher = {International Educational Data Mining Society. e-mail: [email protected]; Web site: http://www.educationaldatamining.org},
 995  journal = {International Educational Data Mining Society},
 996  url = {https://eric.ed.gov/?id=ED608050},
 997  title = {Towards Fair Educational Data Mining: A Case Study on Detecting At-Risk Students.},
 998}
 999
1000@inproceedings{Sapiezynski2017AcademicPP,
1001  year = {2017},
1002  author = {Piotr Sapiezynski and Valentin Kassarnig and Christo Wilson},
1003  title = {Academic performance prediction in a gender-imbalanced environment},
1004}
1005
1006@misc{lundberg2017unified,
1007  primaryclass = {cs.AI},
1008  archiveprefix = {arXiv},
1009  eprint = {1705.07874},
1010  year = {2017},
1011  author = {Scott Lundberg and Su-In Lee},
1012  title = {A Unified Approach to Interpreting Model Predictions},
1013}
1014
1015@misc{ribeiro2016why,
1016  primaryclass = {cs.LG},
1017  archiveprefix = {arXiv},
1018  eprint = {1602.04938},
1019  year = {2016},
1020  author = {Marco Tulio Ribeiro and Sameer Singh and Carlos Guestrin},
1021  title = {"Why Should I Trust You?": Explaining the Predictions of Any Classifier},
1022}
1023
1024@article{ECOA1994,
1025  url = {https://www.fdic.gov/resources/supervision-and-examinations/consumer-compliance-examination-manual/documents/4/iv-1-1.pdf},
1026  year = {1994},
1027  title = {IV. Fair Lending — Fair Lending Laws and Regulations},
1028  author = {FDIC: Federal Deposit Insurance Corporation},
1029}
1030
1031@article{lee2018detecting,
1032  publisher = {Emerald Publishing Limited},
1033  year = {2018},
1034  journal = {Journal of Information, Communication and Ethics in Society},
1035  author = {Lee, Nicol Turner},
1036  title = {Detecting racial bias in algorithms and machine learning},
1037}
1038
1039@article{decamp2020latent,
1040  publisher = {Oxford University Press},
1041  year = {2020},
1042  pages = {2020--2023},
1043  number = {12},
1044  volume = {27},
1045  journal = {Journal of the American Medical Informatics Association},
1046  author = {DeCamp, Matthew and Lindvall, Charlotta},
1047  title = {Latent bias and the implementation of artificial intelligence in medicine},
1048}
1049
1050@article{DBLP:journals/corr/abs-2102-03054,
1051  bibsource = {dblp computer science bibliography, https://dblp.org},
1052  biburl = {https://dblp.org/rec/journals/corr/abs-2102-03054.bib},
1053  timestamp = {Wed, 10 Feb 2021 15:24:31 +0100},
1054  eprint = {2102.03054},
1055  eprinttype = {arXiv},
1056  url = {https://arxiv.org/abs/2102.03054},
1057  year = {2021},
1058  volume = {abs/2102.03054},
1059  journal = {CoRR},
1060  title = {Removing biased data to improve fairness and accuracy},
1061  author = {Sahil Verma and
1062Michael D. Ernst and
1063Ren{\'{e}} Just},
1064}
1065
1066@article{FHA1968,
1067  url = {https://www.fdic.gov/regulations/laws/rules/6000-1400.html},
1068  year = {1968},
1069  title = {Civil Rights Act of 1968},
1070  author = {FDIC: Federal Deposit Insurance Corporation},
1071}
1072
1073@article{gunning2019xai,
1074  publisher = {Science Robotics},
1075  year = {2019},
1076  number = {37},
1077  volume = {4},
1078  journal = {Science Robotics},
1079  author = {Gunning, David and Stefik, Mark and Choi, Jaesik and Miller, Timothy and Stumpf, Simone and Yang, Guang-Zhong},
1080  title = {XAI—Explainable artificial intelligence},
1081}
1082
1083@article{doshivelez2017rigorous,
1084  primaryclass = {stat.ML},
1085  archiveprefix = {arXiv},
1086  eprint = {1702.08608},
1087  year = {2017},
1088  author = {Finale Doshi-Velez and Been Kim},
1089  title = {Towards A Rigorous Science of Interpretable Machine Learning},
1090}
1091
1092@inproceedings{NIPS2016_5680522b,
1093  year = {2016},
1094  volume = {29},
1095  url = {https://proceedings.neurips.cc/paper/2016/file/5680522b8e2bb01943234bce7bf84534-Paper.pdf},
1096  title = {Examples are not enough, learn to criticize! Criticism for Interpretability},
1097  publisher = {Curran Associates, Inc.},
1098  pages = {},
1099  editor = {D. Lee and M. Sugiyama and U. Luxburg and I. Guyon and R. Garnett},
1100  booktitle = {Advances in Neural Information Processing Systems},
1101  author = {Kim, Been and Khanna, Rajiv and Koyejo, Oluwasanmi O},
1102}
1103
1104@misc{hara2016making,
1105  primaryclass = {stat.ML},
1106  archiveprefix = {arXiv},
1107  eprint = {1606.05390},
1108  year = {2016},
1109  author = {Satoshi Hara and Kohei Hayashi},
1110  title = {Making Tree Ensembles Interpretable},
1111}
1112
1113@misc{julia_angwin_2016,
1114  month = {May},
1115  year = {2016},
1116  author = {Julia Angwin, Jeff Larson},
1117  journal = {ProPublica},
1118  url = {https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing},
1119  title = {Machine Bias},
1120}
1121
1122@article{Hong_2020,
1123  pages = {1–26},
1124  month = {May},
1125  year = {2020},
1126  author = {Hong, Sungsoo Ray and Hullman, Jessica and Bertini, Enrico},
1127  publisher = {Association for Computing Machinery (ACM)},
1128  journal = {Proceedings of the ACM on Human-Computer Interaction},
1129  number = {CSCW1},
1130  doi = {10.1145/3392878},
1131  url = {http://dx.doi.org/10.1145/3392878},
1132  issn = {2573-0142},
1133  volume = {4},
1134  title = {Human Factors in Model Interpretability: Industry Practices, Challenges, and Needs},
1135}
1136
1137@article{narayanan2018humans,
1138  year = {2018},
1139  journal = {arXiv preprint arXiv:1802.00682},
1140  author = {Narayanan, Menaka and Chen, Emily and He, Jeffrey and Kim, Been and Gershman, Sam and Doshi-Velez, Finale},
1141  title = {How do humans understand explanations from machine learning systems? an evaluation of the human-interpretability of explanation},
1142}
1143
1144@inproceedings{zejnilovic2021machine,
1145  organization = {Academy of Management Briarcliff Manor, NY 10510},
1146  year = {2021},
1147  pages = {15424},
1148  number = {1},
1149  volume = {2021},
1150  booktitle = {Academy of Management Proceedings},
1151  author = {Zejnilovic, Leid and Lavado, Susana and Soares, Carlos and Mart{\'\i}nez De Rituerto De Troya, {\'I}{\~n}igo and Bell, Andrew and Ghani, Rayid},
1152  title = {Machine Learning Informed Decision-Making with Interpreted Model’s Outputs: A Field Intervention},
1153}
1154
1155@article{10.1525/gp.2020.12908,
1156  eprint = {https://online.ucpress.edu/gp/article-pdf/1/1/12908/462946/12908.pdf},
1157  note = {12908},
1158  url = {https://doi.org/10.1525/gp.2020.12908},
1159  doi = {10.1525/gp.2020.12908},
1160  issn = {2575-7350},
1161  abstract = {{The recent surge of interest in algorithmic decision-making among scholars across disciplines is associated with its potential to resolve the challenges common to administrative decision-making in the public sector, such as greater fairness and equal treatment of each individual, among others. However, algorithmic decision-making combined with human judgment may introduce new complexities with unclear consequences. This article offers evidence that contributes to the ongoing discussion about algorithmic decision-making and governance, contextualizing it within a public employment service. In particular, we discuss the use of a decision support system that employs an algorithm to assess individual risk of becoming long-term unemployed and that informs counselors to assign interventions accordingly. We study the human interaction with algorithms in this context using the lenses of human detachment from and attachment to decision-making. Employing a mixed-method research approach, we show the complexity of enacting the potentials of the data-driven decision-making in the context of a public agency.}},
1162  month = {06},
1163  year = {2020},
1164  number = {1},
1165  volume = {1},
1166  journal = {Global Perspectives},
1167  title = {{Algorithmic Long-Term Unemployment Risk Assessment in Use: Counselors’ Perceptions and Use Practices}},
1168  author = {Zejnilović, Leid and Lavado, Susana and Martínez de Rituerto de Troya, Íñigo and Sim, Samantha and Bell, Andrew},
1169}
1170
1171@article{huysmans2006using,
1172  publisher = {KU Leuven KBI Working Paper},
1173  year = {2006},
1174  author = {Huysmans, Johan and Baesens, Bart and Vanthienen, Jan},
1175  title = {Using rule extraction to improve the comprehensibility of predictive models},
1176}
1177
1178@inproceedings{bell2019proactive,
1179  organization = {IEEE},
1180  year = {2019},
1181  pages = {1--6},
1182  booktitle = {2019 IEEE International Conference on Healthcare Informatics (ICHI)},
1183  author = {Bell, Andrew and Rich, Alexander and Teng, Melisande and Ore{\v{s}}kovi{\'c}, Tin and Bras, Nuno B and Mestrinho, L{\'e}nia and Golubovic, Srdan and Pristas, Ivan and Zejnilovic, Leid},
1184  title = {Proactive advising: a machine learning driven approach to vaccine hesitancy},
1185}
1186
1187@article{stiglic2015comprehensible,
1188  publisher = {Public Library of Science San Francisco, CA USA},
1189  year = {2015},
1190  pages = {e0144439},
1191  number = {12},
1192  volume = {10},
1193  journal = {PloS one},
1194  author = {Stiglic, Gregor and Povalej Brzan, Petra and Fijacko, Nino and Wang, Fei and Delibasic, Boris and Kalousis, Alexandros and Obradovic, Zoran},
1195  title = {Comprehensible predictive modeling using regularized logistic regression and comorbidity based features},
1196}
1197
1198@inproceedings{de2018predicting,
1199  year = {2018},
1200  booktitle = {NeurIPS Workshop on AI for Social Good},
1201  author = {de Troya, Inigo Martinez and Chen, Ruqian and Moraes, Laura O and Bajaj, Pranjal and Kupersmith, Jordan and Ghani, Rayid and Br{\'a}s, Nuno B and Zejnilovic, Leid},
1202  title = {Predicting, explaining, and understanding risk of long-term unemployment},
1203}
1204
1205@article{lamba2021empirical,
1206  publisher = {ACM New York, NY, USA},
1207  year = {2021},
1208  pages = {69--85},
1209  number = {1},
1210  volume = {23},
1211  journal = {ACM SIGKDD Explorations Newsletter},
1212  author = {Lamba, Hemank and Rodolfa, Kit T and Ghani, Rayid},
1213  title = {An Empirical Comparison of Bias Reduction Methods on Real-World Problems in High-Stakes Policy Settings},
1214}
1215
1216@article{gleicher2016framework,
1217  publisher = {Mary Ann Liebert, Inc. 140 Huguenot Street, 3rd Floor New Rochelle, NY 10801 USA},
1218  year = {2016},
1219  pages = {75--88},
1220  number = {2},
1221  volume = {4},
1222  journal = {Big data},
1223  author = {Gleicher, Michael},
1224  title = {A framework for considering comprehensibility in modeling},
1225}
1226
1227@article{wilde_2021,
1228  pages = {e2},
1229  year = {2021},
1230  author = {Wilde, Harrison and Chen, Lucia L. and Nguyen, Austin and Kimpel, Zoe and Sidgwick, Joshua and De Unanue, Adolfo and Veronese, Davide and Mateen, Bilal and Ghani, Rayid and Vollmer, Sebastian and et al.},
1231  publisher = {Cambridge University Press},
1232  journal = {Data and Policy},
1233  doi = {10.1017/dap.2020.23},
1234  volume = {3},
1235  title = {A recommendation and risk classification system for connecting rough sleepers to essential outreach services},
1236}
1237
1238@inproceedings{carton2016identifying,
1239  year = {2016},
1240  pages = {67--76},
1241  booktitle = {Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining},
1242  author = {Carton, Samuel and Helsby, Jennifer and Joseph, Kenneth and Mahmud, Ayesha and Park, Youngsoo and Walsh, Joe and Cody, Crystal and Patterson, CPT Estella and Haynes, Lauren and Ghani, Rayid},
1243  title = {Identifying police officers at risk of adverse events},
1244}
1245
1246@inproceedings{aguiar2015and,
1247  year = {2015},
1248  pages = {93--102},
1249  booktitle = {Proceedings of the Fifth International Conference on Learning Analytics And Knowledge},
1250  author = {Aguiar, Everaldo and Lakkaraju, Himabindu and Bhanpuri, Nasir and Miller, David and Yuhas, Ben and Addison, Kecia L},
1251  title = {Who, when, and why: A machine learning approach to prioritizing students at risk of not graduating high school on time},
1252}
1253
1254@article{holzinger2020measuring,
1255  publisher = {Springer},
1256  year = {2020},
1257  pages = {1--6},
1258  journal = {KI-K{\"u}nstliche Intelligenz},
1259  author = {Holzinger, Andreas and Carrington, Andr{\'e} and M{\"u}ller, Heimo},
1260  title = {Measuring the quality of explanations: the system causability scale (SCS)},
1261}
1262
1263@article{aha1988instance,
1264  year = {1988},
1265  pages = {3--2},
1266  number = {1},
1267  volume = {3},
1268  journal = {University of California},
1269  author = {Aha, D and Kibler, Dennis},
1270  title = {Instance-based prediction of heart-disease presence with the Cleveland database},
1271}
1272
1273@misc{das2020opportunities,
1274  primaryclass = {cs.CV},
1275  archiveprefix = {arXiv},
1276  eprint = {2006.11371},
1277  year = {2020},
1278  author = {Arun Das and Paul Rad},
1279  title = {Opportunities and Challenges in Explainable Artificial Intelligence (XAI): A Survey},
1280}
1281
1282@article{10.1093/jamia/ocz229,
1283  eprint = {https://academic.oup.com/jamia/article-pdf/27/4/592/34153285/ocz229.pdf},
1284  url = {https://doi.org/10.1093/jamia/ocz229},
1285  doi = {10.1093/jamia/ocz229},
1286  issn = {1527-974X},
1287  abstract = {{Implementation of machine learning (ML) may be limited by patients’ right to “meaningful information about the logic involved” when ML influences healthcare decisions. Given the complexity of healthcare decisions, it is likely that ML outputs will need to be understood and trusted by physicians, and then explained to patients. We therefore investigated the association between physician understanding of ML outputs, their ability to explain these to patients, and their willingness to trust the ML outputs, using various ML explainability methods.We designed a survey for physicians with a diagnostic dilemma that could be resolved by an ML risk calculator. Physicians were asked to rate their understanding, explainability, and trust in response to 3 different ML outputs. One ML output had no explanation of its logic (the control) and 2 ML outputs used different model-agnostic explainability methods. The relationships among understanding, explainability, and trust were assessed using Cochran-Mantel-Haenszel tests of association.The survey was sent to 1315 physicians, and 170 (13\\%) provided completed surveys. There were significant associations between physician understanding and explainability (P \\&lt; .001), between physician understanding and trust (P \\&lt; .001), and between explainability and trust (P \\&lt; .001). ML outputs that used model-agnostic explainability methods were preferred by 88\\% of physicians when compared with the control condition; however, no particular ML explainability method had a greater influence on intended physician behavior.Physician understanding, explainability, and trust in ML risk calculators are related. Physicians preferred ML outputs accompanied by model-agnostic explanations but the explainability method did not alter intended physician behavior.}},
1288  month = {02},
1289  year = {2020},
1290  pages = {592-600},
1291  number = {4},
1292  volume = {27},
1293  journal = {Journal of the American Medical Informatics Association},
1294  title = {{Physician understanding, explainability, and trust in a hypothetical machine learning risk calculator}},
1295  author = {Diprose, William K and Buist, Nicholas and Hua, Ning and Thurier, Quentin and Shand, George and Robinson, Reece},
1296}
1297
1298@article{10.1002/isaf.1422,
1299  year = {2018},
1300  abstract = {Summary Recent rapid progress in machine learning (ML), particularly so-called ‘deep learning’, has led to a resurgence in interest in explainability of artificial intelligence (AI) systems, reviving an area of research dating back to the 1970s. The aim of this article is to view current issues concerning ML-based AI systems from the perspective of classical AI, showing that the fundamental problems are far from new, and arguing that elements of that earlier work offer routes to making progress towards explainable AI today.},
1301  eprint = {https://onlinelibrary.wiley.com/doi/pdf/10.1002/isaf.1422},
1302  url = {https://onlinelibrary.wiley.com/doi/abs/10.1002/isaf.1422},
1303  doi = {https://doi.org/10.1002/isaf.1422},
1304  keywords = {artificial intelligence, explainability, interpretability, machine learning},
1305  pages = {63-72},
1306  number = {2},
1307  volume = {25},
1308  journal = {Intelligent Systems in Accounting, Finance and Management},
1309  title = {Asking ‘Why’ in AI: Explainability of intelligent systems – perspectives and challenges},
1310  author = {Preece, Alun},
1311}
1312
1313@article{Ehsan_2021,
1314  month = {May},
1315  year = {2021},
1316  author = {Ehsan, Upol and Liao, Q. Vera and Muller, Michael and Riedl, Mark O. and Weisz, Justin D.},
1317  publisher = {ACM},
1318  journal = {Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems},
1319  doi = {10.1145/3411764.3445188},
1320  url = {http://dx.doi.org/10.1145/3411764.3445188},
1321  isbn = {9781450380966},
1322  title = {Expanding Explainability: Towards Social Transparency in AI systems},
1323}
1324
1325@inproceedings{abdul2020cogam,
1326  year = {2020},
1327  pages = {1--14},
1328  booktitle = {Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems},
1329  author = {Abdul, Ashraf and von der Weth, Christian and Kankanhalli, Mohan and Lim, Brian Y},
1330  title = {COGAM: Measuring and Moderating Cognitive Load in Machine Learning Model Explanations},
1331}
1332
1333@inproceedings{10.1145/1620545.1620576,
1334  series = {UbiComp '09},
1335  location = {Orlando, Florida, USA},
1336  keywords = {explanations, context-aware, satisfaction, intelligibility},
1337  numpages = {10},
1338  pages = {195–204},
1339  booktitle = {Proceedings of the 11th International Conference on Ubiquitous Computing},
1340  abstract = {Intelligibility can help expose the inner workings and inputs of context-aware applications
1341that tend to be opaque to users due to their implicit sensing and actions. However,
1342users may not be interested in all the information that the applications can produce.
1343Using scenarios of four real-world applications that span the design space of context-aware
1344computing, we conducted two experiments to discover what information users are interested
1345in. In the first experiment, we elicit types of information demands that users have
1346and under what moderating circumstances they have them. In the second experiment,
1347we verify the findings by soliciting users about which types they would want to know
1348and establish whether receiving such information would satisfy them. We discuss why
1349users demand certain types of information, and provide design implications on how
1350to provide different intelligibility types to make context-aware applications intelligible
1351and acceptable to users.},
1352  doi = {10.1145/1620545.1620576},
1353  url = {https://doi.org/10.1145/1620545.1620576},
1354  address = {New York, NY, USA},
1355  publisher = {Association for Computing Machinery},
1356  isbn = {9781605584317},
1357  year = {2009},
1358  title = {Assessing Demand for Intelligibility in Context-Aware Applications},
1359  author = {Lim, Brian Y. and Dey, Anind K.},
1360}
1361
1362@inbook{10.1145/1518701.1519023,
1363  numpages = {10},
1364  pages = {2119–2128},
1365  booktitle = {Proceedings of the SIGCHI Conference on Human Factors in Computing Systems},
1366  abstract = {Context-aware intelligent systems employ implicit inputs, and make decisions based
1367on complex rules and machine learning models that are rarely clear to users. Such
1368lack of system intelligibility can lead to loss of user trust, satisfaction and acceptance
1369of these systems. However, automatically providing explanations about a system's decision
1370process can help mitigate this problem. In this paper we present results from a controlled
1371study with over 200 participants in which the effectiveness of different types of
1372explanations was examined. Participants were shown examples of a system's operation
1373along with various automatically generated explanations, and then tested on their
1374understanding of the system. We show, for example, that explanations describing why
1375the system behaved a certain way resulted in better understanding and stronger feelings
1376of trust. Explanations describing why the system did not behave a certain way, resulted
1377in lower understanding yet adequate performance. We discuss implications for the use
1378of our findings in real-world context-aware applications.},
1379  url = {https://doi.org/10.1145/1518701.1519023},
1380  address = {New York, NY, USA},
1381  publisher = {Association for Computing Machinery},
1382  isbn = {9781605582467},
1383  year = {2009},
1384  title = {Why and Why Not Explanations Improve the Intelligibility of Context-Aware Intelligent Systems},
1385  author = {Lim, Brian Y. and Dey, Anind K. and Avrahami, Daniel},
1386}
1387
1388@book{10.5555/3208509,
1389  abstract = {Naomi Klein: "This book is downright scary."Ethan Zuckerman, MIT: "Should be required
1390reading."Dorothy Roberts, author of Killing the Black Body: "A must-read."Astra Taylor,
1391author of The People's Platform: "The single most important book about technology
1392you will read this year."Cory Doctorow: "Indispensable."A powerful investigative look
1393at data-based discriminationand how technology affects civil and human rights and
1394economic equity The State of Indiana denies one million applications for healthcare,
1395foodstamps and cash benefits in three yearsbecause a new computer system interprets
1396any mistake as failure to cooperate. In Los Angeles, an algorithm calculates the comparative
1397vulnerability of tens of thousands of homeless people in order to prioritize them
1398for an inadequate pool of housing resources. In Pittsburgh, a child welfare agency
1399uses a statistical model to try to predict which children might be future victims
1400of abuse or neglect. Since the dawn of the digital age, decision-making in finance,
1401employment, politics, health and human services has undergone revolutionary change.
1402Today, automated systemsrather than humanscontrol which neighborhoods get policed,
1403which families attain needed resources, and who is investigated for fraud. While we
1404all live under this new regime of data, the most invasive and punitive systems are
1405aimed at the poor. In Automating Inequality, Virginia Eubanks systematically investigates
1406the impacts of data mining, policy algorithms, and predictive risk models on poor
1407and working-class people in America. The book is full of heart-wrenching and eye-opening
1408stories, from a woman in Indiana whose benefits are literally cut off as she lays
1409dying to a family in Pennsylvania in daily fear of losing their daughter because they
1410fit a certain statistical profile. The U.S. has always used its most cutting-edge
1411science and technology to contain, investigate, discipline and punish the destitute.
1412Like the county poorhouse and scientific charity before them, digital tracking and
1413automated decision-making hide poverty from the middle-class public and give the nation
1414the ethical distance it needs to make inhumane choices: which families get food and
1415which starve, who has housing and who remains homeless, and which families are broken
1416up by the state. In the process, they weaken democracy and betray our most cherished
1417national values. This deeply researched and passionate book could not be more timely.},
1418  address = {USA},
1419  publisher = {St. Martin's Press, Inc.},
1420  isbn = {1250074312},
1421  year = {2018},
1422  title = {Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor},
1423  author = {Eubanks, Virginia},
1424}
1425
1426@book{10.5555/3002861,
1427  abstract = {A former Wall Street quant sounds an alarm on the mathematical models that pervade
1428modern life and threaten to rip apart our social fabricWe live in the age of the algorithm.
1429Increasingly, the decisions that affect our liveswhere we go to school, whether we
1430get a car loan, how much we pay for health insuranceare being made not by humans,
1431but by mathematical models. In theory, this should lead to greater fairness: Everyone
1432is judged according to the same rules, and bias is eliminated. But as Cathy ONeil
1433reveals in this urgent and necessary book, the opposite is true. The models being
1434used today are opaque, unregulated, and uncontestable, even when theyre wrong. Most
1435troubling, they reinforce discrimination: If a poor student cant get a loan because
1436a lending model deems him too risky (by virtue of his zip code), hes then cut off
1437from the kind of education that could pull him out of poverty, and a vicious spiral
1438ensues. Models are propping up the lucky and punishing the downtrodden, creating a
1439toxic cocktail for democracy. Welcome to the dark side of Big Data. Tracing the arc
1440of a persons life, ONeil exposes the black box models that shape our future, both
1441as individuals and as a society. These weapons of math destruction score teachers
1442and students, sort rsums, grant (or deny) loans, evaluate workers, target voters,
1443set parole, and monitor our health. ONeil calls on modelers to take more responsibility
1444for their algorithms and on policy makers to regulate their use. But in the end, its
1445up to us to become more savvy about the models that govern our lives. This important
1446book empowers us to ask the tough questions, uncover the truth, and demand change.},
1447  address = {USA},
1448  publisher = {Crown Publishing Group},
1449  isbn = {0553418815},
1450  year = {2016},
1451  title = {Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy},
1452  author = {O'Neil, Cathy},
1453}
1454
1455@misc{chouldechova2016fair,
1456  primaryclass = {stat.AP},
1457  archiveprefix = {arXiv},
1458  eprint = {1610.07524},
1459  year = {2016},
1460  author = {Alexandra Chouldechova},
1461  title = {Fair prediction with disparate impact: A study of bias in recidivism prediction instruments},
1462}
1463
1464@article{Goodman_2017,
1465  pages = {50–57},
1466  month = {Oct},
1467  year = {2017},
1468  author = {Goodman, Bryce and Flaxman, Seth},
1469  publisher = {Association for the Advancement of Artificial Intelligence (AAAI)},
1470  journal = {AI Magazine},
1471  number = {3},
1472  doi = {10.1609/aimag.v38i3.2741},
1473  url = {http://dx.doi.org/10.1609/aimag.v38i3.2741},
1474  issn = {0738-4602},
1475  volume = {38},
1476  title = {European Union Regulations on Algorithmic Decision-Making and a “Right to Explanation”},
1477}
1478
1479@inproceedings{Cortez2008UsingDM,
1480  year = {2008},
1481  author = {P. Cortez and A. M. G. Silva},
1482  title = {Using data mining to predict secondary school student performance},
1483}
1484
1485@article{OpenML2013,
1486  address = {New York, NY, USA},
1487  publisher = {ACM},
1488  doi = {10.1145/2641190.2641198},
1489  url = {http://doi.acm.org/10.1145/2641190.2641198},
1490  pages = {49--60},
1491  year = {2013},
1492  number = {2},
1493  volume = {15},
1494  journal = {SIGKDD Explorations},
1495  title = {OpenML: Networked Science in Machine Learning},
1496  author = {Vanschoren, Joaquin and van Rijn, Jan N. and Bischl, Bernd and Torgo, Luis},
1497}
1498
1499@article{miller2019explanation,
1500  publisher = {Elsevier},
1501  year = {2019},
1502  pages = {1--38},
1503  volume = {267},
1504  journal = {Artificial intelligence},
1505  author = {Miller, Tim},
1506  title = {Explanation in artificial intelligence: Insights from the social sciences},
1507}
1508
1509@book{molnar2020interpretable,
1510  publisher = {Lulu. com},
1511  year = {2020},
1512  author = {Molnar, Christoph},
1513  title = {Interpretable machine learning},
1514}
1515
1516@inproceedings{yang2019study,
1517  year = {2019},
1518  booktitle = {IUI Workshops},
1519  author = {Yang, Yiwei and Kandogan, Eser and Li, Yunyao and Sen, Prithviraj and Lasecki, Walter S},
1520  title = {A study on interaction in human-in-the-loop machine learning for text analytics},
1521}
1522
1523@article{ross2017impact,
1524  year = {2017},
1525  journal = {Univerity of Chicago, Harris School of Public Policy Working Paper},
1526  author = {Ross, Robert},
1527  title = {The impact of property tax appeals on vertical equity in Cook County, IL},
1528}
1529
1530@article{amarasinghe2020explainable,
1531  year = {2020},
1532  journal = {arXiv preprint arXiv:2010.14374},
1533  author = {Amarasinghe, Kasun and Rodolfa, Kit and Lamba, Hemank and Ghani, Rayid},
1534  title = {Explainable machine learning for public policy: Use cases, gaps, and research directions},
1535}
1536
1537@inproceedings{jesus2021can,
1538  year = {2021},
1539  pages = {805--815},
1540  booktitle = {Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency},
1541  author = {Jesus, S{\'e}rgio and Bel{\'e}m, Catarina and Balayan, Vladimir and Bento, Jo{\~a}o and Saleiro, Pedro and Bizarro, Pedro and Gama, Jo{\~a}o},
1542  title = {How can I choose an explainer? An Application-grounded Evaluation of Post-hoc Explanations},
1543}
1544
1545@article{fuster2020predictably,
1546  year = {2020},
1547  journal = {The Effects of Machine Learning on Credit Markets (October 1, 2020)},
1548  author = {Fuster, Andreas and Goldsmith-Pinkham, Paul and Ramadorai, Tarun and Walther, Ansgar},
1549  title = {Predictably unequal? the effects of machine learning on credit markets},
1550}
1551
1552@article{dziugaite2020enforcing,
1553  year = {2020},
1554  journal = {arXiv preprint arXiv:2010.13764},
1555  author = {Dziugaite, Gintare Karolina and Ben-David, Shai and Roy, Daniel M},
1556  title = {Enforcing Interpretability and its Statistical Impacts: Trade-offs between Accuracy and Interpretability},
1557}
1558
1559@inproceedings{barocas2020hidden,
1560  year = {2020},
1561  pages = {80--89},
1562  booktitle = {Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency},
1563  author = {Barocas, Solon and Selbst, Andrew D and Raghavan, Manish},
1564  title = {The hidden assumptions behind counterfactual explanations and principal reasons},
1565}
1566
1567@inproceedings{dai2021fair,
1568  year = {2021},
1569  pages = {55--65},
1570  booktitle = {Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society},
1571  author = {Dai, Jessica and Fazelpour, Sina and Lipton, Zachary},
1572  title = {Fair machine learning under partial compliance},
1573}
1574
1575@inproceedings{loi2021towards,
1576  year = {2021},
1577  pages = {757--766},
1578  booktitle = {Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society},
1579  author = {Loi, Michele and Spielkamp, Matthias},
1580  title = {Towards accountability in the use of artificial intelligence for public administrations},
1581}
1582
1583@article{lu2019good,
1584  year = {2019},
1585  journal = {Available at SSRN 3503603},
1586  author = {Lu, Joy and Lee, Dokyun and Kim, Tae Wan and Danks, David},
1587  title = {Good explanation for algorithmic transparency},
1588}
1589
1590@article{jung2017simple,
1591  year = {2017},
1592  journal = {arXiv preprint arXiv:1702.04690},
1593  author = {Jung, Jongbin and Concannon, Connor and Shroff, Ravi and Goel, Sharad and Goldstein, Daniel G},
1594  title = {Simple rules for complex decisions},
1595}
1596
1597@inproceedings{10.1145/3306618.3314274,
1598  series = {AIES '19},
1599  location = {Honolulu, HI, USA},
1600  keywords = {rights, interest theory, right to be forgotten, right to disconnect},
1601  numpages = {1},
1602  pages = {201},
1603  booktitle = {Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society},
1604  abstract = {Over recent decades, technological development has been accompanied by the proposal of new rights by various groups and individuals: the right to public anonymity, the right to be forgotten, and the right to disconnect, for example. Although there is widespread acknowledgment of the motivation behind these proposed rights, there is little agreement about their actual normative status. One potential challenge is that the claims only arise in contingent social-technical contexts, which may affect how we conceive of them ethically (albeit, not necessarily in terms of policy). What sort of morally legitimate rights claims depend on such contingencies? Our paper investigates the grounds on which such proposals might be considered "actual" rights. The full paper can be found at http://www.andrew.cmu.edu/user/cgparker/Parker_Danks_RevealedRights.pdf. We propose the notion of a revealed right, a right that only imposes duties -- and thus is only meaningfully revealed -- in certain technological contexts. Our framework is based on an interest theory approach to rights, which understands rights in terms of a justificatory role: morally important aspects of a person's well-being (interests) ground rights, which then justify holding someone to a duty that promotes or protects that interest. Our framework uses this approach to interpret the conflicts that lead to revealed rights in terms of how technological developments cause shifts in the balance of power to promote particular interests. Different parties can have competing or conflicting interests. It is also generally accepted that some interests are more normatively important than others (even if only within a particular framework). We can refer to this difference in importance by saying that the former interest has less "moral weight" than the latter interest (in that context). The moral weight of an interest is connected to its contribution to the interest-holder's overall well-being, and thereby determines the strength of the reason that a corresponding right provides to justify a duty. Improved technology can offer resources that grant one party increased causal power to realize its interests to the detriment of another's capacity to do so, even while the relative moral weight of their interests remain the same. Such changes in circumstance can make the importance of protecting a particular interest newly salient. If that interest's moral weight justifies establishing a duty to protect it, thereby limiting the threat posed by the new socio-technical context, then a right is revealed. Revealed rights justify realignment between the moral weight and causal power orderings so that people with weightier interests have greater power to protect those interests. In the extended paper, we show how this account can be applied to the interpretation of two recently proposed "rights": the right to be forgotten, and the right to disconnect. Since we are focused on making sense of revealed rights, not any particular substantive theory of interests or well-being, the characterization of 'weights' is a free parameter in this account. Our framework alone cannot provide means to resolve the question of whether specific rights exist, but it can be used to identify empirical questions that need to be answered to decide the existence or non-existence of such rights. The emergence of a revealed right depends on a number of factors, including: whether the plausible uses of the technology could potentially impede another's well-being or interests; whether the technology is sufficiently common to have a wider, social impact; and whether the technology has actually changed the balance of power sufficiently to yield a frequent possibility for misalignment between causal power and moral weight. This approach confronts the question of how, in principle, such rights could be justified, without requiring specific commitments on the ontology of rights. Our account explains why the rhetoric of "new rights" is both accurate (since the rights were not previously recognized) and inaccurate (since the rights were present all along, but without corresponding duties). Further, it explains the rights without grounding their normative status in considerations related to right-holders' capacities to rationally waive or assert claims. This is especially important given that many of the relevant disruptive technological developments pose challenges to understanding by affected parties for the same reasons they pose threats to those parties' well-being. In the course of our discussion, we confront a number of potential objections to the account. We argue that our framework's ability to accommodate highly specific or derivative-seeming rights is un-problematic. We also head off worries that our use of interest theory makes the account likely to recognize absurd rights claims.},
1605  doi = {10.1145/3306618.3314274},
1606  url = {https://doi.org/10.1145/3306618.3314274},
1607  address = {New York, NY, USA},
1608  publisher = {Association for Computing Machinery},
1609  isbn = {9781450363242},
1610  year = {2019},
1611  title = {How Technological Advances Can Reveal Rights},
1612  author = {Parker, Jack and Danks, David},
1613}
1614
1615@article{schmidt2020transparency,
1616  publisher = {Taylor \& Francis},
1617  year = {2020},
1618  pages = {260--278},
1619  number = {4},
1620  volume = {29},
1621  journal = {Journal of Decision Systems},
1622  author = {Schmidt, Philipp and Biessmann, Felix and Teubner, Timm},
1623  title = {Transparency and trust in artificial intelligence systems},
1624}
1625
1626@article{nichols2013consequences,
1627  year = {2013},
1628  journal = {Washington, DC: The Urban Institute},
1629  author = {Nichols, Austin and Mitchell, Josh and Lindner, Stephan},
1630  title = {Consequences of long-term unemployment},
1631}
1632
1633@inproceedings{cech2021tackling,
1634  year = {2021},
1635  pages = {258--268},
1636  booktitle = {C\&T'21: Proceedings of the 10th International Conference on Communities \& Technologies-Wicked Problems in the Age of Tech},
1637  author = {Cech, Florian},
1638  title = {Tackling Algorithmic Transparency in Communal Energy Accounting through Participatory Design},
1639}
1640
1641@inproceedings{eiband2018bringing,
1642  year = {2018},
1643  pages = {211--223},
1644  booktitle = {23rd international conference on intelligent user interfaces},
1645  author = {Eiband, Malin and Schneider, Hanna and Bilandzic, Mark and Fazekas-Con, Julian and Haug, Mareike and Hussmann, Heinrich},
1646  title = {Bringing transparency design into practice},
1647}
1648
1649@article{aizenberg2020designing,
1650  publisher = {SAGE Publications Sage UK: London, England},
1651  year = {2020},
1652  pages = {2053951720949566},
1653  number = {2},
1654  volume = {7},
1655  journal = {Big Data \& Society},
1656  author = {Aizenberg, Evgeni and Van Den Hoven, Jeroen},
1657  title = {Designing for human rights in AI},
1658}
1659
1660@article{gupta2020participatory,
1661  year = {2020},
1662  journal = {arXiv preprint arXiv:2006.00432},
1663  author = {Gupta, Abhishek and De Gasperis, Tania},
1664  title = {Participatory Design to build better contact-and proximity-tracing apps},
1665}
1666
1667@article{preece2018stakeholders,
1668  year = {2018},
1669  journal = {arXiv preprint arXiv:1810.00184},
1670  author = {Preece, Alun and Harborne, Dan and Braines, Dave and Tomsett, Richard and Chakraborty, Supriyo},
1671  title = {Stakeholders in explainable AI},
1672}
1673
1674@article{richards2021human,
1675  year = {2021},
1676  pages = {47--58},
1677  number = {4},
1678  volume = {44},
1679  journal = {IEEE Data Eng. Bull.},
1680  author = {Richards, John T and Piorkowski, David and Hind, Michael and Houde, Stephanie and Mojsilovic, Aleksandra and Varshney, Kush R},
1681  title = {A Human-Centered Methodology for Creating AI FactSheets.},
1682}
1683
1684@article{hind2019explaining,
1685  publisher = {ACM New York, NY, USA},
1686  year = {2019},
1687  pages = {16--19},
1688  number = {3},
1689  volume = {25},
1690  journal = {XRDS: Crossroads, The ACM Magazine for Students},
1691  author = {Hind, Michael},
1692  title = {Explaining explainable AI},
1693}

Attribution

arXiv:2207.01482v1 [cs.CY]
License: cc-by-4.0

Related Posts

Ethics Sheet for Automatic Emotion Recognition and Sentiment Analysis

Ethics Sheet for Automatic Emotion Recognition and Sentiment Analysis

Introduction Emotions play a central role in our lives. Thus affective computing, which deals with emotions and computation (often through AI systems) is atremendously important and vibrant line of work.

How Do AI Timelines Affect Existential Risk?

How Do AI Timelines Affect Existential Risk?

Introduction Recent progress in AI suggests that artificial general intelligence (AGI) that is as capable as humans on a wide variety of tasks is likely to be created this century.

Fairness Assessment for Artificial Intelligence in Financial Industry

Fairness Assessment for Artificial Intelligence in Financial Industry

Introduction Financial intelligence has a fast and accurate machine learning capability to achieve the intellectualization, standardization, and automation of large-scale business transactions.