Join Our Discord (630+ Members)

`It Is Currently Hodgepodge'': Examining AI/ML Practitioners' Challenges During Co-Production of Responsible AI Values

Content License: cc-by-sa

`It is currently hodgepodge'': Examining AI/ML Practitioners' Challenges during Co-production of Responsible AI Values

Papers is Alpha. This content is part of an effort to make research more accessible, and (most likely) has lost some details from the original. You can find the original paper here.


In November 2021, the UN Educational, Scientific, and Cultural Organization (UNESCO) signed a historic agreement outliningshared values needed to ensure the development of Responsible Artificial Intelligence (RAI). RAI is an umbrella term that comprises different human values, principles, and actions to develop AI ethically and responsibly. Through UNESCO’s agreement, for the first time, 193 countries have standardized recommendations on the ethics of AI. While unprecedented, this agreement is just one of several efforts providing recommendations on different RAI values to be implemented within AI/ML systems.

In response, several industry organizations have begun to implement the recommendations, creating cross-functional RAI institutional structures and activities that enable practitioners to engage with the RAI values. For instance, several big-tech companies are implementing common RAI values, such as Fairness, Transparency, Accountability, and Privacy, as part of their RAI initiatives. However, such RAI values have minimal overlap with the values prescribed by UNESCO’s framework that promotes non-maleficence, diversity, inclusion, and harmony. Scholars have attributed the lack of overlap to different business and institutional contexts involved in developing AI/ML [hereafter, we use AI/ML systems to refer to both AI products and ML models] systems. Subsequently, it is essential to understand these contexts by engaging with practitioners across multiple roles who come together to co-produce and enact such RAI values. Co-production is an iterative process through which organizations produce collective knowledge. During co-production, individual practitioners may hold certain values (e.g., social justice), yet their teams might prioritize other values.hints at potential challenges that can arise due to such mismatches in RAI values. Our study builds on this critical gap by giving a detailed analysis of those challenges and strategies (if any) devised to overcome such strains as practitioners co-produce AI/ML systems.

We interviewed 23 practitioners across a variety of roles to understand their RAI value practices and challenges. Our findings show that institutional structures around RAI value co-production contributed to key challenges for the practitioners. We also discovered multiple tensions that arose between roles and organizations during prioritization, deliberation, and implementation. Interestingly, we also observed development of ten different RAI value levers. These are creative activities meant to engage individuals in value conversations that help reduce value tensions. In the remainder of the paper, we first discuss related work about collective values in Responsible AI from an HCI perspective and outline our research questions. We then present the research methodology and results of the study. We conclude with a discussion of our contributions in improving RAI co-production practices of AI practitioners. Overall, this work makes several contributions. First, we describe the experiences and the organizational environment within which AI/ML practitioners co-produce RAI values. Second, we illustrate multiple challenges faced by AI practitioners owing to different organizational structures, resulting in several tensions in co-production. Third, we unpack ten RAI value levers as strategies to overcome challenges and map them on to the RAI values. Lastly, we provide essential strategies at the different levels (individual, and organizational) to better facilitate and sustain RAI value co-production.

The Field of Responsible AI

In the last decade, Responsible AI (RAI) has grown into an overarching field that aims to make AI/ML more accountable to its outcomes. One of the field’s roots lies in Ethical AI, where critical engagement with ethical values in the otherwise traditional AI/ML field have been encouraged. Example studies include engagement with core ethical values to provide nuance in technical AI/ML discourse, translating ethical values into implementation scenarios, and AI/ML guidelines, and seminal studies that brought critical ethical problems to the forefront.

RAI has also drawn its inspiration from AI for Social Good (AI4SG) research to study human values more broadly,going beyond “ethical” values. AI4SG helped the RAI field to translate such values embedded in AI/ML systems into positive community outcomesby eliciting specific values (e.g., solidarity), developing methods (e.g., capabilities approach), and producing representations (e.g., explanations) that strongly align with community goals (e.g., the UN Sustainable Development Goals). For example, studies have explicitly engaged with underserved communities to examine the impact of the embedded values within AI/ML systems in their lives (e.g., in agriculture, health, and educationdomains). More recent studies have shed light on how certain practitioners’ (e.g., data and crowd workers) practices, contributions, and values are often ignored while developing the AI/ML systems. Some have looked at different values (e.g., fairness) that are often ignored by the discriminatory algorithms. More recent work at the intersection of these two byhas also highlighted the impact of data workers of marginalized communities on the AI/ML algorithms to highlight complexity when building for RAI values like equity.

Lastly, RAI has also drawn motivation from recent movements associated with specific value(s). One such movement is around the value of explainability (or explainable AI ) that arose from the need to make AI/ML systems more accountable and trustworthy. A similar movement within the RAI’s purview focused on FATE values (Fairness, Accountability, Transparency, and Ethics/Explainability). While both movements have challenged the notion of universal applicability of RAI values, our study illustrates how these challenges do indeed appear in practice and the strategies used by practitioners on the ground to resolve these challenges.

Taken together, RAI has emerged as an umbrella term that encapsulates the above movements, encouraging critical value discourses to produce a positive impact. At the same time, departing from previous movements that focused on specific issues within AI/ML practices, RAI takes a broad institutional approach that promotes disparate AI/ML practitioners to come together, share, and enact on key values. Our study expands RAI discipline by surfacing on-ground challenges of diverse AI/ML practitioners attempting to engage in shared RAI responsibilities, such as collaborative value discourse and implementation. In the next section, we further unpack the notion of values, examine their roots in HCI and their current role in RAI.

Collective Values in Responsible AI: HCI perspectives

Science & Technology Studies field has long examined how values are embedded in technology systems in various social and political contexts. In recent years, studies within HCI have built on this foundation to bring a critical lens into the development of technology. Initial studies conducted by Nissenbaum and colleagues argued against the previously held belief that technology is “value-neutral”, showcasing how practitioners embed specific values through their deliberate design decisionsValue-sensitive Design_ (VSD) bywas another step in this direction. It has been used as a reflective lens to explore technological affordances (through conceptual and empirical inquiry) as well as an action lens to create technological solutions (technological inquiry). While VSD’s core philosophy remained the same, it has been extended, often in response to its criticisms.

A criticism relevant to this study is practitioners’ ease in applying VSD in the industry contexts. VSD is perceived to have a relatively long turnaround time,often requiring specialists for implementation. To overcome these challenges, Shilton proposed `value levers’ , a low-cost entry point for value-oriented conversations while building technology artifacts in the organization. Value levers are open-ended activities that engage participants in value-oriented discourses to develop common ground. With creative representations, value levers can transform slow and cumbersome value conversations into creative and fruitful engagements. While previous studies have applied and shaped the notion of value levers in a very specific set of contexts, such as showcasing how designers employ them in their practices, this work shows a broader utility of value levers among a diverse set of practitioners while navigating the complex space of RAI value discourse.

Within AI/ML research, initial exploration of values were still primarily computational in nature, such as performance, generalizability, and efficiency. With the advent of HCI and critical studies focusing on discriminatory algorithmsand responsible AI , the discussions shifted to much broader values, such as societal and ethical values, within the AI/ML field. These studies focused on exposing inherent biases in the models due to absence of substantive social and ethical values. For instance,demonstrated how complex ML models have inherent interpretability issues stemming from a lack of transparency about how predictions were achieved. Another set of studies byandscrutinized several algorithms governing the digital infrastructures employed in our daily lives to expose discriminatory behaviors within different situations, especially in the context of fairness,against marginalized populations. In a similar vein, several studies have explored individual values that they felt were critical for models, such as fairness, explainability, non-malfeasance, and justice, reflecting societal norms. A common underlying factor among several of these studies was that they focused on individual values enacted in their own spaces. Recently, however, a few studies have adopted contrasting perspectives which argue that values do not exist in isolation, but often occupy overlapping and contested spaces. Our study aims to provide much-needed deeper insights within this complex space by showing how practitioners engage with and prioritize multiple values in a contested space.

Another value dimension explored is – “whose values should be considered while producing AI/ML algorithms?” Most studies have engaged with end-users values, lending a critical lens to the deployed models and their implications on society. These studies challenged whether developing fair algorithms should be primarily a technical task without considering end-users’ values. Subsequently, researchers leveraged action research (e.g., participatory approaches) to design toolkits, frameworks, and guidelines that accommodated end-user values in producing ML models.

A more relevant set of studies have recognized the importance of understanding the values that different practitioner roles embedded while producing responsible algorithms. Such practitioner-focused studies are critical in understanding “how” and “why” particular values are embedded in AI/ML models early on in the life cycle. However, these studies have explored particular practitioners’ values in silos, leaving much to be learned about their collective value deliberations. A nascent group of studies has answered this call. For examplefocused on controlled settings, where specific practitioners’ values could co-design a fairness checklist as one of their RAI values.explored a broader set of values of practitioners and compared it with end-users in an experimental setting. Another relevant study byhas explored RAI decision-making in an organizational setting, laying a roadmap to get from the current conditions to aspirational RAI practices.

Our study contributes to this developing literature in four ways. First, within the context of Responsible AI practices, our study goes beyond the scenario-based, controlled settings, or experimental setups by focusing on natural work settings, which echoes the sentiment of some of the previous open-ended qualitative studies that were conducted in the organizations, but not in the context of Responsible AI practices. Second, we focus on a diversity of stakeholder roles, who are making an explicit effort to recognize and incorporate RAI values, unlike siloed studies previously. Third, we leverage the lens of co-productionto study RAI values in natural work settings. Fourth, our study extendsby explicitly unpacking the co-production challenges deeply rooted in RAI values. To this end, we answer two research questions:

(RQ-1) : What challenges do AI/ML practitioners face when co-producing and implementing RAI values ?

(RQ-2) : In response, what strategies do practitioners use to overcome challenges as they implement RAI values ?

Co-production as a Lens

To answer our research questions, we employed the conceptual framework of co-production proposed by. She defined co-production as a symbiotic process in which collective knowledge and innovations produced by knowledge societies are inseparable from the social order that governs society. Jasanoff characterized knowledge societies broadly to include both state actors (e.g., governments) and non-state actors (e.g., corporations, non-profits) that have an enormous impact on the communities they serve. Studying co-production can help scholars visualize the relationship between knowledge and practice. Such a relationship offers new ways to not only understand how establishments organize or express themselves but also what they value and how they assume responsibility for their innovations.

To operationalize co-production in our study, we invoke three investigation sites, as Jasanoff proposed. The first site of exploration is the institutions containing different structures that empower or hinder individuals to co-produce. The second site examines different types of discourse that occur as part of the co-production activities. Solving technological problems often involve discourses producing new knowledge and linking such knowledge to practice. The last site of co-production is representations produced both during co-production to facilitate discourses and after co-production in the form of the end-product. The three sites of the co-production framework are appropriate for understanding the current industry challenges around RAI innovation for several reasons. Technological corporations developing AI/ML innovations have a robust bi-directional relationship with their end-user communities. Moreover, for successful RAI value implementation, practitioners need to leverage the complex structures within their organizations that are invisible to external communities. RAI value implementations occur through strategic discourses and deliberations that translate knowledge to effective execution. Lastly, in the process of RAI value deliberations, individuals co-create representations that further the implementation efforts of RAI.


To answer our research questions, we conducted a qualitative study consisting of 23 interviews with active AI/ML practitioners from 10 different organizations that engaged in RAI practices. After receiving internal ethics approval from our organization, we conducted a three month study (April-June, 2022). In this section, we briefly talk about the recruitment methods and participant details.

Participants Recruitment and Demographics

To recruit AI/ML practitioners who actively think and apply RAI values in their day-to-day work, we partnered with a recruitment agency that had strong ties with different types of corporate organizations working in the AI/ML space. We provided diverse recruitment criteria to the agency based on several factors, including gender, role in the company, organization size, sector, type of AI/ML project, and their involvement in different kinds of RAI activities. Using quota sampling technique, the agency advertised and explained the purpose of our study in diverse avenues, such as social media, newsletters, mailing lists, and internal forums of different companies. For the participants that responded with interest, the agency arranged a phone call to capture their AI/ML experience, as well as their experience with different RAI values. Based on the information, we shortlisted and conducted interviews with 23 AI/ML practitioners who fit the diverse criteria mentioned above. The aforementioned factors were used to prioritize diverse participants with experience working on RAI projects within their team in different capacities. For example, while shortlisting, we excluded students working on responsible AI projects as part of their internships and included individuals who were running startup RAI consultancy firms.

Out of the 23 practitioners, 10 identified themselves as women. Participants comprised of product-facing roles, such as UX designers, UX researchers, program/product mangers, content & support executives, model-focused roles, such as engineers, data scientists, and governance focused-roles, such as policy advisors and auditors. Out of 23 practitioners, all but one participant worked for a U.S. based organization. However, participants were geographically based in both Global North and Global South. Participants also worked in a wide variety of domains, including health, energy, social media, personal apps, finance and business among other, lending diversity to the captured experiences. Three participants worked for independent organizations that focused exclusively on RAI initiatives and AI governance. twelve participants had a technical background (e.g., HCI, computer-programming), four had business background, two had law background and one each specialized in journalism and ethics. For more details, please refer to Tabletab-org-participants.


We conducted semi-structured interviews remotely via video calls. Before the start of the each session, we obtained informed consent from the participants. We also familiarized participants with the objective of the study and explicitly mentioned the voluntary nature of the research. The interviews lasted between 40 minutes and 2 hours (avg.= 65 mins.) and were conducted in English. Interviews were recorded, if participants provided consent. Our interview questions covered different co-production practices. First, in order to understand different co-production challenges (RQ-1), we asked questions about (1) how practitioners faced challenges when sharing RAI values across roles (e.g., “Can you describe a situation when you encountered problems in sharing your values? ’’ ) and (2) how practitioners faced challenges when collaborating with different stakeholders (e.g., “What challenges did you face in your collaboration to arrive at shared common responsible values? ’’). Second, to understand different co-production strategies (RQ-2) we asked (3) how practitioners handled conflicts (e.g., “Can you give an example where you resisted opposing peers’ values? ’’) and (4) how practitioners sought assistance to achieve the alignment in RAI values (e.g., “What was the most common strategy you took to resolve the conflict ?’’). To invoke conversations around RAI values, we used a list of RAI values prepared byas an anchor to our conversations. After first few rounds of interviews, we revised the interview script to ask newer questions that provided deeper understanding to our research questions. We stopped our interviews once we reached a theoretical saturation within our data. We compensated participants with a 75$ gift-voucher for participation.

Data collection and Analysis

Out of 23 participants, only three denied permission to record audio. We relied on extensive notes for these users. Overall 25.5 hours of audio-recorded interviews (transcribed verbatim) and several pages of interview notes were captured. We validated accuracy of notes with the respective participants. Subsequently, we engaged in thematic analysis using the NVivo tool. We started the analysis by undertaking multiple passes of our transcribed data to understand the breadth of the interviewee’s accounts. During this stage, we also started creating memos. Subsequently, we conducted open-coding on the transcribed data while avoiding any preconceived notions, presupposed codes, or theoretical assumptions, resulting in 72 codes. We finalized our codes through several iterations of merging the overlapping codes and discarding the duplicate ones. To establish validity and to reduce bias in our coding process, all the authors were involved in prolonged engagement over multiple weeks. Important disagreements were resolved through peer-debriefing. The resultant codebook consisted of 54 codes. Example codes included, social factors’, prior experience’, enablers’, RAI pushback’.As a final step, we used an abductive approachto further map, categorize, and structure the codes under appropriate themes. To achieve this, we used three key instruments of co-production framework developed by, namely, making institutions , making discourses , and making representations . Examples of the resultant themes based on the co-production instruments included value ambiguity’, exploration rigidity’, value conflicts’, and value lever strategies’. Based on the instruments of co-production framework, we have present our resultant findings in the next section.


Our overall findings are divided based on different sites of exploration proposed by. The first section answers RQ-1 by exploring several institutional challenges that hinder the co-production of RAI values among practitioners (Section section-1). The second section explores subsequent knock-on challenges that unstable institutional structures create in co-production discourses (Section section-2). The last section answers the RQ-2 by presenting carefully thought out representations that overcome challenges deliberation and execution of RAI values using the concept of value levers. (Section section-3).

RAI Value Challenges within the Institutional Structures

Institutional structures are essential in enabling co-production of new knowledge. It is these structures that facilitate relationships for deliberation, standardize democratic methods, and validate safety of new technological systems before information is disseminated into the society. We found two key institutional structures that facilitated deliberation around RAI values within AI/ML companies. These structures brought about different RAI challenges.

Bottom-up: Burdened Vigilantes The First type of structures were bottom-up. Within these structures, RAI conversations developed through RAI value sharing in the lower echelons of organizations, often within AI/ML practitioners’ own teams. In our interviews, eight practitioners, namely a UX researcher, designer, content designer, and program manager from two mid-size organizations and two large-size organizations experienced or initiated bottom-up practices that engaged with RAI values. One of the enablers for such bottom-up innovation was individuals’ sense of responsibility towards producing AI/ML models that did not contribute to any harm in society. A few other practitioners paid close attention to `social climate ’ (e.g., LGBTQ month, hate speech incidents) to elicit particular RAI values. For instance, P08, a program manager in a large-scale technology company took responsibility for RAI practices in their team but soon started supporting team members to come together and share RAI values:

We cater to projects that are very self-determined, very bottom-up aligned with our values and priorities within the organization …These are what I call responsible innovation vigilantes around the company. I also started that way but have grown into something more than that. You’ll see this at the product or research team level, where somebody will speak up and say, `Hey, I want to be responsible for these RAI values, make this my job and find solutions’. So you start to see individuals in different pockets of the company popping up to do RAI stuff.

A summary of co-production activities mapped to's co-production sites, along with the themes, RAI values invoked, and key findings and takeaways.fig-1

A summary of co-production activities mapped to's co-production sites, along with the themes, RAI values invoked, and key findings and takeaways.

A key challenge with such bottom-up structures was that the responsibility of engaging with RAI value conversations implicitly fell on a few individual “vigilantes ”. They had to become stalwarts of particular RAI values and take out substantial time out of their work to encourage and convince their teams to engage with RAI values. They also actively seeked out RAI programs available within and outside their organization. When such RAI programs were not available, individuals took it upon themselves to create engagement opportunities with other members within the organization. These bottom-up structures were useful in breaking the norms of “boundary-work” that are often set within AI and similar technical organizational work where only engineers and high-officials in the company maintain control. It allowed non-technical roles such as user experience researchers, product managers, analysts, and content designers to a create safe space and lead the RAI efforts. While such efforts early on in AI/ML lifecycle minimized the potential harm of their ML models or AI products, it often happened at the cost of their overworked jobs.

Bottom-up: Burdened with Educational Efforts Apart from self-motivated vigilantes, the burden of RAI value exploration also fell on a few practitioners who were implicitly curious about RAI innovation. Unlike the vigilantes, these participants were pushed to become the face of their team’s RAI initiatives since there was no-one else who would. P14, a product manager working for two years at a medium size company within the energy sector shared,

When I came in to this team, nobody really believed in it [RAI] or they really didn’t think it [RAI] was important. I was personally interested so I was reading about some of these principles …When there was an indication of a future compliance requirement, people didn’t want to take up this additional work …somebody had to do it.

Similarly P05, a technical manager leading their team on data collection for the development of knowledge graphs, revealed how they were considered “the face of privacy ” for the team. Therefore, P05 was expected to foster awareness and common understanding among internal stakeholders and external partners and ensure they strove towards similar RAI standards and appreciated data-hygiene practices (e.g., data cleaning and de-identification). Practitioners like P14 and P05 had to assume the responsibility of figuring out the RAI space by presenting their team’s needs and asking formative questions even when their objectives around RAI were often not clear, such as which values to consider (e.g., “privacy or transparency? ”), what certain values mean (e.g., “what trustworthiness as an RAI value should mean to the model and the team ”), how to operationalize specific values (e.g., “How does trustworthiness apply to rule-based models? What kind of RAI values to invoke while collecting data?’’ , and how to interpret outcomes and map them on to their team’s objectives.

Participants (n=5) shared how leading such RAI initiatives burdened their professional lives in various ways. Multiple participants reported that the RAI field was still in its infancy and taking up responsibilities in such conditions meant that their efforts were not deemed a priority or sometimes even officially recognized as AI/ML work. Consequently, the practitioners possessed limited understanding of the direction to take to educate their team, convert their efforts into tangible outcomes, and effectively align their educational outcomes to the team’s objectives. P13, an RAI enthusiast and an engineer at a large-scale social media company shared how their RAI effort seemed like an endless effort, “At this point, I probably know more about what things (RAI values) we don’t want in it (model) than what we do want in it …It’s like I am just learning and figuring out what’s missing as I take every step …It is unclear which [RAI] direction will benefit the the team.” Moreover, the burden of educating multiple team members was on the shoulders of a very few practitioners tantamounting to substantial their paper around technology ethics put forward the term _ethic owners_ '. This role shares similarity with the bottom-up vigilantes and the educators as they both are motivated and self-aware practitioners, invested in foregrounding human values by providing awareness and active assistance while institutionalizing the processes. However, Metcalf's ethic owners' responsibilities were clearly defined. Their tasks of working with teams or higher management were perceived as visible, prioritized work for which they would be credited for career growth or otherwise.While bottom-up informal roles in our research performed similar set of tasks, their efforts were seen as tangential,administrative’, and underappreciated. It is not just that there was an additional burden, but even the outcomes of taking on this additional burden for bottom-up informal roles were dissimilar to the ethic owners. Taking up informal RAI work was more problematic when the requirements in the later stages of ML were unprompted, compelling practitioners to focus on these efforts at expense of their own work.

In our findings, one form of the need came as academic criticism or critique around particular values that were seen concerning a particular product (e.g., “what steps are you taking to ensure that your model is equitable?’’ ). Another form of need came from end-users’ behavior who experienced the models through a particular product. P20, a user experience researcher working with deep learning models in finance, shared how user feedback brought about new RAI needs that became their responsibility:

Once the users use our product and we see the feedback, it makes us realize, oh, people are sometimes using this feature in an unintended way that might in turn impact the way we are going about certain values, say transparency' …. Initially we were like, We should strive for transparency by adding a lot of explanations around how our model gave a particular output’. Later we realized too many explanations [for transparency] fostered inappropriate trust over the feature…UXR represents user needs so its on me to update the team on the issues and suggest improvements.

A few practitioners (n=2) also mentioned how the constant juggling between their own role-based work and the unpredictability of the RAI work pushed them to give-up the RAI responsibilities all-together.

Top-down: Rigidity in Open-discovery While the burden of ownership and execution of RAI values in bottom-up structures were on small group of individuals, they had the flexibility to choose RAI values that were contextual and mattered to their team’s ML models or projects. On the contrary, we found that top-down institutional structures limited the teams’ engagement to “key ” RAI values that impacted organization’s core business values. For instance, P15’s company had trust as a key value baked into their business, requiring P15 to focus on RAI values that directly reduced specific model’s biases, thereby increasing the company’s trust among their users. Consequently, several RAI practitioners had to skip RAI value exploration and sharing. Instead they directly implemented predetermined RAI values by the management just before the deployment. P06, an engineer at a large tech company working in conversational analysis models, described this lack of choice:

To be honest, I imagine lots of the conversations, around the sort of values that need to go into the model, happened above my pay grade. By the time the project landed on my desk to execute, the ethics of it was cleared and we had specific values that we were implementing.

Public-oriented legal issues and ethical failures, especially when launching innovative models (e.g., transformer networks), also determined the RAI values that were prioritized and the subsequent formal RAI structures that were established by the organizations. P19, a policy director at a RAI consultation firm facilitating such structures, shared how such impromptu structures were quite common in response to ever-changing laws around AI governance:

Even if you’re conservative, the current climate is such that it’s going to be a year or two max from now, where you will start to have an established, robust regulatory regime for several of these (RAI) issues. So a good way to be prepared is to create the [RAI] programs in whatever capacity that enables companies to comply with the new regulations, even if they are changing because if you have companies doing Responsible AI programs, it eventually gets compliance and executive buy-in.

Instead of investing in RAI structures to comply with different regulations in Tech Ethics such as codes of ethics, statements of principle, checklists, and ethics training as meaningful, organizations perceive them as instruments of risk that they have to mitigate. In line with the previous literature, our findings indicate that practitioners often find false solace in such structures as they run the risk of being superficial and relatively ineffective in making structures and practices accountable and effective in their organizations. However, adding nuance to this argument in the case of RAI practices, we found that practitioners more broadly devoted time and energy to follow established and prioritized values (e.g., trust or privacy) due to directed and concerted focus. It allowed for organization-wide impact since the “buy-in” already existed.

Top-down: Under-developed Centralized Support However, in the case of less clearly defined values (e.g., non-maleficence or safety) we observed a limited scope for nuance and despite best efforts, the centralized concerted direction does not always pan out as intended. Further, while laws continue to evolve in this space, participants felt that pre-mediated RAI values might not longitudinally satisfy the growing complexity of the ML models being implemented (e.g., multimodal models). Hence, while it might seem that setting up a centralized top-down approach might be efficient, the current execution leaves much to be desired. In fact, based on data from over half the participants, we found that five top-down structured companies integrated lesser known RAI values into their workflows in multiple ways without establishing a centralized workflow. Those who did establish centralize workflows created consulting teams to advise on RAI Practices (similar to ethic owners).

However, these top-down centralized RAI consulting teams were not always set up to succeed. As is the nature of consulting, people did not always know the point-of-contact or when and how to reach out. The consulting teams needed to also consider opportunities to advertise about themselves and engagement mechanisms, which was difficult due to the lack of context and nuance around the teams’ projects. Consequently, it was difficult for such teams to generate organic interest, unless the teams were already aware of their RAI requirements and knew a point of contact. P10, a manager who facilitated one such top-down RAI program in a large-scale technology company for AI/ML teams, described lack of fixed ways in which teams engaged with them on RAI values, making it a difficult engagement:

We have a bunch of internal Web pages that point you in all different types of directions. We don’t have a singular voice that the individuals can speak with …. Its currently hodgepodge. Some teams come to us willingly. They had already thought about some harms that could occur. They say, `Here’s our list of harms, here’s some ideas on what we want to do’ They’d already done pre-work and are looking for some feedback. Other teams come to us because they’ve been told to. …They haven’t thought much about RAI and need longer conversations …Other teams were told to go track down an individual or team because they are doing ML stuff that will require RAI assistance, but they don’t know about us ’’

Challenges within RAI value Discourses

Fruitful co-production requires well-established institutional structures that can empower stakeholders to engage in stable democratic discourses with an aim of knowledge production. In the previous section, we uncovered different structural challenges at the institutional level that contributed to knock-on effects, creating further challenges for practitioners during the co-production and implementation of RAI values. Discourse: Insufficient RAI Knowledge A key challenge that many practitioners experienced in co-producing RAI values in team was the difficulty in engaging deeply with new and unfamiliar RAI values deemed important by the team members. P07, a policy advisor in a large technology company, who regularly interacted with those who implemented RAI values, described the “superficial engagement ” with values as an act of “ineffective moralizing ”, wherein practitioners struggled to develop deeper interpretations of the team’s shared values and contextualize them in relation to the ML models they were developing.

P07 mentioned several key critical thinking questions that AI/ML practitioners did not deliberate within their teams, such as “Is this RAI value applicable to our product? ”, “how does this value translate in diverse use cases? ”, or “should this value be enabled through the product? ” The need for deeper engagement becomes particularly important in a high-stakes situation, such as healthcare, where certain conditions have an unequal impact on particular demographics. P12 experienced this complexity while overseeing the development of a ML model focused on health recommendations:

So a lot of models are geared towards ensuring that we have we are predictive about a health event and that almost always depends on different clinical conditions. For example, certain ethnic groups can have more proclivity to certain health risks. So, if your model is learning correctly, it should make positive outcomes for this group more than the other groups. Now, if you blindly apply RAI values without thinking deeply about the context, it might seem that the model is biased against this group when in reality these group of people are just more likely to have this condition, which is a correct conclusion, not a biased one.

Such deeper analysis requires hands-on practice and contextual training in the field and formal RAI education. In our findings, top-down structures were only effective to fill the gap for key values that aligned with the company’s vision, leaving a much needed requirement for contextual, high-quality RAI education for more emergent RAI values that could be modularized for specific teams. P02, a content designer for a large health technology company shared how this gap looked like for their team that was designing content for a machine translation team,

“One thing that would have been beneficial is if I or my team could somehow get more insights on how to think about trustworthiness in the context of the content produced by our machine translation model and probably evaluate it …Often time, I just go to someone who is known to have done some work in this [RAI] and say, `Hey, we want to design and publish the content for the model like in a month from now, what is bare minimum we could do from [RAI] principles point of view?’…Sometimes it’s never-ending because they say I have not thought about this at all and that it is going to take a month or maybe much more longer to get these principles implemented.”

Participants like P02 had no alternative but to reach out to their bottom-up structures to seek assistance, discuss, and reduce gaps in their RAI knowledge. On occasions, such avenues of discussion were non-conclusive. Prior literature in AI/ML and organization studies have shown how such unequal dependence on bottom-up structures over top-down in deliberation can contribute to tensions, and in turn propose an “open, federated system ” linking different actors, resources, and institutions to provide a community based support.

Discourse: Deprioritized Unfamiliar & Abstract Values Naturally, practitioners tried to solve the superficial engagement problem by de-prioritizing values that they found unfamiliar. In our study, most practitioners (n=18) said that they were familiar and comfortable talking about RAI values like privacy and security as they were already “established ” and had “matured over time ”. They sharply contrasted this perceived familiarity with other RAI values like explainability and robustness . The familiar values were well backed with stable top-down structures and dedicated teams, such as compliance departments and dedicate RAI personnel , making it easy for practitioners to develop mental models of deeper engagement. P20 shared their experience in this regard in their organization:

The ideal situation would be like, Oh, I have certain RAI principles that I want to make sure our product has or addresses'. In reality not all the principles are thought out the same way and applied in the first go. It usually happens in layers. First and foremost, people will look at privacy because that's super established, which means everyone knows about it, they already have done probably some work around it, so its easy to implement. And then after that, they're like, Okay, now let’s look at fairness or explainability’ …We usually have to be quick with turnaround like one or two months. Its nice to bring up values that are new but naturally they also require familiarizing and implementation effort within the team and people see that

Other practitioners (n=3) also followed a similar de-prioritization process for RAI values that they felt were abstract and did not have a measurement baseline (benchmarks) as opposed RAI values that could be easily measured quantitatively against a baseline. An example observed in this category was the contrast between RAI values like interpretability, which had concrete implementation techniques and measurements (e.g., LIME) and non-maleficence, which did not have a clear implementation technique or measurements. Similarly, practitioners (n=2) who went out of their way to understand and suggest new interpretability techniques for model debugging techniques (e.g., Integrated Gradients, SHAP) found it disempowering when their team members often negotiated for easier and computationally cheaper values like accuracy (e.g., P/E ratio) for implementation.

Discourse: Value Interpretation Tensions Even in situations, when different practitioners took a similar balanced approach to prioritization, tensions emerged as different roles interpreted and contextualized the RAI values differently during the value deliberations. We found these tensions occurring within practitioners when different practitioners defined RAI values (e.g., equity) and mapped them to RAI features and metric (e.g., skin tone) differently. P18, a senior data scientist leading an AI team in a non-profit institute, shared one such similar tension among their team members working on the same project,

Probably half of my colleagues do believe that there is a cultural, and historical set of RAI values that can be applied to all the products organization wide. Other half are vehemently opposed to that concept and say that [RAI] values are always model and project dependent. So if you are talking about our long-term goal to establish a set of RAI principles, whose perspectives should be considered?…This is an uneasy space that needs careful navigation.

While deliberations might occur between team-members, they might occur within a practitioner, or between the team and the end-consumers of the product/service. Latter two usually surfaced with user-facing roles, e.g., Product Managers or UX Researchers. These roles have the responsibility to understand, internalize, and embody end-user values in addition to their own values. Overall, we found that practitioners in these roles had to invest more effort to tease out their own values from that of the end-users. P04 was a user experience researcher working on interfacing a large language model for natural conversations with users. While P04 was interested in eliciting better insights from the model’s behavior issues (i.e. interpretability), end-users were interested in a simplified understanding of the model’s opaque behavior (i.e. comprehensibility). A UX Researcher is, however, expected to be the voice of the user in the process. Consequently, they had the constant burden to elicit both sets of values appropriately.

Another set of tensions also occurred between practitioners and end-users . P22, an analyst in a financial firm, described how ML practitioners perceived RAI values to be mutable and negotiable, allowing them to implement a particular RAI value in stages instead of all at once. Such a process allowed P22 (and three other participants who reported similar narratives) to build the required experience and embed the value in the ML model or AI product. However, end-users expected these embedded RAI values as absolute and non-negotiable and not on a “sliding spectrum ” because they are “they are often the list of ignored rights ”, leading to practitioner-user RAI tensions.

Our findings show that tensions that arose from non-uniform RAI value knowledge and subsequent disparate value interpretations were unproductive and a significant obstacle in the overall co-production process of RAI values. This can be attributed to a nascent RAI field that has given rise to new forms of values (e.g., explainability, interpretability) whose definitions and contexts which keep changing. This is in contrast with prior value studies in HCI studies where the tensions and conflicts around relatively more established values (e.g., privacy) do not occur until the implementation stage. Our findings show that the majority of value tensions occur much earlier in the value interpretation stage, often contributing to the abandonment of the value discussions altogether.

Implementation: RAI Values and Conflicts within Implementation of RAI values was also not a straight forward process as implementing certain RAI values created conflict with other RAI values. For instance, P1, an engineer working on classification models in VR environments shared how their decision to improve accuracy by excluding instances of objects with sensitive cultural meanings (e.g., objects with LGBTQ references) also had direct repercussions on the diversity and fairness of the model. Implementing RAI values also created cascading dependencies on the inclusion of other RAI values. For instance, P16, a program manager working as an RAI facilitator for a big tech company, shared the issues team members experienced around cascading RAI values:

One common issue I see is with teams that are dealing with model Fairness issues. Most often the solution for them is to improve their datasets or sometimes even collect new forms of demographic data to retrain their model and that opens up another rabbit hole around privacy that the team now has to navigate through and ensure that their data adhere to our privacy standards. More often than not, teams don’t even realize they are creating a new issue while trying to solve their existing problem.

Implementation challenges also occurred when organization’ business values were in tension with those of external clients. In such cases, the team’s commitment to engage with RAI was at odds with clients’ business priorities. P02, a technical program manager for a service company that developed ML models for clients in the energy sector, had a similar issue when their team was building a model for street light automation. After P02’s team received the data and started developing the model, they pushed for the value of safety. However, it was at odds with the company’s value of efficiency,

We should prioritize model optimization in those areas where there are higher crime rates …we don’t want blackouts, right? …Their argument was if there was a very high crime rate, such areas will also have high rate of purposefully damaging the lighting infrastructure. Prioritizing service to such areas will only create high amounts of backlogs as people will just vandalize it again. …So they just have different priorities. After that, our team just stopped following it up as it went into the backlog.

P02’s team gave up RAI value deliberation and implementationaltogether after their clients either deprioritized their multiple attempts to make them RAI ready or took an extremely long time to approve their requests.

Implementation: Unexpected Late-stage Value Changes Another challenge practitioners faced was encountering new RAI values during late stages of implementation. These values were not initially shortlisted. Instead, they were brought out later and sometimes championed by a very vocal practitioner, who felt deeply about it. Such late-stage RAI values also became a part of the discussion when practitioners in the team uncovered last-moment issues (e.g., bias) during implementation that significantly impacted the model. Several participants (n=3) shared how such late-stage RAI values decreased the productivity of their overall RAI discourse and implementation efforts, leading to a negative experience. While such last-minute changes are not welcomed, P12, an engineer shared how it also gives an opportunity to the developers to ship a better product before any harm might have been done. This tension between potentially better outcomes and slower implementation was visible in how the implementation efforts and timelines were impacted.

Such values also disrupted a planned implementation by taking the spotlight and pushing the team in navigating the company’s non-standardized approvals, thereby significantly altering the project timeline. For example, P23, an ML engineer shared how when they received issues around fairness from other stakeholders, it meant “substantial changes to the model from the ground-up, because most of the time, issues with fairness stem from the data ”. It meant revisiting the data and redoing data collection or further debugging to remove the issues. Moreover, when new and untested RAI values assumed prominence (e.g., interpretability), more time and effort was required from the practitioners during implementation. RAI facilitators are essential in easing the tension in such situations by engaging in back-and-forth conversations with the teams to reduce the effort, streamline the process, and help practitioners appreciate the eventual consequences of implementing the RAI values.

Implementation: Perceived Misuse of RAI values Lastly, we also observed tensions between individual efforts in implementing RAI values and their organization’s use of such efforts for the overall business purposes. For instance, P15, research director of a large-scale technology company overseeing research in large language models, shared how he was actively supporting a few teams in his company to co-produce and embed explainability into their models. However, he also expressed his concern about how companies could misrepresent such embedded RAI values,

I worry that explainable AI is largely an exercise in persuasion. This is why you should trust our software' rather than This is why our software is trustworthy’ …I’m not saying everybody who does explainable AI is doing that kind of propaganda work, but it’s a risk. Why do we want our AI to be explainable? Well, we’d like people to accept it and use it …Explainability part is ethically complicated …even for explainability for the practitioners the company wants it to be explainable, transparent, reliable, all those things as a means to an end. And the end is `please like our model, please buy our software’

We found two more practitioners who raised similar concerns with other RAI values, such as privacy and trust. They were concerned that making their product “completely responsible ” could enable companies to market their products as nearly perfect, leading to overtrust and over-reliance. These findings align with the ethics-washing phenomenon within the tech ethics literature which argues that companies sometimes invest in ethics teams and infrastructure, adopting the language of ethics to minimize external controversies and superficiallyengage with the proposed regulations. Practitioners who expressed these sentiments were quite dissatisfied with their RAI implementation work as they felt their actions were merely a “band-aid” solution for the organization, instead of meaningfully altering organization’s culture and practices.

Representational Strategies to Mitigate RAI Challenges

In response to the challenges mentioned in the aforementioned sections, we saw several strategies used by the practitioners to overcome the limitations in RAI value co-production. To present the strategies, we use a form of representations called values levers, a set of activities that facilitate opportunities to share and collaborate around values. We show how different practitioners use value levers to build upon current RAI institutional structures and make their RAI co-productionmanageable. In an ideal situation, value levers can also be employed in any situation of RAI co-production. For example, organizations created several formal RAI structures for practitioners to facilitate sharing and deliberation of values. These included top-down standardized guidelines, such as guidebooks (e.g., PAIR, HAX) around established RAI values, bringing in experts to share their experiences around co-production (lectures), and enabling shared spaces for co-production. However, in this section, we will be looking at value levers specifically developed in response to the challenges experienced in RAI value co-production.

Institutional Value Levers: External Expertise and Certifications to reduce Ambivalence One of the ways in which organizations brought stability to their inconsistent top-down RAI institutional structures was by taking assistance of independent agencies or professionals who specialized in establishing values levers that helped streamline their existing structures. One such value lever was `Responsible AI certifications ’ that were designed to bring different recognized and informal RAI co-production activities under one-roof. These programs act as facilitators between the non-technical and technical workforce by enabling co-production around RAI values to make them compliant with upcoming regulations. Participants reported that different activities were packaged into the RAI certification program, such as getting buy-in for particular RAI values, leveraging trusted partners for running impact assessments, engaging key actors in value discovery and prioritization, and implementing appropriate RAI methods. P19, policy director of one such RAI certification organization, shared how these certifications are effective in sectors, such as energy, mining, and human resources that often have a limited technology workforce. They described effort of facilitating RAI value conversations within their client teams as a key part of the certification process:

It is important to have everybody on board for those [RAI] value conversations. So we try really hard to have all the different teams like internal or external audit, legal, business, data and AI team come together, brainstorm, discuss different [RAI] issues in specific contexts and shortlist [the RAI values], even if we just get a little bit of their time …everyone needs to be brought in early because we conduct a lot of activities likes audit analysis, bias testing…it saves time, addresses several concerns, and establish streamlined [RAI] processes. …For simplicity, we just package all of the different activities we do under RAI certification. …Some times few activities are already being executed by the organization, we just do the job of aligning them in a way that works for the organization.

Such external expertise and certifications can provide an opportunity for open discovery, bolster existing centralized support, and identify RAI values that might otherwise be discovered at the last stages.

Institutional Value Levers: Activities to Distribute RAI burden We also found several nascent but more focused value levers in bottom-up institutions focused on distributing the burden experienced by a few roles more widely within the team. These value levers provided opportunities for increased participation from stakeholders, especially in the starting stages by enabling them to bring-in complementary RAI values into the team. Most commonly used levers in this context included scenario-based narratives and role plays, and open-ended activities that engaged practitioners in opinion formation and sharing. Other value levers included conducting a literature review of specific RAI values and applicable cutting-edge methods, definitions, and guidelines around them to share and invoke feedback from the team. We also observed more experimental value levers that were geared towards bringing-in complementary RAI values of external stakeholders (e.g., end-users) into the team.

For example, P18, a data scientist working in a startup, hosted a panel to capture complementary perspectives around AI explainability. Visibility into how explainability was perceived differently by different community members, such as NGOs and government, contributed to a better understanding and alignment within the team to develop explainable models. In a similar example, P09, an engineer working on a reinforcement learning model in the context of healthcare for low resources in India, facilitated field visits to the end-user communities. Such exposure helped roles that were passive in sharing their values as well as roles that were thinking about new values, such as social justice, in the RAI discourse. Overall, these value levers (narratives, role-plays, literature reviews, panels, and field visits) focused on primarily bottom-up structures, which helped reduce pressure on specific roles and limit superficial value engagements.

Discourse Value Levers: Facilitating Disagreements Moving our focus to RAI value co-production, we saw user-facing practitioners create explicit opportunities for disagreements and healthy conflicts to tackle the problem of superficial value engagement and improve the quality of their teams’ deliberations. Disagreements in co-production phase allowed practitioners like UX researchers and product managers to think inclusively, capture diverse perspectives and expert knowledge, and more importantly predict future value conflicts. For example, P04, a UX researcher, created bottom-up adversarial prioritization framework. In the starting phases of this framework, the UX researcher pushed team members to go broad and co-produce values by wearing other practitioner’s hats and invoking their RAI values. This practice allowed them to bring forward interesting disagreements between different roles that were then resolved and prioritized to achieve a small set of meaningful RAI values. P04 recalled two of the values that received maximum disagreements were diversity and inclusion. Wearing complementary user hats enabled practitioners to familiarize with these values that were otherwise unfamiliar in their own roles. Other top-down RAI programs also facilitated similar structures, explicitly providing narratives that brought out disagreements, e.g.,:

Usually I will write something in the prompts that I think that the team absolutely needs to hear about but is controversial and opposing. But what I do is I put it in the voice of their own team so that it is not coming from us. It is not us scrutinizing them. That promotes interpersonal negotiation that pushes individuals to really defend their values with appropriate reasoning.

According to P19, having such RAI system in place early also allows companies to judge its ML models benchmarks when compared to their competition. Leveraging the adversarial prioritization framework appropriately in both top-down and bottom-up structures can enable open-discovery, and surface the values and related conflicts for resolution.

Discourse Value Levers: Model Cards & Visual Tools to Reduce Abstractness from Values We found that practitioners also created succinct representations and simplified documentation to bring much needed clarity to various RAI values and simply associated models. For instance, engineers shared documentation of model and data cards, making it easier for non-engineering and engineering roles to grasp the information. P23, a senior engineer at an AI startup looking into sustainability, shared the process:

`Even we have introduced this concept of a model card, wherein if a model is developed, the model card has to be filled out. So what is a model card? A model card is a series of questions that captures the basic facts about a model at a model level at an individual model level. What did you use to build a model? What was the population that was used? What is the scoring population? It is like, having all of that in a centralized standard format. Goes a long way to roll it up because the product can be very complex as well, right? With multiple players and whatnot. But having that information collected in this way benefits other roles that own the product to think about different values that are missing '

UI/UX designers, UX researchers, and analysts also used similar documentation tools to initiate discussions and receive feedback from other practitioners in the team. P20, a UX researcher, used presentation slides containing model features to facilitate brainstorming sessions and receive feedback from other roles. They also repurposed tools and methods used in their own work to give shape to their peers’ abstract values. For example, P20 reused online jam-boards containing key RAI values and user findings for affinity diagramming, enabling the team to “categorize the findings and map them to specific RAI values ”. Other RAI levers in this category included designing and sharing infographics and regular RAI standups where practitioners took it upon themselves to be stewards of RAI principles for the team to give updates, receive feedback and learn about team’s perspectives on specific RAI values.

Implementation Value Levers: Office Hours, User stories, Safe Spaces to Reduce Tensions

A few value levers that were part of top-down RAI programs were also effective in reducing various value tensions that occurred between different practitioners(n=2). One such program was RAI office hours that was available for elicitation and production, but were also extremely effective for tension-resolution. A typical office hour was 30 minutes in which practitioners engaged with a relevant expert and an experienced facilitator. One key way experts solved the tensions in these sessions was by collecting and providing concrete case-study examples. For example, P21, an RAI office hour facilitator, shared an example about the use of his office hours. The practitioners were at odds with each other in implementing explainability and trustworthy features. During the office hours, P21 responded by sharing an edge case scenario where even good explanations might backfire, such as, “If a pregnant woman had a miscarriage, showing even good end-user explanations around why they are seeing infant-related content can be very problematic. Explainability should be carefully teased out based on the context in which it is applied.”

Another set of value levers used especially by the roles facing end-users were user stories and scenarios to influence and persuade users to change their value priorities and align with rest of the team. These levers were also used by the roles to converge on key values after engaging in healthy conflicts within the divergence phase. For example, P04, exposed different pain points and key user journeys by “highlighting the clip of a user that is really, really amazing story that is either very painful or poignant ”. Interestingly, P04 was aware how such value levers had to be evoked carefully,

“If that story is not representative, I’m manipulating the system. If it is representative, I’m influencing the system…I will have to be careful not operate on the side of manipulation and try to be very squarely on the side of influence. So, I do like regular checks for myself to make sure that I am operating on influence, not manipulation, in terms of the stories that I am allowing people to amplify.”

Lastly, in order to tackle several types of value conflicts in the co-production of RAI values, we found different RAI value levers that focused on improving value alignment. One key alignment strategy was to create structures and activities that aligned the team’s RAI values pretty early in the process. One such activity that we saw in both practitioner’s and RAI programs was providing a safe space to encourage open discussions among individuals to empathize with other members. P09 shared,

One alignment strategy was open discussion with the safe space, where team members could fail, be called out and to learn from each other as we were developing values. So say someone finds the value of democratization really important, they are made to articulate what they mean by it.…. It is easy if there are different buckets in which they can categorize and explain because then people can easily surface all the different ways they think and prioritize values and that helps with alignment

Discussion and Future Work

Overall, our findings show that co-production of RAI values in practice is complicated by institutional structures that either support top-down decision-making by leadership or are inhabited by bottom-up practitioners exercising voluntary agency (section section-1). In other case, multiple challenges exist when practitioners have to reconcile within their internally held values and RAI values expected from their roles; and between themselves and their team-members. Our findings also show that discourse around alignment and prioritization of RAI values can sometimes be unproductive, non-conclusive, and disempowering when practitioners have to implement said RAI values (section section-2). We observed a lack of transparency, and unequal participation within organizations; and between organizations and end-users of their products/services (section section-2). Despite the relatively complicated lay of the land, practitioners have been pushing ahead and discovering multiple strategies on how to make progress (section section-3). In the subsections below we will unpack these challenges, strategies and potential future work across the three sites of co-production: institutions, discourse, and representations.

Envisioning balanced Institutions: Middle-out RAI Structures

Inequity in Burden According to, strong institutions provide a stable environment for effective knowledge co-production. They can also act as safe spaces for nurturing and transforming contested ideas to effective practices leading to long-lasting impact on the immediate ecosystem. Recent scholarship byhas put faith in an aspirational future where organizations would have deployed strong institutional frameworks for RAI issues. Our findings show that as of today, top-down structures are underdeveloped. Organizations have deployed structures that range from being reactive to external forces (e.g., compliance, public outcry) by tracking teams that implement RAI to proactively establishing structures that make teams RAI-ready (e.g., office hours). Furthermore, stable workflows have been established for a limited number of values or use-cases, restricting the number of teams that could leverage such workflows.

Being in the midst of restrictive structures, particular practitioner roles embraced the persona of bottom-up vigilantes and self-delegated themselves to be champions of lesser known RAI values (e.g., non-maleficence and trustworthiness). They initiate open-ended exploration for value discourses and subsequent value implementation. However, such bottom-up structures also put burden and occupational stress on selective roles, risking the implementation success of such RAI projects. In particular, we have found that roles like UX researchers, designers, product managers, project managers, ethicists have been taking the brunt of this work. These findings build on the previous work, highlighting existing inequity and subsequent individual activism being performed by some - either by volition or due to lack of alternatives. Enabling Equal Participation Going forward, there is a need for a holistic middle-out approach that seeks a balanced synergy between top-down and bottom-up structures while balancing for the challenges that each of the structures provide. For instance, organizations can work with RAI value stalwarts and champions to formalize and streamline bottom-up workflows, making it a standard practice for all teams to engage in open-ended exploration of RAI values. Such a process can enable teams to look beyond loosely applicable organization-recommended RAI values and shortlist those values that actually matter and apply to their team. To standardize the structure, organizations can leverage independent (flat) team/roles that can guide the target team through the team while giving enough room for exploration.

Organizations can also use a middle-out approach to reduce the burden and occupational stress on specific roles through several top-down activities. One such way is to facilitate structures that can lower the barrier for diverse internal stakeholders to engage in RAI value co-production, regardless of their proximity to AI products or ML models. For instance, data-workers and teams/roles that do internal testing of the models (dogfooding) can contribute to the RAI value co-production. Same structures can also enable engagement with external stakeholders, such as end-user communities, policy, experts, and governance agencies in the initial stages of value co-production. Consequently,practitioners’ chances to foresee or anticipate changing requirements could improve, especially in the later stages of AI/ML lifecycle. Better yet, this could potentially improve not just the “user experience” of value discourse, but also the efficiency of implementation - a goal valued by private companies. This could be a win-win situation for multiple stakeholders by helping the top-down RAI structures align with business goals. While, our research uncovered only top-down and bottom-up structures that were mutually exclusive, other structures might exist. For example, while we envisage middle-out structures to be advantageous, future research is needed to operationalize and simulate such structures; and discover existing implementations. There might be some challenges uniquely inherent in those structures. We encourage future researchers to continue this line of enquiry

Envisioning better Discourses: Enabling Critical RAI Value Deliberations

Negativity in Deliberations The ultimate aim of co-production discourse is to engage in competing epistemological questions. Jasanoff calls it interactional co-productionbecause it deals with explicitly acknowledging and bringing conflicts between two competing orders: scientific order brought about by technological innovations and social order brought about by prevalent sociocultural practices within the community. In our findings, several underlying conflicts surfaced between scientific and social order (section section-2). In one instance, practitioners had to choose between socially impactful but lesser explored RAI values (social order) and lesser applicable but established values with measurable scientific benchmark (scientific order). In another instance of tension an RAI value occupied competing social spaces (e.g., equity). The underlying issue was not the conflicts itself but lack of systematic structures that enabled positive conflict resolution around the conflict. Such implicit conflicts often met with deprioritization and conversations ended unresolved. There is an urgent need to transform such implicit conflicts to explicit positive deliberations. Organizations aiming for successful RAI co-production need to be more reflexive, mobilize resources to create safe spaces and encourage explicit disagreements among practitioners positively, enable them to constantly question RAI values or co-production procedures that help embed them. While we saw some instances of these explicit practices in the form of value lever strategies, such instances were sporadic and localized to very few teams.

Acknowledging Differences Safely Our findings around challenges within RAI value discourse also showcase the politically charged space that RAI values occupy in an organization. In their critical piece around implementation of values in organization,bring out a few value characteristics that are applicable to our study. First, values are not universal. Values are recognized, prioritized, and embedded differently based on a bunch of factors, such as practitioner’s role, organizational priorities and business motivations which make RAI as a complex space. In our findings, roles prioritized those RAI values that were incentivized by the organizations and the AI/ML community through computational benchmarks focused on the RAI value outcomes. This is problematic as the RAI values that have computational benchmarks and have implicit organizational incentives might not map onto the RAI issues that are pertinent in the community. One way to solve this mismatch is by rethinking the definition of benchmarks from value outcome to value processes taken up by different roles (or teams). For example, organizations can encourage teams to document their co-production journeys around lesser known RAI values that can act as RAI benchmarks.

The second issue thatbring out is whose value interpretations should be prioritized and considered? In our findings, tensions emerged as same values were interpreted and prioritized differently by different stakeholders. End-users viewed RAI values as immutable and uncompromisable, whereas practitioners viewed them as flexible and iterable. Similar tensions were also observed between theinternal stakeholders in contextualizing the same values. While we saw a few strategic value levers, such as office hours, they were mainly targeted at the stakeholders that were within the same team. Extending this line of value levers, we propose participatory approaches that take a balanced approach for both internal and external stakeholders. In particular, we take inspiration from participatory design fictions , a participatory approach using speculative design scenarios, to find alignment on polyvocal speculations around embedded values in the emergent technologies. Deliberations around the fiction narratives can be used to arrive at a common ground RAI value implementation that are both contextual and implementable.

Shaping Institutional Structures to Improve Discourse Employing co-production as a lens has also allowed us to reveal a very delicate, symbiotic relationship between the practitioners’ discourses and the institutional structures. Several of the discourse challenges observed in our findings, such as issues pertaining to the lack of RAI knowledge and deprioritization of lesser known values, stemmed not only from the lack of mature top-down structures but also dependency on a singular institutional structure. The limitations of top-down structures pushed several practitioners in-turn to create informal bottom-up structures, shifting the pressure from one structure to the other. We argue that RAI-focused organizations need to seek a balance between stability (top-down structures) and flexibility (bottom-up). In addition to new middle-out internal -facing institutional structures, RAI discourses can benefit from external -facing institutional structures that can help such discourses be impactful. This can be achieved by bringing in diverse external collaborators, such as NGOs, social enterprises, governmental institutes, and advocacy groups. It will also help avoid the issue of co-optationfor true meaningful impact of such structures on the practices.

Envisioning better Representations: Value Levers

Value Levers Enabling Progress As the third site of co-production, representations are essential manifestations of knowledge deliberations. They provide surrogate markers for progress on value alignments or even responses to tensions between stakeholders. In our study, we saw several value levers that were deployed within organizations at different stages to tackle co-production challenges. These value levers enabled progress during knowledge deliberations (section section-3). For instance, during deliberation, practitioners used role-specific value levers (e.g., visual tools by UX researchers & designers) as a window into their thought process around specific RAI values. Similarly, enabling safe spaces provided opportunities of RAI value alignment among different roles. As representations, these value levers improved the transitions between internal stages and acted as vital markers to improve alignment. We call them internal representations . Prioritization Framework employed to resolve value related conflicts across individuals was another example of internal representation. In contrast, we also found a few externally represented value levers, such as RAI certifications, that enabled organizations to reduce ambivalence in their structures while showing their RAI readiness and compliance to the outside community. Interestingly, we uncovered limited evidence of external representations that engaged with end-user communities directly. We posit that the lack of external representations can be attributed to the deprioritized perception of the end-users’ role in RAI value conversations. External representations have the potential to act as stable participatory structures that enable participation of external stakeholders, making RAI value co-production successful. This might also be due to lack of sufficient incentives to enable such participatory structures. We hope that recent progress made by UNESCO’s historic agreementmight provide the much needed push for organizations to share and learn together.

Value Levers Enabling Practitioners As we end this discussion, some readers might assume that we now have strategies to manage the multiple challenges coming our way as we deliberate and implement RAI values in products. Our research with 23 practitioners, points to the opposite - we are far from it. As one practitioner said, “It’s currently hodgepodge”. Multiple individuals and organizations are trying to surmount this incredibly complex challenge at institutional and individual level. While value levers discussed in this work were successful in helping practitioners make progress, the discovery of these value levers has at best been serendipitous. Sufficient support structures, skills training, and knowledge dissemination will be needed to enable practitioners to overcome these and unknown challenges yet. One can liken these value levers as a tool-belt for practitioners, as summarized in Table tab-strategies.

There is a subsequent need to design systems that can provide access to these tools in a systematic way. This would require centering research amongst practitioners who are making products. We hope that future educators, researchers, and designers will pursue this opportunity to uncover further challenges, learn from existing strategies, develop better strategies as they iterate, and create design systems to support practitioners. Further, we need to find more tools, and we need to share the tool-belt with other practitioners.

Limitations & Conclusions

Based on qualitative research with 23 AI/ML practitioners, we discovered several challenges in RAI value co-production. Due to their roles and individual value system, practitioners were overburdened with upholding RAI values leading to inequitable workload. Further, we found that implementing RAI values on-the-ground is challenging as sometimes these values conflict within, and with those of team-members. Owing to the nascent stage of RAI values, current institutional structures are learning to adapt. Practitioners are also adapting by discovering strategies serendipitously to overcome the challenges. However, more support is needed by educators to educate RAI/RAI values, researchers to unpack further challenges/strategies, and for community to focus on aiding AI/ML practitioners as we collectively find a common-ground. Our work also has several limitations. First, our overall study is shaped by our prior HCI research experience in responsible AI. While all the participants had rich experience working with ethics in the context of social good in both global north and global south contexts, we acknowledge that our perspectives around responsible AI and ethical values in AI might be constrained. Second, our insights are also limited by the overall methodological limitations of qualitative enquiry, such as sample size of 23 participants across 10 organizations. Future work is required to generalize our insights across different organizational sectors and ML models using mixed-methods approach.


   2  publisher = {Taylor \& Francis},
   3  year = {2000},
   4  pages = {124--130},
   5  number = {3},
   6  volume = {39},
   7  journal = {Theory into practice},
   8  author = {Creswell, John W and Miller, Dana L},
   9  title = {Determining validity in qualitative inquiry},
  13  url = {},
  14  doi = {10.1037/10819-000},
  15  publisher = {University of Chicago Press},
  16  year = {1956},
  17  author = {Morris, Charles},
  18  title = {Varieties of human value.},
  22  year = {2016},
  23  pages = {288--302},
  24  booktitle = {Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work \& Social Computing},
  25  author = {Goyal, Nitesh and Fussell, Susan R},
  26  title = {Effects of sensemaking translucence on distributed collaborative analysis},
  30  url = {},
  31  pdf = {},
  32  publisher = {PMLR},
  33  month = {23--24 Feb},
  34  series = {Proceedings of Machine Learning Research},
  35  volume = {81},
  36  editor = {Friedler, Sorelle A. and Wilson, Christo},
  37  year = {2018},
  38  pages = {134--148},
  39  booktitle = {Proceedings of the 1st Conference on Fairness, Accountability and Transparency},
  40  author = {Chouldechova, Alexandra and Benavides-Prado, Diana and Fialko, Oleksandr and Vaithianathan, Rhema},
  41  title = {A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions},
  45  series = {CHI '22},
  46  location = {New Orleans, LA, USA},
  47  keywords = {crowdsourcing app, mobile crowdsourcing, crowd work, crowdsourcing work, HCI4D, Women},
  48  numpages = {18},
  49  articleno = {298},
  50  booktitle = {Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems},
  51  doi = {10.1145/3491102.3501834},
  52  url = {},
  53  address = {New York, NY, USA},
  54  publisher = {Association for Computing Machinery},
  55  isbn = {9781450391573},
  56  year = {2022},
  57  title = {Feeling Proud, Feeling Embarrassed: Experiences of Low-Income Women with Crowd Work},
  58  author = {Varanasi, Rama Adithya and Siddarth, Divya and Seshadri, Vivek and Bali, Kalika and Vashistha, Aditya},
  62  publisher = {Open university press},
  63  year = {1999},
  64  author = {MacKenzie, Donald and Wajcman, Judy},
  65  title = {The social shaping of technology},
  69  url = {},
  70  doi = {10.1007/978-94-007-7844-3_4},
  71  isbn = {978-94-007-7844-3},
  72  pages = {55--95},
  73  address = {Dordrecht},
  74  publisher = {Springer Netherlands},
  75  year = {2013},
  76  booktitle = {Early engagement and new technologies: Opening up the laboratory},
  77  title = {Value Sensitive Design and Information Systems},
  78  editor = {Doorn, Neelke
  79and Schuurbiers, Daan
  80and van de Poel, Ibo
  81and Gorman, Michael E.},
  82  author = {Friedman, Batya
  83and Kahn, Peter H.
  84and Borning, Alan
  85and Huldtgren, Alina},
  89  month = {Nov},
  90  year = {2020},
  91  author = {González, Felipe and Ortiz, Teresa and Sánchez Ávalos, Roberto},
  92  publisher = {Inter-American Development Bank},
  93  doi = {10.18235/0002876},
  94  url = {},
  95  title = {Responsible use of AI for public policy: Data science toolkit},
  99  publisher = {Mit Press Cambridge, MA},
 100  year = {1992},
 101  pages = {225--258},
 102  volume = {1},
 103  journal = {Shaping technology/building society: Studies in sociotechnical change},
 104  author = {Latour, Bruno and others},
 105  title = {Where are the missing masses? The sociology of a few mundane artifacts},
 109  numpages = {20},
 110  url = {},
 111  pages = {21–40},
 112  publisher = {Cambridge University Press},
 113  year = {1997},
 114  number = {72},
 115  author = {Friedman, Batya},
 116  title = {Human values and the design of computer technology},
 120  howpublished = {\url{}},
 121  year = {2020},
 122  month = {September},
 123  title = {Recommendation on the Ethics of Artificial Intelligence},
 124  author = {UNESCO},
 128  howpublished = {\url{}},
 129  year = {2022},
 130  title = {Responsible AI practices},
 131  author = {Google},
 135  howpublished = {\url{ }},
 136  year = {2021},
 137  month = {June},
 138  title = {Facebook’s five pillars of Responsible AI},
 139  author = {Facebook},
 143  howpublished = {\url{}},
 144  year = {2022},
 145  month = {September},
 146  title = {AI Ethics},
 147  author = {IBM},
 151  howpublished = {\url{}},
 152  year = {2022},
 153  title = {Microsoft Responsible AI Standard (version 2)},
 154  author = {Microsoft},
 158  howpublished = {\url{}},
 159  year = {2021},
 160  month = {November},
 161  title = {Responsible AI Initiative},
 162  author = {Defence Innovation Unit},
 166  howpublished = {\url{}},
 167  year = {2019},
 168  month = {April},
 169  title = {Ethics guidelines for trustworthy AI},
 170  author = {European Union},
 174  howpublished = {\url{}},
 175  year = {2021},
 176  month = {November},
 177  title = {193 countries adopt first-ever global agreement on the Ethics of Artificial Intelligence},
 178  author = {UN},
 182  language = {en},
 183  pages = {1–3},
 184  year = {2021},
 185  author = {Friedman, Batya and Harbers, Maaike and Hendry, David G. and van den Hoven, Jeroen and Jonker, Catholijn and Logler, Nick},
 186  journal = {Ethics and Information Technology},
 187  number = {1},
 188  doi = {10.1007/s10676-021-09585-z},
 189  url = {},
 190  volume = {23},
 191  title = {Introduction to the special issue: value sensitive design: charting the next decade},
 195  language = {en},
 196  pages = {39–44},
 197  year = {2021},
 198  author = {Hendry, David G. and Friedman, Batya and Ballard, Stephanie},
 199  journal = {Ethics and Information Technology},
 200  number = {1},
 201  doi = {10.1007/s10676-021-09579-x},
 202  url = {},
 203  volume = {23},
 204  title = {Value sensitive design as a formative framework},
 208  doi = {10.1017/CBO9780511498725.017},
 209  url = {},
 210  year = {2008},
 211  pages = {24},
 212  volume = {322},
 213  journal = {Information technology and moral philosophy},
 214  author = {Flanagan, Mary and Howe, Daniel C and Nissenbaum, Helen},
 215  title = {Embodying values in technology: Theory and practice},
 219  language = {en},
 220  pages = {100205},
 221  month = {Feb},
 222  year = {2021},
 223  author = {Birhane, Abeba},
 224  journal = {Patterns},
 225  number = {2},
 226  doi = {10.1016/j.patter.2021.100205},
 227  url = {},
 228  issn = {26663899},
 229  volume = {2},
 230  title = {Algorithmic injustice: a relational ethics approach},
 234  bibsource = {dblp computer science bibliography,},
 235  biburl = {},
 236  timestamp = {Sat, 30 Jan 2021 18:02:51 +0100},
 237  eprint = {2101.09869},
 238  eprinttype = {arXiv},
 239  url = {},
 240  year = {2021},
 241  volume = {abs/2101.09869},
 242  journal = {CoRR},
 243  title = {Black Feminist Musings on Algorithmic Oppression},
 244  author = {Lelia Marie Hampton},
 248  bibsource = {dblp computer science bibliography,},
 249  biburl = {},
 250  timestamp = {Mon, 13 Aug 2018 16:47:25 +0200},
 251  eprint = {1609.07236},
 252  eprinttype = {arXiv},
 253  url = {},
 254  year = {2016},
 255  volume = {abs/1609.07236},
 256  journal = {CoRR},
 257  title = {On the (im)possibility of fairness},
 258  author = {Sorelle A. Friedler and
 259Carlos Scheidegger and
 260Suresh Venkatasubramanian},
 264  series = {FAccT ’22},
 265  location = {Seoul, Republic of Korea},
 266  keywords = {algorithmic systems, data-driven governance, accountability theory, algorithmic accountability},
 267  booktitle = {Proceedings of the 2022 Conference on Fairness, Accountability, and Transparency},
 268  doi = {10.1145/3531146.3533083},
 269  url = {},
 270  address = {New York, NY, USA},
 271  publisher = {Association for Computing Machinery},
 272  year = {2022},
 273  title = {The Values Encoded in Machine Learning Research},
 274  author = {Abeba Birhane and
 275Pratyusha Kalluri and
 276Dallas Card and
 277William Agnew and
 278Ravit Dotan and
 279Michelle Bao},
 283  series = {CHI '22},
 284  location = {New Orleans, LA, USA},
 285  keywords = {data quality, ML, field workers, domain expertise, labour studies, data collection, deskilling, Taylorism},
 286  numpages = {14},
 287  articleno = {587},
 288  booktitle = {Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems},
 289  doi = {10.1145/3491102.3517578},
 290  url = {},
 291  address = {New York, NY, USA},
 292  publisher = {Association for Computing Machinery},
 293  isbn = {9781450391573},
 294  year = {2022},
 295  title = {The Deskilling of Domain Expertise in AI Development},
 296  author = {Sambasivan, Nithya and Veeraraghavan, Rajesh},
 300  language = {en},
 301  pages = {689–707},
 302  month = {Dec},
 303  year = {2018},
 304  author = {Floridi, Luciano and Cowls, Josh and Beltrametti, Monica and Chatila, Raja and Chazerand, Patrice and Dignum, Virginia and Luetge, Christoph and Madelin, Robert and Pagallo, Ugo and Rossi, Francesca and Schafer, Burkhard and Valcke, Peggy and Vayena, Effy},
 305  journal = {Minds and Machines},
 306  number = {4},
 307  doi = {10.1007/s11023-018-9482-5},
 308  url = {},
 309  issn = {0924-6495, 1572-8641},
 310  volume = {28},
 311  title = {AI4People—An Ethical Framework for a Good AI Society: Opportunities, Risks, Principles, and Recommendations},
 315  language = {en},
 316  pages = {3},
 317  month = {Dec},
 318  year = {2019},
 319  author = {Ghallab, Malik},
 320  journal = {AI Perspectives},
 321  number = {1},
 322  doi = {10.1186/s42467-019-0003-z},
 323  url = {},
 324  issn = {2523-398X},
 325  volume = {1},
 326  title = {Responsible AI: requirements and challenges},
 330  series = {ICML '09},
 331  location = {Montreal, Quebec, Canada},
 332  numpages = {8},
 333  pages = {129–136},
 334  booktitle = {Proceedings of the 26th Annual International Conference on Machine Learning},
 335  doi = {10.1145/1553374.1553391},
 336  url = {},
 337  address = {New York, NY, USA},
 338  publisher = {Association for Computing Machinery},
 339  isbn = {9781605585161},
 340  year = {2009},
 341  title = {Multi-View Clustering via Canonical Correlation Analysis},
 342  author = {Chaudhuri, Kamalika and Kakade, Sham M. and Livescu, Karen and Sridharan, Karthik},
 346  url = {},
 347  pdf = {},
 348  publisher = {PMLR},
 349  month = {10--15 Jul},
 350  series = {Proceedings of Machine Learning Research},
 351  volume = {80},
 352  editor = {Dy, Jennifer and Krause, Andreas},
 353  year = {2018},
 354  pages = {1407--1416},
 355  booktitle = {Proceedings of the 35th International Conference on Machine Learning},
 356  author = {Espeholt, Lasse and Soyer, Hubert and Munos, Remi and Simonyan, Karen and Mnih, Vlad and Ward, Tom and Doron, Yotam and Firoiu, Vlad and Harley, Tim and Dunning, Iain and Legg, Shane and Kavukcuoglu, Koray},
 357  title = {{IMPALA}: Scalable Distributed Deep-{RL} with Importance Weighted Actor-Learner Architectures},
 361  year = {2008},
 362  volume = {21},
 363  journal = {Advances in neural information processing systems},
 364  author = {Jain, Prateek and Kulis, Brian and Dhillon, Inderjit and Grauman, Kristen},
 365  title = {Online metric learning and fast similarity search},
 369  series = {CHI '21},
 370  location = {Yokohama, Japan},
 371  keywords = {Social Good, AI, India, Qualitative, HCI4D, Healthcare},
 372  numpages = {21},
 373  articleno = {598},
 374  booktitle = {Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems},
 375  doi = {10.1145/3411764.3445130},
 376  url = {},
 377  address = {New York, NY, USA},
 378  publisher = {Association for Computing Machinery},
 379  isbn = {9781450380966},
 380  year = {2021},
 381  title = {AI in Global Health: The View from the Front Lines},
 382  author = {Ismail, Azra and Kumar, Neha},
 386  series = {CHI '21},
 387  location = {Yokohama, Japan},
 388  keywords = {HCI4D, mHealth, Community health worker, ICTD, CHW, Artificial Intelligence, AI},
 389  numpages = {20},
 390  articleno = {701},
 391  booktitle = {Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems},
 392  doi = {10.1145/3411764.3445420},
 393  url = {},
 394  address = {New York, NY, USA},
 395  publisher = {Association for Computing Machinery},
 396  isbn = {9781450380966},
 397  year = {2021},
 398  title = {“It Cannot Do All of My Work”: Community Health Worker Perceptions of AI-Enabled Mobile Health Applications in Rural India},
 399  author = {Okolo, Chinasa T. and Kamath, Srujana and Dell, Nicola and Vashistha, Aditya},
 403  series = {AIES '18},
 404  location = {New Orleans, LA, USA},
 405  keywords = {accountable ai, deontology, kantian ai, explainable ai, artificial intelligence ethics, machine ethics, autonomous machine ethics, modal logic},
 406  numpages = {7},
 407  pages = {130–136},
 408  booktitle = {Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society},
 409  doi = {10.1145/3278721.3278753},
 410  url = {},
 411  address = {New York, NY, USA},
 412  publisher = {Association for Computing Machinery},
 413  isbn = {9781450360128},
 414  year = {2018},
 415  title = {Toward Non-Intuition-Based Machine and Artificial Intelligence Ethics: A Deontological Approach Based on Modal Logic},
 416  author = {Hooker, John N. and Kim, Tae Wan N.},
 420  language = {en},
 421  pages = {341–362},
 422  year = {2022},
 423  author = {Häußermann, Johann Jakob and Lütge, Christoph},
 424  journal = {AI and Ethics},
 425  number = {2},
 426  doi = {10.1007/s43681-021-00047-2},
 427  url = {},
 428  volume = {2},
 429  title = {Community-in-the-loop: towards pluralistic value creation in AI, or—why AI needs business ethics},
 433  series = {FAccT '22},
 434  location = {Seoul, Republic of Korea},
 435  keywords = {value-sensitive design, empirical ethics, Responsible AI},
 436  numpages = {14},
 437  pages = {310–323},
 438  booktitle = {2022 ACM Conference on Fairness, Accountability, and Transparency},
 439  doi = {10.1145/3531146.3533097},
 440  url = {},
 441  address = {New York, NY, USA},
 442  publisher = {Association for Computing Machinery},
 443  isbn = {9781450393522},
 444  year = {2022},
 445  title = {How Different Groups Prioritize Ethical Values for Responsible AI},
 446  author = {Jakesch, Maurice and Bu\c{c}inca, Zana and Amershi, Saleema and Olteanu, Alexandra},
 450  copyright = { perpetual, non-exclusive license},
 451  year = {2020},
 452  publisher = {arXiv},
 453  title = {Algorithmic Fairness in Education},
 454  keywords = {Computers and Society (cs.CY), Artificial Intelligence (cs.AI), Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences},
 455  author = {Kizilcec, René F. and Lee, Hansol},
 456  url = {},
 457  doi = {10.48550/ARXIV.2007.05443},
 461  numpages = {14},
 462  pages = {1–14},
 463  booktitle = {Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems},
 464  url = {},
 465  address = {New York, NY, USA},
 466  publisher = {Association for Computing Machinery},
 467  isbn = {9781450367080},
 468  year = {2020},
 469  title = {Keeping Community in the Loop: Understanding Wikipedia Stakeholder Values for Machine Learning-Based Systems},
 470  author = {Smith, C. Estelle and Yu, Bowen and Srivastava, Anjali and Halfaker, Aaron and Terveen, Loren and Zhu, Haiyi},
 474  series = {CHI '21},
 475  location = {Yokohama, Japan},
 476  keywords = {mechanism design, value sensitive design, student assignment},
 477  numpages = {14},
 478  articleno = {589},
 479  booktitle = {Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems},
 480  doi = {10.1145/3411764.3445748},
 481  url = {},
 482  address = {New York, NY, USA},
 483  publisher = {Association for Computing Machinery},
 484  isbn = {9781450380966},
 485  year = {2021},
 486  title = {Modeling Assumptions Clash with the Real World: Transparency, Equity, and Community Challenges for Student Assignment Algorithms},
 487  author = {Robertson, Samantha and Nguyen, Tonya and Salehi, Niloufar},
 491  series = {FAccT '22},
 492  location = {Seoul, Republic of Korea},
 493  numpages = {12},
 494  pages = {440–451},
 495  doi = {10.1145/3531146.3533110},
 496  url = {},
 497  address = {New York, NY, USA},
 498  publisher = {Association for Computing Machinery},
 499  isbn = {9781450393522},
 500  year = {2022},
 501  title = {The Model Card Authoring Toolkit: Toward Community-Centered, Deliberation-Driven AI Design},
 502  author = {Shen, Hong and Wang, Leijie and Deng, Wesley H. and Brusse, Ciell and Velgersdijk, Ronald and Zhu, Haiyi},
 506  language = {en},
 507  pages = {99–120},
 508  month = {Mar},
 509  year = {2020},
 510  author = {Hagendorff, Thilo},
 511  doi = {10.1007/s11023-020-09517-8},
 512  url = {},
 513  issn = {0924-6495, 1572-8641},
 514  volume = {30},
 515  title = {The Ethics of AI Ethics: An Evaluation of Guidelines},
 519  copyright = {Creative Commons Attribution 4.0 International},
 520  year = {2019},
 521  publisher = {arXiv},
 522  title = {Impossibility and Uncertainty Theorems in AI Value Alignment (or why your AGI should not have a utility function)},
 523  keywords = {Artificial Intelligence (cs.AI), FOS: Computer and information sciences, FOS: Computer and information sciences},
 524  author = {Eckersley, Peter},
 525  url = {},
 526  doi = {10.48550/ARXIV.1901.00064},
 530  language = {en},
 531  pages = {719–735},
 532  month = {Jun},
 533  year = {2019},
 534  author = {van Wynsberghe, Aimee and Robbins, Scott},
 535  journal = {Science and Engineering Ethics},
 536  number = {3},
 537  doi = {10.1007/s11948-018-0030-8},
 538  url = {},
 539  issn = {1353-3452, 1471-5546},
 540  volume = {25},
 541  title = {Critiquing the Reasons for Making Artificial Moral Agents},
 545  language = {eng},
 546  collection = {Artificial Intelligence foundations, theory, and algorithms},
 547  year = {2019},
 548  author = {Dignum, Virginia},
 549  publisher = {Springer},
 550  isbn = {9783030303716},
 551  title = {Responsible artificial intelligence: how to develop and use AI in a responsible way},
 552  series = {Artificial Intelligence foundations, theory, and algorithms},
 553  address = {Cham},
 557  series = {CHI '17},
 558  location = {Denver, Colorado, USA},
 559  keywords = {mHealth, virtue ethics, psychological wellbeing, human flourishing, maternal mental health, phronesis, ethical design},
 560  numpages = {49},
 561  pages = {2708–2756},
 562  booktitle = {Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems},
 563  doi = {10.1145/3025453.3025918},
 564  url = {},
 565  address = {New York, NY, USA},
 566  publisher = {Association for Computing Machinery},
 567  isbn = {9781450346559},
 568  year = {2017},
 569  title = {MHealth for Maternal Mental Health: Everyday Wisdom in Ethical Design},
 570  author = {Barry, Marguerite and Doherty, Kevin and Marcano Belisario, Jose and Car, Josip and Morrison, Cecily and Doherty, Gavin},
 574  keywords = {values work, ethics, values, soft resistance, ethics work, work practice, UX practice, values in design, UX professionals},
 575  numpages = {28},
 576  articleno = {355},
 577  month = {oct},
 578  journal = {Proc. ACM Hum.-Comput. Interact.},
 579  doi = {10.1145/3479499},
 580  url = {},
 581  number = {CSCW2},
 582  volume = {5},
 583  address = {New York, NY, USA},
 584  publisher = {Association for Computing Machinery},
 585  issue_date = {October 2021},
 586  year = {2021},
 587  title = {Tactics of Soft Resistance in User Experience Professionals' Values Work},
 588  author = {Wong, Richmond Y.},
 592  language = {en},
 593  pages = {157–171},
 594  year = {2019},
 595  editor = {Isotani, Seiji and Millán, Eva and Ogan, Amy and Hastings, Peter and McLaren, Bruce and Luckin, Rose},
 596  author = {Holstein, Kenneth and McLaren, Bruce M. and Aleven, Vincent},
 597  publisher = {Springer International Publishing},
 598  booktitle = {Artificial Intelligence in Education},
 599  doi = {10.1007/978-3-030-23204-7_14},
 600  url = {},
 601  isbn = {9783030232030},
 602  volume = {11625},
 603  title = {Designing for Complementarity: Teacher and Student Needs for Orchestration Support in AI-Enhanced Classrooms},
 604  address = {Cham},
 608  bibsource = {dblp computer science bibliography,},
 609  biburl = {},
 610  timestamp = {Wed, 17 Jul 2019 10:27:36 +0200},
 611  eprint = {1907.04534},
 612  eprinttype = {arXiv},
 613  url = {},
 614  year = {2019},
 615  volume = {abs/1907.04534},
 616  journal = {CoRR},
 617  title = {The Role of Cooperation in Responsible {AI} Development},
 618  author = {Amanda Askell and
 619Miles Brundage and
 620Gillian K. Hadfield},
 624  language = {en},
 625  pages = {59–64},
 626  month = {Nov},
 627  year = {2018},
 628  author = {Awad, Edmond and Dsouza, Sohan and Kim, Richard and Schulz, Jonathan and Henrich, Joseph and Shariff, Azim and Bonnefon, Jean-François and Rahwan, Iyad},
 629  journal = {Nature},
 630  number = {7729},
 631  doi = {10.1038/s41586-018-0637-6},
 632  url = {},
 633  issn = {0028-0836, 1476-4687},
 634  volume = {563},
 635  title = {The Moral Machine experiment},
 639  numpages = {13},
 640  pages = {253–265},
 641  booktitle = {Android Epistemology},
 642  address = {Cambridge, MA, USA},
 643  publisher = {MIT Press},
 644  isbn = {0262061848},
 645  year = {1995},
 646  title = {The Ethics of Autonomous Learning Systems},
 647  author = {Khan, A. F. Umar},
 651  language = {en},
 652  month = {Jan},
 653  year = {2021},
 654  author = {Dignum, Virginia},
 655  journal = {London Review of Education},
 656  number = {1},
 657  doi = {10.14324/LRE.19.1.01},
 658  url = {},
 659  issn = {1474-8479},
 660  volume = {19},
 661  title = {The role and challenges of education for responsible AI},
 665  publisher = {MIT Press},
 666  year = {1994},
 667  author = {James Gips},
 668  title = {Toward the Ethical Robot},
 669  editor = {Kenneth M. Ford and C. Glymour and Patrick Hayes},
 670  booktitle = {Android Epistemology},
 671  pages = {243--252},
 675  year = {2017},
 676  author = {Abrassart, Christophe and Bengio, Yoshua and Chicoisne, Guillaume and De Marcellis-warin, Nathalie and Dilhac, Marc-Antoine and Gamb, Sebastien and Gautrais, Vincent and Giber, Martin And Langlois, Lyse And Laviolett, François},
 677  institution = {Université de Montréal},
 678  journal = {Montréal declaration for responsible development of artifcial intelligence},
 679  url = {},
 683  year = {2009,
 685  title = {Elinor Ostrom Prize Lecture},
 686  author = {Ostrom Elinor},
 690  series = {CHI '21},
 691  location = {Yokohama, Japan},
 692  keywords = {Human AI Interaction, Workflow, CDSS, Healthcare, Rural Clinic, Developing Country, Decision Making, Human AI Collaboration, AI, Clinical Decision Making, Future of Work, Trust AI, Collaborative AI, China, Implementation, AI Deployment},
 693  numpages = {18},
 694  articleno = {697},
 695  doi = {10.1145/3411764.3445432},
 696  url = {},
 697  address = {New York, NY, USA},
 698  publisher = {Association for Computing Machinery},
 699  isbn = {9781450380966},
 700  year = {2021},
 701  title = {“Brilliant AI Doctor” in Rural Clinics: Challenges in AI-Powered Clinical Decision Support System Deployment},
 702  author = {Wang, Dakuo and Wang, Liuping and Zhang, Zhan and Wang, Ding and Zhu, Haiyi and Gao, Yvonne and Fan, Xiangmin and Tian, Feng},
 706  language = {en},
 707  pages = {82–115},
 708  month = {Jun},
 709  year = {2020},
 710  author = {Barredo Arrieta, Alejandro and Díaz-Rodríguez, Natalia and Del Ser, Javier and Bennetot, Adrien and Tabik, Siham and Barbado, Alberto and Garcia, Salvador and Gil-Lopez, Sergio and Molina, Daniel and Benjamins, Richard and Chatila, Raja and Herrera, Francisco},
 711  journal = {Information Fusion},
 712  doi = {10.1016/j.inffus.2019.12.012},
 713  url = {},
 714  issn = {15662535},
 715  volume = {58},
 716  title = {Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI},
 720  publisher = {Frontiers Media SA},
 721  year = {2019},
 722  pages = {32},
 723  url = {},
 724  volume = {2},
 725  journal = {Frontiers in Big Data},
 726  author = {Moore, Jared},
 727  title = {AI for not bad},
 731  language = {en},
 732  pages = {277–284},
 733  month = {Sep},
 734  year = {2019},
 735  author = {Shin, Donghee and Park, Yong Jin},
 736  journal = {Computers in Human Behavior},
 737  doi = {10.1016/j.chb.2019.04.019},
 738  url = {},
 739  issn = {07475632},
 740  volume = {98},
 741  title = {Role of fairness, accountability, and transparency in algorithmic affordance},
 745  pages = {633},
 746  volume = {165},
 747  journal = {University of Pennsylvania Law Review},
 748  author = {Kroll, Joshua A and Huey, Joanna and Barocas, Solon and Felten, Edward W and Reidenberg, Joel R and Robinson, David G and Yu, Harlan},
 749  title = {Accountable algorithms’(2017)},
 753  language = {en},
 754  pages = {917–926},
 755  month = {Dec},
 756  year = {2020},
 757  author = {de Fine Licht, Karl and de Fine Licht, Jenny},
 758  journal = {AI \& SOCIETY},
 759  number = {4},
 760  doi = {10.1007/s00146-020-00960-w},
 761  url = {},
 762  issn = {0951-5666, 1435-5655},
 763  volume = {35},
 764  title = {Artificial intelligence, transparency, and public decision-making: Why explanations are key when trying to produce perceived legitimacy},
 768  year = {2019},
 769  volume = {17},
 770  booktitle = {Proceedings of the AI for Social Good workshop at NeurIPS},
 771  url = {},
 772  author = {Green, Ben},
 773  title = {Good” isn’t good enough},
 777  series = {FAccT '22},
 778  location = {Seoul, Republic of Korea},
 779  numpages = {12},
 780  pages = {2057–2068},
 781  booktitle = {2022 ACM Conference on Fairness, Accountability, and Transparency},
 782  doi = {10.1145/3531146.3533781},
 783  url = {},
 784  address = {New York, NY, USA},
 785  publisher = {Association for Computing Machinery},
 786  isbn = {9781450393522},
 787  year = {2022},
 788  title = {Disentangling the Components of Ethical Research in Machine Learning},
 789  author = {Ashurst, Carolyn and Barocas, Solon and Campbell, Rosie and Raji, Deborah},
 793  language = {en},
 794  pages = {141–163},
 795  month = {Mar},
 796  year = {2021},
 797  author = {Mitchell, Shira and Potash, Eric and Barocas, Solon and D’Amour, Alexander and Lum, Kristian},
 798  journal = {Annual Review of Statistics and Its Application},
 799  number = {1},
 800  doi = {10.1146/annurev-statistics-042720-125902},
 801  url = {},
 802  issn = {2326-8298, 2326-831X},
 803  volume = {8},
 804  title = {Algorithmic Fairness: Choices, Assumptions, and Definitions},
 808  series = {AIES '21},
 809  location = {Virtual Event, USA},
 810  keywords = {artificial intelligence for social good, participatory design, capabilities approach},
 811  numpages = {12},
 812  pages = {425–436},
 813  booktitle = {Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society},
 814  doi = {10.1145/3461702.3462612},
 815  url = {},
 816  address = {New York, NY, USA},
 817  publisher = {Association for Computing Machinery},
 818  isbn = {9781450384735},
 819  year = {2021},
 820  title = {Envisioning Communities: A Participatory Approach Towards AI for Social Good},
 821  author = {Bondi, Elizabeth and Xu, Lily and Acosta-Navas, Diana and Killian, Jackson A.},
 825  series = {COMPASS '19},
 826  location = {Accra, Ghana},
 827  keywords = {commodities, anomaly, time series, analysis, agriculture, prices},
 828  numpages = {13},
 829  pages = {52–64},
 830  doi = {10.1145/3314344.3332488},
 831  url = {},
 832  address = {New York, NY, USA},
 833  publisher = {Association for Computing Machinery},
 834  isbn = {9781450367141},
 835  year = {2019},
 836  title = {Price Forecasting \& Anomaly Detection for Agricultural Commodities in India},
 837  author = {Madaan, Lovish and Sharma, Ankur and Khandelwal, Praneet and Goel, Shivank and Singla, Parag and Seth, Aaditeshwar},
 841  series = {ICTD2020},
 842  location = {Guayaquil, Ecuador},
 843  numpages = {5},
 844  articleno = {24},
 845  doi = {10.1145/3392561.3397577},
 846  url = {},
 847  address = {New York, NY, USA},
 848  publisher = {Association for Computing Machinery},
 849  isbn = {9781450387620},
 850  year = {2020},
 851  title = {DIA: A Human AI Hybrid Conversational Assistant for Developing Contexts},
 852  author = {Cannanure, Vikram Kamath and Brown, Timothy X and Ogan, Amy},
 856  language = {en},
 857  pages = {44–65},
 858  month = {Jan},
 859  year = {2019},
 860  author = {Berendt, Bettina},
 861  journal = {Paladyn, Journal of Behavioral Robotics},
 862  number = {1},
 863  doi = {10.1515/pjbr-2019-0004},
 864  url = {},
 865  issn = {2081-4836},
 866  volume = {10},
 867  title = {AI for the Common Good?! Pitfalls, challenges, and ethics pen-testing},
 871  series = {CHI '22},
 872  location = {New Orleans, LA, USA},
 873  keywords = {future of work, AI labour, data annotation, qualitative study},
 874  numpages = {16},
 875  articleno = {582},
 876  booktitle = {Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems},
 877  doi = {10.1145/3491102.3502121},
 878  url = {},
 879  address = {New York, NY, USA},
 880  publisher = {Association for Computing Machinery},
 881  isbn = {9781450391573},
 882  year = {2022},
 883  title = {Whose AI Dream? In Search of the Aspiration in Data Annotation.},
 884  author = {Wang, Ding and Prabhat, Shantanu and Sambasivan, Nithya},
 888  bibsource = {dblp computer science bibliography,},
 889  biburl = {},
 890  timestamp = {Fri, 10 Jan 2020 13:10:19 +0100},
 891  eprint = {2001.01818},
 892  eprinttype = {arXiv},
 893  url = {},
 894  year = {2020},
 895  volume = {abs/2001.01818},
 896  journal = {CoRR},
 897  title = {Artificial Intelligence for Social Good: {A} Survey},
 898  author = {Zheyuan Ryan Shi and
 899Claire Wang and
 900Fei Fang},
 904  language = {en},
 905  pages = {501–507},
 906  month = {Nov},
 907  year = {2019},
 908  author = {Mittelstadt, Brent},
 909  journal = {Nature Machine Intelligence},
 910  number = {11},
 911  doi = {10.1038/s42256-019-0114-4},
 912  url = {},
 913  issn = {2522-5839},
 914  volume = {1},
 915  title = {Principles alone cannot guarantee ethical AI},
 919  copyright = {Creative Commons Attribution 4.0 International},
 920  year = {2022},
 921  publisher = {arXiv},
 922  title = {Towards a multi-stakeholder value-based assessment framework for algorithmic systems},
 923  keywords = {Machine Learning (cs.LG), Human-Computer Interaction (cs.HC), FOS: Computer and information sciences, FOS: Computer and information sciences},
 924  author = {Yurrita, Mireia and Murray-Rust, Dave and Balayn, Agathe and Bozzon, Alessandro},
 925  url = {},
 926  doi = {10.48550/ARXIV.2205.04525},
 930  eprint = {},
 931  url = { 
 934  doi = {10.1177/2053951715622512},
 935  year = {2016},
 936  pages = {2053951715622512},
 937  number = {1},
 938  volume = {3},
 939  journal = {Big Data \& Society},
 940  title = {How the machine ‘thinks’: Understanding opacity in machine learning algorithms},
 941  author = {Jenna Burrell},
 945  series = {CHI '09},
 946  location = {Boston, MA, USA},
 947  keywords = {fieldwork, values, empirical methods, methodology, value sensitive design, photo elicitation},
 948  numpages = {10},
 949  pages = {1141–1150},
 950  booktitle = {Proceedings of the SIGCHI Conference on Human Factors in Computing Systems},
 951  doi = {10.1145/1518701.1518875},
 952  url = {},
 953  address = {New York, NY, USA},
 954  publisher = {Association for Computing Machinery},
 955  isbn = {9781605582467},
 956  year = {2009},
 957  title = {Values as Lived Experience: Evolving Value Sensitive Design in Support of Value Discovery},
 958  author = {Le Dantec, Christopher A. and Poole, Erika Shehan and Wyche, Susan P.},
 962  pages = {91–104},
 963  year = {2015},
 964  author = {JafariNaimi, Nassim and Nathan, Lisa and Hargraves, Ian},
 965  journal = {Design Issues},
 966  number = {4},
 967  url = {},
 968  issn = {07479360, 15314790},
 969  volume = {31},
 970  title = {Values as Hypotheses: Design, Inquiry, and the Service of Values},
 974  language = {en},
 975  pages = {377–391},
 976  year = {1999},
 977  author = {Star, Susan Leigh},
 978  journal = {American Behavioral Scientist},
 979  number = {3},
 980  doi = {10.1177/00027649921955326},
 981  url = {},
 982  volume = {43},
 983  title = {The Ethnography of Infrastructure},
 987  series = {CHI '16},
 988  location = {San Jose, California, USA},
 989  keywords = {maintenance, repair, ethnography, values, design},
 990  numpages = {12},
 991  pages = {1403–1414},
 992  booktitle = {Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems},
 993  doi = {10.1145/2858036.2858470},
 994  url = {},
 995  address = {New York, NY, USA},
 996  publisher = {Association for Computing Machinery},
 997  isbn = {9781450333627},
 998  year = {2016},
 999  title = {Values in Repair},
1000  author = {Houston, Lara and Jackson, Steven J. and Rosner, Daniela K. and Ahmed, Syed Ishtiaque and Young, Meg and Kang, Laewoo},
1004  copyright = {Creative Commons Attribution 4.0 International},
1005  year = {2021},
1006  publisher = {arXiv},
1007  title = {Stakeholder Participation in AI: Beyond "Add Diverse Stakeholders and Stir"},
1008  keywords = {Artificial Intelligence (cs.AI), Computers and Society (cs.CY), Human-Computer Interaction (cs.HC), FOS: Computer and information sciences, FOS: Computer and information sciences},
1009  author = {Delgado, Fernando and Yang, Stephen and Madaio, Michael and Yang, Qian},
1010  url = {},
1011  doi = {10.48550/ARXIV.2111.01122},
1015  series = {CHI '22},
1016  location = {New Orleans, LA, USA},
1017  keywords = {Data work, valuation, public health, India},
1018  numpages = {16},
1019  articleno = {322},
1020  booktitle = {Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems},
1021  doi = {10.1145/3491102.3501868},
1022  url = {},
1023  address = {New York, NY, USA},
1024  publisher = {Association for Computing Machinery},
1025  isbn = {9781450391573},
1026  year = {2022},
1027  title = {When is Machine Learning Data Good?: Valuing in Public Health Datafication},
1028  author = {Thakkar, Divy and Ismail, Azra and Kumar, Pratyush and Hanna, Alex and Sambasivan, Nithya and Kumar, Neha},
1032  pages = {1–6},
1033  month = {Jun},
1034  year = {2018},
1035  author = {Vakkuri, Ville and Abrahamsson, Pekka},
1036  publisher = {IEEE},
1037  booktitle = {2018 IEEE International Conference on Engineering, 
1038Technology and Innovation (ICE/ITMC)},
1039  doi = {10.1109/ICE.2018.8436265},
1040  url = {},
1041  isbn = {9781538614693},
1042  title = {The Key Concepts of Ethics of Artificial Intelligence},
1043  address = {Stuttgart},
1047  language = {en},
1048  pages = {167–186},
1049  month = {Sep},
1050  year = {2012},
1051  author = {Timmermans, Stefan and Tavory, Iddo},
1052  journal = {Sociological Theory},
1053  number = {3},
1054  doi = {10.1177/0735275112457914},
1055  url = {},
1056  issn = {0735-2751, 1467-9558},
1057  volume = {30},
1058  title = {Theory Construction in Qualitative Research: From Grounded Theory to Abductive Analysis},
1062  language = {en},
1063  pages = {53},
1064  month = {Aug},
1065  year = {2021},
1066  author = {Harris, Jamie and Anthis, Jacy Reese},
1067  journal = {Science and Engineering Ethics},
1068  number = {4},
1069  doi = {10.1007/s11948-021-00331-8},
1070  url = {},
1071  issn = {1353-3452, 1471-5546},
1072  volume = {27},
1073  title = {The Moral Consideration of Artificial Entities: A Literature Review},
1077  language = {en},
1078  pages = {169–169},
1079  month = {Jul},
1080  year = {2020},
1081  author = {Kalluri, Pratyusha},
1082  journal = {Nature},
1083  number = {7815},
1084  doi = {10.1038/d41586-020-02003-2},
1085  url = {},
1086  issn = {0028-0836, 1476-4687},
1087  volume = {583},
1088  title = {Don’t ask if artificial intelligence is good or fair, ask how it shifts power},
1092  language = {en},
1093  pages = {659–684},
1094  month = {Dec},
1095  year = {2020},
1096  author = {Mohamed, Shakir and Png, Marie-Therese and Isaac, William},
1097  journal = {Philosophy \& Technology},
1098  number = {4},
1099  doi = {10.1007/s13347-020-00405-8},
1100  url = {},
1101  issn = {2210-5433, 2210-5441},
1102  volume = {33},
1103  title = {Decolonial AI: Decolonial Theory as Sociotechnical Foresight in Artificial Intelligence},
1107  language = {en},
1108  pages = {389–399},
1109  year = {2019},
1110  author = {Jobin, Anna and Ienca, Marcello and Vayena, Effy},
1111  journal = {Nature Machine Intelligence},
1112  number = {9},
1113  doi = {10.1038/s42256-019-0088-2},
1114  url = {},
1115  volume = {1},
1116  title = {The global landscape of AI ethics guidelines},
1120  copyright = { perpetual, non-exclusive license},
1121  year = {2020},
1122  publisher = {arXiv},
1123  title = {What If I Don't Like Any Of The Choices? The Limits of Preference Elicitation for Participatory Algorithm Design},
1124  keywords = {Computers and Society (cs.CY), Human-Computer Interaction (cs.HC), Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences},
1125  author = {Robertson, Samantha and Salehi, Niloufar},
1126  url = {},
1127  doi = {10.48550/ARXIV.2007.06718},
1131  numpages = {14},
1132  pages = {1–14},
1133  booktitle = {Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems},
1134  url = {},
1135  address = {New York, NY, USA},
1136  publisher = {Association for Computing Machinery},
1137  isbn = {9781450367080},
1138  year = {2020},
1139  title = {Co-Designing Checklists to Understand Organizational Challenges and Opportunities around Fairness in AI},
1140  author = {Madaio, Michael A. and Stark, Luke and Wortman Vaughan, Jennifer and Wallach, Hanna},
1144  keywords = {race and ethnicity, facial analysis, facial recognition, training and evaluation data, classification, gender, facial classification, computer vision, facial detection, identity},
1145  numpages = {35},
1146  articleno = {58},
1147  month = {may},
1148  journal = {Proc. ACM Hum.-Comput. Interact.},
1149  doi = {10.1145/3392866},
1150  url = {},
1151  number = {CSCW1},
1152  volume = {4},
1153  address = {New York, NY, USA},
1154  publisher = {Association for Computing Machinery},
1155  issue_date = {May 2020},
1156  year = {2020},
1157  title = {How We've Taught Algorithms to See Identity: Constructing Race and Gender in Image Databases for Facial Analysis},
1158  author = {Scheuerman, Morgan Klaus and Wade, Kandrea and Lustig, Caitlin and Brubaker, Jed R.},
1162  series = {FAT* '20},
1163  location = {Barcelona, Spain},
1164  keywords = {ethics, regulation, technology ethics, moral philosophy, technology law, self-regulation, AI},
1165  numpages = {10},
1166  pages = {210–219},
1167  booktitle = {Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency},
1168  doi = {10.1145/3351095.3372860},
1169  url = {},
1170  address = {New York, NY, USA},
1171  publisher = {Association for Computing Machinery},
1172  isbn = {9781450369367},
1173  year = {2020},
1174  title = {From Ethics Washing to Ethics Bashing: A View on Tech Ethics from within Moral Philosophy},
1175  author = {Bietti, Elettra},
1179  author = {Katie Shilton},
1180  pages = {107-171},
1181  number = {2},
1182  issn = {1551-3955},
1183  doi = {10.1561/1100000073},
1184  title = {Values and Ethics in Human-Computer Interaction},
1185  journal = {Foundations and Trends in Human–Computer Interaction},
1186  volume = {12},
1187  year = {2018},
1188  url = {},
1192  series = {C\&T '13},
1193  location = {Munich, Germany},
1194  keywords = {virtual community formation, values in design, internet architecture, technology ethics, ethnography},
1195  numpages = {10},
1196  pages = {110–119},
1197  booktitle = {Proceedings of the 6th International Conference on Communities and Technologies},
1198  doi = {10.1145/2482991.2482993},
1199  url = {},
1200  address = {New York, NY, USA},
1201  publisher = {Association for Computing Machinery},
1202  isbn = {9781450321044},
1203  year = {2013},
1204  title = {Making Space for Values: Communication Values Levers in a Virtual Team},
1205  author = {Shilton, Katie and Koepfler, Jes A.},
1209  language = {en},
1210  pages = {40–61},
1211  month = {Jan},
1212  year = {2019},
1213  author = {Chen, Jiawei and Hanrahan, Benjamin V. and Carroll, John M},
1214  journal = {International Journal of Mobile Human Computer Interaction},
1215  number = {1},
1216  doi = {10.4018/IJMHCI.2019010103},
1217  url = {},
1218  issn = {1942-390X, 1942-3918},
1219  volume = {11},
1220  title = {Withshare: A Mobile Application to Support Communzity Coproduction Activities},
1224  year = {1999},
1225  booktitle = {Workshop in Political Theory and Policy Analysis. Ann Arbor MI: University of Michigan Press},
1226  author = {Ostrom, Elinor},
1227  title = {Crossing the great divide. Co-production, synergy \& development, polycentric governance and development},
1231  language = {en},
1232  pages = {565–577},
1233  year = {2019},
1234  author = {Chen, Jiawei and Doryab, Afsaneh and Hanrahan, Benjamin V. and Yousfi, Alaaeddine and Beck, Jordan and Wang, Xiying and Bellotti, Victoria and Dey, Anind K. and Carroll, John M.},
1235  publisher = {Springer International Publishing},
1236  booktitle = {Information in Contemporary Society},
1237  doi = {10.1007/978-3-030-15742-5_54},
1238  url = {},
1239  isbn = {9783030157418},
1240  volume = {11420},
1241  title = {Context-Aware Coproduction: Implications for Recommendation Algorithms},
1242  address = {Cham},
1246  year = {2009,
1248  title = {Elinor Ostrom Prize Lecture},
1249  author = {Ostrom Elinor},
1253  year = {2019,
1255  title = {People + AI Guidebook},
1256  author = {PAIR},
1260  year = {2022,
1262  title = {HAX Toolkit},
1263  author = {Microsoft},
1267  year = {2019,
1269  title = {Ethically Aligned Design - A Vision for Prioritizing Human Well-being with Autonomous and Intelligent System},
1270  author = {IEEE},
1274  publisher = {Routledge},
1275  year = {2013},
1276  pages = {163--186},
1277  booktitle = {New public governance, the third sector, and co-production},
1278  author = {Porter, David O},
1279  title = {Co-production and network structures in public education},
1283  pages = {358–382},
1284  month = {Apr},
1285  year = {2014},
1286  author = {van Eijk, C. J. A. and Steen, T. P. S.},
1287  journal = {Public Management Review},
1288  number = {3},
1289  doi = {10.1080/14719037.2013.841458},
1290  url = {},
1291  issn = {1471-9037},
1292  volume = {16},
1293  title = {Why People Co-Produce: Analysing citizens’ perceptions on co-planning engagement in health care services},
1297  publisher = {Routledge},
1298  year = {2013},
1299  pages = {147--162},
1300  booktitle = {New public governance, the third sector, and co-production},
1301  author = {Cahn, Edgar S and Gray, Christine},
1302  title = {Co-production from a normative perspective},
1306  language = {eng},
1307  year = {2004},
1308  author = {Jasanoff, Sheila},
1309  publisher = {Routledge},
1310  url = {},
1311  isbn = {9781134328345},
1312  title = {States of knowledge: the co-production of science and the social order},
1313  address = {London},
1317  copyright = { perpetual, non-exclusive license},
1318  year = {2019},
1319  publisher = {arXiv},
1320  title = {Enabling Value Sensitive AI Systems through Participatory Design Fictions},
1321  keywords = {Human-Computer Interaction (cs.HC), Artificial Intelligence (cs.AI), FOS: Computer and information sciences, FOS: Computer and information sciences},
1322  author = {Liao, Q. Vera and Muller, Michael},
1323  url = {},
1324  doi = {10.48550/ARXIV.1912.07381},
1328  language = {en},
1329  pages = {769–778},
1330  month = {Apr},
1331  year = {2014},
1332  author = {Dourish, Paul and Bell, Genevieve},
1333  journal = {Personal and Ubiquitous Computing},
1334  number = {4},
1335  doi = {10.1007/s00779-013-0678-7},
1336  url = {},
1337  issn = {1617-4909, 1617-4917},
1338  volume = {18},
1339  title = {“Resistance is futile”: reading science fiction alongside ubiquitous computing},
1343  series = {CHI '12},
1344  location = {Austin, Texas, USA},
1345  keywords = {value sensitive design, collaborative ethnography, participatory design, feminism, design, culturally-specific values, universal values, qualitative research, voice},
1346  numpages = {10},
1347  pages = {1125–1134},
1348  doi = {10.1145/2207676.2208560},
1349  url = {},
1350  address = {New York, NY, USA},
1351  publisher = {Association for Computing Machinery},
1352  isbn = {9781450310154},
1353  year = {2012},
1354  title = {Next Steps for Value Sensitive Design},
1355  author = {Borning, Alan and Muller, Michael},
1359  year = {1996},
1360  pages = {44},
1361  volume = {40},
1362  journal = {Risk, environment and modernity: Towards a new ecology},
1363  author = {Wynne, Brian},
1364  title = {May the sheep safely graze? A reflexive view of the expert-lay knowledge divide},
1368  publisher = {Sage Publications London},
1369  year = {2001},
1370  pages = {461--467},
1371  number = {3},
1372  volume = {31},
1373  journal = {Social Studies of Science},
1374  author = {Jasanoff, Sheila},
1375  title = {Election 2000: mechanical error or system failure},
1379  isbn = {9780203413845},
1380  publisher = {Routledge},
1381  year = {2004},
1382  pages = {1--12},
1383  booktitle = {States of knowledge},
1384  author = {Jasanoff, Sheila},
1385  title = {The idiom of co-production},
1389  series = {CSCW '16},
1390  location = {San Francisco, California, USA},
1391  keywords = {online knowledge collaboration, socio-technical system, routines, organizational practice, sequence analysis, peer production, Wikipedia},
1392  numpages = {15},
1393  pages = {1065–1079},
1394  doi = {10.1145/2818048.2819962},
1395  url = {},
1396  address = {New York, NY, USA},
1397  publisher = {Association for Computing Machinery},
1398  isbn = {9781450335928},
1399  year = {2016},
1400  title = {Analyzing Organizational Routines in Online Knowledge Collaborations: A Case for Sequence Analysis in CSCW},
1401  author = {Keegan, Brian C. and Lev, Shakked and Arazy, Ofer},
1405  numpages = {7},
1406  pages = {26–32},
1407  month = {jul},
1408  journal = {Computer},
1409  abstract = {Service coproductions are reciprocal activities with no sharp boundary between providers and recipients; instead, all participants collaborate to enact the service. User participation is validated by recognition instead of extrinsic economic exchange. Mobility and context awareness can be crucial for successful coproductions.},
1410  doi = {10.1109/MC.2016.194},
1411  url = {},
1412  issn = {0018-9162},
1413  number = {7},
1414  volume = {49},
1415  address = {Washington, DC, USA},
1416  publisher = {IEEE Computer Society Press},
1417  issue_date = {July 2016},
1418  year = {2016},
1419  title = {In Search of Coproduction: Smart Services as Reciprocal Activities},
1420  author = {Carroll, John M. and Chen, Jiawei and Yuan, Chien Wen Tina and Hanrahan, Benjamin V.},
1424  copyright = { perpetual, non-exclusive license},
1425  year = {2018},
1426  publisher = {arXiv},
1427  title = {Building Ethics into Artificial Intelligence},
1428  keywords = {Artificial Intelligence (cs.AI), FOS: Computer and information sciences, FOS: Computer and information sciences},
1429  author = {Yu, Han and Shen, Zhiqi and Miao, Chunyan and Leung, Cyril and Lesser, Victor R. and Yang, Qiang},
1430  url = {},
1431  doi = {10.48550/ARXIV.1812.02953},
1435  language = {en},
1436  pages = {374–397},
1437  year = {2013},
1438  author = {Shilton, Katie},
1439  journal = {Science, Technology, \& Human Values},
1440  number = {3},
1441  doi = {10.1177/0162243912436985},
1442  url = {},
1443  volume = {38},
1444  title = {Values Levers: Building Ethics into Design},
1448  year = {1939},
1449  journal = {International encyclopedia of unified science},
1450  author = {Dewey, John},
1451  title = {Theory of valuation.},
1455  series = {DIS '22},
1456  location = {Virtual Event, Australia},
1457  keywords = {values levers, design ethnography, values in design, digital civics},
1458  numpages = {11},
1459  pages = {643–653},
1460  booktitle = {Designing Interactive Systems Conference},
1461  doi = {10.1145/3532106.3533570},
1462  url = {},
1463  address = {New York, NY, USA},
1464  publisher = {Association for Computing Machinery},
1465  isbn = {9781450393584},
1466  year = {2022},
1467  title = {Critically Engaging with Embedded Values through Constrained Technology Design},
1468  author = {Richardson, Dan and Cumbo, Bronwyn J. and Bartindale, Tom and Varghese, Delvin and Saha, Manika and Saha, Pratyasha and Ahmed, Syed Ishtiaque and Oliver, Gillian C. and Olivier, Patrick},
1472  series = {CHI '22},
1473  location = {New Orleans, LA, USA},
1474  keywords = {design, artificial intelligence, User experience, machine learning},
1475  numpages = {13},
1476  articleno = {483},
1477  booktitle = {Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems},
1478  doi = {10.1145/3491102.3517491},
1479  url = {},
1480  address = {New York, NY, USA},
1481  publisher = {Association for Computing Machinery},
1482  isbn = {9781450391573},
1483  year = {2022},
1484  title = {How Experienced Designers of Enterprise Applications Engage AI as a Design Material},
1485  author = {Yildirim, Nur and Kass, Alex and Tung, Teresa and Upton, Connor and Costello, Donnacha and Giusti, Robert and Lacin, Sinem and Lovic, Sara and O'Neill, James M and Meehan, Rudi O'Reilly and \'{O} Loide\'{a}in, Eoin and Pini, Azzurra and Corcoran, Medb and Hayes, Jeremiah and Cahalane, Diarmuid J and Shivhare, Gaurav and Castoro, Luigi and Caruso, Giovanni and Oh, Changhoon and McCann, James and Forlizzi, Jodi and Zimmerman, John},
1489  language = {eng},
1490  year = {2018},
1491  author = {Noble, Safiya Umoja},
1492  publisher = {New York university press},
1493  isbn = {9781479837243},
1494  title = {Algorithms of oppression: how search engines reinforce racism},
1495  address = {New York},
1499  bibsource = {dblp computer science bibliography,},
1500  biburl = {},
1501  timestamp = {Fri, 30 Jul 2021 13:03:06 +0200},
1502  eprint = {2107.13509},
1503  eprinttype = {arXiv},
1504  url = {},
1505  year = {2021},
1506  volume = {abs/2107.13509},
1507  journal = {CoRR},
1508  title = {The Who in Explainable {AI:} How {AI} Background Shapes Perceptions
1509of {AI} Explanations},
1510  author = {Upol Ehsan and
1511Samir Passi and
1512Q. Vera Liao and
1513Larry Chan and
1514I{-}Hsiang Lee and
1515Michael J. Muller and
1516Mark O. Riedl},
1520  year = {2017},
1521  author = {Eubanks, Virginia},
1522  publisher = {St. Martin’s Press},
1523  callnumber = {HC79.P6 E89 2017},
1524  isbn = {9781250074317},
1525  title = {Automating inequality: how high-tech tools profile, police, and punish the poor},
1526  edition = {First Edition},
1527  address = {New York, NY},
1531  series = {FAccT '22},
1532  location = {Seoul, Republic of Korea},
1533  keywords = {Trends, AI Ethics, Justice, FAccT, AIES},
1534  numpages = {11},
1535  pages = {948–958},
1536  booktitle = {2022 ACM Conference on Fairness, Accountability, and Transparency},
1537  doi = {10.1145/3531146.3533157},
1538  url = {},
1539  address = {New York, NY, USA},
1540  publisher = {Association for Computing Machinery},
1541  isbn = {9781450393522},
1542  year = {2022},
1543  title = {The Forgotten Margins of AI Ethics},
1544  author = {Birhane, Abeba and Ruane, Elayne and Laurent, Thomas and S. Brown, Matthew and Flowers, Johnathan and Ventresque, Anthony and L. Dancy, Christopher},
1548  copyright = { perpetual, non-exclusive license},
1549  year = {2017},
1550  publisher = {arXiv},
1551  title = {What Does Explainable AI Really Mean? A New Conceptualization of Perspectives},
1552  keywords = {Artificial Intelligence (cs.AI), FOS: Computer and information sciences, FOS: Computer and information sciences},
1553  author = {Doran, Derek and Schulz, Sarah and Besold, Tarek R.},
1554  url = {},
1555  doi = {10.48550/ARXIV.1710.00794},
1559  language = {en},
1560  month = {Jan},
1561  year = {2020},
1562  author = {Fjeld, Jessica and Achten, Nele and Hilligoss, Hannah and Nagy, Adam and Srikumar, Madhulika},
1563  institution = {Social Science Research Network},
1564  number = {3518482},
1565  url = {},
1566  title = {Principled Artificial Intelligence: Mapping Consensus in Ethical and Rights-Based Approaches to Principles for AI},
1567  type = {SSRN Scholarly Paper},
1568  address = {Rochester, NY},
1572  series = {FAT* '19},
1573  location = {Atlanta, GA, USA},
1574  keywords = {Machine Learning, Data Science, Problem Formulation, Target Variable, Fairness},
1575  numpages = {10},
1576  pages = {39–48},
1577  booktitle = {Proceedings of the Conference on Fairness, Accountability, and Transparency},
1578  doi = {10.1145/3287560.3287567},
1579  url = {},
1580  address = {New York, NY, USA},
1581  publisher = {Association for Computing Machinery},
1582  isbn = {9781450361255},
1583  year = {2019},
1584  title = {Problem Formulation and Fairness},
1585  author = {Passi, Samir and Barocas, Solon},
1589  bibsource = {dblp computer science bibliography,},
1590  biburl = {},
1591  timestamp = {Tue, 17 Sep 2019 11:23:44 +0200},
1592  eprint = {1909.03012},
1593  eprinttype = {arXiv},
1594  url = {},
1595  year = {2019},
1596  volume = {abs/1909.03012},
1597  journal = {CoRR},
1598  title = {One Explanation Does Not Fit All: {A} Toolkit and Taxonomy of {AI}
1599Explainability Techniques},
1600  author = {Vijay Arya and
1601Rachel K. E. Bellamy and
1602Pin{-}Yu Chen and
1603Amit Dhurandhar and
1604Michael Hind and
1605Samuel C. Hoffman and
1606Stephanie Houde and
1607Q. Vera Liao and
1608Ronny Luss and
1609Aleksandra Mojsilovic and
1610Sami Mourad and
1611Pablo Pedemonte and
1612Ramya Raghavendra and
1613John T. Richards and
1614Prasanna Sattigeri and
1615Karthikeyan Shanmugam and
1616Moninder Singh and
1617Kush R. Varshney and
1618Dennis Wei and
1619Yunfeng Zhang},
1623  series = {FAccT '21},
1624  location = {Virtual Event, Canada},
1625  keywords = {surveillance, accountability, algorithmic equity, Participatory design, algorithmic justice, participatory action research, regulation},
1626  numpages = {10},
1627  pages = {772–781},
1628  booktitle = {Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency},
1629  doi = {10.1145/3442188.3445938},
1630  url = {},
1631  address = {New York, NY, USA},
1632  publisher = {Association for Computing Machinery},
1633  isbn = {9781450383097},
1634  year = {2021},
1635  title = {An Action-Oriented AI Policy Toolkit for Technology Audits by Community Advocates and Activists},
1636  author = {Krafft, P. M. and Young, Meg and Katell, Michael and Lee, Jennifer E. and Narayan, Shankar and Epstein, Micah and Dailey, Dharma and Herman, Bernease and Tam, Aaron and Guetler, Vivian and Bintz, Corinne and Raz, Daniella and Jobe, Pa Ousman and Putz, Franziska and Robick, Brian and Barghouti, Bissan},
1640  url = {},
1641  address = {New York, NY, USA},
1642  publisher = {Association for Computing Machinery},
1643  isbn = {9781450369367},
1644  year = {2020},
1645  title = {Toward Situated Interventions for Algorithmic Equity: Lessons from the Field},
1646  author = {Katell, Michael and Young, Meg and Dailey, Dharma and Herman, Bernease and Guetler, Vivian and Tam, Aaron and Bintz, Corinne and Raz, Daniella and Krafft, P. M.},
1650  year = {2019},
1651  author = {Gray, Mary L. and Suri, Siddharth},
1652  publisher = {Houghton Mifflin Harcourt},
1653  callnumber = {HD6331},
1654  isbn = {9781328566287},
1655  title = {Ghost work: how to stop Silicon Valley from building a new global underclass},
1656  address = {Boston},
1660  language = {en},
1661  pages = {449–476},
1662  month = {Jun},
1663  year = {2019},
1664  author = {Metcalf, Jacob and Moss, Emanuel and boyd, danah},
1665  journal = {Social Research: An International Quarterly},
1666  number = {2},
1667  doi = {10.1353/sor.2019.0022},
1668  url = {},
1669  issn = {1944-768X},
1670  volume = {86},
1671  title = {Owning Ethics: Corporate Logics, Silicon Valley, and the Institutionalization of Ethics},
1675  year = {2019},
1676  author = {Friedman, Batya and Hendry, David},
1677  publisher = {The MIT Press},
1678  callnumber = {QA76.9.S88 F75 2019},
1679  isbn = {9780262039536},
1680  title = {Value sensitive design: shaping technology with moral imagination},
1681  address = {Cambridge, Massachusetts},
1685  pages = {830736},
1686  month = {Feb},
1687  year = {2022},
1688  author = {Eriksson, Eva and Nilsson, Elisabet M. and Hansen, Anne-Marie and Bekker, Tilde},
1689  journal = {Frontiers in Computer Science},
1690  doi = {10.3389/fcomp.2022.830736},
1691  url = {},
1692  issn = {2624-9898},
1693  volume = {4},
1694  title = {Teaching for Values in Human–Computer Interaction},
1698  bibsource = {dblp computer science bibliography,},
1699  biburl = {},
1700  timestamp = {Sat, 02 May 2020 19:17:26 +0200},
1701  eprint = {2004.13676},
1702  eprinttype = {arXiv},
1703  url = {},
1704  year = {2020},
1705  volume = {abs/2004.13676},
1706  journal = {CoRR},
1707  title = {Value-based Engineering for Ethics by Design},
1708  author = {Sarah Spiekermann and
1709Till Winkler},
1713  language = {eng},
1714  collection = {Topics in social psychology},
1715  year = {2006},
1716  author = {Weick, Karl E.},
1717  publisher = {McGraw-Hill},
1718  isbn = {9780075548089},
1719  title = {The social psychology of organizing},
1720  series = {Topics in social psychology},
1721  edition = {2. ed., [Nachdr.]},
1722  address = {New York},
1726  language = {eng},
1727  year = {2016},
1728  author = {Scott, W. Richard and Davis, Gerald F.},
1729  publisher = {Routledge},
1730  url = {},
1731  isbn = {9781317345923},
1732  title = {Organizations and organizing: rational, natural and open systems perspectives},
1733  edition = {1st ed},
1734  address = {Abingdon, Oxon},
1738  pages = {209–225},
1739  month = {Sep},
1740  year = {2021},
1741  author = {Green, Ben},
1742  journal = {Journal of Social Computing},
1743  number = {3},
1744  doi = {10.23919/JSC.2021.0018},
1745  url = {},
1746  issn = {2688-5255},
1747  volume = {2},
1748  title = {The Contestation of Tech Ethics: A Sociotechnical Approach to Technology Ethics in Practice},
1752  pages = {84–89},
1753  month = {Dec},
1754  year = {2019},
1755  editor = {Bayamlioglu, Emre and Baraliuc, Irina and Janssens, Liisa Albertha Wilhelmina and Hildebrandt, Mireille},
1756  author = {Wagner, Ben},
1757  publisher = {Amsterdam University Press},
1758  booktitle = {BEING PROFILED},
1759  doi = {10.1515/9789048550180-016},
1760  url = {},
1761  isbn = {9789048550180},
1762  title = {Ethics As An Escape From Regulation. From “Ethics-Washing” To Ethics-Shopping?},
1766  pages = {781},
1767  month = {Dec},
1768  year = {1983},
1769  author = {Gieryn, Thomas F.},
1770  journal = {American Sociological Review},
1771  number = {6},
1772  doi = {10.2307/2095325},
1773  url = {},
1774  issn = {00031224},
1775  volume = {48},
1776  title = {Boundary-Work and the Demarcation of Science from Non-Science: Strains and Interests in Professional Ideologies of Scientists},
1780  keywords = {organizational structure, industry practice, responsible ai},
1781  numpages = {23},
1782  articleno = {7},
1783  month = {apr},
1784  journal = {Proc. ACM Hum.-Comput. Interact.},
1785  doi = {10.1145/3449081},
1786  url = {},
1787  number = {CSCW1},
1788  volume = {5},
1789  address = {New York, NY, USA},
1790  publisher = {Association for Computing Machinery},
1791  issue_date = {April 2021},
1792  year = {2021},
1793  title = {Where Responsible AI Meets Reality: Practitioner Perspectives on Enablers for Shifting Organizational Practices},
1794  author = {Rakova, Bogdana and Yang, Jingying and Cramer, Henriette and Chowdhury, Rumman},
1798  publisher = {ACM New York, NY, USA},
1799  year = {2022},
1800  pages = {1--28},
1801  number = {CSCW2},
1802  volume = {6},
1803  journal = {Proceedings of the ACM on Human-Computer Interaction},
1804  author = {Goyal, Nitesh and Kivlichan, Ian D and Rosen, Rachel and Vasserman, Lucy},
1805  title = {Is Your Toxicity My Toxicity? Exploring the Impact of Rater Identity on Toxicity Annotation},


arXiv:2307.10221v1 [cs.AI]
License: cc-by-sa-4.0

Related Posts

The Fallacy of AI Functionality

The Fallacy of AI Functionality

Introduction As one of over 20,000 cases falsely flagged for unemployment benefit fraud by Michigan’s MIDAS algorithm, Brian Russell had to file for bankruptcy, undermining his ability to provide for his two young children.

The Right Tool for the Job: Open-Source Auditing Tools in Machine Learning

The Right Tool for the Job: Open-Source Auditing Tools in Machine Learning

Introduction As a society, we are tossed around on a sea of competing ideas, backed by statistics, data, and algorithms.

Institutionalising Ethics in AI through Broader Impact Requirements

Institutionalising Ethics in AI through Broader Impact Requirements

Introduction Growing concerns about the ethical, societal, environmental, and economic impacts of artificial intelligence (AI) have led to a wealth of governance initiatives.