Join Our Discord (940+ Members)

U.S. Public Opinion on the Governance of Artificial Intelligence

Content License: cc-by

U.S. Public Opinion on the Governance of Artificial Intelligence

Papers is Alpha. This content is part of an effort to make research more accessible, and (most likely) has lost some details from the original. You can find the original paper here.

Introduction

Advances in artificial intelligence (AI) could impact nearly all aspects of society, including the labor market, transportation, healthcare, education, and national security. AI’s effects may be profoundly positive, but the technology entails risks and disruptions that warrant attention. While technologists and policymakers have begun to discuss the societal implications of machine learning and AI more frequently, public opinion has not shaped much of these conversations. Given AI’s broad impact, civil society groups argue that the public, particularly those underrepresented in the tech industry, should have a role in shaping the technology.

In the U.S., public opinion has shaped policy outcomes, including those concerning immigration, free trade, international conflicts, and climate change mitigation. As in these other policy domains, we expect the public to become more influential over time in impacting AI policy. It is thus vital to have a better understanding of how the public thinks about AI and the governance of AI. Such understanding is essential to crafting informed policy and identifying opportunities to educate the public about AI’s character, benefits, and risks.

Agreement with statement that robots and AI require careful management (EU data from 2017 Special Eurobarometer #460)fig-manage

Agreement with statement that robots and AI require careful management (EU data from 2017 Special Eurobarometer #460)

Using an original, large-scale survey ($N$=2000), we studied how the American public perceives AI governance. The overwhelming majority of Americans (82%) believe that AI and/or robots should be carefully managed. This statistic is comparable to survey results from respondents residing in European Union countries, as seen in Figure fig-manage. Furthermore, Americans consider all of the 13 AI governance challenges presented in the survey to be important for tech companies and governments to manage carefully.

At the same time, Americans have only low to moderate levels of trust in governmental, corporate, and multistakeholder institutions to develop and manage AI in the public’s interest. Public trust varies greatly between organizations. Broadly, the public puts the most trust in university researchers (50% reporting “a fair amount of confidence” or “a great deal of confidence”) and the U.S. military (49%); followed by scientific organizations, the Partnership on AI, tech companies (excluding Facebook), and intelligence organizations; followed by U.S. federal or state governments, and the UN; followed by Facebook. Contrary to existing research on attitudes toward other emerging technologies, our study finds that individual-level trust in various actors to responsibly develop and manage AI does not predict one’s general support for developing AI.

Background

Public Opinion and AI Governance

Studying public opinion allows us to anticipate how electoral politics could shape AI governance as policymakers seek to regulate applications of the technology. Over the past two years, corporations, governments, civil society groups, and multistakeholder organizations have published dozens of high-level AI ethics principles. These documents borrow heavily from core principles in bioethics or medical ethics. But unlike medical ethics, which is guided by the common aim of promoting the health of the patient, AI development does not share one common goal but many goals. Furthermore, many of these AI ethics principles are at tension with each other. For instance, how does one improve the accuracy of algorithmic predictions while ensuring fair and equal treatment of those impacted by the algorithm? As corporations and governments attempt to translate these principles into practice, tensions between principles could lead to political contestation.

In the U.S., voters are divided on how to regulate facial recognition technology and algorithms used to display social media content, with some of these divisions reflecting existing disagreements between partisans. For instance, Democrats, compared with Republicans, are much more opposed to law enforcement using facial recognition technology. While several states and cities that are heavily Democrat have adopted or proposed moratoriums on law enforcement using facial recognition technology, progress to enact federal regulation has been slow. Industry self-regulation is not immune to partisan politics. For example, Google’s AI ethics board dissolved after the company’s employees and external civil society groups protested the inclusion of Heritage Foundation president Kay Coles James and drone company CEO Dyan Gibbens on the board. Recognizing the public’s divergent policy preferences on AI governance issues is necessary to have a productive public policy deliberation.

How Trust in Institutions Affects Regulation of Emerging Technology

The public’s trust in various institutions to develop and manage AI could affect the regulation of the technology. General social trust is negatively correlated with perceived risk from technological hazards. One observational study finds that that general social distrust is positively correlated with support for government regulation, even when the public perceives the government to be corrupt. A follow-up study using more extensive survey data suggests that the public prefers governmental regulation if they trust the government more than they do major companies. This latter finding is supported by how the public has reacted to genetically modified (GM) foods and nanotechnology. These two emerging technologies are similar to AI in that the public has to rely on those with scientific expertise when evaluating potential risks.

Distrust in institutions producing and regulating GM foods is a compelling explanation for the widespread opposition to GM foods in developed countries. Those with a high-level trust of scientists and regulators are more accepting of GM foods; in contrast, distrust of the agricultural/food industry, contrasted with trust in environmental watchdogs, predicts opposition to GM foods. Although scientists are among the most trusted group in the U.S., Americans have cynical views toward scientists when considering GM foods. Only 19% thinks that scientists understand the health consequences of GM foods very well, even though scientists have formed a consensus that GM foods are safe to eat. Furthermore, the American public believes that scientists are more motivated by concerns for their industry than concerns for the public.

Nanotechnology, though less salient than GM foods in the media, is the subject of extensive public opinion research. A meta-analysis of 11 surveys conducted in developed countries finds that the public perceives that the use of nanotechnology has greater benefits than risks; nevertheless, a large subset of the public is uncertain of the consequences of nanotechnology. As in perceptions of GM foods, trust in institutions seems to play a significant role in shaping attitudes toward nanotechnology. Americans who have lower confidence in business leaders within the nanotechnology industry also perceive the technology to be riskier. A higher level of distrust in government agencies to protect the public from nanotechnology hazards is associated with a higher perceived risk of the technology. Consumers who are less trustful of the food industry indicate they are more reluctant to buy foods produced or packaged using nanotechnology.

Existing Survey Research on AI Governance and Trust in Tech Companies

Our survey builds on existing public opinion research on attitudes toward AI and trust in tech companies. Past survey research related to AI tends to focus on specific governance challenges, such as lethal autonomous weapons, algorithmic fairness, or facial recognition technology. A few large-scale surveys have taken a more comprehensive approach by asking about a range of AI governance challenges. We improved upon these surveys by asking respondents to consider a variety of AI governance challenges using the same question-wording. This consistency in question-wording allowed us to compare respondents’ perceptions toward major issues in AI governance.

Another innovation of our research project is that we connect AI governance with trust in tech companies. Trust in tech companies has become a growing topic in survey research. While Americans perceive tech companies to have a more positive impact on society than other institutions, including the news media and the government, their feelings toward tech companies have declined dramatically since 2015. Nevertheless, there exists heterogeneity in how the public views individual tech companies; one consistent finding in non-academic surveys is that the public strongly distrusts Facebook, particularly in its handling of personal data. In our survey, we examined whether this general distrust of tech companies extend to their management of AI development and applications.

Survey Data and Analysis

We conducted an original online survey ($N$=2000) through YouGov between June 6 and 14, 2018. YouGov drew a random sample from the U.S. adult population (i.e., the target sample) and selected respondents from its online respondent panel that matched the target sample on key demographic variables. The details of YouGov’s sample matching methodology can be found in the Appendix.

We pre-registered nearly all of the analysis on Open Science Framework (pre-analysis plan URL: https://osf.io/7gqvm/ ). Pre-registration increases research transparency by requiring researchers to specify their analysis before analyzing the data. Doing so prevents researchers from misusing data analysis to come up with statistically significant results when they do not exist, otherwise known as $p$-hacking. Survey weights provided by YouGov were used in our primary analysis. We followed the Standard Operating Procedures for Don Green’s Lab at Columbia University when handling missing data or “don’t know” responses. Heteroscedasticity-consistent standard errors were used to generate the margins of error at the 95% confidence level. We report cluster-robust standard errors whenever there is clustering by respondent. In figures, each error bar shows the 95% confidence intervals.

Perceptions of AI Governance Challenges

Methodology

We sought to understand how Americans prioritize policy issues associated with AI. Respondents were asked to consider five AI governance challenges, randomly selected from a set of 13 (see the Appendix for the full text); the order these five were given to each respondent was also randomized. After considering each governance challenge, respondents were asked how likely they think the challenge will affect large numbers of people both 1) in the U.S. and 2) around the world within 10 years using a seven-point scale that displayed both numerical likelihoods and qualitative descriptions (e.g., “Very unlikely: less than 5% chance (2.5%)”). Respondents were also asked to evaluate how important it is for tech companies and governments to carefully manage each challenge presented to them using a four-point scale.

General Results

Perceptions of AI governance challenges in the U.S. and around the worldfig-aigovchallenges

Perceptions of AI governance challenges in the U.S. and around the world Perceptions of AI governance challenges in the U.S. and around the world

We use scatterplots to visualize our survey results. In the top graph in Figure fig-aigovchallenges, the x-axis is the perceived likelihood of the problem happening to large numbers of people in the U.S. In the bottom graph, the x-axis is the perceived likelihood of the problem happening to large numbers of people around the world. The y-axes on both graphs in Figure fig-aigovchallenges represent respondents’ perceived issue importance, from 0 (not at all important) to 3 (very important). Each dot represents the mean perceived likelihood and issue importance. The correspondent ellipse represents the 95% confidence region of the bivariate means assuming the two variables are distributed multivariate normal.

Americans consider all the AI governance challenges we present to be important: the mean perceived issues importance of each governance challenge is between “somewhat important” (2) and “very important” (3), though there is meaningful and discernible variation across items.

The AI governance challenges Americans think are most likely to impact large numbers of people, and are important for tech companies and governments to tackle, are found in the upper-right quadrant of the two plots. These issues include data privacy as well as AI-enhanced cyber attacks, surveillance, and digital manipulation. We note that the media have widely covered these issues during the time of the survey.

There are a second set of governance challenges that are perceived on average, as about 7% less likely, and marginally less important. These include autonomous vehicles, value alignment, bias in using AI for hiring, the U.S.-China arms race, disease diagnosis, and technological unemployment. Finally, the third set of challenges are perceived on average another 5% less likely, and about equally important, including criminal justice bias and critical AI systems failures.

We also note that Americans predict that all of the governance challenges mentioned in the survey, besides protecting data privacy and ensuring the safety of autonomous vehicles, are more likely to impact people around the world than to affect people in the U.S. While most of these statistically significant differences are substantively small, one difference stands out: Americans think that autonomous weapons are 7.6 percentage points more likely to impact people around the world than Americans (two-sided (p)-value 0.001).

We want to reflect on one result. “Value alignment” consists of an abstract description of the alignment problem and a reference to what sounds like individual level harms: “while performing jobs they could unintentionally make decisions that go against the values of its human users, such as physically harming people.” “Critical AI systems failures,” by contrast, references military or critical infrastructure uses, and unintentional accidents leading to “10 percent or more of all humans to die.” The latter was weighted as less important than the former: we interpret this as a probability-weighted assessment of importance, since presumably the latter, were it to happen, is much more important. We thus think the issue importance question should be interpreted in a way that down-weights low probability risks.

Subgroup Analysis

AI governance challenges: issue importance by demographic subgroupsfig-subgroup

AI governance challenges: issue importance by demographic subgroups

We performed further analysis by calculating the percentage of respondents in each subgroup who consider each governance challenge to be “very important” for governments and tech companies to manage. In general, differences in responses are more salient across demographic subgroups than across governance challenges. In a linear multiple regression predicting perceived issue importance using demographic subgroups, governance challenges, and the interaction between the two, we find that the stronger predictors are demographic subgroup variables, including age group and having CS or programming experience.

Two highly visible patterns emerge from our data visualization. First, a higher percentage of older respondents, compared with younger respondents, consider nearly all AI governance challenges to be “very important.” In another part of the survey (see Figure 7 in the Appendix), we find that older Americans, compared with younger Americans, are less supportive of developing AI. Our results here might explain this age gap: older Americans see each AI governance challenge as substantially more important than do younger Americans. Whereas 85% of Americans older than 73 consider each of these issues to be very important, only 40% of Americans younger than 38 do.

Second, those with CS or engineering degrees, compared with those who do not, rate all AI governance challenges as less important. This result could explain another finding in our survey that shows those with CS or engineering degrees tend to express greater support for developing AI. In Table 4 in the Appendix, we report the results of a saturated linear model using demographic variables, governance challenges, and the interaction between these two types of variables to predict perceived issue importance. We find that those who are 54-72 or 73 and older, relative to those who are below 38, view the governance challenges as more important (two-sided (p)-value 0.001 for both comparisons). Furthermore, we find that those who have CS or engineering degrees, relative to those who do not, view the governance challenges as less important (two-sided (p)-value 0.001).

Trust in Actors to Develop and Manage AI

Methodology

Respondents were asked how much confidence they have in various actors to develop AI. They were randomly assigned five actors out of 15 to evaluate. We provided a short description of actors that are not well-known to the public (e.g., NATO, CERN, and OpenAI). Also, respondents were asked how much confidence, if any, they have in various actors to manage the development and use of AI in the best interests of the public. They were randomly assigned five out of 15 actors to evaluate. Again, we provided a short description of actors that are not well-known to the public (e.g., AAAI and Partnership on AI). Confidence was measured using the same four-point scale described above. The two sets of 15 actors differed slightly because, for some actors, it seemed inappropriate to ask one or the other question. See the Appendix for the exact wording of the questions and descriptions of the actors.

Results

Trust in various actors to develop and manage AI in the interest of the publicfig-trust

Trust in various actors to develop and manage AI in the interest of the public

Americans do not express great confidence in most actors to develop or to manage AI, as seen in Figures 5 and 6 in the Appendix. A majority of Americans do not have a “great deal” or even a “fair amount” of confidence in any institution, except university researchers, to develop AI. Furthermore, Americans place greater trust in tech companies and non-governmental organizations (e.g., OpenAI) than in governments to manage the development and use of the technology.

University researchers and the U.S. military are the most trusted groups to develop AI: about half of Americans express a “great deal” or even a “fair amount” of confidence in them. Americans express slightly less confidence in tech companies, non-profit organizations (e.g., OpenAI)[At the time the survey was conducted (June 2018), OpenAI was a 501(c)(3) nonprofit organization. In March 2019, OpenAI announced it is restructuring into “capped-profit” company to attract investments.], and American intelligence organizations. Nevertheless, opinions toward individual actors within each of these groups vary. For example, while 44% of Americans indicated they feel a “great deal” or even a “fair amount” of confidence in tech companies, they rate Facebook as the least trustworthy of all the actors. More than four in 10 indicate that they have no confidence in the company.

Our survey was conducted between June 6 and 14, 2018, shortly after the Facebook/Cambridge Analytica scandal became highly salient in the media. On April 10-11, 2018, Facebook CEO Mark Zuckerberg testified before the U.S. Congress regarding the Cambridge Analytica data leak. On May 2, 2018, Cambridge Analytica announced its shutdown. Nevertheless, Americans’ distrust of the company existed before the Facebook/Cambridge Analytica scandal. In a pilot survey that we conducted on Mechanical Turk during July 13–14, 2017, respondents indicated a substantially lower level of confidence in Facebook, compared with other actors, to develop and regulate AI.

The results on the public’s trust of various actors to manage the develop and use of AI provided are similar to the results discussed above. Again, a majority of Americans do not have a “great deal” or even a “fair amount” of confidence in any institution to manage AI. In general, the public expresses greater confidence in non-governmental organizations than in governmental ones. Indeed, 41% of Americans express a “great deal” or even a “fair amount” of confidence in “tech companies,” compared with 26% who feel that way about the U.S. federal government. But when presented with individual big tech companies, Americans indicate less trust in each than in the broader category of “tech companies.” Once again, Facebook stands out as an outlier: respondents give it a much lower rating than any other actor. Besides “tech companies,” the public places relatively high trust in intergovernmental research organizations (e.g., CERN)[We asked about trust in intergovernmental research organizations, such as CERN, because some policy researchers have proposed creating a politically neutral AI development hub similar to CERN to avoid risks associated with competition in AI development between rival states.], the Partnership on AI, and non-governmental scientific organizations (e.g., AAAI). Nevertheless, because the public is less familiar with these organizations, about one in five respondents give a “don’t know” response.

Similar to our findings, recent survey research suggests that while Americans feel that AI should be regulated, they are unsure who the regulators should be. When asked who “should decide how AI systems are designed and deployed,” half of Americans indicated they do not know or refused to answer. Our survey results also seem to reflect Americans’ general attitudes toward public institutions. According to a 2016 Pew Research Center survey, an overwhelming majority of Americans have “a great deal” or “a fair amount” of confidence in the U.S. military and scientists to act in the best interest of the public. In contrast, public confidence in elected officials is much lower: 73% indicated that they have “not too much confidence” or “no confidence”. Less than one-third of Americans thought that tech companies do what’s right “most of the time” or “just about always”; moreover, more than half indicate that tech companies have too much power and influence in the U.S. economy. Nevertheless, Americans’ attitudes toward tech companies are not monolithic but varies by company. For instance, our research findings reflect the results from several non-academic surveys that find the public distrusts Facebooks significantly more than other major tech companies.

Predicting Support for Developing AI Using Institutional Trust

Regression results: predicting support for developing AI using respondents’ trust in different types of actors to develop AI

Table Label: tab-trustdevsupport

Download PDF to view table

Regression result: predicting support for developing AI using respondents’ trust in different types of actors to manage AI

Table Label: tab-trustmanageupport

Download PDF to view table

The American public’s support for developing AI is not predicted by their trust in various actors to develop and manage AI. Our null results stand in contrast to studies of GM foods and nanotechnology, where trust in institutions that develop and regulate the technology is associated with greater support for the technology.

For our analysis, we use multiple linear regression to whether individual-level trust in various types of institutions can predict support for developing AI. Note that this analysis is not included in our pre-registration and is therefore exploratory in nature. There are four categories of actors (the same ones shown in Figure fig-trust): U.S. government, international bodies, corporate, and others (e.g., universities and non-profits). In our regressions, the group “others” is the reference group. Support for developing AI is measured using a five-point Likert scale, with -2 meaning “strongly oppose” and 2 meaning “strongly support.” Our regressions controlled for all the demographic variables shown in Table 3 in the Appendix. Trust in none of the actor types predicts support for developing AI, as seen in Tables tab-trustdevsupport and tab-trustmanageupport.

Conclusion

To understand what the American public thinks about the regulation of the technology, we conducted a large-scale survey. The survey reveals that while Americans consider all AI governance challenges to have high issue importance, they do not necessarily trust the actors who have the power to develop and manage the technology to act in the public’s interest. Nevertheless, as our exploratory analysis from the previous section shows, institutional distrust does not necessarily predict opposition to AI development.

One direction for future research is to examine other factors that shape the public’s preferences toward AI governance. Technological knowledge, moral and psychological attributes, and perceptions of risks versus benefits are associated with support for GM foods and nanotechnology. Another direction for future research is to improve the measurement of institutional trust. Our existing survey questions focused on the public interest component of trust; future studies could investigate other components of trust, such as competence, transparency, and honesty. Finally, we plan to investigate whether the public’s perceived lack of political power makes them more distrustful of institutions to govern AI. In a recent survey of the British public, a large majority felt that they were unable to influence AI development, and many felt tech companies and governments have disproportionate influence. Research on public perceptions of AI is nascent but could have a broad impact as AI governance moves from the realm of abstract principles into the world of mass politics.

YouGov Sampling and Weights

YouGov interviewed 2387 respondents who were then matched down to a sample of 2000 to produce the final dataset. The respondents were matched to a sampling frame on gender, age, race, and education. The frame was constructed by stratified sampling from the full 2016 American Community Survey (ACS) one-year sample with selection within strata by weighted sampling with replacements (using the person weights on the public use file).

The matched cases were weighted to the sampling frame using propensity scores. The matched cases and the frame were combined and a logistic regression was estimated for inclusion in the frame. The propensity score function included age, gender, race/ethnicity, years of education, and geographic region. The propensity scores were grouped into deciles of the estimated propensity score in the frame and post-stratified according to these deciles.

The weights were then post-stratified on 2016 U.S. presidential vote choice, and a four-way stratification of gender, age (four-categories), race (four-categories), and education (four-categories), to produce the final weight.

Text of the Questions

Below, we present the survey text as shown to respondents. The numerical codings are shown in parentheses following each answer choice.

Should AI and/or robots should be carefully managed

Please tell me to what extent you agree or disagree with the following statement.

[Respondents were presented with one statement randomly selected from the list below.]

  • AI and robots are technologies that require careful management.
  • AI is a technology that requires careful management.
  • Robots are technologies that require careful management.

ANSWER CHOICES:

  • Totally agree (2)
  • Tend to agree (1)
  • Tend to disagree (-1)
  • Totally disagree (-2)
  • I don’t know

Support for developing AI

[All respondents were presented with the following prompt.]

Next, we would like to ask you questions about your attitudes toward artificial intelligence.

Artificial Intelligence (AI) refers to computer systems that perform tasks or make decisions that usually require human intelligence. AI can perform these tasks or make these decisions without explicit human instructions. Today, AI has been used in the following applications:

[Respondents were shown five items randomly selected from the list below.]

  • Translate over 100 different languages
  • Predict one’s Google searches
  • Identify people from their photos
  • Diagnose diseases like skin cancer and common illnesses
  • Predict who are at risk of various diseases
  • Help run factories and warehouses
  • Block spam email
  • Play computer games
  • Help conduct legal case research
  • Categorize photos and videos
  • Detect plagiarism in essays
  • Spot abusive messages on social media
  • Predict what one is likely to buy online
  • Predict what movies or TV shows one is likely to watch online

QUESTION: How much do you support or oppose the development of AI?

ANSWER CHOICES:

  • Strongly support (2)
  • Somewhat support (1)
  • Neither support nor oppose (0)
  • Somewhat oppose (-1)
  • Strongly oppose (-2)
  • I don’t know

AI governance challenges

We would like you to consider some potential policy issues related to AI. Please consider the following:

[Respondents were shown five randomly-selected items from the list below, one item at a time. For ease of comprehension, we include the shorten labels used in the figures in bold.]

  • Hiring bias: Fairness and transparency in AI used in hiring: Increasingly, employers are using AI to make hiring decisions. AI has the potential to make less biased hiring decisions than humans. But algorithms trained on biased data can lead to lead to hiring practices that discriminate against certain groups. Also, AI used in this application may lack transparency, such that human users do not understand what the algorithm is doing, or why it reaches certain decisions in specific cases.
  • Criminal justice bias: Fairness and transparency in AI used in criminal justice: Increasingly, the criminal justice system is using AI to make sentencing and parole decisions. AI has the potential to make less biased hiring decisions than humans. But algorithms trained on biased data could lead to discrimination against certain groups. Also, AI used in this application may lack transparency such that human users do not understand what the algorithm is doing, or why it reaches certain decisions in specific cases.
  • Disease diagnosis: Accuracy and transparency in AI used for disease diagnosis: Increasingly, AI software has been used to diagnose diseases, such as heart disease and cancer. One challenge is to make sure the AI can correctly diagnose those who have the disease and not mistakenly diagnose those who do not have the disease. Another challenge is that AI used in this application may lack transparency such that human users do not understand what the algorithm is doing, or why it reaches certain decisions in specific cases.
  • Data privacy: Protect data privacy: Algorithms used in AI applications are often trained on vast amounts of personal data, including medical records, social media content, and financial transactions. Some worry that data used to train algorithms are not collected, used, and stored in ways that protect personal privacy.
  • Autonomous vehicles: Make sure autonomous vehicles are safe: Companies are developing self-driving cars and trucks that require little or no input from humans. Some worry about the safety of autonomous vehicles for those riding in them as well as for other vehicles, cyclists, and pedestrians.
  • Ditigal manipulation: Prevent AI from being used to spread fake and harmful content online: AI has been used by governments, private groups, and individuals to harm or manipulate internet users. For instance, automated bots have been used to generate and spread false and/or harmful news stories, audios, and videos.
  • Cyber attacks: Prevent AI cyber attacks against governments, companies, organizations, and individuals: Computer scientists have shown that AI can be used to launch effective cyber attacks. AI could be used to hack into servers to steal sensitive information, shut down critical infrastructures like power grids or hospital networks, or scale up targeted phishing attacks.
  • Surveillance: Prevent AI-assisted surveillance from violating privacy and civil liberties: AI can be used to process and analyze large amounts of text, photo, audio, and video data from social media, mobile communications, and CCTV cameras. Some worry that governments, companies, and employers could use AI to increase their surveillance capabilities.
  • U.S.-China arms race: Prevent escalation of a U.S.-China AI arms race: Leading analysts believe that an AI arms race is beginning, in which the U.S. and China are investing billions of dollars to develop powerful AI systems for surveillance, autonomous weapons, cyber operations, propaganda, and command and control systems. Some worry that a U.S.-China arms race could lead to extreme dangers. To stay ahead, the U.S. and China may race to deploy advanced military AI systems that they do not fully understand or can control. We could see catastrophic accidents, such as a rapid, automated escalation involving cyber and nuclear weapons.
  • Value alignment: Make sure AI systems are safe, trustworthy, and aligned with human values: As AI systems become more advanced, they will increasingly make decisions without human input. One potential fear is that AI systems, while performing jobs they are programmed to do, could unintentionally make decisions that go against the values of its human users, such as physically harming people.
  • Autonomous weapons: Ban the use of lethal autonomous weapons (LAWs): Lethal autonomous weapons (LAWs) are military robots that can attack targets without control by humans. LAWs could reduce the use of human combatants on the battlefield. But some worry that the adoption of LAWs could lead to mass violence. Because they are cheap and easy to produce in bulk, national militaries, terrorists, and other groups could readily deploy LAWs.
  • Technological unemployment: Guarantee a good standard of living for those who lose their jobs to automation: Some forecast that AI will increasingly be able to do jobs done by humans today. AI could potentially do the jobs of blue-collar workers, like truckers and factory workers, as well as the jobs of white-collar workers, like financial analysts or lawyers. Some worry that in the future, robots and computers can do most of the jobs that are done by humans today.
  • Critical AI systems failure: Prevent critical AI systems failures: As AI systems become more advanced, they could be used by the military or in critical infrastructure, like power grids, highways, or hospital networks. Some worry that the failure of AI systems or unintentional accidents in these applications could cause 10 percent or more of all humans to die.

QUESTION: In the next 10 years, how likely do you think it is that this AI governance challenge will impact large numbers of people in the U.S.?

ANSWER CHOICES:

  • Very unlikely: less than 5% chance (2.5%)
  • Unlikely: 5-20% chance (12.5%)
  • Somewhat unlikely: 20-40% chance (30%)
  • Equally likely as unlikely: 40-60% chance (50%)
  • Somewhat likely: 60-80% chance (70%)
  • Likely: 80-95% chance (87.5%)
  • Very likely: more than 95% chance (97.5%)
  • I don’t know

QUESTION: In the next 10 years, how likely do you think it is that this AI governance challenge will impact large numbers of people around the world?

ANSWER CHOICES:

  • Very unlikely: less than 5% chance (2.5%)
  • Unlikely: 5-20% chance (12.5%)
  • Somewhat unlikely: 20-40% chance (30%)
  • Equally likely as unlikely: 40-60% chance (50%)
  • Somewhat likely: 60-80% chance (70%)
  • Likely: 80-95% chance (87.5%)
  • Very likely: more than 95% chance (97.5%)
  • I don’t know

QUESTION: In the next 10 years, how important is it for tech companies and governments to carefully manage the following challenge?

ANSWER CHOICES:

  • Very important (3)
  • Somewhat important (2)
  • Not too important (1)
  • Not at all important (0)
  • I don’t know

Trust of actors to develop AI

QUESTION: How much confidence, if any, do you have in each of the following to develop AI in the best interests of the public?

[Respondents were shown five items randomly selected from the list below. We included explainer text for actors not well known to the public; respondents could view the explainer text by hovering their mouse over the actor’s name. The items and the answer choices were shown in a matrix format.]

  • The U.S. military
  • The U.S. civilian government
  • National Security Agency (NSA)
  • Federal Bureau of Investigation (FBI)
  • Central Intelligence Agency (CIA)
  • North Atlantic Treaty Organization (NATO);explainer text for NATO: NATO is a military alliance that includes 28 countries including most of Europe, as well as the U.S. and Canada.
  • An international research organization (e.g., CERN); explainer text for CERN: The European Organization for Nuclear Research, known as CERN, is a European research organization that operates the largest particle physics laboratory in the world.
  • Tech companies
  • Google
  • Facebook
  • Apple
  • Microsoft
  • Amazon
  • A non-profit AI research organization (e.g., OpenAI); explainer text for OpenAI: Open AI is an AI non-profit organization with backing from tech investors that seeks to develop safe AI. University researchers

ANSWER CHOICES:

  • A great deal of confidence (3)
  • A fair amount of confidence (2)
  • Not too much confidence (1)
  • No confidence (0)
  • I don’t know

Trust of actors to manage AI

QUESTION: How much confidence, if any, do you have in each of the following to manage the development and use of AI in the best interests of the public?

[Respondents were shown five items randomly selected from the list below. We included explainer text for actors not well known to the public; respondents could view the explainer text by hovering their mouse over the actor’s name. The items and the answer choices were shown in a matrix format.]

  • U.S. federal government
  • U.S. state governments
  • International organizations (e.g., United Nations, European Union)
  • The United Nations (UN)
  • An intergovernmental research organization (e.g., CERN); explainer text for CERN: The European Organization for Nuclear Research, known as CERN, is a European research organization that operates the largest particle physics laboratory in the world.
  • Tech companies
  • Google
  • Facebook
  • Apple
  • Microsoft
  • Amazon
  • Non-government scientific organizations (e.g., AAAI); explainer text for AAAI: Association for the Advancement of Artificial Intelligence (AAAI) is a non-government scientific organization that promotes research in, and responsible use of AI.
  • Partnership on AI, an association of tech companies, academics, and civil society groups

ANSWER CHOICES:

  • A great deal of confidence (3)
  • A fair amount of confidence (2)
  • Not too much confidence (1)
  • No confidence (0)
  • I don’t know

Additional Tables and Figures

Size of demographic subgroups

Table Label: tab-subgroups

Download PDF to view table

Results from a saturated regression predicting perceived issue importance using demographic variables, AI governance challenge, and interactions between the two types of variables; the coefficients for the interactions variables are not shown due to space constraints

Table Label: tab-aigovregsat

Download PDF to view table
Trust in various actors to develop AI: distribution of responsesfig-trustdev

Trust in various actors to develop AI: distribution of responses

Trust in various actors to manage AI: distribution of responsesfig-trustmanage

Trust in various actors to manage AI: distribution of responses

Support for developing AI across demographic characteristics: distribution of responsesfig-supportdemo

Support for developing AI across demographic characteristics: distribution of responses

Bibliography

  1@article{floridi2019unified,
  2  day = {23},
  3  month = {6},
  4  year = {2019},
  5  date = {2019-06-23},
  6  author = {Floridi, Luciano and Cowls, Josh},
  7  volume = {1},
  8  url = {https://hdsr.mitpress.mit.edu/pub/l0jsh9d1},
  9  title = {A Unified Framework of Five Principles for AI in Society},
 10  note = {https://hdsr.mitpress.mit.edu/pub/l0jsh9d1},
 11  number = {1},
 12  doi = {10.1162/99608f92.8cd550d1},
 13  journal = {Harvard Data Science Review},
 14}
 15
 16@article{scott2018overview,
 17  publisher = {Annual Reviews},
 18  year = {2018},
 19  pages = {459--479},
 20  volume = {38},
 21  journal = {Annual Review of Nutrition},
 22  author = {Scott, Sydney E and Inbar, Yoel and Wirz, Christopher D and Brossard, Dominique and Rozin, Paul},
 23  title = {An overview of attitudes toward genetically engineered food},
 24}
 25
 26@unpublished{mittelstadt2019ai,
 27  year = {2019},
 28  url = {https://dx.doi.org/10.2139/ssrn.3391293},
 29  note = {Forthcoming in \textit{Nature Machine Intelligence}},
 30  author = {Mittelstadt, Brent},
 31  title = {AI Ethics--Too Principled to Fail?},
 32}
 33
 34@article{caughey2018policy,
 35  publisher = {Cambridge University Press},
 36  year = {2018},
 37  pages = {249--266},
 38  number = {2},
 39  volume = {112},
 40  journal = {American Political Science Review},
 41  author = {Caughey, Devin and Warshaw, Christopher},
 42  title = {Policy preferences and policy change: Dynamic responsiveness in the American states, 1936--2014},
 43}
 44
 45@techreport{west2019,
 46  url = {https://perma.cc/85VP-HL8R},
 47  year = {2019},
 48  institution = {AI Now Institute},
 49  title = {Discriminating Systems: Gender, Race and Power in AI},
 50  author = {Sarah Myers West and Meredith Whittaker and Kate Crawford},
 51}
 52
 53@book{oecdai,
 54  doi = {https://doi.org/https://doi.org/10.1787/eedfee77-en},
 55  year = {2019},
 56  title = {Artificial Intelligence in Society},
 57  author = {{OECD}},
 58}
 59
 60@inproceedings{whittlestone2019role,
 61  organization = {ACM},
 62  year = {2019},
 63  pages = {195--200},
 64  booktitle = {Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society},
 65  author = {Whittlestone, Jess and Nyrup, Rune and Alexandrova, Anna and Cave, Stephen},
 66  title = {The role and limits of principles in AI ethics: towards a focus on tensions},
 67}
 68
 69@article{cave2019bridging,
 70  publisher = {Nature Publishing Group},
 71  year = {2019},
 72  pages = {5},
 73  number = {1},
 74  volume = {1},
 75  journal = {Nature Machine Intelligence},
 76  author = {Cave, Stephen and {\'O}h{\'E}igeartaigh, Se{\'a}n S},
 77  title = {Bridging near-and long-term concerns about AI},
 78}
 79
 80@techreport{zhang2019regulation,
 81  url = {https://perma.cc/B3FY-K6JH},
 82  year = {2019},
 83  institution = {Brookings Institution},
 84  title = {Public Opinion Lessons for {AI} Regulation},
 85  author = {Baobao Zhang},
 86}
 87
 88@techreport{smith2019facial,
 89  url = {https://perma.cc/FUV7-5BDJ},
 90  year = {2019},
 91  institution = {Pew Research Center},
 92  title = {More Than Half of U.S. Adults Trust Law Enforcement to Use Facial Recognition Responsibly},
 93  author = {Aaron Smith},
 94}
 95
 96@article{siegrist2005perception,
 97  publisher = {Taylor \& Francis},
 98  year = {2005},
 99  pages = {145--156},
100  number = {2},
101  volume = {8},
102  journal = {Journal of Risk Research},
103  author = {Siegrist, Michael and Gutscher, Heinz and Earle, Timothy C},
104  title = {Perception of risk: the influence of general trust, and general confidence},
105}
106
107@article{aghion2010regulation,
108  publisher = {MIT Press},
109  year = {2010},
110  pages = {1015--1049},
111  number = {3},
112  volume = {125},
113  journal = {The Quarterly Journal of Economics},
114  author = {Aghion, Philippe and Algan, Yann and Cahuc, Pierre and Shleifer, Andrei},
115  title = {Regulation and distrust},
116}
117
118@techreport{smith2018,
119  url = {https://perma.cc/KSN6-6FRW},
120  type = {Survey report},
121  institution = {Pew Research Center},
122  year = {2018},
123  author = {Aaron Smith},
124  title = {Public Attitudes Toward Technology Companies},
125}
126
127@article{pitlik2015does,
128  publisher = {Springer},
129  year = {2015},
130  pages = {355--377},
131  number = {3-4},
132  volume = {163},
133  journal = {Public Choice},
134  author = {Pitlik, Hans and Kouba, Ludek},
135  title = {Does social distrust always lead to a stronger support for government intervention?},
136}
137
138@article{siegrist1999causal,
139  publisher = {Wiley Online Library},
140  year = {1999},
141  pages = {2093--2106},
142  number = {10},
143  volume = {29},
144  journal = {Journal of Applied Social Psychology},
145  author = {Siegrist, Michael},
146  title = {A causal model explaining the perception and acceptance of gene technology 1},
147}
148
149@article{siegrist2000influence,
150  publisher = {Wiley Online Library},
151  year = {2000},
152  pages = {195--204},
153  number = {2},
154  volume = {20},
155  journal = {Risk Analysis},
156  author = {Siegrist, Michael},
157  title = {The influence of trust and perceptions of risks and benefits on the acceptance of gene technology},
158}
159
160@article{lang2005does,
161  publisher = {Wiley Online Library},
162  year = {2005},
163  pages = {1241--1252},
164  number = {5},
165  volume = {25},
166  journal = {Risk Analysis: An International Journal},
167  author = {Lang, John T and Hallman, William K},
168  title = {Who does the public trust? The case of genetically modified food in the United States},
169}
170
171@article{olofsson2006attitudes,
172  publisher = {Taylor \& Francis},
173  year = {2006},
174  pages = {601--624},
175  number = {4},
176  volume = {8},
177  journal = {European Societies},
178  author = {Olofsson, Anna and {\"O}hman, Susanna and Rashid, Saman},
179  title = {Attitudes to gene technology: the significance of trust in institutions},
180}
181
182@article{marques2015attitudes,
183  publisher = {SAGE Publications Sage UK: London, England},
184  year = {2015},
185  pages = {601--618},
186  number = {5},
187  volume = {24},
188  journal = {Public Understanding of Science},
189  author = {Marques, Mathew D and Critchley, Christine R and Walshe, Jarrod},
190  title = {Attitudes to genetically modified food over time: How trust in organizations and the media cycle predict support},
191}
192
193@techreport{funk2016new,
194  url = {https://perma.cc/TJ2L-K9JU},
195  year = {2016},
196  institution = {Pew Research Center},
197  title = {The New Food Fights: U.S. Public Divides Over Food Science},
198  author = {Cary Funk and Brian Kennedy},
199}
200
201@article{satterfield2009anticipating,
202  publisher = {Nature Publishing Group},
203  year = {2009},
204  pages = {752},
205  number = {11},
206  volume = {4},
207  journal = {Nature Nanotechnology},
208  author = {Satterfield, Terre and Kandlikar, Milind and Beaudrie, Christian EH and Conti, Joseph and Harthorn, Barbara Herr},
209  title = {Anticipating the perceived risk of nanotechnologies},
210}
211
212@article{macoubrie2004public,
213  publisher = {Springer},
214  year = {2004},
215  pages = {395--405},
216  number = {4},
217  volume = {6},
218  journal = {Journal of Nanoparticle Research},
219  author = {Michael D. Cobb and Jane Macoubrie},
220  title = {Public perceptions about nanotechnology: Risks, benefits and trust},
221}
222
223@article{macoubrie2006nanotechnology,
224  publisher = {Sage Publications Sage CA: Thousand Oaks, CA},
225  year = {2006},
226  pages = {221--241},
227  number = {2},
228  volume = {15},
229  journal = {Public Understanding of Science},
230  author = {Macoubrie, Jane},
231  title = {Nanotechnology: public concerns, reasoning and trust in government},
232}
233
234@article{siegrist2007laypeople,
235  publisher = {Wiley Online Library},
236  year = {2007},
237  pages = {59--69},
238  number = {1},
239  volume = {27},
240  journal = {Risk Analysis: An International Journal},
241  author = {Siegrist, Michael and Keller, Carmen and Kastenholz, Hans and Frey, Silvia and Wiek, Arnim},
242  title = {Laypeople's and experts' perception of nanotechnology hazards},
243}
244
245@article{siegrist2007public,
246  publisher = {Elsevier},
247  year = {2007},
248  pages = {459--466},
249  number = {2},
250  volume = {49},
251  journal = {Appetite},
252  author = {Siegrist, Michael and Cousin, Marie-Eve and Kastenholz, Hans and Wiek, Arnim},
253  title = {Public acceptance of nanotechnology foods and food packaging: The influence of affect and trust},
254}
255
256@article{horowitz2016public,
257  url = {https://doi.org/10.1177/2053168015627183},
258  publisher = {SAGE Publications Sage UK: London, England},
259  year = {2016},
260  number = {1},
261  volume = {3},
262  journal = {Research \& Politics},
263  author = {Horowitz, Michael C},
264  title = {Public opinion and the politics of the killer robots debate},
265}
266
267@article{young2018does,
268  publisher = {Oxford University Press},
269  year = {2018},
270  pages = {562--576},
271  number = {3},
272  volume = {62},
273  journal = {International Studies Quarterly},
274  author = {Young, Kevin L and Carpenter, Charli},
275  title = {Does science fiction affect political fact? Yes and no: A survey experiment on “Killer Robots”},
276}
277
278@inproceedings{saxena2019fairness,
279  organization = {ACM},
280  year = {2019},
281  pages = {99--106},
282  booktitle = {Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society},
283  author = {Saxena, Nripsuta Ani and Huang, Karen and DeFilippis, Evan and Radanovic, Goran and Parkes, David C and Liu, Yang},
284  title = {How Do Fairness Definitions Fare?: Examining Public Attitudes Towards Algorithmic Definitions of Fairness},
285}
286
287@inproceedings{cave2019scary,
288  organization = {ACM},
289  year = {2019},
290  pages = {331--337},
291  booktitle = {Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society},
292  author = {Cave, Stephen and Coughlan, Kate and Dihal, Kanta},
293  title = {Scary robots: examining public responses to AI},
294}
295
296@techreport{politicallyneutral2019,
297  url = {https://perma.cc/JA6M-RDUS},
298  year = {2019},
299  institution = {Center for Security Studies (CSS) at ETH Zurich},
300  title = {A Politically Neutral Hub for Basic AI Research},
301  author = {Sophie-Charlotte Fischer and Andreas Wenger},
302}
303
304@techreport{facevalue2019,
305  url = {https://perma.cc/M4EG-N5MY},
306  year = {2019},
307  institution = {Ada Lovelace Institute},
308  title = {Beyond face value: public attitudes to facial recognition technology},
309  author = {{Ada Lovelace Institute}},
310}
311
312@techreport{eurobarometer460,
313  url = {https://perma.cc/9FRT-ADST},
314  institution = {Eurobarometer},
315  year = {2017},
316  author = {Eurobarometer},
317  title = {Special Eurobarometer 460: Attitudes towards the impact of digitisation and automation on daily life},
318}
319
320@techreport{rs2018,
321  url = {https://perma.cc/79FE-TEHH},
322  type = {Survey report},
323  institution = {The Royal Society},
324  year = {2018},
325  author = {{Ipsos MORI}},
326  title = {Public Views of Machine Learning: Findings from public research and engagement conducted on behalf of the Royal Society},
327}
328
329@techreport{smith2017,
330  url = {https://perma.cc/WU6B-63PZ},
331  institution = {Pew Research Center},
332  year = {2016},
333  author = {Aaron Smith and Monica Anderson},
334  title = {Automation in Everyday Life},
335}
336
337@techreport{doherty2019americans,
338  url = {https://perma.cc/PHQ3-LD8D},
339  institution = {Pew Research Center},
340  year = {2019},
341  author = {Carroll Doherty and Jocelyn Kiley},
342  title = {Americans have become much less positive about tech companies’ impact on the U.S.},
343}
344
345@techreport{gecewicz2019americans,
346  url = {https://perma.cc/F75M-ZRYA},
347  institution = {Pew Research Center},
348  year = {2019},
349  author = {Claire Gecewicz and Lee Rainie},
350  title = {Why Americans Don’t Fully Trust Many Who Hold Positions of Power and Responsibility},
351}
352
353@techreport{zhang2019artificial,
354  url = {http://dx.doi.org/10.2139/ssrn.3312874},
355  year = {2019},
356  institution = {Centre for the Governance of AI, University of Oxford},
357  author = {Zhang, Baobao and Dafoe, Allan},
358  title = {Artificial intelligence: American attitudes and trends},
359}
360
361@article{newton2017verge,
362  url = {https://perma.cc/F75M-ZRYA},
363  journal = {The Verge},
364  year = {2017},
365  author = {Casey Newton, Nick Statt, and Michael Zelenko},
366  title = {The Verge Tech Survey},
367}
368
369@article{molla2018facebook,
370  url = {https://perma.cc/F75M-ZRYA},
371  journal = {Recode},
372  year = {2018},
373  author = {Rani Molla},
374  title = {Facebook is the least-trusted major tech company},
375}
376
377@article{kahn2018americans,
378  url = {https://perma.cc/5KSL-2253},
379  journal = {Reuters},
380  year = {2018},
381  author = {Chris Kahn and David Ingram},
382  title = {Americans less likely to trust Facebook than rivals on personal data: Reuters/Ipsos poll},
383}
384
385@techreport{west2018divided,
386  url = {https://perma.cc/3NJV-5GV4},
387  institution = {Brookings Institution},
388  type = {Survey report},
389  year = {2018},
390  author = {Darrell M. West},
391  title = {Brookings survey finds divided views on artificial intelligence for warfare, but support rises if adversaries are developing it},
392}
393
394@article{nosek2018preregistration,
395  year = {2018},
396  number = {11},
397  volume = {115},
398  pages = {2600--2606},
399  journal = {Proceedings of the National Academy of Sciences},
400  author = {Nosek, Brian A and Ebersole, Charles R and DeHaven, Alexander C and Mellor, David T},
401  title = {The preregistration revolution},
402}
403
404@article{lin2016standard,
405  year = {2016},
406  pages = {495--500},
407  number = {3},
408  volume = {49},
409  journal = {PS: Political Science \& Politics},
410  author = {Lin, Winston and Green, Donald P},
411  title = {Standard operating procedures: A safety net for pre-analysis plans},
412}
413
414@article{funk2017,
415  url = {https://perma.cc/UF9P-WSRL},
416  year = {2017},
417  number = {1},
418  volume = {34},
419  journal = {Issues in Science and Technology},
420  author = {Cary Funk},
421  title = {Real Numbers: Mixed Messages about Public Trust in Science},
422}

Attribution

arXiv:1912.12835v1 [cs.CY]
License: cc-by-4.0

Related Posts

Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society

Beyond Near- and Long-Term: Towards a Clearer Account of Research Priorities in AI Ethics and Society

Introduction There is a growing community of researchers focused on ensuring that advances in AI are safe and beneficial for humanity (which we call the AI ethics and society community, or AI E&S for short.

How to design an AI ethics board

How to design an AI ethics board

Introduction It becomes increasingly clear that state-of-the-art artificial intelligence (AI) systems pose significant societal risks.

The Right Tool for the Job: Open-Source Auditing Tools in Machine Learning

The Right Tool for the Job: Open-Source Auditing Tools in Machine Learning

Introduction As a society, we are tossed around on a sea of competing ideas, backed by statistics, data, and algorithms.