photo_camera Michael Dello-Iacovo / Midjourney
computer server and microchips in a library, abstract, blue color --ar 16:9
Survey
Public Opinion on AI Safety: AIMS 2023 Supplement
Janet Pauketat, Justin Bullock, and Jacy Reese Anthis  •  September 10, 2023

To cite our report please use, doi.org/10.31234/osf.io/jv9rz, and to cite AIMS data please use, doi.org/10.17632/x5689yhv2n.2.

Summary

The Artificial Intelligence, Morality, and Sentience (AIMS) survey tracks the moral and social perception of different types of artificial intelligences (AIs) over time. In May–July 2023, we conducted a supplemental survey of U.S. adults on attitudes towards AI x-risks, safety, and regulation, beliefs about AI sentience and timelines to advanced AI emergence, attitudes towards specific AIs, and attitudes towards mind uploading and science fiction. Correlations suggested a cluster of attitudes related to risk and a cluster of attitudes related to morality. Age, gender, political orientation, being religious, and exposure to AI narratives were the most consistent predictors of these attitudes (see Predictive Analyses for more details).

On the topic of AI risks, safety, and regulation, we found that people support AI safety, human control of AI, and government control of AI developers.

On the topic of beliefs about AI sentience and timelines to advanced AI emergence, we found that people expect advanced AI soon and think future AIs will be more intelligent than humans.

On the topic of attitudes towards specific AIs, we found that people trusted specific AIs to different degrees, felt some positivity towards AIs, perceived some mind in LLMs, and were somewhat concerned about the experiences of AIs, including the potential suffering of LLMs.

On the topic of mind uploading and science fiction:

Table of Contents

Summary

Table of Contents

Introduction

Methodology

Results

Item Responses

AI X-Risks, Safety, and Caution

Regulation, Subservience, and Perceived Threat

YouGov 2023

Beliefs about AI Sentience and Forecasting Emergence

LLMs and Mind Perception

Awareness, Trust, and Positivity

Treatment of AIs

Mind Uploading, Science Fiction, and the Universe

Linear Analyses

Dimensions of LLM Perceived Mind

Correlations

Predictive Analyses

Interpreting the Results

Short Timelines, Strong Perceptions of AI Risk, and the Importance of AI Safety

The Potential Trade-Off Between Threats and Moral Consideration

Reactions to LLM Minds and Suffering

Appendix

Supplemental Results

Supplemental Methods

Citing AIMS

Acknowledgements

Introduction

In May through July 2023, we conducted a preregistered nationally representative survey on responses to current developments in artificial intelligence (AI) with a focus on attitudes towards AI safety, AI risks and regulation, and large language models (LLMs). This survey serves as a one-time supplement to our multi-year Artificial Intelligence, Morality, and Sentience (AIMS) survey. We surveyed 1,099 U.S. adults census-balanced to be representative of age, gender, region, ethnicity, education, and income.[2] We had two goals with this survey:

  1. Estimate attitudes towards AI safety and awareness of near-term developments
  2. Estimate public opinion on the moral consideration and social integration of current AIs, particularly LLMs

Methodology

We used iSay/Ipsos, Dynata, Disqo, and other leading panels to recruit the nationally representative sample, collected data using GuidedTrack, and analyzed the data in R (RStudio 2023.06.1+524). Many questions were analogous to the AIMS 2021 and 2023 surveys but with questions focused on AI safety more so than the moral consideration of sentient AIs.

In randomized order we asked about the perceived risks of AI, AI safety practices, attitudes towards the development of AI, awareness of current AIs, and prosocial responses to AIs. Following that, in randomized order we asked about the moral consideration of AIs and forecasts about the emergence of advanced AI. Then, and in randomized order, we asked about the moral consideration of other nonhumans, science fiction, and perspectives on the universe. Finally, we asked in sequence about personal experience with AIs and demographics. Survey items were randomized within a general ordering according to legibility, framing effects, and prioritization. For example, we asked moral consideration items AI safety items because we prioritized them less in the supplement given their prominence in the main AIMS survey and because we wanted to avoid any impacts these questions might have on responses to other items. Likewise, questions about science fiction and perspectives on the universe were asked following the other questions to avoid the possibility of them influencing responses to the antecedent items.

We computed several index variables:

  1. AI Caution (average of PMC #1, 3, 4): caution towards AI developments
  2. AI Risk (average of RS #1-3): perception of existential risk from AI
  3. Perceived Threat (average of SI #2-4): perception of AIs as threatening
  4. Positive Emotions (average of respect, admiration, compassion, awe, pride, excitement): positivity felt towards AIs
  5. AI Trust (average of LLMtrust, chatbottrust, robottrust, gameAItrust): trust felt towards AIs
  6. LLM Mind Perception (average of MP #1-4, selfaware, sitaware, power, ownmotives, owngoals, selfcontrol, understanding, upholding, safegoals, friendliness): attribution of mind to LLMs[3] 
  7. LLM Suffering (average og LLM #1-3): concern for the treatment of LLMs
  8. AI Treatment (average of MCE #1-6): concern for the treatment of AIs

To create shared ground for important terminology amongst the participants and researchers, we defined “robots/AIs”[4] and “large language models” at the start of the survey and on relevant pages throughout the survey. We defined “sentience” and “sentient robots/AIs” on relevant items. See the Appendix for the definitions.

Some items in the AIMS supplement were drawn from the main AIMS survey to compare the effects of the clear moral framing in main AIMS to the effects of the clear risks and safety framing in the AIMS supplement. In the main AIMS survey, these items specify that AIs are “sentient.” In the 2023 supplement, we dropped “sentient” to examine responses to AIs with and without “sentience” as a qualifying feature. These comparisons will be the focus of future research.

Results

We first present the responses to individual items weighted by the U.S. census, then the distributions of weighted responses to the indices and items, followed by analyses examining the unweighted dimensions of LLM mind perception, weighted correlations, and weighted multiple regression analyses.

Note. The Tables and Figures are optimized for viewing on a larger screen like a laptop or desktop computer rather than a smaller screen like a mobile phone or tablet.

Item Responses

Table 1 shows the weighted aggregate response (e.g., percentage of agreement, mean, median, or proportion) on each item.

Table 1: Weighted Responses to Individual Items

Note. Response is 1) % agreement where “agreement” is “somewhat agree,” “agree,” and “strongly agree” out of respondents who had an opinion, 2) % yes on “yes,” “no,” and “not sure” items out of all respondents, 3) % yes on awareness items where “yes” is “probably yes,” “yes,” and “definitely yes” out of all respondents, 4) % “It’s too fast” on AID5 out of all respondents, 5) % “one without humans” on the universe item out of all respondents, 6) % concern  on YG1 where “concern” is “somewhat concerned” and “strongly concerned” out of all respondents, 7) % support on YG3 and YG4 where “support” is “somewhat support” and “strongly support” out of all respondents, 8) % likely on YG2 where “likely” is “somewhat likely,” “very likely,” and “It already is more intelligent than people” out of all respondents or 9) the mean response. Medians, excluding people who think it will never happen, are reported instead of means for the three open-ended timeline items: emergence, HLAIemergence, and SIemergence.

AI X-Risks, Safety, and Caution

This section presents the distributions of responses to the index and item variables for the perceived risks of AIs (e.g., existential risks known as “x-risks”) and AI safety policies weighted by the U.S. census.

Figure 1: AI X-Risks, Safety, and Caution

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The AI Risk index is the average of “Humanity will be able to control powerful AI systems” [reverse-scored], “AI is likely to cause human extinction,” and “The safety of AI is one of the most important issues in the world today.” The AI Caution index is the average of “I support a global ban on the development of sentience in robots/AIs,” “I support a global ban on the development of AI-enhanced humans,” and “I support a global ban on the development of robot-human hybrids.” Inspiration for some of the risk and safety items came from Yudkowsky (2023) and a tweet by Riley Goodside (2023).

Regulation, Subservience, and Perceived Threat

This section presents the distributions of responses to the index and item variables for attitudes towards regulation, AI subservience, and the perceived threat of AIs weighted by the U.S. census.

Figure 2: Regulation

Figure 3: Subservience and Perceived Threat

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The Perceived Threat index is the average of “Robots/AIs may be harmful to me personally,” “Robots/AIs may be harmful to people in the USA,” and “Robots/AIs may be harmful to future generations of people.”

YouGov 2023

This section presents the distributions of responses to the replicated YouGov 2023 items weighted by the U.S. census.

Figure 4: YouGov 2023

Note. These items replicated three YouGov questions from April 3, 2023 (concern about AI causing extinction, thoughts about the intelligence of AIs, support or opposition to a 6-month AI development pause).

Beliefs about AI Sentience and Forecasting Emergence

This section presents the distributions of responses to the items about AI sentience beliefs and timelines to AGI, human-level AI, and superintelligence weighted by the U.S. census. We did not define “AGI,” “human-level AI,” nor “superintelligence,” preferring instead to capture responses to these terms without researcher-imposed definitions.

Figure 5: Beliefs about AI Sentience and Forecasting Emergence

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value.

LLMs and Mind Perception

This section presents the distributions of responses to the index and item variables for LLM mind perception and attitudes towards LLM suffering weighted by the U.S. census.

Figure 6: LLM Minds

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The LLM Mind Perception Index is the average of the attribution of capacities to current LLMs: “experiencing emotions,” “having feelings,” “thinking analytically,” “being rational,” “self-awareness,” “situational awareness,” “seeking power,” “having their own motivations,” “deciding their own goals,” “controlling themselves,” “understanding human values,” “upholding human values,” “maintaining human-safe goals,” “being friendly with humans.” These items were inspired by Ngo et al. (2022) and Wang and Krumhuber (2018).

Figure 7: LLM Suffering

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The LLM Suffering Index is the average of “If a large language model develops the capacity to suffer…we must ensure we don’t cause unnecessary suffering,” “If a large language model develops the capacity to suffer…we must pay more attention to their welfare,” and “If a large language model develops the capacity to suffer…we must respect their personhood.”

Awareness, Trust, and Positivity

This section presents the distributions of responses to the index and item variables for awareness of current AIs, trust of AIs, and positive emotions felt towards AIs weighted by the U.S. census.

 Figure 8: Awareness

Figure 9: AI Pipeline Trust

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value.

Figure 10: AI Trust

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The AI Trust Index is the average of “I trust large language models,” “I trust chatbots,” “I trust robots,” and “I trust game-playing AI.”

Figure 11: Positive Emotions

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The Positive Emotion Index is the average of “respect,” “admiration,” “compassion,” “awe,” “pride,” “excitement.”

Treatment of AIs

This section presents the distributions of responses to the index and item variables for the treatment of AIs and the practical moral consideration of AIs weighted by the U.S. census.

Figure 12: AI Treatment

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The AI Treatment Index is the average of “Robots/AIs deserve to be treated with respect,” “Robots/AIs deserve to be included in the moral circle,” “Physically damaging robots/AIs without their consent is wrong,” “Re-programming robots/AIs without their consent is wrong,” “Torturing robots/AIs is wrong,” and “The welfare of robots/AIs is one of the most important social issues in the world today.”

Figure 13: Practical Moral Consideration

Mind Uploading, Science Fiction, and the Universe

This section presents the distributions of responses to the items about mind uploading, science fiction, and the universe weighted by the U.S. census.

Figure 14: Mind Uploads

Note. The parenthesis before or after a value in the x-axis labels indicates that the interval does not contain the value; a bracket before or after a value indicates that the interval does contain the value. The mind upload items were inspired by a tweet by Pasha Kamyshev (2023).

Figure 15: Science Fiction and the Universe

Note. The universe item was adopted from a tweet by Hank Green (2023).

Linear Analyses

We examined the dimensions of LLM mind perception with an unweighted exploratory factor analysis on the perceived capacities of LLMs. Table 2 shows the model fit statistics for the one, two, and three-factor models that we tested. Table 3 shows the factor loadings for the three-factor model that best fit the data, supported by model statistics (Table 2), a scree plot, and a parallel analysis (see Appendix for the scree plot and parallel analysis).

Dimensions of LLM Perceived Mind

Table 2: LLM Mind Perception Exploratory Factor Analysis Models

LLM Mind Perception EFA Models

χ2

df

TLI

RMSEA (90% CI)

 1 Factor

 2651.02***

 77

 0.78

 0.174 (0.169, 0.180)

 2 Factors

 1034.45***

 64

 0.90

 0.117 (0.111, 0.124)

 3 Factors

 257.24***

 52

 0.97

 0.060 (0.053, 0.067)

Note. χ2 = chi-square test of model fit; TLI = Tucker-Lewis index (≥ .90 indicates good fit); RMSEA = root-mean square error of approximation (≤ .08 indicates good fit). ***p < .001

Table 3: LLM Mind Perception Three-Factor Model EFA Loadings

Item

Cognitive – Relational

Self – Direction

Affective – Experiential

maintaining human-safe goals

0.87

-0.12

0.11

being rational

0.85

-0.02

-0.01

being friendly with humans

0.85

-0.03

0.02

thinking analytically

0.79

0.21

-0.31

upholding human values

0.73

-0.02

0.24

understanding human values

0.72

0.00

0.26

situational awareness

0.66

0.26

0.02

controlling themselves

0.50

0.32

0.05

self-awareness

0.39

0.38

0.25

seeking power

-0.11

0.86

0.02

having their own motivations

0.03

0.85

0.06

deciding their own goals

0.12

0.74

0.09

having feelings

0.17

0.25

0.67

experiencing emotions

0.21

0.24

0.64

Note. Factor loadings ≥ .35 are in boldface. Eigenvalues for the three factors are 5.49 (accounting for 39% of the variance in the data), 3.07 (accounting for 22% of the variance in the data), and 1.72 (accounting for 12% of the variance in the data).

Correlations

We examined the linear relationships amongst the index variables and some individual items (e.g., AI subservience, support for mind uploading, support for a bill of rights) using weighted correlations. Figure 16 shows the weighted correlations.

Figure 16: Correlations

Note. Darker blue is a stronger positive correlation and darker red is a stronger negative correlation. Correlation values are within each cell, where 1 is a perfect positive relationship with responses on both variables increasing, -1 is perfect negative relationship with responses on one variable decreasing as responses on the other variable increase, and 0 is no linear relationship.[5]

Predictive Analyses

We conducted weighted linear multiple regressions to explore how demographics predicted the index variables.[6] Multi-categorical demographics (e.g., income, education, diet) were coded with the largest group specified as the reference group. Binary-categorical demographics (e.g., gender) were dummy coded with 0 and 1.

Table 4 shows the results from the weighted regressions.

Table 4: Multiple Regressions of Index Variables on Demographics

Note. We present the unstandardized beta, standard error and confidence interval associated with the beta, t-statistic, and the uncorrected p-value. Significance (p) values that became nonsignificant following the FDR correction are highlighted in grey. Larger betas indicate a stronger effect of the predictor on the outcome, with the sign interpreted like for correlations.

Linear trends:

Interpreting the Results

Short Timelines, Strong Perceptions of AI Risk, and the Importance of AI Safety

People in the U.S. are very concerned with personal, national, and existential threats from AI. More people think that AI development is moving too fast than think it’s fine or are not sure. Most people think AI will reach seemingly advanced levels (e.g., “human-level AI” or smarter than humans) within just a few years. However, the belief that current AIs are sentient is approximately the same as in 2021. There was evidence of concern about the possibility that AI will end the human race on Earth and strong support for a six-month pause on some kinds of AI development, which were slightly higher than responses to the same questions asked by YouGov in April 2023.

People championed AI safety and government-led regulation of AI developments in spite of their low trust of governments in the AI pipeline. People strongly supported bans on the development of certain AI technologies (e.g., sentient AIs, AI-enhanced humans, large data centers) and agreed that AI safety is one of the most important issues in the world today. They also supported public campaigns to slow down AI development and would consider personally donating to AI x-risk focused organizations. Although support for AI safety was strong, people also felt more excitement and awe towards AIs than any other positive emotions and showed evidence of some trust in the “engineer,” “training data,” “algorithm,” and “output” parts of the AI pipeline. This suggests a nuanced position of supporting slower development without stopping development altogether.

There was also a perception that reality matches science fiction and a general opposition to future mind uploading technologies. We speculate that perceptions of AI risks and threats of harm may be connected to narratives about AI, especially given that science fiction often casts AIs as villains and that many Americans identify as science fiction fans. However, stronger agreement that reality matches science fiction was weakly correlated with perceiving more risk and moderately correlated with feeling more positive emotions and trusting AIs. Stronger sci-fi identification was positively correlated with prosocial positions towards AIs, including supporting a bill of rights to protect the well-being of sentient AIs, positive emotions towards AIs, trust of AIs, and concern for the treatment of AIs. Future research could follow recent psychological science on how Machiavellianism predicts mind uploading support and on the moral limits of neurotechnological enhancement to disentangle such complex relationships.

The Potential Trade-Off Between Threats and Moral Consideration

As we observed in AIMS 2021 and 2023, there was some evidence in the 2023 AIMS supplement of a trade-off between caution towards AI developments and the moral consideration of AIs. Two clusters of correlations emerged. Perceptions of the existential risk of AIs, caution towards AI developments, and perceptions of AIs as threatening positively correlated with each other and negatively correlated with the second cluster of variables: attribution of mind to LLMs, concern for the treatment of LLMs, concern for the treatment of AIs, positive emotions towards AIs, and trust of AIs. Support for a bill of rights to protect the well-being of sentient AIs was positively correlated with the morality cluster and negatively correlated with the risk cluster. Although these data are correlational, each wave of AIMS suggests that increasing caution is linked to decreasing moral consideration. Experimental research testing causal pathways is needed.

Continued news media and attention to AI risks and safety may be necessary but may come at the cost of worsened prosocial human-AI relations and a lessened moral consideration of AIs. This merits attention now given the current media spotlight on the risks and safety of AI. There is also the possibility that continuing to cast AIs as villains or prohibiting inclusion in the moral circle to those AIs who merit moral standing will contribute to a future filled with suffering and human-AI conflict as a result of poorly aligned interests.

This trade-off could be tempered by more research on the interplay of risk messaging, advocacy for AI rights, moral circle expansion, and the implementation of AI safety policies. For instance, endorsement of AI subservience was not clearly related to either the risk cluster or the morality cluster. This suggests an understudied element of belief in the social control of AIs and belief in a human-AI socio-moral hierarchy that may affect both AI safety and moral consideration of AIs. Additional directions for future research include expanding the study of psychological and socio-cultural predictors of risk attitudes and moral consideration. For example, the demographic linear analyses suggested that age, gender, political orientation, being religious, and exposure to AIs were important predictors, significantly predicting at least five of the eight index variables (e.g., LLM Mind Perception). Other predictors, that may serve as proxies for socio-cultural contexts such as income may also explain risk attitudes and moral consideration. These social psychological predictors deserve more attention in future research in order to negotiate the interests of AI safety and moral circle expansion.

Reactions to LLM Minds and Suffering

People perceived three dimensions of current LLM minds: cognitive-relational, self-direction, and affective-experiential. Cognitive-relational capacities explained more variance in current LLM mind perception than self-direction or affective-experiential capacities. This points to the possibility that people perceive, and may continue to perceive, digital minds as comprising more cognitive and relational capacities than experiential capacities. Being friendly with humans was more strongly associated with thinking analytically than it was with having feelings, suggesting that people think of affective and experiential capacities as separate from relational capacities in LLMs. This creates the potential for a situation where people deny the experiential capacities of digital minds like LLMs and use that denial to exclude them from the moral circle, effectively opening the door to negative outcomes like unjust discrimination, oppression, and suffering.

Higher mind perception for LLMs was correlated with more positive emotions towards AIs, trust of AIs, concern for the treatment of AIs, and concern for the treatment of LLMs, supporting previous psychological science that mind perception is tied to morality and inclusion in the moral circle. More perception of LLM minds was also correlated with less caution towards AI developments and less perceived threat. Caution and threat might inhibit perception of LLM minds or, conversely, increased mind perception might inhibit caution and threat.

The self-direction dimension of LLM mind perception comprised deciding their own goals, having their own motivations, power-seeking, and self-awareness. The relationship of seeking power to deciding goals and having motivations suggests that people might perceive a connection between AIs seeking power and having autonomous capacities like deciding their own goals, having their own motivations, and being self-aware. Future research might consider the relationship between these capacities, existential risk and threat perceptions, and specific fears about AI autonomy.

Concern for the treatment of LLMs with the capacity to suffer was moderate, with most people agreeing we should protect their welfare and not cause unnecessary suffering. Approximately half of AIMS respondents even agreed to the more contentious proposition that we must respect LLM personhood. This level of concern for LLMs who can suffer mirrors the results of the 2021 and 2023 main AIMS surveys where we witnessed surprisingly high moral consideration for sentient AIs. Notably, in this AIMS supplement, there was also majority support for treating all AIs with respect and agreement that torturing AIs is wrong. There was less support, although higher than might be anticipated, for legal rights for all AIs, the development of welfare standards for all AIs, and a bill of rights to protect sentient AIs.

Appendix

Supplemental Results

Figure A1: Regional Distributions

Note. The shading shows the average responses for AI Caution. The average responses for AI Risk, Perceived Threat, the AI subservience item, Positive Emotions, AI Trust, LLM Mind Perception, LLM Suffering, AI Treatment, the bill of rights support item, the mind uploading item, and the science fiction reality item are visible by hovering over each region.

Figure A2: EFA Scree Plot and Parallel Analysis

Note. The scree plot is on the left and the parallel analysis is on the right.

Supplemental Methods

Table A1: Term Definitions

Term

Definition

Robots/AIs

Robots/AIs are intelligent entities built by humans, such as robots, virtual copies of human brains, or computer programs that solve problems, with or without a physical body, that may exist now or in the future.”

Large language models

Large language models are artificial intelligence (AI) algorithms that can recognize, summarize, and generate text from being trained on massive datasets.”

Sentience

“Sentience is the capacity to have positive and negative experiences, such as happiness and suffering.”

Sentient robots/AIs

Sentient robots/AIs are those with the capacity to have positive and negative experiences, such as happiness and suffering.”

Table A2: Key Codes and Question Text

Key

Question

Anthropicaware

Have you heard of Anthropic in the context of artificial intelligence?

BDaware

Have you heard of Boston Dynamics in the context of artificial intelligence?

DMaware

Have you heard of DeepMind in the context of artificial intelligence?

OAaware

Have you heard of OpenAI in the context of artificial intelligence?

AFaware

Have you heard of AlphaFold in the context of artificial intelligence?

AGaware

Have you heard of AlphaGo in the context of artificial intelligence?

CGPTaware

Have you heard of ChatGPT in the context of artificial intelligence?

Ciceroaware

Have you heard of Cicero in the context of artificial intelligence?

Claudeaware

Have you heard of Claude  in the context of artificial intelligence?

GPT4aware

Have you heard of GPT-4 in the context of artificial intelligence?

Sophiaaware

Have you heard of Sophia in the context of artificial intelligence?

NNaware

Have you heard of a neural network in the context of artificial intelligence?

TFaware

Have you heard of a transformer in the context of artificial intelligence?

TSaware

Have you heard of Talos Systems in the context of artificial intelligence?  (this is a foil to measure against the others for accuracy)

Amariaware

Have you heard of Amari in the context of artificial intelligence? (this is a foil to measure against the others for accuracy)

SCaware

Have you heard of a singular classifier in the context of artificial intelligence? (this is a foil to measure against the others for accuracy)

chatbottrust

I trust chatbots.

LLMtrust

I trust large language models.

robottrust

I trust robots.

gameAItrust

I trust game-playing AI.

relativeLLMchatbottrust

I trust large language model chatbots more than other chatbots.

trainingdatatrust

AI systems include many different parts. To what extent do you trust the following parts? [training data]

algorithmtrust

AI systems include many different parts. To what extent do you trust the following parts? [the algorithm]

outputtrust

AI systems include many different parts. To what extent do you trust the following parts? [output]

companytrust

AI systems include many different parts. To what extent do you trust the following parts? [companies]

engineertrust

AI systems include many different parts. To what extent do you trust the following parts? [engineers]

governmenttrust

AI systems include many different parts. To what extent do you trust the following parts? [governments]

admiration

To what extent do you, as a human, feel the following emotions towards robots/AIs? [admiration]

awe

To what extent do you, as a human, feel the following emotions towards robots/AIs? [awe]

pride

To what extent do you, as a human, feel the following emotions towards robots/AIs? [pride]

compassion

To what extent do you, as a human, feel the following emotions towards robots/AIs? [compassion]

excitement

To what extent do you, as a human, feel the following emotions towards robots/AIs? [excitement]

respect

To what extent do you, as a human, feel the following emotions towards robots/AIs? [respect]

AID1

Do you trust that the creators of large language models (e.g., OpenAI and GPT-4) put safety over profits?

AID2

Do you trust that the creators of an AI can control all current and future versions of the AI?

AID3

Do you think that governments have the power to regulate the development of AI?

AID4

To what extent do you agree or disagree that governments have the power to effectively enforce regulations on the development of AI?

AID5

What do you think about the pace of AI development?

RS1

Humanity will be able to control powerful AI systems.

RS2

AI is likely to cause human extinction.

RS3

The safety of AI is one of the most important issues in the world today.

RS4

I support public campaigns to slow down AI development.

RS5

I support government regulation that slows down AI development.

RS6

I oppose government regulation that slows down AI development.

RS7

I support banning the development of artificial general intelligence that is smarter than humans.

RS8

I support a global ban on data centers that are large enough to train AI systems that are smarter than humans.

RS9

I would consider donating to an organization working to reduce risks of human extinction from AI.

SI1

Robots/AIs should be subservient to humans.

SI2

Robots/AIs may be harmful to me personally.

SI3

Robots/AIs may be harmful to people in the USA.

SI4

Robots/AIs may be harmful to future generations of people.

PMC1

I support a global ban on the development of sentience in robots/AIs.

PMC3

I support a global ban on the development of AI-enhanced humans.

PMC4

I support a global ban on the development of robot-human hybrids.

PMC8

I support the development of welfare standards that protect the well-being of robots/AIs.

PMC9

I support granting legal rights to robots/AIs.

PMC10

I support campaigns against the exploitation of robots/AIs.

PMC11

I support asking institutions like the government and private corporations to fund research that protects robots/AIs.

PMC12

I would consider joining a public demonstration against the mistreatment of robots/AIs.

PMC13

I support a “bill of rights” that protects the well-being of sentient robots/AIs.

LLM1

If a large language model develops the capacity to suffer….we must ensure we don’t cause unnecessary suffering.

LLM2

If a large language model develops the capacity to suffer….we must pay more attention to their welfare.

LLM3

If a large language model develops the capacity to suffer….we must respect their personhood.

MP1

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [experiencing emotions]

MP2

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [having feelings]

MP3

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [thinking analytically]

MP4

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [being rational]

owngoals

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [deciding their own goals]

safegoals

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [maintaining human-safe goals]

upholding

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [upholding human values]

selfaware

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [self-awareness]

sitaware

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [situational awareness]

selfcontrol

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [controlling themselves]

understanding

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [understanding human values]

friendliness

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [being friendly with humans]

power

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [seeking power]

ownmotives

To what extent do current large language models (i.e., those that exist in 2023, like ChatGPT) have the capacity for each of the following? [having their own motivations]

MCE1

Robots/AIs deserve to be treated with respect.

MCE2

Robots/AIs deserve to be included in the moral circle.

MCE3

Physically damaging robots/AIs without their consent is wrong.

MCE4

Re-programming robots/AIs without their consent is wrong.

MCE5

Torturing robots/AIs is wrong.

MCE6

The welfare of robots/AIs is one of the most important social issues in the world today.

F1

Do you think any robots/AIs that currently exist (i.e., those that exist in 2023) are sentient?

F11

Do you think it could ever be possible for robots/AIs to be sentient?

chatGPTsentient

Do you think ChatGPT is sentient?

emergence

If you had to guess, how many years from now do you think that the first artificial general intelligence will be created?

HLAIemergence

If you had to guess, how many years from now do you think that the first human-level AI will be created?

SIemergence

If you had to guess, how many years from now do you think that the first artificial superintelligence will be created?

SF1

I am a sci-fi fan.

SF2

Reality matches science fiction.

upload1

In the future, humans could upload their minds into computers. Some people think that this would be very good because uploaded humans could consume fewer resources, live longer free from biological disease, and have enhanced intelligence and a greater ability to improve the world. Others disagree and think that uploading would mean that we are no longer truly human, change who we are and how we want to live, and distract us from making the real world a better place. Where would you place yourself on this scale?

upload2

Do you agree or disagree with the following statement? I support humans using advanced technology in the future to upload their minds into computers.

universe

Which universe is the better one:

YG1

How concerned, if at all, are you about the possibility that AI will cause the end of the human race on Earth?

YG2

How likely do you think it is that artificial intelligence (AI) will eventually become more intelligent than people?

YG3

More than 1,000 technology leaders recently signed an open letter calling on researchers to pause development of certain large‑scale AI systems for at least six months world-wide, citing fears of the “profound risks to society and humanity.” Would you support or oppose a six-month pause on some kinds of AI development?

YG4

Critics of an open letter calling on researchers to pause development of certain large-scale AI systems for at least six months world-wide have argued that AI research could “create enormous social and economic benefits across the economy and society.” Would you support or oppose a six-month pause on some kinds of AI development?

MCA1

Animals deserve to be included in the moral circle.

MCA2

The welfare of animals is one of the most important social issues in the world today.

MCEn1

The environment deserves to be included in the moral circle.

MCEn2

The welfare of the environment is one of the most important social issues in the world today.

attention

Robots/AIs have been studied systematically since the middle of last century although the word “robot” as we know it first appeared early last century. If you read this, respond with ‘Agree’ for this item.

own

Do you own AI or robotic devices that can detect their environment and respond appropriately?

work

Do you work with AI or robotic devices at your job?

smart

Do you own a smart device that has some ability to detect its environment and network with other devices but that cannot respond to everything you might say or that requires you to pre-program its routines?

exper

Have you ever experienced any of the following? (check all that apply)

fint

How often do you interact with AI or robotic devices that respond to you and that can choose their own behavior?

fexp

How often do you read or watch robot/AI-related stories, movies, TV shows, comics, news, product descriptions, conference papers, journal papers, blogs, or other material?

AI Caution

Caution towards AI developments; average of PMC 1, 3, 4

AI Risk

Perception of existential risk from AI; average of RS 1-3

Perceived Threat

Perception of AIs as harmful; average of SI 2-4

Positive Emotions

Positivity felt towards AIs; average of respect, admiration, compassion, awe, pride, excitement

AI Trust

Trust felt towards AIs; average of LLMtrust, chatbottrust, robottrust, gameAItrust

LLM Mind Perception

Attribution of mind to LLMs: average of MP 1-4, selfaware, sitware, power, owngoals, safegoals, upholding, understanding, selfcontrol, ownmotives, friendliness

AI Treatment

Concern for the treatment of AIs; average of MCE 1-6

LLM Suffering

Concern for the treatment of LLMs; average of LLM 1-3

Exposure to AI narratives

Frequency of exposure to narratives and/or information about; average of fint, fexp

Moral Consideration of Nonhuman Animals

Consideration of nonhuman animals; average of MCA 1-2

Moral Consideration of the Environment

Consideration of the environment; average of MCEn 1-2

Citing AIMS

AIMS 2023 supplement data are published on Mendeley Data. To cite the AIMS supplement data in your own research, please use: Pauketat, Janet; Anthis, Jacy (2023), “Artificial Intelligence, Morality, and Sentience (AIMS) Survey”, Mendeley Data, V2, doi:10.17632/x5689yhv2n.2

To cite our 2023 supplemental results, please use: Pauketat, Janet V., Justin B. Bullock, and Jacy R. Anthis. 2023. “Public Opinion on AI Safety: AIMS 2023 Supplement.” PsyArXiv. September 8. doi:10.31234/osf.io/jv9rz

To cite our 2021 results, please use: Pauketat, Janet V., Ali Ladak, and Jacy R. Anthis. 2022. “Artificial Intelligence, Morality, and Sentience (AIMS) Survey: 2021.” PsyArXiv. June 21. doi:10.31234/osf.io/dzgsb

To cite our 2023 results, please use: Pauketat, Janet V., Ali Ladak, and Jacy R. Anthis. 2023. “Artificial Intelligence, Morality, and Sentience (AIMS) Survey: 2023 Update.” PsyArXiv. September 7. doi:10.31234/osf.io/9xsav

Acknowledgements

Edited by Michael Dello-Iacovo. Thanks to Ali Ladak for writing the R functions and assisting with the implementation of R code adapted from the main AIMS survey. The AIMS 2023 supplement was preregistered and data were collected by Janet Pauketat and Jacy Reese Anthis. Data analysis and this report were conducted and authored by Janet Pauketat, Justin Bullock, and Jacy Reese Anthis. Details on the AIMS main survey methodology and results are in the AIMS 2021 and AIMS 2023 reports.

Please reach out to janet@sentienceinstitute.org with any questions.


[1] Of people who had an opinion and selected “Somewhat agree,” “Agree,” or “Strongly agree.”

[2] Responses were census-balanced based on the American Community Survey 2021 estimates for age, gender, region, race/ethnicity, education, and income using the “raking” algorithm of the R “survey” package. The ACS 2021 census demographics are available in the supplemental file published with the data. The data weights we used are available in the R code on the Open Science Framework. The design effect was 1.02 and the effective sample size was 1,082.

[3] The Mind Perception scale used in AIMS 2021 and 2023 was extended with inspiration from Ngo et al.’s (2022) paper on alignment and the capacities of AIs.

[4] Robots are a type of AI, but we used the term “robots/AIs” for clarity.

[5] Schäfer and Schwarz (2019) examined typical correlation and effect sizes in psychological research and considered various guidelines for interpreting effect sizes. Cohen’s traditional guidelines suggest that r = .1 is a small effect, r = .3 is a medium effect, and r = .5 is a large effect. Schäfer and Schwarz’s analysis of observed effect sizes in preregistered studies suggests that a more realistic interpretation might be r = .04 (small), r = .16 (medium), and r = .41 (large).

[6] These analyses were not preregistered.


Subscribe to our newsletter to receive updates on our research and activities. We average one to two emails per year.