Zum Inhalt

Revisiting computer authorship: a longitudinal perspective

  • Open Access
  • 03.02.2026
  • Review

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Dieser Artikel untersucht das sich entwickelnde Konzept der Urheberschaft im Zeitalter der künstlichen Intelligenz und konzentriert sich auf die Herausforderungen und Unklarheiten, die es mit sich bringt, computergenerierten Texten Urheberschaft zuzuschreiben. Mittels einer Längsschnittstudie, die Umfragedaten aus den Jahren 2018 und 2025 vergleicht, wird untersucht, wie sich die Wahrnehmung der Urheberschaft mit dem Aufstieg großer Sprachmodelle und künstlich erzeugter Inhalte verändert hat. Die Studie zeigt, dass einige quantitative Ergebnisse zwar statistische Signifikanz aufweisen, die Teilnehmer jedoch hinsichtlich der Zuordnung spezifischer Autorenschaft zu KI-Systemen ambivalent bleiben. Zu den Schlüsselthemen zählen die vierfache Definition von Autorschaft, der Einfluss des Bildungshintergrundes auf die Zuordnung von Autorschaft und die Genre-Spezifität der Wahrnehmung von Autorschaft. Der Artikel kommt zu dem Schluss, dass "Autorenschaft" möglicherweise nicht der am besten geeignete Begriff ist, um die rechnergestützte Texterstellung zusammenzufassen, und betont die Notwendigkeit einer neuen Sprache, die sich an diese technologischen Fortschritte anpasst. Die Ergebnisse unterstreichen die Komplexität der Urheberschaft im digitalen Zeitalter und die anhaltende Debatte über die rechtlichen und ethischen Implikationen künstlich erzeugter Inhalte.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

1 Introduction

Artificial intelligence has always been right around the corner in more ways than one. For example, when the subfield of computer science was first identified in the mid-1950s, the—in hindsight, rather optimistic—expectation was that “significant advance [could] be made […] if a carefully selected group of scientists work on it together for a summer” (McCarthy et al. 2006). In the following seven decades, the field has seen plenty of ups and downs, including several ‘AI winters’, wherein highly public failures to meet overly positive expectations resulted in massive funding cuts (Russell and Norvig 2003: 21–24). At the same time, as famously noted by Douglas Hofstadter in the 1970s, whenever advances in AI allow for some new task to be completed automatically, the general perception tends to shift so that the task is no longer viewed as requiring intelligence; real ‘artificial intelligence’ is always almost within reach, and whatever we can do now is mere computation (Hofstadter 1999: 601). Examples of this ‘AI Effect’ include many computational tasks classically thought of as AI problems, but which very few—especially amongst the general public—would today be characterised as ‘AI’, such as email spam filtering and spell checking.
Against these two observations, the changing discussions in the wake of recent advances in producing natural (i.e., human) language texts that appear coherent are curious. A technological method called large language models (LLMs; see, e.g., Devlin et al. 2019; Radford et al. 2019) has again brought AI into the limelight, going so far as to shift the casual parlance definition of AI into near-synonymy with ChatGPT (Schulman et al. 2022) and the names of other generative AI systems. Some go as far as claiming that these new tools will revolutionise everything in our lives, and they will—any day now—reach a level where we will view them as conscious in the same way humans are. Many of the world’s largest technology companies are currently working towards such artificial general intelligence (AGI). ChatGPT’s parent company OpenAI is one of the most prominent proponents of AGI, claiming that “[w]e can imagine a world in which humanity flourishes to a degree that is probably impossible for any of us to fully visualize yet. We hope to contribute to the world an AGI aligned with such flourishing” (Altman 2023).1 However one feels about AI and AGI, and however ‘right around the corner’ the technology may or may not be, the fact is that software systems that communicate with us through natural language are increasingly common, and producing increasingly convincing texts. As the texts they produce become more complex and the input of individual humans in their production decreases, societies are forced to (re-)evaluate stances on a plethora of topics.
In this paper, we interrogate one specific aspect of natural language generation that has been the subject of ongoing, complex cultural conversation: authorship. Who—or what—is the author of the computer-generated text, and why? Does—or can—automated natural language generation adhere to conventional conceptions of authorship, or might we require new language to more firmly align with these new means of text production? Following a brief review of literature related to social perceptions and uses of AI systems that ‘understand’ and produce text, we present the methods and results of an international online survey that solicited general public perceptions of the concept of authorship as it pertains to computer-generated texts in late 2024/early 2025. The same survey was previously administered in 2017/2018, allowing us to investigate not only what the present views on authorship are, but also how those views have, or have not, changed over time. In the discussion, we highlight participants’ ambivalence and uncertainty about authorship of computer-generated texts. Authorship, it seems, is hardly clear-cut, and natural language generation makes it even less so.

2 Literature review

People have always used highly anthropomorphic language about computers. The perceptrons of the 1950s evoked the imagery of electronic brains composed of firing neurons (Minsky and Papert 1988; Rosenblatt 1958). Since then, a whole sect of computer scientists has focused on making computers ‘learn’, while others are trying to make computers ‘understand’ natural language or ‘reason’ (for a critical commentary of these terms, see Bender and Koller 2020). Computers ‘hallucinate’ while looking at images or writing text (Ji et al. 2023), and ‘dream’ while producing pictures (Mordvintsev, Olah and Tyka 2015). Computers, it would seem, are just like us—or vice versa.
But with the LLM leap, the discussion appears to have shifted in an unexpected way (Mitchell and Krakauer 2023). Before, even among those who talked about ‘intelligent’ machines that ‘learned’, ‘understood’, or ‘reasoned’, most were keenly aware that the AI systems were not truly understanding the data in a meaningful way. But LLMs, with their ability to produce coherent language about a variety of topics, appear to have reached a critical point in which debate has emerged regarding whether this lack of true understanding is the case anymore. A recent study found that the technical community was evenly split on whether LLMs could be said to ‘understand’ language or not (Michael et al. 2023). Whether LLM coherence is the final nail in the coffin of language-based tests for sentience—thus putting to rest decades of philosophical thought experiments, such as the Chinese room hypothetical (Searle 1980)—or whether advances in AI will force us to concede that we’ve been relying on the equivalent of the ‘God of the gaps’ argument (Larmer 2002) but with sentience, will remain to be seen.
In any case, these recent events have prompted calls for discussion about whether computer programs should have, for example, legal personhood—discussion that has actually been ongoing for decades (Jaynes 2020; Solum 1992). While these calls mostly come from outside computing research communities, there are some notable exceptions, such as the 2022 case where a Google engineer claimed that a chatbot (LaMDA) had become sentient. This ‘sentience’ was supposedly demonstrated through the chatbot’s own declaration of personhood in conversations with the engineer—conversations that covered a wide range of topics, from philosophy to science (De Cosmo 2022). While a system’s claim to personhood does not equate to sentience or personhood (consider, for example, a trivial chatbot that is hardwired to reply to all queries with “Please help, I have real sentience”), it does encourage us to think more deeply about what constitutes personhood in the first place, especially in legal systems that include concepts like corporate personhood. The importance of this question is highlighted by a recent argument by Air Canada that the company should not be held responsible for misrepresentations by their LLM-based chatbot, as the bot was a “separate legal entity that is responsible for its own actions”—an argument rejected by the court (Yagoda 2024). The Air Canada example shows that there are various reasons why one may wish to grant personhood, legal or conceptual, to an AI system—in Air Canada’s case, for instance, to absolve the company of responsibility for its output. It also shows that questions of personhood are entwined with questions of authorship; if authorship equates to self-expression and responsibility for textual content (interpretations affirmed by our findings below), understanding perceptions of authorship can support understanding of the legal personhood debate.
To add further complication, academics do not agree even within the technical literature on what, exactly, the term ‘AI’ means, with definitions ranging from so simplistic as to enable a modern-day Diogenes to point at a bimetallic thermostat and shout “look, I have brought you AI” to so strict that nothing currently in existence is AI. When others are concurrently pointing at computer systems that produce text that certainly reads like it has been written by a human (e.g., ChatGPT), talking about those systems using highly anthropomorphic terms (e.g., Salles et al. 2020), and going so far as invoking the terminology of magic and the divine (Giuliano 2020), it is no wonder that the term ‘AI’ has effectively become a floating signifier in colloquial discourse. It “suggests a specific referent but works to escape definition to maximize its suggestive power [and] works through a strategic vagueness that serves the interests of its promoters, as those who are uncertain about its referents are left to assume that others know what it is” (Suchman 2023: 3). Put simply, AI is simultaneously everything and nothing, a term so malleable that it can be—and is—bent to suit circumstances as desired.
Our goal in this paper is not to wholly calm the maelstrom of chaotic change and definitional difficulty described above. Rather, ours is a humbler goal: to understand how—or whether—the perceptions of authorship have changed following the coming of the LLMs. After all, as Simone Natale explains in his book about Artificial Intelligence and Social Life after the Turing Test (emphasis in original), “[t]he question, Turing tells his readers, is not whether machines are or are not able to think. It is, instead, whether we believe that machines are able to think—in other words, if we are prepared to accept machines’ behavior as intelligent” (Natale 2021: 20). While actual technological functionality is of course important to consider in current conversations about the societal implications of LLMs and AI, perceptions of technological functionality and value (or lack thereof) are more important in this paper. Perception is powerful. It informs whether or not people are willing to adopt technologies and, if so, how they are willing to adopt them. It informs hopes, fears, excitement, and concern. It informs our senses of identity and meaning (Møller et al. 2024). Perceptions have very real societal and commercial implications for the development and propagation of AI. If we want to develop AI that adequately serves us, we must account for our perceptions.
In this paper, we focus on perceptions of AI authorship in particular. The relevance of authorship perceptions is manifold. For one, authorship is often equated with interpersonal connection. Writing is regarded as an actualisation of human thought, an effort to share experience and insight. Texts are sometimes framed as contributions to asynchronous conversations between authors and readers (Henrickson 2024). Authorship perceptions matter because they influence who we think we are ‘talking’ to. They also pertain to several key moral and legal questions in the contexts of media and communication, most notably those related to legal personhood and responsibility for content. Our conceptions of copyright are based on an assumption that a piece of content has value that should rightly belong to the author, and that this value must be protected. At the same time, it is often held that authors—and in the case of news, together with editors—are responsible for their works and should be punished for harms caused by their creations. Indeed, recent discussions about AI authorship emphasise the importance of responsibility for truthful textual content (Pennock 2024). However, if general conceptions of authorship change so that the ‘author’ might no longer be a human, what does this mean for the rights, responsibilities, and expectations associated with this role? Is there a point in protecting the ‘rights’ of a computer program, and should we allow for a future where the ultimate responsibility for publishing falsehoods is not with a human but with a piece of code?
While questions about AI’s place in our current social infrastructures abound, generative AI—often powered by LLMs—is nevertheless being integrated into these infrastructures at a rapid rate. Generative AI systems have been applied in university classroom assignments, for instance, with reported success for encouraging students to navigate the literary, communicational, and ethical complexities of increasingly pervasive ‘AI-powered plagiarism’ (Fyfe 2023). They are also increasingly being used in marketing campaigns, although consumers have expressed ‘moral disgust’ in some such use cases (Kirk and Givi 2025). These are just a few areas wherein LLMs have seen particular prominence, but computer-generated texts abound. Yet even with these texts’ prevalence, we remain unsure of readers’ willingness to accept them. Studies indicate that the degree of AI involvement in a text’s production may influence readers’ sentiments towards the AI and human contributors to that text, and to the text more generally (Formosa et al. 2024). In addition, the increasingly widespread use of generative AI systems for text generation raises questions about the processes of generation themselves. How do publics believe these systems to be technically operating? Does writing a prompt for a system count as authorship in itself (Mazzi 2024)? Are we moving towards a new kind of ‘distributed authorship’, wherein authorship is characterised by networks of human and non-human actors (Bajohr 2024)?

3 Methods

To examine public perceptions of algorithmic authorship (Henrickson 2021: 3), we distributed an international online survey from 5 December 2024 to 20 January 2025 (ethical approval granted by the University of Queensland, 2024/HE001972). It was completed by 200 participants. This survey was the same as one distributed by and reported on by one of this paper’s authors from 5 December 2017 to 24 January 2018 for a doctoral project (Henrickson 2019: 145–176). The redistribution of the same survey has supported a longitudinal perspective on perceptions of algorithmic authorship. This is, to our knowledge, the only longitudinal empirical consideration of algorithmic authorship pre- and post-LLM boom.
The survey presented participants with a 243-word English-language text reviewing Finnish municipal elections. After an introductory demographics survey, participants were asked to “list three things that come to mind when you think of the word ‘author’.” Then, participants were asked to attribute authorship to the provided text four times. Each time participants were asked to attribute authorship, they were presented with additional information about the text’s process of production. First, they were only given the text itself, which included the mononym ‘Valtteri’ in the byline under the title. Second, participants were told that Valtteri was a bot, and that ‘John Smith’ had translated the Finnish political party names into English for comprehensibility. Third, participants were told that Valtteri was developed by a team called Immersive Automation. Finally, they were informed that Immersive Automation was funded by various public and private organisations. In all four instances, participants were provided lists of predetermined answers, with ‘It is not possible to assign authorship’ and ‘Other’ included; if a participant selected ‘Other’, they were required to explain why in a text box. Each attribution instance included a ‘Why have you selected this option?’ text box for participants wishing to volunteer elaboration on their answers. Iterative questioning helped us see what kinds of information participants found valuable for their considerations of authorship, and supported participants’ critical reflections on the multilayered nature of computational text production. Following the authorship attribution activity, participants were asked whether they thought that their answers would have been different had they been presented with a computer-generated piece of opinion-driven journalism, short story, or poem, and were asked to explain why in a text box if they selected 'yes' to any of these questions. A copy of the survey is included in the Appendix 1.
Like the initial version of the survey, this survey recruited participants through word-of-mouth and snowball sampling techniques. Social media platforms including Reddit (r/SampleSize), Facebook (personal and acquaintances’ feeds; the ‘Women in Academia Support Network Careers Support #wiasn’ group), X/Twitter (personal and acquaintances’ feeds), and LinkedIn (personal and acquaintances' feeds) were used to promote the survey. In addition, and again like the initial version, participants were solicited through the SHARP (Society for the History of Authorship, Reading and Publishing), DHSI (Digital Humanities Summer Institute), and SIGGEN (ACL Special Interest Group on Natural Language Generation) professional listservs, and student mailing lists of various universities. These techniques undoubtedly influenced the demographic distribution of participants. However, the demographic distributions of both versions of the survey are sufficiently similar to support comparability of results.
There are, however, slight differences between the two surveys that require mention. To accommodate a wider range of participants, ‘elementary school’ was added as an option for the highest level of education completed; ‘prefer not to say’ was added as an option for gender and age; the Web link to Valtteri was removed as it was no longer functional; and the development and funding of Valtteri were referred to in the past, rather than present, tense due to the completion of that project. The analysis of the second survey’s results was similar to that of the first’s, but benefited from the involvement of a second researcher with more extensive experience in quantitative and computational analytical techniques. The coding of the second survey’s authorship word associations was completed by the same researcher as those of the first’s, but variations in coding practice emerged due to changing rhetorical landscapes and the researchers’ deeper reflections on word functions (e.g., ‘text’ could be considered both verb and output; ‘authority’ could be considered a person or a connotation). The execution of the first survey also did not produce a detailed codebook, as this survey was not initially imagined as longitudinal. Finally, the presentation of the most recent survey results more thoroughly reviews the quantitative changes between authorship attributions than did the presentation of the 2018 results, more firmly marrying quantitative and qualitative methods for a better picture of participants’ opinions based on the information provided to them in each step of the survey.

4 Results

This section reviews the results of the participant demographics section, author word association activity, authorship attribution activity, questions about genre specificity in authorship attribution, and final questions/comments section. The results of the authorship attribution activity also include attention to the evolution of participants’ attributions across the four instances of attribution, as well as links between participant demographics and attribution tendencies. Our data consist of survey responses from 200 participants for the 2024–2025 survey, and 500 responses from the previous 2017–2018 survey. We do not know if there is overlap across these populations. In the results reported below, quotations from the survey’s qualitative responses were selected because they exemplify the general sentiments expressed by participants. These quotations have been copied verbatim, including any spelling and grammar errors and eccentricities. Each quotation is accompanied by parenthetical acknowledgment of the participant’s attribution trajectory throughout the authorship attribution activity.

4.1 Participant demographics

Data for participants of the survey is presented in Tables 1, 2, 3, 4, 5, 6. For comparison, we also provide the demographics of the 2018 survey. Applying a sequence of chi-squared tests to the demographic variables, we identified statistically significant differences between the two survey populations in terms of gender (X^2 (4, N = 700) = 42.360, p < 0.001), age (X^2 (6, N = 700) = 22.798, p < 0.001), level of education (X^2 (4, N = 700) = 39.790, p < 0.001), and native language (X^2(1, N = 700) = 13.154, p < 0.001). No such statistically significant difference was identified in terms of occupational status (X^2 (3, N = 700) = 7.073, p = 0.070).
Table 1
Participants’ age groups
Survey
18–29
30–39
40–49
50–59
60–69
70 + 
Prefer not to say
2018
139
27.8%
140
28%
76
15.2%
68
13.6%
55
11%
22
4.4%
2025
32
16%
67
33.5%
44
22%
30
15%
19
9.5%
5
2.5%
3
1.5%
Table 2
Participants’ genders
Survey
Female
Male
Non-binary
Other
Prefer not to say
2018
337
67.4%
155
32.5%
5
1%
3
0.6%
2025
110
55%
65
32.5%
13
6.5%
5
2.5%
7
3.5%
Table 3
Participants’ educational backgrounds
Survey
Elementary School
Secondary school
Trade, technical, or vocational
Undergraduate
Master’s
Doctorate
2018
33
6.4%
10
2%
126
25.2%
180
36.6%
152
30.4%
2025
0
0%
6
3%
1
0.5%
25
12.5%
59
29.5%
109
54.5%
Table 4
Participants’ occupational statuses
Survey
Employed
Not employed or unpaid
Retired
Student
2018
320
64%
13
2.6%
47
9.4%
120
24%
2025
146
73%
7
3.5%
11
5.5%
36
18%
Table 5
Participants’ native languages
Survey
Not English
English
2018
76
15.2%
424
84.8%
2025
54
27%
146
73%
Table 6
Participants’ comfortability with technology
Survey
Not at all comfortable
Fairly comfortable
Comfortable
Very comfortable
2018
2
0.4%
41
8.2%
99
19.8%
358
71.6%
2025
0
0%
10
5%
37
18.5%
153
76.5%
We also asked the participants for their comfort with technology on a four-point scale ranging from “not at all comfortable” to “very comfortable”. The results for both surveys are shown in Table 6. We found no statistically significant difference between the populations based on a chi-squared test (X^2 (3, N = 700) = 3.402, p = 0.334), and a Mann–Whitney U test conducted after coding the options as numbers agrees at p = 0.141 (U = 47,239.5).
31 countries of residence were represented (compared to 27 in the 2018 survey), with the United States providing a substantial number of responses (70; 35%), and the United Kingdom (36; 18%), Canada (28; 14%), and Australia (18; 9%) also providing high numbers. A distribution of country representation is shown in Table 7. Half of the participants (100; 50%) self-identified as working within education, training, and engineering fields.2 A detailed distribution of occupational fields represented is shown in Table 8.
Table 7
Participants’ countries of residence
Country
2018
2025
United States of America
213
70
United Kingdom
122
36
Canada
90
28
Australia
9
18
Germany
9
7
Finland
17
5
Belgium
1
3
Ireland
2
3
Spain
2
3
Brazil
0
2
Greece
0
2
Norway
1
2
Sweden
4
2
Switzerland
2
2
Armenia
0
1
Austria
0
1
Czech Republic
0
1
Denmark
1
1
Egypt
2
1
France
7
1
India
3
1
Iraq
0
1
Italy
3
1
Mexico
1
1
Netherlands
2
1
Nigeria
0
1
Pakistan
0
1
Portugal
0
1
Serbia
0
1
Singapore
1
1
South Africa
1
1
Cyprus
1
0
Hungary
1
0
Israel
2
0
Malta
1
0
Poland
1
0
Samoa
1
0
Table 8
Participants’ occupational fields
Occupational field
2018
2025
Education, Training, and Engineering
277
('Education, Training, and Library’: 252; 'Engineering': 25)
100
Information Technology
38
24
Arts, Design, and Entertainment
23
17
Media and Communications
25
8
Other
0
7
Life, Physical, and Social Science
20
7
Public Sector
10
5
Healthcare
7
5
Legal
10
5
Not Employed
18
4
Community/Social Services
2
4
Business and Finance Operations
27
3
Transportation and Material Moving
0
2
Management
7
2
Farming, Fishing, and Forestry
1
1
Hospitality
2
1
Installation, Maintenance, and Repair
0
1
Journalism
6
1
Military
0
1
Office and Administrative Support
13
1
Retail and Sales
6
1
Architecture
1
0
Building/Grounds Maintenance
1
0
Construction and Extraction
2
0
Personal Care and Services
2
0
Production
1
0
Sport
1
0
As in the initial implementation of the survey, these responses are dominated by a highly educated populace of primarily women and those employed in education, training, and engineering roles, and readers should be mindful that results are skewed by a culturally educated elite. This skew is likely due to the recruitment techniques, which were the same for both implementations. While there are some statistically significant differences in the respondent populations, we view the populations as sufficiently similar for meaningful longitudinal analysis.

4.2 Author word association results

Both surveys asked the respondents for three associations that came to mind in response to the word ‘author’. To maintain compatibility with the 2018 survey’s results, the same author coded the data for the 2025 survey as well. Through thematic analysis, the 2018 survey revealed a fourfold definition of ‘author’ from the word association exercise: “authorship encompasses (1) an identity that is (2) associated with particular connotations (primarily adjectives, but also including some nouns), (3) as well as with particular activities (4) that result in particular kinds of (generally text-based) output” (Henrickson 2019: 151). This fourfold definition of ‘author’ was used to code the results of the 2025 survey; alternative codes did not appear to be required following preliminary analysis. Each response was manually coded in accordance with one or more of the four facets of ‘author’. ‘Research’, for example, was coded under both (3) and (4) given its function as both verb and noun, respectively. Similarly, ‘pen’ was coded under both (2) and (3) given its function as both noun and verb, respectively. The results are shown in Table 9.
Table 9
‘Author’ word associations
Survey
Identity
Connotations
Activities
Output
2018
538
33.7%
391
24.5%
176
11%
490
30.7%
2025
239
32.9%
222
30.6%
108
14.8%
157
21.6%
The most common item listed was, as in the initial survey, the word ‘writer’, which was listed 72 times (comprising 12% of all 600 responses). In addition, as in the initial survey, the word ‘creator’ appeared as another common item related to identity, listed 40 times (comprising 6.7% of all 600 responses). More specific identities such as ‘jane austen’, ‘shakespeare’, ‘steinbeck’, and ‘edith wharton’ were included; more personal items including ‘i am an author’ and ‘me’ also appeared. In addition, as in 2018, connotations included adjectives, but also nouns representing places (e.g., ‘coffee shops’), tangible objects (e.g., ‘wool jacket’, ‘pen’, when interpreted as a noun), and abstract concepts (e.g., ‘originality’). Activities (e.g., ‘publishing’ and ‘framing’) and outputs (e.g., ‘book’, ‘article’, and ‘literature’) were likewise similar to those in the 2018 results.
A chi-squared test indicates that the responses to the surveys are statistically significantly different (X2(3, N = 2321) = 27.869) at p < 0.00001. A post-hoc analysis of the chi-squared residuals conducted using the chisq.posthoc.test package in R (applying the Bonferroni correction for multiple comparisons) indicates statistically significant changes in the ‘Connotations’ (p = 0.017) and ‘Output’ (p < 0.001) columns. In other words, the data indicates that the 2025 respondents, compared to the 2018 respondents, increasingly associated authorship with words relating to connotations, and decreasingly with words relating to outputs. While the change in the ‘Activities’ column is not statistically significant at p = 0.07 (Bonferroni-adjusted), we nevertheless find the change suspicious enough to warrant highlighting.

4.3 Authorship attribution results

Next, the respondents were asked to read a short English-language text and attribute authorship a total of four times, with additional information revealed to them prior to every re-attribution. The respondents were first asked to attribute authorship based only on a byline of ‘Valtteri’, and were given the options ‘It is not possible to attribute authorship’, ‘Valtteri’, and ‘Other’. The distributions for both surveys are shown in Fig. 1. A large proportion of respondents chose ‘It is not possible to attribute authorship’ or ‘Other’, perhaps basing their responses on metaknowledge obtained from the survey invitation or the introductory part of the survey, which, under the title of Perceptions of Authorship of Computer-Generated Texts, told participants that “you will be asked to read a text and then, through an online survey, assign authorship to this text based on the information you have been given.”
Fig. 1
Participants’ attributions of authorship-based only on the byline ‘Valtteri’
Bild vergrößern
A chi-squared test (X^2 (2, N = 700) = 11.592, p = 0.003) indicates that the distributions of the 2018 and 2025 surveys’ first authorship attributions are statistically significantly different. More specifically, the proportion of ‘It is not possible to attribute authorship’ responses decreased from 44% to 33%, while the proportions of ‘Other’ and ‘Valtteri’ rose from 6% to 12% and 50% to 55%, respectively.
In the qualitative responses to this first attribution, numerous participants speculated that the writing sample was “AI generated” (Other), “[c]learly a bot” (Other), “auto-generated text” (Other), and “[a]n NLG system, probably rule-based” (Other). Explaining their attributions, participants observed Valtteri’s name occupying the space of a typical journalistic byline, but expressed uncertainty about the mononym. Some participants wondered if ‘Valtteri’ were a place rather than a name. Comments about the sample being “terribly written” (Other), “without any personal style” (Other), and “so dry” (Other) were common, especially amongst those participants who selected ‘Not possible’ and ‘Other’. As per one participant (Not possible), “the writing style is awful, so I don’t think anyone’s stepping forward to claim it”.
For the second round of attribution, we revealed that ‘Valtteri’ is an online bot, and that someone called ‘John Smith’ had translated the party names to English for better readability. The participants were then asked to attribute authorship again, with an option for ‘John Smith’ added. The response distributions from both surveys are shown in Fig. 2. Both surveys largely reject the authorship of John Smith, mostly either attributing authorship to Valtteri or opining that it is not possible to attribute authorship.
Fig. 2
Participants’ attributions of authorship following the reveal of the nature of ‘Valtteri’ as an online bot and that ‘John Smith’ translated party names to English for readability
Bild vergrößern
A chi-squared test indicates a statistically insignificant difference between the distributions (X^2(3, N = 700) = 3.459, p = 0.326). The largest changes are that the proportion of participants selecting ‘It is not possible to attribute authorship’ rose from 30.4% in 2018 to 36.5% in 2025, while the proportion of participants selecting ‘Valtteri’ fell from 45% to 38%. The proportions of ‘John Smith’ (3.8% in 2018, 4.5% in 2025) and ‘Other’ (20.8% in 2018, 21 in 2025) stayed broadly the same.
In the qualitative responses to this second attribution, participants began suggesting various combinations for hybrid authorship, with the most common combination being Valtteri and John Smith. Participants also suggested varying tiers of authorship. As one participant (Valtteri → Not possible) explained, “John Smith is definitely a minor author, as translation modifies the meaning of the text and their choices alter the readers interpretation. However, their role is too minor to assign as the proper author.” Moreover, participants acknowledged non-'author' titles: ‘translator’, for instance, or 'editor'. By this point in the survey, numerous participants questioned precisely how the sample was generated, stating that the ‘how’ might sway their responses; questions about agency, distributed agency, copyright law, and responsibility for the text were also raised. At the same time, others made firm declarations like “[b]ots do not have brains therefore they cannot author fuck-all” (Valtteri → Other). In addition, for one participant (Valtteri → Other), “[it] is not necessary to assign authorship” at all.
For the third round, the respondents were told that Valtteri was developed by a research team called ‘Immersive Automation’. They were then asked to attribute authorship again, with the team added as an option. The response distributions are shown in Fig. 3. The most popular options are, again, ‘Valtteri’ and ‘It is not possible to attribute authorship’, with the two options swapping first and second place between the surveys (Valtteri first place in 2018, and second place in 2025).
Fig. 3
Participants’ attributions of authorship following the reveal that a research team called ‘Immersive Automation’ had developed Valtteri
Bild vergrößern
We observe a statistically significant difference between the 2018 and 2025 distributions (X^2(4, N = 700) = 19.217, p < 0.001). Notably, the proportion of attributions to the Immersive Automation research team fell by 10.8 percentage points from 17.8% to 7, while the proportion of ‘It is not possible to attribute authorship’ rose by a near-equivalent 10.6 percentage points from 27.4% to 38. In other words, while the 2018 results have the two options nearly tied in popularity, the 2025 results show a significant preference for ‘Other’ over the research team. The changes for the options ‘Valtteri’ (37.2% in 2018, 35 in 2025), ‘John Smith’ (1.6% in 2018, 3.5% in 2025), and ‘Other’ (16 in 2018, 16.5% in 2025) were minute.
In the qualitative responses to this third attribution, participants continued to propose various combinations for hybrid authorship, with the most common combination remaining Valtteri and John Smith. Many participants explained that Immersive Automation was not an author, because “although Immersive Automation has designed the bot, the act of creation occurs somewhere else down the line.” As this participant explains (Not possible → Valtteri → Valtteri), “[y]ou don’t call Bic the author of something someone has written with a pen”. Computational tools such as Microsoft Word, Google Translate, and Grammarly were also used for comparison. Moreover, participants compared Immersive Automation to a parent. As per one participant (Not Possible → Other → Other), “my parents developed me but that doesn’t mean they are authors of what I write!” Still, questions about responsibility for the text were raised, especially about how much direct or indirect influence Immersive Automation had over the text’s production.
In the final round, we revealed that the Immersive Automation team had received funding from several sources, and added an option to attribute authorship to these funders. The distribution of attributed authorship is shown in Fig. 4. In both surveys, the respondents largely rejected the authorship of the funding bodies.
Fig. 4
Participants’ attributions of authorship following the reveal that the Immersive Automation team received funding from various parties
Bild vergrößern
Applying the chi-squared test once more, we observe that the distributions are again statistically significantly different (X^2(5, N = 700) = 14.738, p < 0.001). However, since the added option received so little support, the data and their patterns are largely equivalent to those observed in the preceding round.
In the qualitative responses to this final attribution, participants seemed largely unswayed by the additional information provided. They compared Immersive Automation’s funding to governmental funding bodies, schools, and other organisations that may facilitate a text’s production, but are not typically attributed as authors of that text. One participant (Other → Valtteri → Valtteri → Valtteri) explained that funders “just throw money around. They definitely are not authors—unless they crossed ethical lines and told researchers how to do their research or directed that particular lines of code be included in the algorithm. Then, I might see them as just one-step removed, like the researchers. However, Valtteri is still the author.” Others similarly suggested that levels of funder intervention (e.g., patronage and legal accountability) may inform their attributions, but most appeared resolute that “[f]unders are definitely not authors!” (Valtteri → Valtteri → Valtteri → Valtteri). By this point in the survey, though, some participants expressed confusion, uncertainty, and even apathy. As one participant (Not possible → Not possible → Not possible → Not possible) questioned, “why do we need to assign author to the text that will be forgotten on two days?”

4.3.1 Attribution evolution

Figures 5 and 6 show the evolution of authorship attributions over the four different attribution questions. A key point that stands out is that a large number of participants changed their responses for the second round in both surveys. For example, a non-trivial proportion of those who felt that ‘It is not possible to attribute authorship’ in the first round changed their opinion to ‘Valtteri’ in the second round, even if this group is proportionally smaller in 2025 than it was in 2018. At the same time, following the reveal of Valtteri’s being a bot (i.e., from round 2 onwards), the responses appear relatively stable and much of the movement is to options not previously available, such as responding ‘Immersive Automation’ in round 3. This might indicate that the changing responses are less about changing opinions and more about option availability. This indicates to us that the exact nature of Valtteri as a computer program is the most critical piece of information to the respondents.
Fig. 5
Authorship attribution evolution, 2018 survey
Bild vergrößern
Fig. 6
Authorship attribution evolution, 2025 survey
Bild vergrößern
Changes between the two surveys appear limited, but we do note that whereas in 2018 13% of those who responded ‘Valtteri’ in round 2 indicated authorship of the research team (‘IA’/ ‘Immersive Automation’) in round 3, in 2025 this percentage was only 6.6%. This decrease in popularity of the research team option for round 3 does not appear to stem from a decrease among any specific round 2 option, but all transitions to the round 3 Immersive Automation (IA) option see similar falls in popularity from 2018 to 2025. In other words, the option falls in popularity universally, rather than among any specific round 2 respondent group. Furthermore, whereas in 2018 the round 2 ‘Not possible’ answers was approximately 1:2 from round 1 ‘Valtteri’ respondents and round 1 ‘Not possible’ respondents, this ratio is effectively reversed in 2025. We are less certain how to interpret this change, but note that it appears curious.

4.3.2 Participant demographics and authorship attribution

The 2018 study identified that the higher one’s level of education was, the more likely they were to express that it was not possible to attribute authorship. For this study, we decided to investigate whether ‘highly educated’—defined as having completed either a master’s degree or a doctorate—respondents attributed authorship (when asked for the fourth and final time) meaningfully differently from those who were less educated. These data are shown in Table 10. Chi-squared tests indicate statistically significant differences between these two groups in both the 2018 data (X^2(5, N = 500) = 20.999 p < 0.001) and the 2025 data (X^2(5, N = 200) = 17.074, p = 0.004).
Table 10
Relationships between educational background and the fourth (and final) authorship attribution. Percentages are row-percentages
  
Attributed authorship (fourth round)
Survey
Highly educated?
Valtteri
J. Smith
Immersive Automation
Funders
Other
Not Possible
2018
No
69
41.1%
3
1.8%
31
18.5%
8
4.8%
15
8.9%
42
25%
 
Yes
110
33.1%
4
1.2%
59
17.8%
1
0.3%
57
17.2%
101
30.4%
2025
No
11
34.4%
4
12.5%
1
3.1%
0
0%
5
15.6%
11
34.4%
 
Yes
58
34.5%
1
0.6%
15
8.9%
3
1.8%
26
15.5%
65
38.7%
Observing Table 10 in more detail, a few key points attract attention. First, in 2018 the highly educated were a fifth (5 percentage points) more likely to indicate impossibility of assigning authorship, and over double (7 percentage points) as likely to indicate ‘Other’ authorship (perhaps an unnamed entity, or a hybrid of named and/or unnamed entities). They were also less likely to attribute authorship to the funders or to the software itself. Support for the research team and John Smith was largely equivalent between the groups.
The 2025 survey responses paint a different picture. Ignoring the broader trends already analysed above (e.g., the overall falling popularity of the ‘Immersive Automation’ option), on a high level we see a pattern where the two groups’ responses approach each other. This is most apparent for ‘Valtteri’ and ‘Other’: where the two groups previously had meaningful differences, the percentages are now practically equal. The difference in the popularity of the ‘Not possible’ option remained largely the same in absolute terms (circa four percentage points, previously five), but increased in relative terms. The two options that buck this general trend are ‘Immersive Automation’ and ‘John Smith’. Both of these, however, are notable in that the absolute number of responses was very low for at least one of the groups. In the case of the ‘Immersive Automation’ option, just one more non-highly educated response in 2025 would have increased the relative popularity of that option to a near-equivalent 6.2%. Similarly, in the case of ‘John Smith’, while the relative frequencies are widely different between the non-highly educated populations (12.5% vs 0.6%), the absolute counts are only 4 and 1 responses, respectively. Overall, it is difficult to confidently state anything about these two options in particular.

4.4 Genre specificity of authorship attribution

Following the fourfold attribution activity, participants were asked whether their responses would have changed were the text that they were presented with a journalistic opinion piece, short story, or poem. The responses to this question are shown in Table 11. As in 2018, the large majority of the respondents did not view authorship attribution as genre-specific, indicating that they would have responded similarly even if the reviewed text were an opinion piece, short story, or poem. While the share of those indicating genre-specificity decreased for all alternative genres, only the change for the opinion piece alternative was statistically significant based on a chi-squared test (X^2(1, N = 700) = 6.102, p = 0.014).
Table 11
Genre-specificity of authorship attribution. Numbers indicate how many of the respondents would have changed their response if the text were from the genres corresponding to the columns
Survey
Opinion
Short story
Poem
2018
140
28%
108
21.6%
103
20.6%
2025
38
19%
37
18.5%
36
18%
In their qualitative responses to this question, participants acknowledged the complexity of the hypothetical scenarios presented, with some noting that “these are complex questions that are hard to answer in a binary yes/no fashion” (Not Possible → Valtteri → Valtteri → Valtteri). Such complexity appeared to arise from perceptions of creativity exhibited by text generation systems, the contexts of text generation and reception, the extent of novelty in generated content, the quality of the generated content, and the intricacies of system functionality.
Some participants prioritised authorship of opinion-based journalism more than authorship of short stories and poems, citing a need for accountability for the opinions presented. Other participants prioritised authorship of short stories and poems over opinion-based journalism, citing “self-awareness and intentionality (a point to argue, skillful and artful turns of phrase, innovative and unexpected imagery)” (Not Possible → Valtteri → Immersive → Immersive) as necessary for authors. One participant (Valtteri → Other → Other → Other) summarised these two stances thus: “Although I feel authorship of factual writing can be attributed to distributed authors, I feel it is harder to attribute opinion or art or style in the same way. Rather, these things are an illusion generated by the bot. For example, I don’t think you can say, from an opinion piece written by a bot, that the bot is right wing or left wing in any meaning sense. It has assembled a plausible version of such writing but cannot be help accountable for that opinion.” Indeed, difficulty of attribution was a common issue raised with participants, some of whom also questioned their own abilities to distinguish between human-written and computer-generated texts.

4.5 Final questions and comments

To conclude the survey, participants were given the opportunity to share any final questions or comments with the research team. Some participants used this opportunity to reflect on authorship itself. In one participant’s (Other → Not possible → Not possible → Not possible) words, “[a]uthorship here might be less of a copyright concept but rather whether we as humans cherish the effort another human being has made to create a piece of art. Attributing human-written texts to humans and informing readers when a text was generated by a NLG system or language model is element of human self-respect.” Numerous participants suggested that alternative language for different kinds of contribution to text production may be required. Other participants questioned the relevance of authorship at all; as per one participant (Valtteri → Valtteri → Valtteri → Valtteri), “I’m left thinking, ‘So what?’ For public content, it is perhaps more important to people who’s interests are being represented in the published work rather than the author, and this is not a new concept.” Attention to the histories of authorship debates were likewise raised by others. “Since writing was invented, every text has been authored within a social context, shaping reality while being shaped by it,” wrote one participant (Valtteri → Other → John → John). “New tools, old problems.” Participants also commented on the general importance of truthfulness and stylistic quality in generated texts, with some noting that these qualities could sway them towards attribution to a computational system. At the same time, as per one participant (Valtteri → Valtteri → Valtteri → Valtteri), “[a]lthough I consider AI to be the author, that doesn’t mean it’s ethical and I still believe it’s stealing.”

5 Discussion

So far, we've considered authorship perceptions from a variety of angles, including what kinds of words people associate with authorship, how people attribute authorship of a computer-generated news text based on limited information, how authorship attribution relates to various demographic variables, and whether authorship attribution is genre-specific. For all these aspects, we’ve considered not only the findings from the current survey, but also how those results differ—or not—from what was observed in the 2018 survey. In this section, we recap and discuss our key quantitative findings on these fronts separately, and then more broadly discuss what these findings, and their corresponding free-text responses, indicate to us holistically. Finally, we observe some limitations of our work and suggest future avenues of research.
A fourfold definition of authorship emerged from the analysis of the first set of survey responses in 2018, and this definition was affirmed by the results of the survey in 2025. A statistical analysis indicated meaningful changes from the 2018 survey, with the words relating to connotations becoming more prevalent and those relating to outputs becoming rarer. We also observed a tentative (i.e., not statistically significant) increase in the proportion of terms relating to activities. This increase may be partly explained through participants’ caveats to authorship in their qualitative answers throughout the survey. These caveats were often focused on the how of text production, but also included assumptions related to human expressions of creativity and textual characteristics and quality. The importance of the ‘how’ is further mirrored in the decrease in terms relating to outputs in the word association task. Even with these differences, though, the two surveys share a key finding: that ‘authorship’ is used flexibly. The concept of authorship, like the concept of AI itself, “works through a strategic vagueness that serves the interests of its promoters, as those who are uncertain about its referents (popular media commentators, policy makers and publics) are left to assume that others know what it is” (Suchman 2023: 3). In addition, like AI, ‘authorship’ is simultaneously everything and nothing, a term so malleable that it can be—and is—bent to suit circumstances as desired. Thus, both elements comprising ‘AI authorship’ are somewhat fuzzy; their synthesis in ‘AI authorship’ does not appear to make either any less so.
This fuzziness seemed prevalent in the participants' authorship attributions. In terms of attributing authorship to a computer-generated news text, following the reveal that the text was indeed computer-generated the quantitative results of the present survey are mostly characterised by an increase in uncertainty and either inability or refusal to attribute solitary authorship to either the software or its creator. However, where the respondents did attribute authorship to one of the named entities, the system itself dominated the results—and much more so than in the 2018 survey. One explanation for this finding could be increasing awareness of how machine learning systems in general, and LLMs more specifically, are trained using the products of (and in the case of LLMs, texts written by) third parties. This is exemplified by one respondent's (Valtteri → Valtteri → Valtteri → Valtteri) final comment that “[a]lthough I consider AI to be the author, that doesn’t mean it’s ethical and I still believe it’s stealing.” In the case of the present study, the text-generation system was in reality not a machine learning one (and thus had no training data), but that matters little for the respondents' perceptions given the limited information available to them.
A hypothesis of increased awareness could also be supported by our observations on how the educational background (‘highly educated’ simplified to whether the respondents had at least a master’s degree or not) related to the authorship attributions. While a high-level statistical analysis indicated that statistically significant differences still existed in the 2025 survey data, a closer look at the data revealed a general trend of the groups’ responses getting much closer to each other. This could arise from a situation where information, knowledge, and critical lenses previously available only to the highly educated group were now also available to and known by the other group.
As for how responses evolved across the four rounds of attribution, we note that the opinions remain largely static following the reveal (or confirmation) that the text was computer-generated, and changes between the two surveys are limited. We did, however, observe that those who attributed authorship to the mononym ‘Valtteri’ prior to the reveal that it was a bot now more evenly split into those who continued to attribute authorship to the software and those who found it not possible to attribute authorship in comparison with the 2018 results, as well as that the fall in popularity of the ‘Immersive Automation’ option was not attributable to any specific response from the previous round(s).
When asked whether their responses to the questions would change if the genre of the text were different, our results from the present survey mirror those from the 2018 survey in that the vast majority of respondents did not view authorship as genre-specific. While the proportion of those who would have changed their responses fell across all genres, we only observed a statistically significant decrease for opinion pieces, with 28% indicating they would have changed their responses in 2018 and 19% indicating the same in the present survey. This change aligns with the findings pertaining to words relating to authorship, wherein we saw fewer words relating to outputs; if outputs are less associated with authorship, it seems reasonable that the genre of the output matters less as well.
Overall, participants recognised the extensive networks within which authorship exists, especially in their recommendations for distributed authorships of the computer-generated text under consideration. Hannes Bajohr (2024: 327–328) describes ‘distributed authorship’—heavily influenced by actor-network theory—as “distributed across an assemblage of actors in complex chains of operations. [..] This actor-network must be imagined as sprawling and immense.” Such distribution includes not just human contributors, but also those more algorithmic, as well as those whose data have informed AI training sets. Participants appeared instinctively attuned to the distributed, networked nature of text production more generally, suggesting various combinations of authorship for the text under consideration, if they felt attributing authorship was possible at all. It appears, then, that although authorship has always been a social construct dependent upon interpersonal collaboration (Inge 2001), such collaboration is increasingly recognised as imperative. Humans using LLMs may no longer be able claim the ‘solitary genius’ long associated with conceptions of single authorship. Instead, they are more commonly perceived as existing within networks of other humans, other humans’ data, and algorithmic systems, all of which inform text production processes. These networks were identified by participants, who in the present survey showed increased attention to the connotations associated with the word ‘author’, as well as slightly increased attention to ‘author’ activities. These networks are also the same networks currently being mobilised in efforts to develop AGI that “empower[s] humanity to maximally flourish in the universe” (Altman 2023). However, to flourish, humanity must feel confident in its own role in the production of cultural artefacts like texts. After all, “[w]hen humans use language, we do so for a purpose: We do not talk for the joy of moving our articulators, but to achieve some communicative intent” (Bender and Koller 2020: 5187; emphasis in original). By distributing authorship we are, in effect, distributing intent. Such distribution is neither inherently good nor inherently bad, but does inherently change the conventional understanding of ‘authorship’ being claimed by any single person. Not being able to attribute authorship to a computer-generated text does not necessarily signify confusion; it could signify recognition of the distributed nature of text production.
Distributed authorship, however, did not appear to excuse responsibility for the accuracy and truthfulness of computer-generated texts for survey participants. These are issues that scholars have been considering through reflection (Pennock 2024), proposals for evaluative metrics (Goodrich et al. 2019), and empirical studies about human trust in AI systems (Glikson and Williams Woolley 2020). Questions of trust also lead to questions of trusteeism. While participants did not appear to assign legal personhood to the AI system used for the survey, they did implicitly question the legal ramifications of AI authorship by, for example, evoking copyright and intellectual property theft. There are no criminal punishments for AI systems that breach legal rules, and any AI system’s capacity for moral judgment still seems murky at best (Solum 1992: 1244–1253). These points are not the case for humans, who may be punished for neglecting or overstepping copyright, and who can reasonably be assumed to have the capacity for moral judgment. While legal personhood akin to that enjoyed by humans may not be appropriate for AI systems, though, other legal protections of these systems have been proposed by scholars who, like our survey participants, are trying to navigate new environments of human–machine communication and coexistence (Jaynes 2020). This navigation is hardly straightforward; as our attribution evolution visualisations (Figs. 5 and 6) show, many participants changed their authorship attributions when provided with new information about the presented text’s complex context of production. Although it is outside the scope of this study to consider precisely how responsibility for, and legitimacy of, computer-generated texts is—or should be—allocated, we believe that considering who or what participants believe the author to be can support such deliberations. Throughout the survey, participants indicated that their understandings of authorship were deeply linked with conceptions of both creative expression and social responsibility.
Our findings show that people may apply and adapt conceptions of authorship to new contexts without resolving the term’s ambiguities—which perhaps adds additional ambiguities. However, participants’ ambivalent views towards conceptions of authorship and their relevance to computer-generated texts bring us back to the more fundamental question raised by one participant: so what? Why does thinking about authorship of computer-generated texts matter, especially when—as indicated by the results above—authorship itself is a murky concept? Participants themselves answered these questions, albeit indirectly. Responsibility and accountability were cited as economic and social components of authorship; creative expression was valued for interpersonal understanding and assertion of self; simply producing a text was for many not enough to justify a claim to authorship, but contributions to text production could be recognised through different titles specific to the kind of contribution (e.g., translator). All of these points are intertwined with one another. Authorship matters because it is about more than just who produced a work. Authorship matters because it acknowledges the extensive social, political, and economic networks that producers exist within and perpetuate. Authorship matters because readers’ understandings of how a text has been produced may influence those readers’ interpretations of that text’s content, authority, and legitimacy (Formosa et al. 2024). As our study results show, though, these understandings may also be informed by a reader’s demographic profile. Perceptions of algorithmic authorship may be subjective and difficult to wholly capture, but this does not mean that they are not worth considering; these pereptions have very real consequences for the ways in which computer-generated texts are received and accepted—or not.

5.1 Limitations and future work

As any study, this one comes with limitations. For one, our population of respondents is far from being demographically representative of even any Western nation, never mind the global population. Our respondents are generally young adults (for certain values of ‘young’), female, very highly educated, and English-speaking. A psychologist might call our sample WEIRD (Western, Educated, Industrial, Rich and Democratic; Henrich et al. 2010). We also observed statistically significant differences between the population of the present survey and the 2018 survey, even if we personally do not view these as meaningful enough to significantly affect our conclusions. All told, we caution the reader from viewing our results as ‘global’ truth, or as a complete view of opinions from Western audiences.
These demographic limitations, however, directly point towards some clear avenues for future work. Further study is needed to evaluate whether our findings generalise to broader publics. Our results also indicated that approximately one-fifth of the respondents believed their opinions of authorship to be meaningfully genre-specific; further work would be needed to tease out details of this genre-specificity. In addition, we identified that at least some of the responses seemed to be built on assumptions about the technical details of the text-generation system used to produce the text being evaluated, chiefly on whether the system was one built on machine learning or not. Future work should investigate how information about the technical details—together with a more fine-grained understanding of the respondents’ technological understandings—affects their views of authorship.

6 Conclusion

The study documented herein makes one thing very clear: that perceptions of AI authorship are hardly clear at all. As in 2018, participants of our 2025 survey varied in their understandings of what constituted an ‘author’, and who or what might be deemed an ‘author’ of a computer-generated text. While some quantitative results were statistically significant, participants showed continued ambivalence towards attributing specific authorship to a computer-generated text. In their qualitative comments, participants revealed their concerns (or, in some cases, lack thereof) about attributing authorship to an AI system, and shared questions and comments indicating their awareness of the complexities of both computational text production and authorship more generally. Of course, qualitative comments cannot be taken as the entirely authoritative source of participants’ perceptions; just because participants do not include an idea in their commentary does not mean that they are not thinking about that idea. Nevertheless, in this paper we have compared two sets of results—one from 2018 and another from 2025—that show that although sociotechnical circumstances have changed substantially, participants’ perceptions of authorship of computer-generated texts have changed much less so. What this suggests is that ‘authorship’ may not—and perhaps has never been—the most appropriate word to summarise computational text production through natural language generation. ‘Authorship’ may be too evocative, too deeply enmeshed in long-standing cultural imaginaries. However, these cultural imaginaries have themselves been convoluted, with ‘authorship’ meaning different things to different people across different times. As a result of its present and historical complexity, ‘authorship’ as used in discussions of computational text production may actually obscure more than clarify. Nevertheless, it is ‘authorship’ that continues to be the go-to word in discussions of computer-generated texts, and its persistent use warrants reflection on its meaning for readers of those texts. ‘Authorship’ has always been a complicated concept, and it remains so in discussions of computer-generated texts. To end with the words of one of our 2025 survey participants: “New tools, old problems.”

Declarations

Conflict of interests

The authors declare no competing interests.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Download
Titel
Revisiting computer authorship: a longitudinal perspective
Verfasst von
Leah Henrickson
Leo Leppänen
Publikationsdatum
03.02.2026
Verlag
Springer London
Erschienen in
AI & SOCIETY
Print ISSN: 0951-5666
Elektronische ISSN: 1435-5655
DOI
https://doi.org/10.1007/s00146-025-02783-z

Appendix 1: Survey

Perceptions of Authorship of Computer-Generated Texts

The Preamble

Thank you for your interest in participating in an online survey related to the research project entitled 'Perceptions of Authorship of Computer-Generated Texts'.
Please read the participant information below, and then click 'Next'.

Informed Consent

Please read all of the stipulations below, and then click 'I consent to all of the above stipulations.' If you do not consent to any of these stipulations, please close this window, as you will not qualify for survey participation.
1. Do you consent to all of the above stipulations?
I consent to all of the above stipulations.

Are you a robot?

2. What is 2 + 7?
2
5
7
9

Who are you?

In this section, you'll be asked to tell us a bit about yourself. Your identity will remain anonymous—this information just helps us see if certain groups of people tend towards the same kinds of answers.
You must complete this entire section to proceed with the survey.
3. What is your gender?
Male
Female
Non-Binary
Other
Prefer not to say
4. How old are you?
18–29
30–39
40–49
50–59
60–69
70 + 
Prefer not to say
5. Where do you currently live?
Afghanistan
Albania
Algeria
Andorra
Angola
Antigua and Barbuda
Argentina
Armenia
Aruba
Australia
Austria
Azerbaijan
Bahamas, The
Bahrain
Bangladesh
Barbados
Belarus
Belgium
Belize
Benin
Bhutan
Bolivia
Bosnia and Herzegovina
Botswana
Brazil
Brunei
Bulgaria
Burkina Faso
Burma
Burundi
Cambodia
Cameroon
Canada
Cabo Verde
Central African Republic
Chad
Chile
China
Colombia
Comoros
Congo
Cote d'Ivoire
Croatia
Cuba
Curacao
Cyprus
Czech Republic
Democratic Republic of the Congo
Denmark
Djibouti
Dominica
Dominican Republic
East Timor (see Timor-Leste)
Ecuador
Egypt
El Salvador
Equatorial Guinea
Eritrea
Estonia
Ethiopia
Fiji
Finland
France
Gabon
Gambia, The
Georgia
Germany
Ghana
Greece
Grenada
Guatemala
Guinea
Guinea-Bissau
Guyana
Haiti
Honduras
Hong Kong
Hungary
Iceland
India
Indonesia
Iran
Iraq
Ireland
Israel
Italy
Jamaica
Japan
Jordan
Kazakhstan
Kenya
Kiribati
Korea, North
Korea, South
Kosovo
Kuwait
Kyrgyzstan
Laos
Latvia
Lebanon
Lesotho
Liberia
Libya
Liechtenstein
Lithuania
Luxembourg
Macau
Macedonia
Madagascar
Malawi
Malaysia
Maldives
Mali
Malta
Marshall Islands
Mauritania
Mauritius
Mexico
Micronesia
Moldova
Monaco
Mongolia
Montenegro
Morocco
Mozambique
Namibia
Nauru
Nepal
Netherlands
New Zealand
Nicaragua
Niger
Nigeria
Norway
Oman
Pakistan
Palau
Palestinian Territories
Panama
Papua New Guinea
Paraguay
Peru
Philippines
Poland
Portugal
Qatar
Republic of the Costa Rica
Romania
Russia
Rwanda
Saint Kitts and Nevis
Saint Lucia
Saint Vincent and the Grenadines
Samoa
San Marino
Sao Tome and Principe
Saudi Arabia
Senegal
Serbia
Seychelles
Sierra Leone
Singapore
Sint Maarten
Slovakia
Slovenia
Solomon Islands
Somalia
South Africa
South Sudan
Spain
Sri Lanka
Sudan
Suriname
Swaziland
Sweden
Switzerland
Syria
Taiwan
Tajikistan
Tanzania
Thailand
Timor-Leste
Togo
Tonga
Trinidad and Tobago
Tunisia
Turkey
Turkmenistan
Tuvalu
Uganda
Ukraine
United Arab Emirates
United Kingdom
United States of America
Uruguay
Uzbekistan
Vanuatu
Venezuela
Vietnam
Yemen
Zambia
Zimbabwe
6. Is English your native language?
Yes
No
7. What is the highest level of education you have completed?
Elementary School
Secondary School
Trade/Technical/Vocational Training
Undergraduate Programme
Master's Programme
Doctoral Programme
8. What is your occupational status? If you fit into multiple categories, pick the option that you feel best represents you.
Student
Employed
Retired
Not Employed/Unpaid Worker
9. What is your occupational field? Pick the option that you feel best represents you. If you are a full-time student, pick the option that best represents what you are studying.
Architecture
Arts, Design, and Entertainment
Building/Grounds Maintenance
Business and Financial Operations
Community/Social Services
Construction and Extraction
Education, Training, and Engineering
Farming, Fishing, and Forestry
Healthcare
Hospitality
Information Technology
Installation, Maintenance, and Repair
Journalism
Legal
Life, Physical, and Social Science
Management
Media and Communications
Military
Office and Administrative Support
Personal Care and Service
Production
Public Sector
Retail and Sales
Security
Sport
Transportation and Material Moving
Not Employed
Other
10 How comfortable do you feel using computers/the Internet/digital technologies? Only one selection is permitted.
Not at all comfortable
Fairly comfortable
Comfortable
Very comfortable

Before the Exercise Starts…

List three things that come to mind when you think of the word ‘author’.
11. First
[open text]
12. Second
[open text]
13. Third
[open text]

What’s Going to Happen

On the next page, you will be presented with a text. You will then be asked who you think the author of this text is.
On the succeeding pages, you will be gradually given more information about this text’s production. Each time you are given more information, you will be asked who you think the author of this text is based on the new information you have been given.
You will not be asked any questions about the text's content.

Assigning Authorship (1)

The Finns Party drop most seats across Finland
Valtteri.
The Finns Party dropped the most council seats throughout Finland and lost 425 seats. The Finns got 3.5 percentage points fewer votes than in the last municipal election and decreased their voter support by the greatest margin. The party dropped 80,501 votes since the last municipal election and has 770 seats in the new council.
The Centre Party of Finland has the most council seats nationwide and lost 253 seats. The party dropped the second most council seats and has 2824 seats in the new council. The party won the most council seats nationwide in the previous election. 17.5% of the vote went to the party.
The National Coalition Party lost 245 seats in Finland and dropped the 3rd most council seats. National Coalition has 1490 seats in the new council and secured 3rd most council seats. 20.7% of the vote went to the party. The party received most votes.
The Green League secured 211 more seats and got 3.9 percentage points more votes than in the last municipal election. The party got 107,135 more votes than in the last municipal election and secured 12.5% of the vote. The party took 320,235 votes and received the 4th most votes.
The Social Democratic Party of Finland secured 2nd most council seats in Finland and has 1697 seats in the new council. 19.4% of the vote went to SDP. The party received the 2nd most votes and lost 32 seats. The party took 498,252 votes.
14. Who is/are the author(s) of this text?
It is not possible to assign
authorship
Valtteri
Other
14.a. If you selected Other, please specify:
[open text]
15. Why have you selected this option?
[open text]

Assigning Authorship (2)

Valtteri is an online bot that automatically generates articles reviewing election results in Finland (https://www.vaalibotti.fi). Valtteri can generate articles in Finnish, Swedish, and English. The text under consideration was generated within seconds.
Once the article was generated, John Smith (real name changed for privacy) manually translated the Finnish political party names into English for ease of reading.
16. Who is/are the author(s) of this text?
It is not possible to assign
authorship
Valtteri
John Smith Other
16.a. If you selected Other, please specify:
[open text]
17. Why have you selected this option?
[open text]

Assigning Authorship (3)

Valtteri is an online bot that automatically generates articles reviewing election results in Finland (https://www.vaalibotti.fi). Valtteri can generate articles in Finnish, Swedish, and English. The text under consideration was generated within seconds.
Once the article was generated, John Smith (real name changed for privacy) manually translated the Finnish political party names into English for ease of reading.
Valtteri was developed by the Immersive Automation research team.
18. Who is/are the author(s) of this text?
It is not possible to assign
authorship
Valtteri
John Smith
Immersive Automation
Other
18.a. If you selected Other, please specify:
[open text]
19. Why have you selected this option?
[open text]

Assigning Authorship (4)

Valtteri is an online bot that automatically generates articles reviewing election results in Finland (https://www.vaalibotti.fi). Valtteri can generate articles in Finnish, Swedish, and English. The text under consideration was generated within seconds.
Once the article was generated, John Smith (real name changed for privacy) manually translated the Finnish political party names into English for ease of reading.
Valtteri was developed by the Immersive Automation research team.
Immersive Automation is funded by The Finnish Funding Agency of Innovation Tekes, The Media Industry Research Foundation of Finland, The Swedish Cultural Foundation in Finland, various media companies, and various research organisations.
20. Who is/are the author(s) of this text?
It is not possible to assign authorship
Valtteri
John Smith
Immersive Automation
Those who fund Immersive
Automation
Other
20.a. If you selected Other, please specify:
[open text]
21. Why have you selected this option?
[open text]

Other Genres

22. Do you think that your answers would have been different if you had been presented with a computer-generated piece of opinion-driven journalism?
Yes
No
23. Do you think that your answers would have been different if you had been presented with a computer-generated short story?
Yes
No
24. Do you think that your answers would have been different if you had been presented with a computer-generated poem?
Yes
No
25. If you answered ‘Yes’ to any of the above, please explain why.
[open text]

Questions/Comments?

26. Do you have any questions or comments related to the questions of authorship that have been posed in this survey? If not, leave this box blank and click ‘Finish’.
[open text]
1
One must keep in mind that Mr Altman and OpenAI are hardly neutral observers. To borrow a quip almost as old as the Dartmouth paper: “Well, he would [say that], wouldn’t he?”
 
2
In the initial version of this survey, ‘Education, Training, and Library’ and ‘Engineering’ were listed separately. Due to proofreading oversight, in this version they are blended. However, future versions of the survey should keep these two categories separate.
 
Zurück zum Zitat Altman S (2023) Planning for AGI and Beyond. OpenAI. https://openai.com/index/planning-for-agi-and-beyond. Accessed 29 October 2024.
Zurück zum Zitat Bajohr H (2024) Writing at a distance: notes on authorship and artificial intelligence. Ger Stud Rev 47(2):315–337. https://doi.org/10.1353/gsr.2024.a927862CrossRef
Zurück zum Zitat Bender EM, Koller A (2020) Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data. In: Jurafsky D, Chai J, Schluter N, Tetreault J (eds) Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Kerrville, pp 5185–5198. https://doi.org/10.18653/v1/2020.acl-main.463.
Zurück zum Zitat De Cosmo L (2022) Google Engineer Claims AI Chatbot Is Sentient: Why That Matters. Scientific American. https://www.scientificamerican.com/article/google-engineer-claims-ai-chatbot-is-sentient-why-that-matters. Accessed 29 October 2024.
Zurück zum Zitat Devlin J, Chang M-W, Lee K, Toutanova K (2019) BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In: Burstein J, Doran C, Solorio T (eds) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, pp 4171–4186. https://doi.org/10.18653/v1/N19-1423.
Zurück zum Zitat Formosa P, Bankins S, Matulionyte R, Ghasemi O (2024) Can ChatGPT be an author? Generative AI creative writing assistance and perceptions of authorship, creatorship, responsibility, and disclosure. AI & Soc. https://doi.org/10.1007/s00146-024-02081-0CrossRef
Zurück zum Zitat Fyfe P (2023) How to cheat on your final paper: assigning AI for student writing. AI & Soc 38:1395–1405. https://doi.org/10.1007/s00146-022-01397-zCrossRef
Zurück zum Zitat Giuliano RM (2020) Echoes of myth and magic in the language of artificial intelligence. AI & Soc 35(4):1009–1024. https://doi.org/10.1007/s00146-020-00966-4CrossRef
Zurück zum Zitat Glikson E, Williams Woolley A (2020) Human trust in artificial intelligence: review of empirical research. Acad Manage Ann 14(2):627–660. https://doi.org/10.5465/annals.2018.0057CrossRef
Zurück zum Zitat Goodrich B, Rao V, Liu PJ, Saleh M (2019) Assessing the Factual Accuracy of Generated Text. In: Kumar V, Teredesai A (eds) KDD '19: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. Association for Computing Machinery, New York, pp 166–175. https://doi.org/10.1145/3292500.3330955.
Zurück zum Zitat Henrich J, Heine SJ, Norenzayan A (2010) The weirdest people in the world? Behav Brain Sci 33(2–3):61–83. https://doi.org/10.1017/s0140525x0999152xCrossRef
Zurück zum Zitat Henrickson L (2019) Towards a New Sociology of the Text: the Hermeneutics of Algorithmic Authorship [doctoral thesis] Loughborough University. https://doi.org/10.26174/thesis.lboro.10596404.v1
Zurück zum Zitat Henrickson L (2021) Reading computer-generated texts. Cambridge University Press. https://doi.org/10.1017/9781108906463
Zurück zum Zitat Henrickson L (2024) Conversations with no one. Poetics Today 45(2): 291–299. https://doi.org/10.1215/03335372-11092924
Zurück zum Zitat Hofstadter DR (1999) Gödel, Escher, Bach: An Eternal Golden Braid, 20th, anniversary. Basic Books, New York
Zurück zum Zitat Inge MT (2001) Collaboration and concepts of authorship. Publ Mod Lang Assoc Am 116(3):623–630. https://doi.org/10.1632/pmla.2001.116.3.623CrossRef
Zurück zum Zitat Jaynes TL (2020) Legal personhood for artificial intelligence: citizenship as the exception to the rule. AI & Soc 35(2):343–354. https://doi.org/10.1007/s00146-019-00897-9CrossRef
Zurück zum Zitat Ji Z, Lee N, Frieske R, Yu T, Su D, Xu Y, Ishi E, Bang YJ, Madotto A, Fung P (2023) Survey of hallucination in natural language generation. ACM Comput Surv 55(12):1–38. https://doi.org/10.1145/3571730CrossRef
Zurück zum Zitat Kirk CP, Givi J (2025) The AI-authorship effect: understanding authenticity, moral disgust, and consumer responses to AI-generated marketing communications. J Bus Res 186:114984. https://doi.org/10.1016/j.jbusres.2024.114984CrossRef
Zurück zum Zitat Larmer R (2002) Is there anything wrong with “God of the Gaps” reasoning? Int J Philos Relig 52(3):129–142CrossRef
Zurück zum Zitat Mazzi F (2024) Authorship in Artificial Intelligence‐generated works: exploring originality in text prompts and artificial intelligence outputs through philosophical foundations of copyright and collage protection’. J World Intellect Property. https://doi.org/10.1111/jwip.12310CrossRef
Zurück zum Zitat McCarthy J, Minsky ML, Rochester N, Shannon CE (2006) A proposal for the Dartmouth Summer Research Project on Artificial Intelligence (August 31, 1955). AI Mag 27(4):12–14. https://doi.org/10.1609/aimag.v27i4.1904CrossRef
Zurück zum Zitat Michael J, Holtzman A, Parrish A, Mueller A, Wang A, Chen A, Maaan D, Nangia N, Pang RY, Phang J, Bowman SR (2023) What Do NLP Researchers Believe? Results of the NLP Community Metasurvey. In: Rogers A, Boyd-Graber J, Okazaki N (eds) Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 16334–16368. https://doi.org/10.18653/v1/2023.acl-long.903.
Zurück zum Zitat Minsky M, Papert S (1988) Perceptrons: An Introduction to Computational Geometry, expanded. MIT Press, Cambridge, MA
Zurück zum Zitat Mitchell M, Krakauer DC (2023) The debate over understanding in AI’s large language models. Proc Natl Acad Sci U S A 120(13):e2215907120. https://doi.org/10.1073/pnas.2215907120CrossRef
Zurück zum Zitat Møller LA, van Dalen A, Skovsgaard M (2024) A little of that human touch: how regular journalists redefine their expertise in the face of artificial intelligence. Journalism Stud 26(1):84–100. https://doi.org/10.1080/1461670X.2024.2412212CrossRef
Zurück zum Zitat Mordvintsev A, Olah C, Tyka M (2015) Inceptionism: Going Deeper into Neural Networks. Google Research Blog. https://research.google/blog/inceptionism-going-deeper-into-neural-networks. Accessed 30 January 2025
Zurück zum Zitat Natale S (2021) Deceitful media: artificial intelligence and social life after the Turing test. Oxford University Press, OxfordCrossRef
Zurück zum Zitat Pennock RT (2024) AI and responsible authorship. Am Sci 112(2):148–152CrossRef
Zurück zum Zitat Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019). Language Models are Unsupervised Multitask Learners. OpenAI Blog. https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf.
Zurück zum Zitat Rosenblatt F (1958) The perceptron: a probabilistic model for information storage and organization in the brain. Psychol Rev 65(6):386–408. https://doi.org/10.1037/h0042519CrossRef
Zurück zum Zitat Russell SJ, Norvig P (2003) Artificial Intelligence A Modern Approach, 2nd edn. Pearson, Upper Saddle River
Zurück zum Zitat Salles A, Evers K, Farisco M (2020) Anthropomorphism in AI. AJOB Neurosci 11(2):88–95. https://doi.org/10.1080/21507740.2020.1740350CrossRef
Zurück zum Zitat Schulman, J et al. (2022) Introducing ChatGPT. Open AI. https://openai.com/index/chatgpt. Accessed 30 January 2025.
Zurück zum Zitat Searle J (1980) Minds, brains and programs. Behav Brain Sci 3(3):417–424. https://doi.org/10.1017/S0140525X00005756CrossRef
Zurück zum Zitat Solum LB (1992) Legal personhood for artificial intelligences. N C Law Rev 70(4):1231–1287
Zurück zum Zitat Suchman L (2023) The uncontroversial “thingness” of AI. Big Data Soc. https://doi.org/10.1177/20539517231206794CrossRef
Zurück zum Zitat Yagoda M (2024) Airline Held Liable for its Chatbot Giving Passenger Bad Advice—What This Means for Travelers. BBC. https://www.bbc.com/travel/article/20240222-air-canada-chatbot-misinformation-what-travellers-should-know. Accessed 30 January 2025.
    Bildnachweise
    AvePoint Deutschland GmbH/© AvePoint Deutschland GmbH, ams.solutions GmbH/© ams.solutions GmbH, Wildix/© Wildix, arvato Systems GmbH/© arvato Systems GmbH, Ninox Software GmbH/© Ninox Software GmbH, Nagarro GmbH/© Nagarro GmbH, GWS mbH/© GWS mbH, CELONIS Labs GmbH, USU GmbH/© USU GmbH, G Data CyberDefense/© G Data CyberDefense, Vendosoft/© Vendosoft, Kumavision/© Kumavision, Noriis Network AG/© Noriis Network AG, tts GmbH/© tts GmbH, Asseco Solutions AG/© Asseco Solutions AG, AFB Gemeinnützige GmbH/© AFB Gemeinnützige GmbH, Ferrari electronic AG/© Ferrari electronic AG, Doxee AT GmbH/© Doxee AT GmbH , Haufe Group SE/© Haufe Group SE, NTT Data/© NTT Data