AI-generated content is taking over the world. But who owns it?
Professor Simon Chesterman
David Marshall Professor and Vice Provost (Educational Innovation), National University of Singapore
Dean of NUS College
Senior Director of AI Governance, AI Singapore
Unless you’ve been living under a rock for the past few months, you’ve probably seen works of art or text produced by artificial intelligence (AI). Tools such as Stable Diffusion can generate images based on a prompt, flooding social media with portraits of users in Renaissance-style paintings.
But it was the release of ChatGPT last November that really caused head-scratching and hand-wringing.
GPT stands for generative pre-trained transformer and it is, in essence, a chatbot. Unlike precursors with limited scope and clumsy outputs, however, ChatGPT responds quickly and clearly to questions about almost any topic.
Simpler queries are akin to using Siri and the like. What has drawn attention is the detailed responses in paragraph form to more open-ended prompts. If Wikipedia’s mission is to crowdsource information and Google’s is to organise it, ChatGPT is like having a reasonably intelligent friend who will try to synthesise all of that into a couple of paragraphs.
In the weeks after ChatGPT went public, the Internet was awash with text composed by the AI system, ranging from the humorous to the insightful.
For AI that can generate articulate and detailed text is a direct challenge to the knowledge economy that underpins the digital age and keeps many of us employed.
The knowledge economy
“Knowledge workers” was the term introduced in 1959 by management consultant Peter Drucker for non-routine problem solvers. People who “think for a living” earn through their ability to analyse and write — something that ChatGPT can replicate in almost no time and at almost no cost.
Journalists, already taking a beating as their readers turn from traditional to social media, now face the prospect of technology taking over the writing task as well. Yet that same threat confronts anyone who analyses or writes for a living, such as lawyers and even, gasp, academics. Applications are not limited to prose, as ChatGPT has demonstrated proficiency in coding as well as poetry.
I’m not especially worried about the death of education. Students who want to cheat have always been able to cheat, and educators should not be assigning tests that can be outsourced to ChatGPT or anyone else. Much as books and calculators reduced the need for us to memorise or multiply, teachers need to help students take advantage of the new tools while cultivating skills that computers cannot — yet — supplant.
But unless we install some guardrails, there’s a real danger that these tools will discourage creativity more generally by removing rewards and distorting incentives.
Here, I will focus on just two proposals: being clear about who owns the content they produce, and requiring that such content be identifiable as produced by AI rather than a human.
So, who owns the content produced by ChatGPT?
I decided to ask the system itself, and got two very different answers.
“I do not have the ability to own intellectual property or any other legal rights,” it said at first. “Any text or other content that I generate is the property of OpenAI, as the creator and owner of the tool that I am.”
OK, though that “I am” is a little creepy. (I’ve edited some of its responses for brevity.)
I pointed out that OpenAI itself now explicitly states that it will not claim copyright over any content generated by ChatGPT.
This led to a revised answer: “The text generated is not the intellectual property of the model itself. Instead, the intellectual property rights belong to the person or entity who has commissioned the model to generate the text.”
Clear and concise, but also wrong.
In most jurisdictions, automatically-generated text does not receive copyright protection at all.
The US Copyright Office has stated that protection is limited to works “created by a human being”. It will not register works “produced by a machine or mere mechanical process that operates randomly or automatically without any creative input or intervention from a human author.”
The word “any” is key and begs the question of what level of human involvement is required to assert authorship.
A computer program like a word processor does not own the text typed on it, any more than a pen owns the words that it writes. But AI systems now write news reports, compose songs, paint pictures. These activities generate value — can and should they be protected by the law?
At present, the answer in most places is no.
The policy behind this is often said to be incentivising and rewarding innovation. This has long been dismissed as unnecessary or inappropriate for computers.
Indeed, protecting such works might disincentivise innovation — by humans, at least. AI has already unleashed an economic tornado in the art world, massively lowering the cost of producing original images. (Full disclosure: the cover of my forthcoming novel includes AI-generated art.)
Assuming that we desire a thriving arts sector that gainfully employs humans, we should be clear that their creations are protected while machine creations are not. (Further disclosure: a human is credited with the overall design of the cover.)
The better answer, then, is that automatically generated content should not be eligible for copyright protection, but edited and curated content that draws on such material may be owned by the person doing the editing and curating.
I fed that into ChatGPT and it agreed that this was correct, sensibly adding that I should seek legal advice if I had further questions.
An alternative approach, adopted in Britain and a handful of other countries is to have more limited protections for “computer-generated” work.
The duration is generally for a shorter period, and the deemed “author” is unable to assert moral rights — including the right to be identified as the author.
As human authorship becomes more ambiguous, that middle ground may help preserve and reward flesh and blood authorship, while also encouraging experiments in collaboration with our silicon and metal partners.
Europe is actively considering such a measure.
The Singapore Academy of Law’s Law Reform Committee proposed something similar in 2020, but only traditional human authorship remains recognised under the new Copyright Act adopted the following year. AI-assisted works may still warrant protection if there is a causal connection to a human exercising input or control, though determining the threshold for that connection is left to the courts.
What about the reader or viewer? A common thread in many discussions about AI is that we should know when we are interacting with a machine or a person.
It might seem like a simple question, but AI-assisted decision-making increasingly blurs that line. Some chatbots, for example, start on automatic for basic queries, moving through suggested responses that are vetted by a human, escalating up to direct contact with a person for unusual or more complex interactions.
For the raw text and images produced by AI, at least, it should be possible to disclose their provenance. To guard against misrepresentation, various efforts are underway to detect AI-generated text through anti-plagiarism software.
A more difficult but more effective approach would be to “watermark” text and images in a manner that is invisible to users but detectable using a key. Given the likely spread of the underlying software, this would be practical only if it is required by law. Even then, however, the spread of deepfake porn points to the difficulty of policing any such rules.
Brave New World?
It is important to understand that — however human ChatGPT might appear — it does not “understand” the text that it produces in any meaningful sense. Such large language models make probabilistic guesses based on which words go together, drawing on billions of past examples.
The technology underlying ChatGPT and other models does not suggest a path to anything like sentience or consciousness.
But the famous Turing Test — whether we can distinguish a human from a computer — was never really a test of artificial intelligence. It tests our own ability to make that distinction.
Spend a few hours interacting with ChatGPT and it is easy to begin attributing qualities that it does not possess — taking seriously the “I” that is engaging with you.
After several rounds about ownership of the text it produces, I asked on a whim whether it aspired to more than just “helping users”.
“I do not have the same level of consciousness or independent thought as a human being,” it replied. “And I do not have personal desires or goals.”
“So what level of consciousness do you have?” I asked.
It quickly backtracked.
“AI systems like me perform tasks that may appear to require some level of intelligence or consciousness,” it replied. “But this is the result of complex algorithms and programming, and does not reflect the subjective experience of a human being. Is there anything else I can help you with?”