Aiven Blog

Sep 25, 2024

Navigating the Creative Commons: AI, Ownership, and Software Development

The debate surrounding AI-generated content ownership has raged since the technology's inception. Despite significant advancements, the question of creative rights remains unresolved. This is particularly problematic in software development, where AI's influence can often be subtle and overlooked. To fully grasp the implications, it's essential to delve into the complex landscape of intellectual property and licensing.

josep-prat

Josep Prat

|RSS Feed

Open Source Engineering Director

The difference between working on open source software (OSS) and a commercial or proprietary product comes down to how that software is licensed. The latter is the intellectual property of an individual or company and cannot be tampered with by external developers, additionally, more often than not, its source code is only available to its creator. There is a long list of licenses that govern open source software. We have licenses that are, so called, permissive and others, so called copyleft. The General Public License (GPL) family of licenses belongs to this latter subgroup, the copyleft one. As with all open source licenses, it is intended to guarantee freedom to use, share and change all versions of a program. However, the license has a condition which means any modification of the software needs to be released under the same license as the original. If a developer is modifying a piece of code under a GPL license, it means that whatever comes out the other end also needs to be covered by the same license. This is a way to guarantee that the software under this license will remain forever open source.

In the days before AI, this was well known and many companies had policies on which type of licensed open source software was fine to include or copy over. And this worked fine because code snippets had a usually traceable lineage. However today, that is not the case which has created a legal landmine. A developer might be using AI for assistance, sometimes sharing a problem statement that the AI should provide some code for and possibly making minor tweaks to what AI suggests but the problem is, where is this code coming from? This is challenging on a number of levels. We looked at the practicalities in an earlier blog but, from a moral and ethical point of view, it is one of the main drivers fanning the flames of the ownership of creativity debate.

A monkey called Naruto

We can chart this back to a monkey. Not Darwin (we’re way ahead of him now) but a monkey called Naruto. You may not know him by name but you’ll likely have seen his face. He’s famous for taking a (brilliant) selfie in 2011 with a wildlife photographer’s camera. It also started a years-long saga in which the U.S., The Copyright Office and even Wikipedia weighed in over the issue of ownership. The upshot being that it was agreed, by law, that the only group able to own creativity are us humans - not animals and certainly not machines.

Bringing it back to AI, what this sentence tells us is that anything that is not human cannot create something new. That they do not have the capacity for creativity but, instead, can only tweak existing materials. Because of this, it means copyright can only be attributed to something new and therefore any modification is opening the door to copyright infringement.

That means that, legally speaking, machines cannot own copyrights because they are not creating things but taking things that already exist, making this output a derived work and not a novel one. Even the world’s most powerful wizard isn't immune. There are stories of AI regurgitating elements of the Harry Potter script as new ideas while, only last year, Stephen Fry said his voice was stolen from the Harry Potter audiobooks and replicated by AI. This isn’t to say that AI is the source of all copyright issues. We humans have been dealing with this controversy for years, for example in music where authors get inspiration from others (most recently with Ed Sheeran v Marvin Gaye) or the replication in storylines from Dune to Star Wars. The difference now is that machines are involved and cannot be judged by the same standards purely because we have deemed that only humans are capable of innovating and creating.

The law of unintended consequences

The technology world is in a state of permanent reflection regarding this matter. What is creativity? Can machines be creative and do we want them to be? Are we opening the door to Skynet and is the fear real? Professor Stephen Hawking told the BBC in 2014 that, “The development of full AI could spell the end of the human race….It would take off on its own, and re-design itself at an ever increasing rate. Humans, who are limited by slow biological evolution, couldn’t compete, and would be superseded.”

What we can gather from this is that, if we agree that machines can create, it solves today’s problem. They can create new work based on code they learned from GPL license software and be influenced by multiple sources, much like we do as humans. Unfortunately this is almost the embodiment of ‘the law of unintended consequences’. We might solve copyright infringement today but it might come at great expense in the long run.

Balancing innovation with integrity

In order to move forward, what is required is traceability. Users of AI, and in this particular case, developers, need to know exactly where code originates from so they know what they’re working with. The most common OS licenses already require attribution but, if AI learns from projects licensed under those ones and outputs a piece of code that is almost verbatim to it, the traceability and attribution gets lost which is where the problems start. Maybe licensing will come that will prevent source codes to be used in training so we can mitigate this issue but, even still, attribution would still be a problem that remains to be solved.

We also addressed ways that organizations can help protect themselves and developers in a piece that explored the set-up and remit of an AI Tactical Discovery team - more on that here. The bigger picture is that we need to start imposing guardrails on what AI can and cannot do. This is happening or being considered at a regulatory level. For example, the UK recently released the outcome of a consultation with an ‘approach that is strongly pro-innovation and pro-safety’. Most of the software foundations are coming up with their own set of guidelines too because, as AI creativity capabilities grow, we'll have to grapple with how to balance innovation with integrity.

How can we use AI to spark human creativity while still cherishing that special, intangible touch only a person can bring?


Subscribe to the Aiven newsletter

All things open source, plus our product updates and news in a monthly newsletter.

Related resources