In Opposition to Copyright Protection for AI Works

This is a response to “Paradise Rejected: A Conversation about AI and Authorship with Dr. Ryan Abbott” hosted by Professor Sandra Aistars at the Center for the Protection of Intellectual Property (CPIP) at George Mason University School of Law. It was first published on the CPIP blog in conjunction with Professor Aistars’s post. 


On February 14, the U.S. Copyright Office confirmed its rejection of an application for a claim of copyright in a 2D artwork called “A Recent Entrance to Paradise.” The image, created by an AI designed by Dr. Stephen Thaler, was rejected by the Office on the longstanding doctrine which holds that in order for copyright to attach, a work must be the product of human authorship. Among the examples cited in the Copyright Office Compendium as ineligible for copyright protection is “a piece of driftwood shaped by the ocean,” a potentially instructive analog as the debate about copyright and AI gets louder in the near future.

What follows assumes that we are talking about autonomous AI machines producing creative works that no human envisions at the start of the process, other than perhaps the medium. So, the human programmers might know they are building a machine to produce music or visual works, but they do not engage in co-authorship with the AI to produce the expressive elements of the works themselves. Code and data go in, and something unpredictable comes out, much like nature forming the aesthetic piece of driftwood.

As a cultural question, I have argued many times that AI art is a contradiction in terms—not because an AI cannot produce something humans might enjoy, but because the purpose of art, at least in the human experience so far, would be obliterated in a world of machine-made works. It seems that what the AI would produce would be literally and metaphorically bloodless, and after some initial astonishment with the engineering, we may quickly become uninterested in most AI works that attempt to produce more than purely decorative accidents.

In this regard, I would argue that the question presented is not addressed by the “creative destruction” principle, which demands that we not stand in the way of machines doing things better than humans. “Better” is a meaningful concept if the job is microsurgery but meaningless in the creation or appreciation of art. Regardless, the copyrightability question does not need to delve too deeply into the nature or purpose of art because the human element in copyright is not just a paragraph about registration in the USCO Compendium but, in fact, runs throughout application of the law.

Doctrinal Oppositions to Copyright in AI Works

In the United States and elsewhere, copyright attaches automatically to the “mental conception” of a work the moment the conception is fixed in a tangible medium such that it can be perceived by an observer. So, even at this fundamental stage, separate from the Copyright Office approving an application, the AI is ineligible because it does not engage in “mental conception” by any reasonable definition of that term. We do not protect works made by animals, who possess consciousness that far exceeds anything that can be said to exist in the most sophisticated AI. (And if an AI attains true consciousness, we humans may have nothing to say about laws and policies on the other side of that event horizon.)

Next, the primary reason to register a claim of copyright with the USCO is to provide the author with the opportunity, if necessary, to file a claim of infringement in federal court. But to establish a basis for copying, a plaintiff must prove that the alleged infringer had access to the original work and that the secondary work is substantially or strikingly similar to the work allegedly copied. The inverse ratio rule applied by the courts holds that the more that access can be proven, the less similarity weighs in the consideration and vice-versa. But in all claims of copying, independent creation (i.e., the principle that two authors might independently create nearly identical works) nullifies any complaint. These are considerations not just about two works, but about human conduct.

If AIs do not interact with the world, listen to music, read books, etc. in the sense that humans do these things, then, presumably, all AI works are works of independent creation. If multiple AIs are fed the same corpus of works (whether in or out of copyright works) for the purpose of machine learning, and any two AIs produce two works that are substantially, or even strikingly, similar to one another, the assumption should still be independent creation. Not just independent, but literally mindless, unless again, the copyright question must first be answered by establishing AI consciousness.

In principle, AI Bob is not inspired by, or even aware of, the work of AI Betty. So, if AI Bob produces a work strikingly similar to a work made by AI Betty, any court would have to toss out BettyBot v. BobBot on a finding of independent creation. Alternatively, do we want human juries considering facts presented by human attorneys describing the alleged conduct of two machines?

If, on the other hand, an AI produces a work too similar to one of the in-copyright works fed into its database, this begs the question as to whether the AI designer has simply failed to achieve anything more than an elaborate Xerox machine. And hypothetical facts notwithstanding, it seems that there is little need to ask new copyright questions in such a circumstance.

The factual copying complication raises two issues. One is that if there cannot be a basis for litigation between two AI creators, then there is perhaps little or no reason to register the works with the Copyright Office. But more profoundly, in a world of mixed human and AI works, we could create a bizarre imbalance whereby a human could infringe the rights of a machine while the machine could potentially never infringe the rights of either humans or other machines. And this is because the arguments for copyright in AI works unavoidably dissociate copyright from the underlying meaning of authorship.

Authorship, Not Market Value, is the Foundation of Copyright

Proponents of copyright in AI works will argue that the creativity applied in programming (which is separately protected by copyright) is coextensive to the works produced by the AIs they have programmed. But this would be like saying that I have claim of co-authorship in a novel written by one of my children just because I taught them things when they were young. This does not negate the possibility of joint authorship between human and AI, but as stated above, the human must plausibly argue his own “mental conception” in the process as a foundation for his contribution.

Commercial interests vying for copyright in AI works will assert that the work-made-for-hire (WMFH) doctrine already implicates protection of machine-made works. When a human employee creates a protectable work in the course of his employment, the corporate entity, by operation of law, is automatically the author of that work. Thus, the argument will be made that if non-human entities called corporations may be legal authors of copyrightable works, then corporate entities may be the authors of works produced by the AIs they own. This analogizes copyrightable works to other salable property, like wines from a vineyard, but elides the fact that copyright attaches to certain products of labor, and not to others, because it is a fiction itself whose medium is the “personality of the author,” as Justice Holmes articulated in Bleistein.

The response to the WMFH argument should be that corporate-authored works are only protected because they are made by human employees who have agreed, under the terms of their employment, to provide authorship for the corporation. Authorship by the fictious entity does not exist without human authorship, and I maintain that it would be folly to remove the human creator entirely from the equation. We already struggle with corporate personhood in other areas of law, and we should ask ourselves why we believe that any social benefit would outweigh the risk of allowing copyright law to potentially exacerbate those tensions.

Alternatively, proponents of copyright for AI works may lobby for a sui generis revision to the Copyright Act with, perhaps, unique limitations for AI works. I will not speculate about the details of such a proposal, but it is hard to imagine one that would be worth the trouble, no matter how limited or narrow. If the purpose of copyright is to proscribe unlicensed copying (with certain limitations), we still run into the independent creation problem and the possible result that humans can infringe the rights of machines while machines cannot infringe the rights of humans. How does this produce a desirable outcome which does not expand the outsize role giant tech companies already play in society?

Moreover, copyright skeptics and critics, many with deep relationships with Big Tech, already advocate a rigidly utilitarian view of copyright law, which is then argued to propose new limits on exclusive rights and protections. The utilitarian view generally rejects the notion that copyright protects any natural rights of the author beyond the right to be “paid something” for the exploitation of her works, and this cynical, mercenary view of authors would likely gain traction if we were to establish a new framework for machine authorship.

Registration Workaround (i.e., lying)

In the meantime, as Stephen Carlisle predicts in his post on this matter, we may see a lot of lying by humans registering works that were autonomously created by their machines. This is plausible, but if the primary purpose of registration is to establish a foundation for defending copyrights in federal court, the prospect of a discovery process could militate against rampant falsification of copyright applications. Knowing misrepresentation on an application is grounds for invalidating the registration, subject to a fine of up to $2,500, and further implies perjury if asserted in court.

Of course, that’s only if the respondent can defend himself. A registration and threat of litigation can be enough to intimidate a party, especially if it is claimed by a big corporate tech company. So, instead of asking whether AI works should be protected, perhaps we should be asking exactly the opposite question: How do we protect human authorship against a technology experiment, which may have value in the world of data science, but which has nothing to do with the aim of copyright law?

 About the IP Clause

And with that statement, I have just implicated a constitutional argument because the purpose of copyright law, as stated in Article I Clause 8, is to “promote science.” Moreover, the first three subjects of protection in 1790—maps, charts, and books—suggest a view at the founding period that copyright’s purpose, twinned with the foundation for patent law, was more pragmatic than artistic.

Of course, nobody could reasonably argue that the American framers imagined authors as anything other than human or that copyright law has not evolved to encompass a great deal of art which does not promote the endeavor we ordinarily call “science.” So, we may see AI copyright proponents take this semantic argument out for a spin, but I do not believe it should withstand scrutiny for very long.

Perhaps, the more compelling question presented by the IP clause, with respect to this conversation, is what it means to “promote progress.” Both our imaginations and our experiences reveal technological results that fail to promote progress for humans. And if progress for people is not the goal of all law and policy, then what is? Surely, against the present backdrop in which algorithms are seducing humans to engage in rampant, self-destructive behavior, it does seem like a mistake to call these machines artists.

Is “Machine Learning” Copying or Reading?

machine reading

I recently attended a round-table discussion on the subject of artificial intelligence and copyright.  The first of several engaging topics I thought warranted a post was the question of “machine learning,” which I put in quotes here with respect to one scholar who admonished against anthropomorphizing AI by using words for human activities to describe the actions of computers.  I think that view is fundamentally correct, though there is also grounds for analogy, as will be made clear by the following premise:

When you read a book, even if we might say, by way of analogy, that you are “copying” the content of that book onto your brain, this clearly does not infringe §106(1) of the copyright law proscribing unauthorized copying.  Since the author naturally hopes that you will read her book, such a prohibition would be absurd, even if you had an eidetic memory and could, if prompted, recite the entire work verbatim.  But if you used that gift to type from memory the entire book and made that document available, you would then violate more than one statute under the copyright law.

So, the question raised in regard to “machine learning” is whether the computer scientist who wishes to feed a corpus of books—say the anthology of American literature—into an AI should be required to obtain licenses for the works still under copyright.  Thus, the first analysis is whether the act of “copying” can be said to occur in this circumstance any more than it would be for the human reader who consumes the same body of literature.

It strikes me that if what the AI does in this case is ingest the corpus of books and almost instantly deconstructs those works by synthesizing them through a neural network, then the computer scientist has a pretty solid argument that no copying has taken place.   If the machine does not retain intact copies of works—or even large sections of works—-with the purpose of making those intact copies available to the human market, then this “machine reading” process is arguably analogous to the human whose reading does not infringe §106(1) of the copyright law.

That said, intent of the computer scientist may be a significant factor.  For instance, if the training of the AI will have a commercial purpose, this may suggest a requirement to license the works under copyright.  But intent can be very tricky on the leading edge of science because it is neither realistic, nor even desirable, to insist that every researcher know exactly where his experiments will lead.  This would nullify the process of discovery whence many great achievements have been made; hence, discovery is justification itself, and I suspect the tech companies would appeal to this rationale in regard to “machine learning.”

If the computer scientist’s goal is to see whether he can get his AI to “learn” about the American experience through literature, but he does not have a particular product or service in mind at the outset, it seems that copyright owners would be on fairly shaky ground to enjoin his use of the books.  As long as nothing that comes out the other end looks like any of the products that went in, it strikes me that this experiment exists beyond the statutory framework of copyright law.

Of course this portrait of the individual scientist beavering away in his modest lab to see what he may discover is not what is taking place in reality. We know perfectly well that major AI experimentation occurs in the R&D labs of companies like Google and Facebook, who are well shielded by trade-secret law from divulging what they are working on or for what purpose.  Like any other corporations, they are free to announce a new product or service without telling the public how they arrived at the latest result.

So, even if the use of copyrighted works as source material resulting in a commercial end might recommend some type of licensing regime, it may be very difficult to identify the threshold when the blind process of scientific discovery becomes a clear intent to exploit a commercial opportunity.  And, as mentioned, these companies would be under no obligation to divulge that eureka moment to anyone.  

On the other hand, the moment Google or Facebook did announce that new product, rightsholders could justifiably complain that a massive, highly-profitable corporation has used potentially billions of dollars worth of material without paying for any of it.  As one scholar at the round-table noted, tech companies may not use raw silicon for free, so why should they get to exploit millions of creative works for free, no matter what they’re turning that data into?

It’s a good question.  One that would seem to suggest a new subsection of the copyright law, and this would certainly be consistent with the fact that new forms of exploitation of works may demand equally new forms of compensation.  If nothing else, that type of statutory response could spare us all the tedious and false harangue that insists “copyright owners just want to stand in the way of innovation.”

That argument prevailed for far too long, and now the so-called innovators have a lot of splainin’ to do about their culture of blind disruption for the sake of disruption. Especially in light of the fact that AI may have some very profound effects on society as we know it, maybe this time around the copyright owners should be treated like experienced voices in the conversation rather than canaries wasting their breath in the proverbial coal mine.

Discovered Wiener Paper Warns about Smart Machines

John Markoff, writing for The New York Times, reports on the discovery of a long lost paper by M.I.T mathematician Norbert Wiener.  Written more than sixty years ago, the final paragraphs of the paper resonate loud and clear as we now flirt with the realities of artificial intelligence and, one hopes, consider carefully what it is we wish for from our machines.

“Moreover, if we move in the direction of making machines which learn and whose behavior is modified by experience, we must face the fact that every degree of independence we give the machine is a degree of possible defiance of our wishes. The genie in the bottle will not willingly go back in the bottle, nor have we any reason to expect them to be well disposed to us.”  

See full article here.