Major Film Studios File Copyright Suit Against Midjourney

midjourney

“Midjourney is the quintessential copyright free-rider and a bottomless pit of plagiarism.”

Well, there it is. If you had been wondering whether and when the major studios would file a copyright infringement suit against a developer of generative AI (GAI), it finally happened on June 11. Disney and its subsidiaries, along with Universal Studios, filed a complaint against developer Midjourney alleging copyright infringement of many of the studios’ most famous and valuable intellectual properties. In the broader AI saga, the lawsuit is a big deal, though what it means for creators in general is hard to say. The studios imply that their complaint is a slam dunk, and it probably is. As the brief states:

This case is not a “close call” under well-settled copyright law. Midjourney set up a lucrative commercial service by making countless, unauthorized copies of Plaintiffs’ copyrighted works, and now sells subscriptions to consumers so that they can view and download copies and derivatives of Plaintiffs’ valuable copyrighted characters. That is textbook copyright infringement.

The 110-page complaint hardly expounds upon legal arguments and, instead, presents substantial evidence that Midjourney willfully violates the reproduction, derivative works, public display, and public distribution rights of the plaintiffs. Both as a matter of input (model training) and output of prompted materials, the studios compellingly show that their highly valuable works went into the system and that those same works come out of the system with simple prompting by users. Outputs also include expressive details like lighting and production design copied from the motion picture environments associated with famous characters.

For example, the user doesn’t just produce Darth Vader but can obtain a rendering of Vader on a Star Wars ship with lighting and production values that result in a highly detailed unlicensed derivative. The studios also allege that Midjourney itself publicly displays these outputs as a means of promoting its product’s capabilities. And finally, the studios anticipate that Midjourney will claim not to be the direct infringer and, therefore, allege that the developer is liable for secondary copyright infringement by facilitating mass infringement among its customers.

Aaron Moss, on his blog Copyright Lately, contends this case stands out among the roughly 40 active lawsuits against AI developers because the studios present so much compelling visual evidence of mass copying of protected works. While other GAI cases like Suno, Udio, and NY Times also present substantial evidence of infringing outputs, I think Moss is right that in the visual arts cases, plaintiffs rely more on the assumed infringement by means of model training more than they show obviously infringing outputs. In Disney et al., any ordinary observer can see that the characters and worlds produced are precise copies of iconic IP belonging to the studios.

Personally, I don’t see how Midjourney defends itself and, therefore, assume this case will be settled rather than fully litigated. Whatever comes of that settlement, motion picture studios generally have an interest in the development of GAI, which raises questions about independent creators and workers in the industry, as well as the future of filmmaking itself.

GAI and the Future of Filmmaking

Historically, the independent creator and the “line worker” in the industry both benefit from copyright enforcement actions of the institutional creator. For instance, when the studios go after pirate sites, the indie filmmaker benefits from a legal action she cannot afford to take, and the many crew professionals have their livelihoods protected. But with GAI, not only are the studios not seeking to shut down the developers, but they are presumably interested in the prospect of using the technology to produce motion pictures with fewer workers. This longstanding ambition of various film executives may not be attainable, but many professionals are certain that it will be attempted.

In conversations with friends in the industry, opinions vary, including those who find that GAI tools present intriguing opportunities for independent creators to produce new kinds of work at lower cost and greater speed. But at the same time, anxieties are high that GAI will mean job loss in every department of a typical production, including writers, directors, and actors.

There is no question that many motion picture jobs are threatened by GAI, but concerns that the future of filmmaking will be reduced to a few executives overseeing a roomful of programmers may be unwarranted. While GAI motion visuals are impressive and improving rapidly, it is also a shiny new toy that forecasts various cultural, financial, legal, and technological questions yet to be answered. And that’s just for film production.

My long-held view is that it was Star Wars (1977) that short-circuited the era known as the American New Wave in cinema because suddenly the prospect of the mega-franchise was much bigger than the movie itself. The end of the studio contract system resulted in greater creative independence, and a group of young directors, including George Lucas, produced what many consider a brief golden age of American motion pictures that were simultaneously box office hits, critical successes, and award winners.

When Star Wars demonstrated the potential of the film as franchise, the concurrent influx of MBA types into Hollywood amplified a new tension between film as art and film as product—i.e., the tension between filmmaker and film executive. Within that tug-o-war, it is only natural for the “suits” to want to produce as much product as they can with as little labor as possible—let alone expensive labor with opinions! Now, GAI theoretically presents that opportunity, though maybe not to the extent that many seem to either pessimistically or optimistically assume.

Usually, when the “suits” overreach with their analytics and try to predict what the market wants, the results are unimpressive. A theme I have repeated on this blog many times is that audiences want works that surprise them, not works that have been analyzed to death—and to produce fresh work takes artists, not algorithms. This rule, if it is a rule, implies a boundary that rejects the most stark predictions that major motion pictures will soon be made by three guys and a computer.

Potential Limits of GAI in Motion Pictures

Clearly, GAI output will continue to improve, and subtle renderings of naturalism will be attainable, which does imply that a whole motion picture with “human” characters can be produced without a camera or actor being involved. That this describes the future of all cinema seems unlikely, though it is notable that among the evidence presented in Disney et al., only one image is a depiction of a natural person (Mark Hamill as Skywalker), while every other infringing image is either a masked character or was originally made with pen and ink or computer illustration. Thus, the GAI’s ability to render these particular derivative works implies precisely the franchise material that could be produced without anyone building a set or rigging lights or pointing a camera at an actor. Still, there are limits.

For instance, the current Marvel franchise was primed with the first Iron Man (2008), the success of which owes a LOT to the performance by Robert Downey, Jr. This implies a caveat that films made without human artists can become what we might soon call the slop-flops of this dawning GAI era. Still, without a crystal ball or room to explore all the implications of eliminating one type of creative professional or another, a limiting factor for overuse of GAI may be copyright itself.

Because the human authorship requirement is, and will likely remain, a bedrock principle of copyright protection, GAI enables the production of a very large volume of unprotectable expression. Additionally, if two creators are using the same product, the likelihood of substantially similar, but independently created,[1] works may increase as well. In this light, creators, large and small, might want to be wary of overreliance on producing GAI material that may lack copyright protection. This concern would apply with greater force to newer characters, interpretations, and/or imagined worlds, if authorship in these works could be challenged on the basis that the expressions are the result of machine interpretation of the idea rather than human artists expressing the idea.

Personally, I would love to see the GAI genie stuffed back into its bottle because I believe that on balance the technology produces more social harms than benefits—and because the Techbros have zero credibility when it comes to ethical development or application of any of their products. But knowing that genie’s bottle has been shattered, I recognize how the technology can be used as a tool for new creative expression and am hopeful that lawsuits like this one at least push the application of GAI in that direction.


[1] Works that are independently created are, by definition, non-infringing even if they are substantial similar to other works.

Finding Fair Use for GAI Training is Highly Problematic

fair use

Although I have expressed aspects of these views in several posts over the past couple of years, I will try to consolidate my opinion as to why GAI training with protected creative works is a more problematic fair use consideration than many, even the courts, seem to believe. I acknowledge that even fellow copyright advocates will disagree with some of this analysis, but here it goes:

For the sake of narrowing the focus to the question of whether training generative AI (GAI) with protected works favors a fair use exception, the following assumes that the training requires unlicensed copying of protected expression. Further, even if the GAI maker limits the product’s capacity to output infringing copies, this does not alter the fact that considering fair use for this purpose is, at best, troubling and, at worst, so disturbing to case law that the AI developers are begging the courts to articulate doctrine out of whole cloth.

A GAI’s Purpose is Not Analogous to Past Fair Use Factor One Findings

The courts have largely rejected the overbroad opinion that making “something new” is a sufficient justification for unlicensed use of protected works. Thus, it is difficult to see where any court finds an authority to support the argument that making a “creator robot,” however revolutionary its developers proclaim it to be, is a transformative purpose under a factor one analysis.

Typically, a GAI’s purpose neither expresses “critical bearing” on the works used (AWF v. Warhol) nor provides information about the works to human readers (Authors Guild v. Google) nor fosters interoperability in computer devices (Google v. Oracle). Instead, a GAI’s most widely applied and widely promoted purpose is artificial “authorship” without authors—a purpose which forecasts myriad negative effects that may prove to dramatically overwhelm any benefits promised by the developers.

Naturally, certain GAIs (e.g., ChatGPT) can be used for various purposes, about which more below, but if the courts are distracted by the sheer novelty, scope, and hype around the “importance” of GAI and, therefore, presume transformativeness, they may be persuaded to articulate a rationale that would be tantamount to a blanket exception for GAI training. If the court adopts this carve-out in the context of fair use factor one, the result would be a reversal of its own reluctance to favor the broad “something new” argument for transformativeness so recently rejected in Warhol.

Notably, it is not unprecedented for the court to articulate rationales beyond the four-factor analysis. In the Google Books case, the court found that the search tool provides a “social benefit,” and a similar sentiment was articulated in Google v. Oracle regarding consumer benefit in advancing mobile products. Or looking back at the Betamax VCR case, the concept of “time shifting” the viewing schedule served the public interest by expanding flexibility in the consumption of copyrighted material that was lawfully obtained.

But if the courts look for a rationale beyond the case law (e.g., a clear social benefit of GAI), not only will they be making a wild guess, but any conclusion in favor of the developers will probably be wrong—perhaps dangerously so. While it is understandable that the courts may be reluctant to hobble technological development in principle, the available facts militate against disturbing fair use jurisprudence for the sake of nurturing GAI in general.

Put differently, if the courts are going to take a wait-and-see approach, there is ample evidence that GAIs already cause harm to individuals—from CSAM and defamation to cheating and psychological issues—to say nothing of the well-founded anxieties—social, political, economic, and environmental—associated with this multi-trillion-dollar gamble being played by the same people who unrepentantly accrued wealth and power from the darkest results of Web 2.0.

GAI as a Tool for Creators

To the extent that a given GAI product may be considered a tool for producing creative works, a fair use holding should at least find that the tool “promotes the progress” of authorship with respect to copyright’s purpose. But this is difficult because the same GAI in the hands of one skilled creator offers little insight about its ultimate purpose in the hands of 100-million unskilled users.

At the positive end of considering GAI’s purpose, my friend David Bolinsky, a medical illustrator and animator, recently made a series of 8 dozen topically and stylistically distinct ten second animations, introducing speakers and segment topics for a scientific conference, a daunting assignment. GAI collapsed well over a year of work (if using his standard 3D animation tools) into a matter of weeks. He was surprised at the breadth and depth of creative latitude GAI enabled. Further, he explained that although these presentations allowed more creativity than his typical discrete medical and scientific educational animations, an amateur lacking his experience still could not have used the same GAI tools to achieve the same results. Consequently, Bolinsky sees GAI as an opportunity to do more and different kinds of work and not as a threat to his creativity or livelihood.

In this example, the technology is socially beneficial and arguably “promoting the progress” of authorship, which may favor a finding that the tool is transformative. That said, due to the human authorship requirement, we are years away from guidance as to the degree of copyright protection on those animations; and if GAI tools are used to produce millions of works that have no “authors” as a matter of law, it is contrary to find that this “promotes progress” in regard to copyright’s purpose.

Further, the difficulty for the court in considering fair use is that Bolinsky and his colleagues who specialize in medical work are unique among professional creators, to say nothing of the many millions of non-creator customers that GAI developers need—because they are leveraged into the stratosphere—to make their products profitable. This scale implies an analysis reminiscent of Sony—i.e., a question of whether the purpose of the GAI is substantially beneficial or substantially harmful. But knowing that requires time travel.

If a court could see a few years into the future and find, for instance, that the GAI at issue will be used substantially for nonconsensual pornography, disinformation, and scams, it would presumably decline to find these purposes are social benefits that favor an expansive transformativeness finding. Instead, at the moment, the courts simply have no idea what the true “purposes” are of various GAIs, which is unprecedented in fair use jurisprudence. The VTR, Google Books, Android phones, et al. did not serve materially different purposes years after they were presented to the courts in their respective cases. By contrast, GAIs present an incomplete and dynamic set of facts; and in my view, this alone should militate against finding that factor one favors any of these products.

The Threat to Authorship Itself

As stated in other posts and in comments to the Copyright Office, one unique challenge of GAI is that it poses a potential threat to authorship (i.e., that it will shrink the number of creative workers), which is clearly destructive to the progress clause and copyright law. Although my own view is that a party who poses an existential threat to copyright’s purpose should not be allowed to invoke one of copyright law’s affirmative defenses, I recognize the difficulty in that opinion.

Under U.S. law, copyright protects authors indirectly by protecting certain exclusive rights to use their works. Consequently, there is little foundation for arguing generalized harm to authorship itself, despite the overwhelming recognition that diversity in authorship has benefitted the United States both culturally and economically for almost two centuries. In this context, GAI provokes the question as to whether U.S. policy might shift toward a “moral rights” approach akin to Europe, but that’s a discussion for a different post.

Instead, the general threat to authorship is considered, to an extent, under fair use factor four, which weighs the potential threat to the market value of the works used. The key difficulty, however, is that if the GAI does not output the song “Ordinary” but instead outputs music in the style of Alex Warren, then the output is not, strictly speaking, a threat to the market value of “Ordinary” itself. While proposals like the NO FAKES Act would prohibit unauthorized replication of Warren’s voice, copyright law does not clearly prevent a GAI that makes Warren-like music that could theoretically obviate the need for Warren himself.[1]

For now, several plaintiffs in the roughly 40 active lawsuits GAI developers have presented evidence of outputs that are substantially similar to the works used in training, and this should disfavor fair use for the GAI developers under factor four. More broadly, plaintiffs in these cases argue that licensing works for the purpose of AI training is itself a market opportunity exclusive to the copyright owner, and therefore, the failure to license constitutes market harm under factor four.

Some courts may be reluctant to agree with the lost licensing opportunity claim, but that reluctance is unfounded—even if a developer successfully prevents its product from outputting copies of works used in training. So long as one of the exclusive copyright rights is implicated (and here, it would be the reproduction right), then a requirement to license exists. Consequently, failure to license, especially at such an extraordinary scale for unprecedented commercial venture, is unquestionably market harm to the copyright owner.

Even where there may be a close call on factor four, because the GAI developer should lose on factor one, and because factors two and three decidedly favor creator plaintiffs, factor four should not reasonably control in many of these cases. Moreover, the courts should pay scant attention to the claim by developers that the cost of licensing is existentially prohibitive to the development of GAI. In addition to the fact that this plea is barely tolerable from parties wildly spending billions on high-risk ventures, any claim that a license is “too costly” for any venture is no defense under copyright law. The copyright owner sets the terms for the use of her work, and the prospective user can accept those terms or not before using the work. If that rule applies to the bootstrapping indie filmmaker, surely it applies to Microsoft, Meta, Google, et al.

Conclusion

Fair use is a mixed question of fact and law, and I maintain that what should be most fatal to the developers’ fair use defense is that, like the public, the courts have insufficient facts about the ultimate purpose of GAI products. Just as with Web 2.0 in the late 1990s, we are witnessing unfounded political sentiment to once again let Big Tech do what it wants, preaching to the public that this time, the technology really will “solve the world’s problems.”

Of course, there is no rational basis for that belief beyond the self-interest of the developers and the investors losing billions every year. If past is prologue, Congress would live to regret the folly of allowing AI to run amok, just as Members of both parties now rue the unconditioned immunity of Section 230. In the meantime, while licensing copyrighted works for GAI training will not address all, or most, of the potential hazards of artificial intelligence, the courts should decline to adopt strained fair use rationales in the name of assumed progress that may turn out to be a complete disaster.


[1] I believe there are cultural reasons that militate against this result, but those predictions do not influence the fair use consideration.

Questions But Not Chaos at the Copyright Office

copyright office

I have not commented on developments since May 13 because in this instance, caution is more important than keeping up with every rumor, of which there are plenty. I stand by my general views articulated in that last post but am not quite ready to agree with Digital Music News reporting on May 23 that the Copyright Office has “plunged into total chaos.” In fact, it is both premature and self-defeating for creators to go there.

What is certain is that the administration’s unprecedented attempts to appoint the acting heads of both the Library of Congress and the Copyright Office invite statutory and constitutional conflict. These were presented in the lawsuit filed by Shira Perlmutter over what she argues was her unlawful and ineffective dismissal by the White House on May 10 from her position as Register of Copyrights and Director of the U.S. Copyright Office. As the complaint describes, the President’s concurrent and unilateral naming of DOJ attorney Todd Blanche as acting Librarian triggers a cascade of questions that are both legally uncertain and politically fraught.

That the President may dismiss the Librarian of Congress is well founded, but the process of installing a lawful acting Librarian pending a new nominee is another matter. In essence, Perlmutter’s argument rests on the foundation that the Library of Congress is not an executive agency as a matter of statutory or constitutional law. Under Title 2, the Librarian is nominated by the President and confirmed by the Senate, but “The Library of Congress is, in name and function, Congress’s Library,” Perlmutter’s complaint states.

Perlmutter enumerates both statutory and case law examples to support her claim that because the Library is not an executive agency, the President had no authority to name Blanche as acting Librarian under any provision that might be construed to give him that power. On that basis, because the Register of Copyrights is undeniably an appointee of the Librarian under Title 17, Perlmutter argues that the absence of a lawfully appointed acting Librarian nullifies both her dismissal and the attempted appointment of DOJ Associate Deputy AG Paul Perkins as acting Register. Further, as a constitutional matter, Perlmutter alleges that the President has attempted to arrogate to himself powers that rest solely with Congress.

While Perlmutter’s allegations read to this layman as compelling, I do not have sufficient knowledge about administrative law, let alone the relevant case law, to anticipate the counterarguments to her claims. On May 28, the DC district court denied Perlmutter’s request for a temporary restraining order (TRO) that would have reinstated her as Register pending the court addressing the merits of her claim. Notably, in denying the TRO, colleagues who attended the hearing say the court focused on the fact that the Office, not Perlmutter, would suffer the harm. The court also opined that it was compelling Congress did not intervene and noted that the Library of Congress is a “unicorn” that serves both legislative and executive functions.

And therein lies the rub—a kerfuffle that is legally uncertain but also ripe for substantial political haggling because not even Republicans on the Hill want the White House mucking about in the Library of Congress. Specifically, the Congressional Research Service (CRS) is a non-partisan agency that provides confidential reports to Members, and nobody in Congress wants that agency to be directed by whichever party is in the White House.

Meanwhile, all speculation as to the role of Big Tech and the timing of the Office’s third report on copyright and artificial intelligence is just that. The fact that an early draft of the report was made public one day before Perlmutter’s “dismissal” supports the theory that tech interests sought to quash or amend the conclusions of the report through its influence with Trump. Other reports implied that “tech” was synonymous with Elon Musk as the driving force and that his “abolish all IP” view ran afoul of right-wing media’s interest in its copyright-protected material. And, of course, that was before this past week’s fireworks between Musk and Trump.

Pick your favorite narrative, and it’s probably mostly wrong. But as a practical matter, I do think it is premature and unhelpful to say that the Copyright Office is in a state of utter chaos while both the legal and political difficulties triggered by the White House are addressed. Registration applications are still being processed, though it is safe to assume that the Office has paused at least some of its work as a consulting agency, including the anticipated fourth report on AI.

With the Library, it certainly appears that Trump may have stepped in a pile of WTF on the Hill because of the CRS. With the Copyright Office, creators should want a restoration of the normal, non-partisan function of the agency, maintaining the registration process and advising Congress, the courts, and the public on copyright law and policy. For now, I wouldn’t panic just yet.