Computer-Written Scripts & Deepfakes: AI is Part of the Hollywood Strikes

Image: Unsplash

Law

Computer-Written Scripts & Deepfakes: AI is Part of the Hollywood Strikes

For the first time in 60 years, the Writers Guild of America (“WGA”) and the Screen Actors Guild (“SAG”) are simultaneously facing off against the Alliance of Motion Picture and Television Producers. Among the key points of contention for the individuals ...

August 1, 2023 - By Jasmin Pfefferkorn

Computer-Written Scripts & Deepfakes: AI is Part of the Hollywood Strikes

Image : Unsplash

Case Documentation

Computer-Written Scripts & Deepfakes: AI is Part of the Hollywood Strikes

For the first time in 60 years, the Writers Guild of America (“WGA”) and the Screen Actors Guild (“SAG”) are simultaneously facing off against the Alliance of Motion Picture and Television Producers. Among the key points of contention for the individuals participating in the strike: Working conditions, adequate pay, and the increasing encroachment of artificial intelligence (“AI”) into their professions.

The use of AI in the film and television industry is not new. In fact, many common post-production techniques use AI technology in special effects, color grading, animation and video editing. Not only was the Lord of the Rings trilogy a defining moment of the early 2000s, it also illuminated how AI could be used in film production. The Battle of Helm’s Deep, for instance, features computer-generated AI armies to create one of the most memorable scenes in cinematic history. But in the current strike, the specific concern is a subset of AI known as Generative AI, and it is crucial that an equilibrium is reached between protections for creative professionals, and the application of Generative AI as a useful tool.

What is Generative AI?

Like all AI, a Generative AI model is fed existing data (content), using algorithms to process this data, identify patterns and produce outputs – such as an image or a piece of writing. What is significant about Generative AI, in particular, is the capacity to undertake the so-called “learning” process relatively autonomously and to generate original content. 

Many of us are most familiar with Generative AI as the technical process that gives us increasingly sophisticated deepfakes. The viral image of Pope Francis wearing an oversized puffer jacket? Courtesy of a 31-year-old construction worker using the AI image generator Midjourney. Hardly a coincidence, the rise of Generative AI, complete with fake images of the pope, has taken off in the mainstream as a result of the offerings of companies, such as Midjourney, Stable Diffusion, Meta and OpenAI, with the latter now infamous for – and on the receiving end of a growing number of lawsuits over – its large image model Dall-E and large language model ChatGPT. 

What’s happening in Hollywood?

Hollywood workers have valid reason for their unease. The fear is that AI will not only be used for supporting technical jobs, such as color grading or adding characters in the far background, but it will also replace creative jobs. For both WGA and SAG, there is also a legitimate worry that entry level jobs (such as writers’ assistants and background extras on sets) will be largely replaced by AI, which would significantly reduce opportunities for people entering the workforce to gain necessary expertise in their craft.

With the staggering improvements in each ChatGPT iteration, screenwriters have also been grappling with the possibility they will be sharing creative control over scripts with large language models. Questions arise around how these works would be attributed, who or what would be given credit, and consequently, how payment would be allocated. 

It is worth noting that the unions are not entirely against the use of AI. The WGA has proposed a model for human-AI collaboration where generative AI could produce early versions of a script that human screenwriters will then refine. But many experts and industry professionals see this proposal as alienating writers from the creative process, repositioning writers as copy editors. One of the most dystopian scenarios to be put on the table by big studios has been termed “performance cloning,” which involves paying background actors a one-off fee to scan their likeness. This likeness can then be owned and used by companies in perpetuity. 

While creating a regressive payment model, this scenario also raises issues of consent: What happens, for example, if your AI body double is used in a way you would never agree to? It also raises questions from a contract point of view: What if your AI body double is used to promote a product or service that clashes with existing – and exclusivity-based – deals your human person currently maintains? And the list goes on. 

It’s also a question of copyright

With generative AI, consent is closely bound together with issues of copyright. Comedian Sarah Silverman is currently suing OpenAI and Meta for direct and vicarious copyright infringement, violations of section 1202(b) of the Digital Millennium Copyright Act (“DMCA”), unjust enrichment, violations of the California and common law unfair competition laws, and negligence. Silverman and the other plaintiffs allege that the tech company defendants’ AI models were trained on their work (including Silverman’s copyright-protected book, called “The Bedwetter”) without their consent, and were consequently able to roughly reproduce their literary/comedy styles. That their works are part of the machine learning dataset is unsurprising; this dataset encompasses billions of data points – essentially all that has made its way onto the internet. 

Though Generative AI is said to produce original content, a better way to view this content is as a remix. These models regurgitate what they have been trained on. If they become foundational to the film and television industry, the originality of our cultural products is up for debate. Streaming services have already primed audiences in the algorithmic curation of taste. Generative AI extends this existing trajectory. If studios become overly reliant on these technologies, chances are the “new” content offered to us will only echo what has come before. It may even move us further away from equality in representation, with the bias of these AI models being well-documented

Collaboration without exploitation

As workers fight for industry regulation to ban the replacement of humans by AI, it is important to reiterate this is not a call to ban the technology outright. Generative AI has already been used in valuable and compelling ways in film. An early example is David France’s 2020 documentary Welcome to Chechnya, which explores the persecution of LGBTQ+ people in Russia. France did extensive post-production work using AI, producing synthetic voices, and superimposed faces to protect his subjects’ anonymity while retaining their humanity. 

The question at the heart of copyright – how we balance protecting the rights of creatives with the openness needed for cultural production – resurfaces in this context. We need regulatory measures that enable creative collaboration with Generative AI while ensuring creative workers are not exploited to further centralized power. In June, the Directors Guild of America won protection against being replaced by AI tools in a new labor contract with producers. The AI-centric hope among those involved in the ongoing strike is that protections will be extended to screenwriters and actors. Otherwise, in Hollywood, AI might just steal the show. 


Jasmin Pfefferkorn is a Postdoctoral Research Fellow at the School of Culture and Communication at the University of Melbourne. (This article was initially published by The Conversation.)

related articles