Newsletter Subject

A lesson from the VCR

From

vox.com

Email Address

newsletter@vox.com

Sent On

Wed, Jan 17, 2024 09:01 PM

Email Preheader Text

A 1980s copyright case shows how OpenAI could survive its lawsuits. How copyright lawsuits could kil

A 1980s copyright case shows how OpenAI could survive its lawsuits. How copyright lawsuits could kill OpenAI If you’re old enough to remember watching [the hit kid’s show Animaniacs](, you probably remember Napster, too. The peer-to-peer file-sharing site, which made it easy to download music for free in an era before Spotify and Apple Music, took college campuses by storm in the late 1990s. This did not escape the notice of the record companies, and in 2001, a federal court [ruled]( that Napster was liable for copyright infringement. The content producers fought back against the technology platform and won. But that was 2001 — before the iPhone, before YouTube, and before generative AI. This generation’s big copyright battle is pitting journalists against artificially intelligent software that has learned from and can regurgitate their reporting. Late last year, the New York Times [sued OpenAI and Microsoft](, alleging that the companies are stealing its copyrighted content to train their large language models and then profiting off of it. [In a point-by-point rebuttal]( to the lawsuit’s accusations, OpenAI claimed no wrongdoing. Meanwhile, the Senate Judiciary Subcommittee on Privacy, Technology, and Law [held a hearing]( in which news executives implored lawmakers to force AI companies to pay publishers for using their content. Depending on who you ask, what’s at stake is either the future of the news business, [the future of copyright law](, the future of innovation, or, specifically, the future of OpenAI and other generative AI companies. Or all of the above. Ideally, Congress would step in to settle the debate, but as James Grimmelmann, a professor of digital and information law at Cornell Law School, told me: “Congress does not like to legislate on copyright unless there’s a consensus of most of the players in the room — and there’s not anything resembling that consensus right now. So Congress may hold hearings and talk about it, but we’re really far from any legislative action.” So which is it? Advocates of technological innovation would say that AI technology is full of promise and we’d better not stifle that while it’s in the early days of development. Media companies would say that even exciting technology companies need to pay when they use copyrighted content, and if we give AI a free pass, journalism as we know it could eventually cease to exist. The consensus of [casual]( [observers]( and [legal experts alike]( is that this New York Times lawsuit is a big deal. Not only does the Times appear to have a solid case, but OpenAI [has a lot to lose]( — [perhaps its very existence](. The case against OpenAI, briefly explained If you ask ChatGPT a question about, say, the fall of the Berlin Wall, there’s a good chance some of the information in the answer has been culled from New York Times articles. That’s because the large language model, or LLM, that powers ChatGPT has been trained on over 500 gigabytes of data, [including newspaper archives](. Generative AI tools only work because this training data helps them know how to effectively respond to prompts. In other words, copyrighted data, in part, is what makes this new technology powerful and what makes OpenAI such a [valuable company](. The New York Times claims that OpenAI trained its model with copyrighted Times content and did not pay proper licensing fees. That, [the lawsuit says](, enables OpenAI to “compete with and closely mimic” the New York Times, perhaps by summing up a news story based on Times reporting or summing up a product recommendation based on Wirecutter reviews. Even worse is what the lawsuit calls “regurgitation,” which is when OpenAI spits out text that matches Times articles verbatim. The Times provides 100 examples of such “regurgitation” in the lawsuit. In its rebuttal, OpenAI said that regurgitation is a “rare bug” that the company is “working to drive to zero.” It also claims that the Times “intentionally manipulated prompts” to get this to happen and “cherry-picked their examples from many attempts.” But at the end of the day, the New York Times argues that OpenAI is making money off of content and costing the newspaper “billions of dollars in statutory and actual damages.” [By one estimate](, given the millions of articles potentially implicated and the cost per instance of copying, the New York Times might be looking for $450 billion in damages. OpenAI has a clear solution to this conflict: Pay the copyright owners upfront. The company has already announced licensing deals with folks [like the Associated Press]( and [Axel Springer](. OpenAI also claims that it was negotiating a deal with the New York Times right before the newspaper filed its lawsuit. Just how much OpenAI is willing to pay news outlets is unclear. A January 4 report [in the Information]( said that OpenAI has offered some media firms “as little as between $1 million and $5 million to license their articles for use in training its large language models,” which seems like a small amount of money to OpenAI, currently aiming for a [valuation]( as high as $100 billion. But the mounting lawsuits, should they go against the company, could be far more expensive than paying heftier licensing fees. The New York Times is also not the only party suing OpenAI and other tech companies over copyright infringement. A growing list of authors and entertainers have been filing lawsuits [since ChatGPT made its splashy debut in the fall of 2022](, accusing these companies of copying their works in order to train their models. The copyright holders filing these lawsuits extend well beyond writers, too. Developers have sued OpenAI and Microsoft [for allegedly stealing software code](, while Getty Images is embroiled in a lawsuit against Stability AI, the makers of image-generating model Stable Diffusion, over its copyrighted photos. “When you’re talking about copyright and you get statutory damages,” said [Corynne McSherry](, legal director at the Electronic Frontier Foundation, “if you lose, the downside and the financial risk is massive.” The case for innovation While [it’s easy]( to compare the Times case to the Napster one, the better precedent involves the VCR, according to McSherry. In 1984, a years-long copyright case between Sony and Universal Studios over the practice of using VCRs to record TV shows [made it all the way to the United States Supreme Court](. The studio alleged that Sony’s Betamax video tapes could be used for copyright infringement, while Sony’s lawyers argued that taping shows was [fair use](, which is the doctrine that allows copyrighted material to be reused without permission or payment. Sony won. The judge’s decision, which has never been overturned, [said that]( if machines, including the VCR, have non-infringing uses then the company that makes them can’t be held liable if customers use them to infringe upon copyrights. The entertainment industry [was forever changed]( by this case. The VCR let people watch whatever was broadcast on TV whenever they wanted, and in just a few years, Hollywood studios actually ended up [seeing their profits grow]( in the VCR era. The machine got people more excited about watching movies, and they watched more of them, both at home and in theaters. “If you have to go to copyright owners for permission for technological innovation, you’re going to get a lot less innovation,” McSherry told Vox. That in mind, there’s one more copyright lawsuit worth looking at: [the Google Books case](. In 2004, Google started scanning books, including copyrighted works, so that “snippets” of their text would show up in search results. It partnered with libraries at places like Harvard, Stanford, and the University of Michigan, [as well as magazines](, like New York Magazine and Popular Mechanics, that wanted their archives digitized. Then came the lawsuits, including a 2005 class action suit from the Authors Guild. The authors cried copyright infringement, and Google claimed that making books searchable [amounted to fair use](. As Judge Denny Chin [said]( in a 2013 decision dismissing the authors’ lawsuit, Google Books is transformative because, thanks to the tool, “words in books are being used in a way they have not been used before.” It took about a decade, but Google eventually won, and [Google Books is now legal](. Like Sony and Napster before it, the Google Books case is ultimately about the battle between new technology platforms and copyright holders. It also raises the question of innovation. Is it possible that giving copyright holders too much power could stifle technological progress? In that 2013 decision, Judge Chin said its technology “advances the progress of the arts and sciences, while maintaining respectful consideration for the rights of authors and other creative individuals, and without adversely impacting the rights of copyright holders.” And a 2023 economics [study of the effects of Google Books]( found that “digitization significantly boosts the demand for physical versions” and “allows independent publishers to introduce new editions for existing books, further increasing sales.” So consider that another point in favor of giving tech platforms room to innovate. Few would disagree that technological progress has shaped the media business since the invention of the printing press. That’s basically why the [earliest copyright laws]( were written over 300 years ago: Technology made copying easier, and authors needed some way to protect their intellectual property. But AI is a bigger leap forward, technologically speaking, than the VCR, Napster, and Google Books combined. We don’t know yet, but AI seems destined to transform our understanding of copyright and how content creators get paid for their work. It will take a while, too. A ruling in the New York Times’s case against OpenAI will take years, and even then, questions will remain. “I think generative AI could be as transformational for copyright as the printing press,” said Grimmelmann, the Cornell law professor. “But that will probably take a little bit longer to play out.” —Adam Clark Estes, senior correspondent [Three speech bubbles representing the OpenAI GPT chatbot store are floating above a horizon in an etched drawing of a countryside.]( Paige Vickers/Vox; Getty Images [There are too many chatbots]( [Will OpenAI’s new chatbot store finally make AI useful?](   [The letters AI glowing on a dark ground surrounded by tiny dots of light like stars.]( Getty Images [Thousands of AI experts are torn about what they’ve created, new study finds]( [The very confusing landscape of advanced AI risk, briefly explained.](   [A vehicle that resembles a large drone with propellors extending from its base on four sides hovers in the air. ]( VCG/VCG via Getty Images [Are flying cars finally here?]( [The world had “flying cars” in the 1930s. We could be getting them again.](    [Learn more about RevenueStripe...](   [A hand puts a ballot into a box with a digital code on it.]( Moor Studio/Getty Images [You thought 2023 was a big year for AI? Buckle up.]( [AI will change the world this year. We just don’t know how yet.](   [An illustrated “2024” is seen through a binocular field of view.]( Paige Vickers/Vox [24 things we think will happen in 2024]( [From Trump to Tesla, how 2024 will shake out, according to the Future Perfect team.](   Support our work Vox Technology is free for all, thanks in part to financial support from our readers. Will you join them by making a gift today? [Give](   [Listen To This] [Listen to This]( [Hollywood's secret musicals]( The studios promoting Mean Girls, Wonka, and The Color Purple are hiding something from you. [Listen on Apple Podcasts](   [This is cool] [Astronomers spotted something perplexing near the beginning of time](  [Learn more about RevenueStripe...](   [Facebook]( [Twitter]( [YouTube]( This email was sent to {EMAIL}. Manage your [email preferences]( , or [unsubscribe](param=tech)  to stop receiving emails from Vox Media. View our [Privacy Notice]( and our [Terms of Service](. Vox Media, 1201 Connecticut Ave. NW, Washington, DC 20036. Copyright © 2024. All rights reserved.

EDM Keywords (215)

zero youtube yet year written world works working work willing well way watched wanted vehicle vcr valuation using used use university understanding unclear ultimately trump transformative transformational transform training trained train torn took think theaters thanks text tesla terms talking talk take summing storm stifle stealing statutory stake spotify specifically sony snippets shaped shake settle sent seen seeing sciences say ruling room rights revenuestripe resembles remain regurgitation regurgitate readers questions question protect prompts promise progress profiting professor practice possible point players play permission peer pay partnered part order openai one offered notice never negotiating napster money models model mind millions microsoft michigan mcsherry massive manage making makes makers made lot lose looking llm little listen like license libraries liable lesson legislate learned learn lawsuits lawsuit know judge join iphone invention innovation innovate information horizon home hollywood high hearing happen going go getting get generation future full free floating favor far fall expensive existence exist excited examples even escape era entertainers end embroiled email either effects easy drive downside dollars doctrine digital developers demand decision decade debate deal day culled could costing copyright copying content consider consensus congress compete compare company companies change case came broadcast box books better beginning battle basically base ballot authors ask arts articles answer also ai advocates according 2024 2001 1984 1930s

Marketing emails from vox.com

View More
Sent On

25/05/2024

Sent On

24/05/2024

Sent On

24/05/2024

Sent On

24/05/2024

Sent On

23/05/2024

Sent On

22/05/2024

Email Content Statistics

Subscribe Now

Subject Line Length

Data shows that subject lines with 6 to 10 words generated 21 percent higher open rate.

Subscribe Now

Average in this category

Subscribe Now

Number of Words

The more words in the content, the more time the user will need to spend reading. Get straight to the point with catchy short phrases and interesting photos and graphics.

Subscribe Now

Average in this category

Subscribe Now

Number of Images

More images or large images might cause the email to load slower. Aim for a balance of words and images.

Subscribe Now

Average in this category

Subscribe Now

Time to Read

Longer reading time requires more attention and patience from users. Aim for short phrases and catchy keywords.

Subscribe Now

Average in this category

Subscribe Now

Predicted open rate

Subscribe Now

Spam Score

Spam score is determined by a large number of checks performed on the content of the email. For the best delivery results, it is advised to lower your spam score as much as possible.

Subscribe Now

Flesch reading score

Flesch reading score measures how complex a text is. The lower the score, the more difficult the text is to read. The Flesch readability score uses the average length of your sentences (measured by the number of words) and the average number of syllables per word in an equation to calculate the reading ease. Text with a very high Flesch reading ease score (about 100) is straightforward and easy to read, with short sentences and no words of more than two syllables. Usually, a reading ease score of 60-70 is considered acceptable/normal for web copy.

Subscribe Now

Technologies

What powers this email? Every email we receive is parsed to determine the sending ESP and any additional email technologies used.

Subscribe Now

Email Size (not include images)

Font Used

No. Font Name
Subscribe Now

Copyright © 2019–2024 SimilarMail.