The father, the son, and the holy fake puffer jacket.
How unbelievably realistic fake images could take over the internet Last week, a 31-year-old construction worker [took a few psychedelics]( and thought it might be fun to use AI image generator Midjourney to create a photorealistic image of Pope Francis wearing a big white Balenciaga-style puffer jacket. A lot of people who saw it thought it was fun, too, so they spread it around social media. Most of them probably had no idea that it wasnât real. Now, the Pope [having that drip]( isnât the worst nor most dangerous deployment of photorealistic [AI-generated art](, in which new images are created from text prompts. But it is an example of just how good this technology is becoming, to the point that it can even [trick people]( who are usually more discerning about spreading misinformation online. You might even call it a turning point in the war against mis- and disinformation, which the people fighting were, frankly, already losing simply because [social]( [media]( [exists](. Now we have to deal with the prospect that even the people who are fighting that war may inadvertently help spread the disinformation theyâre trying to combat. And then what? Itâs not just Coat Pope. In the last two weeks, weâve seen several ominous AI-image stories. We had Trumpâs [fake arrest]( and attempted escape from the long AI-generated arm of the law, which was capped by a set of [poorly rendered]( fingers. We had Leviâs [announcing]( it would âsupplementâ its human models with AI-generated ones in the name of diversity (hiring more diverse human models was apparently not an option). Microsoft [unleashed]( its Bing Image Creator in its new AI-powered Bing and Edge browser, and Midjourney, known for its photorealistic images, released its [latest version](. Finally, [thereâs the news]( that AI image generators are getting better at drawing hands, which had been one of the tell-tale signs to detect if an image is fake. Even as convincing as Coat Pope appeared, a close look at his right hand would have revealed its AI origins. But soon, we may not even have that. Leviâs will be able to use AI models to show off its gloves, while the rest of us might be thrown into a new world where we have absolutely no idea what we can trust â one thatâs even worse than the world we currently inhabit. âWeâve had this issue with text and misinformation on social platforms. People are conditioned to be skeptical with text,â said Ari Lightman, a professor of digital media and marketing at Carnegie Mellon University. âAn image ... adds some legitimacy in the user's mind. An image of video creates more resonance. I don't think our blinders are up yet.â In just a few short years, AI-generated images have come a long way. In a more innocent time (2015) Google [released]( âDeepDream,â which used Googleâs artificial neural network programs â that is, artificial intelligence thatâs been trained to learn in a way that mimics a human brain's neural networks â to recognize patterns in images and make new images from them. Youâd feed it an image, and it would spit back something that resembled it but with a bunch of new images weaved in, often things approximating eyeballs and fish and dogs. It wasnât meant to create images so much as to show, visually, how the artificial neural networks detected patterns. [The results]( looked like a cross between a [Magic Eye drawing]( and my junior year of college. Not particularly useful in practice, but pretty cool (or creepy) to look at. These programs got better and better, training on billions of images that were usually scraped from the internet [without their original creatorsâ knowledge or permission](. In 2021, OpenAI released DALL-E, which could make photorealistic images from text prompts. It was a âbreakthrough,â says Yilun Du, a PhD student at MITâs Computer Science and Artificial Intelligence Laboratory who studies generative models. Soon, not only was photorealistic AI-generated art shockingly good, but it was also very much available. OpenAIâs Dall-E 2, Stability AIâs Stable Diffusion, and Midjourney were all released to the general public in the second half of 2022. The expected ethical concerns followed, from copyright issues to allegations of racist or sexist bias to the possibility that these programs could put a lot of artists out of work to what weâve seen more recently: convincing deepfakes used to spread disinformation. And while the images are very good, they still arenât perfect. But given how quickly this technology has advanced so far, itâs safe to assume that weâll soon be hitting a point where AI-generated images and real images are nearly impossible to tell apart. Take Nick St. Pierreâs work, for example. St. Pierre, a New York-based 30-year-old who works in product design, has spent the last few months [showing off]( his super-realistic AI art creations and explaining how he got them. He may not have the artistic skills to compose these images on his own, but he has developed a skill for getting them out of Midjourney, which he says he uses because he thinks itâs the best one out there. St. Pierre says he dedicated the month of January to 12-hour days of working in Midjourney. Now he can create something [like this]( in just about two hours. âWhen you see a digital image on the internet and it's AI generated, it can be cool, but it doesn't, like, shock you,â St. Pierre said. âBut when you see an image that's so realistic and you're like, âwow, this is a beautiful imageâ and then you realize it's AI? It makes you question your entire reality.â But St. Pierre doesnât usually put real people in his work (his [rendering]( of Brad Pitt and John Oliver as female Gucci models from the '90s is an exception, though few people would look at either and think they were actually Brad Pitt or John Oliver). He also thinks social media companies will continue to develop better tools to detect and moderate problematic content like AI-generated deepfakes. âI'm not as concerned about it as a lot of people are,â he said. âBut I do see the obvious dangers, especially in the Facebook world.â Two of Nick St. Pierre's AI-generated images, made using Midjourney V5. You can find the prompts St. Pierre used to create them [here]( (man on the left) and [here]( (woman on the right). Du, from MIT, thinks weâre at least a few years away from AI being able to produce images and videos that flood our world with fake information. Itâs worth noting that, as realistic as St. Pierreâs images are, theyâre also the end product of hours and hours of training. Coat Pope was made by someone who said heâd been playing around with Midjourney since last November. So these arenât yet images that anyone can just spin up with no prior experience. Lightman, from Carnegie Mellon, says the question now is whether weâll be ready for that possibility. Of course, a lot of this depends on the companies that make these programs, the platforms that host them, and the people who create the images to act responsibly and do everything possible to prevent this from happening. There are plenty of signs that they wonât. Bing Creator wonât generate an image of a real person, but Midjourney â the source of both Coat Pope and Fugitive Trump â clearly does (it has since banned the creators of both images from the platform but did not respond to request for comment). They all have their own rules for what is or isnât allowed. Sometimes, there arenât any rules at all. Stable Diffusion is open source, so anyone with any motives can build their own thing on top of it. Social media platforms have struggled for years to figure out what to do about the disinformation campaigns that run wild through them, or if and how they should curb the spread of misinformation. They donât seem very [well-equipped]( to deal with deepfakes either. Expecting all of humanity to do the right thing and not try to trick people or use AI images for malicious purposes is impossibly naive. And while many leaders of the AI movement [signing a letter]( from an effective altruism-linked nonprofit that urged a six-month moratorium on developing more advanced AI models is better than nothing, itâs also not legally compelling. Nor has it been signed by everyone in the industry. This all assumes that most people care a lot about not being duped by deepfakes or other lies on the internet. If the past several years have taught us anything, itâs that, while a lot of people [think]( fake news is a real issue, they often donât care or [donât know]( how to check that what theyâre consuming is real â [especially]( when that information conforms to their beliefs. And there are people who are happy enough to take what they see at face value because they donât have the time or perhaps the knowledge to question everything. As long as it comes from a trusted source, they will assume itâs true. Which is why itâs important that those trusted sources are able to do the work of vetting the information they distribute. But there are also people who do care and see the potential damage that deepfakes that are indistinguishable from reality pose. The race is on to come up with some kind of solution to this problem before AI-generated images get good enough for it to be one. We donât yet know who will win, but we have a pretty good idea of what we stand to lose. Until then, if you see an image of Pope Francis strolling around Rome in Gucci jeans on Twitter, you might want to think twice before you hit retweet. âSara Morrison, senior reporter [A photo of Starbucks interim CEO Howard Schultz is posted on a tree as striking Starbucks workers picket outside of a Starbucks coffee shop during a national strike on November 17, 2022, in San Francisco, California.]( Justin Sullivan/Getty Images [Starbucks wonât admit to breaking the law by union busting]( [Longtime CEO Howard Shultz testified before a Senate committee as the companyâs labor violations stack up.]( SOPA Images/LightRocket via Gett [How to test what an AI model can â and shouldnât â do]( [Inside the labs that helps evaluate AI safety for models like GPT-4]( [From right to left, Senators Joe Manchin, John Thune, Mark Warner, Tammy {NAME}, and Michael Bennet standing behind a lectern.]( Bill Clark/CQ-Roll Call, Inc via Getty Images [The RESTRICT Act is more bad news for TikTok]( [Sens. Mark Warner and John Thune have a bipartisan bill to deal with TikTok and beyond.](
Â
[Learn more about RevenueStripe...]( [Elon Musk sits onstage holding a microphone.]( Hector Retamal/AFP via Getty Images [AI leaders (and Elon Musk) urge all labs to press pause on powerful AI]( [We got GPT-4. We could stop there for now, placing a moratorium on new AI systems more powerful than that.]( [A personâs hands hold a smartphone with the TikTok app open on it. In the background, two other people are looking at their phones.]( AFP via Getty Images [Has TikTok made us better? Or much, much worse?]( [The case for and against TikTok as a cultural force.]( Support our work Vox Technology is free for all, thanks in part to financial support from our readers. Will you join them by making a gift today? [Give]( [Listen To This] [Listen to This]( [A podcasting pioneer's new co-op]( Jesse Thorn turned his podcast company, Maximum Fun, into a worker-owned co-op. Thorn still owns a piece of the company he built over two decades, but so do his former employees, who are now his co-workers. And he thinks other media companies should do this, too. [Listen to Apple Podcasts]( [This is cool] [How video game sound effects are made](
Â
[Learn more about RevenueStripe...]( [Facebook]( [Twitter]( [YouTube]( This email was sent to {EMAIL}. Manage yourâ¯[email preferences]( , orâ¯[unsubscribe](param=tech) â¯to stop receiving emails from Vox Media. View our [Privacy Notice]( and our [Terms of Service](. Vox Media, 1201 Connecticut Ave. NW, Washington, DC 20036. Copyright © 2023. All rights reserved.