Home

The Year Ahead in Generative AI

At this very moment, I could boot up OpenAI’s popular ChatGPT and ask it to write this article for me. Or I could ask it to “write a lullaby about a cool goose who just wants to honk.” I could pull up Midjourney, Stable Diffusion, DALL-E 2, or any number of AI art generators and order it to craft me an portrait of this cool goose.

Then comes the ethical conundrums: Could AI write and illustrate a children’s book about the goose for me? Could I sell that book, even though I put much less effort into it than any children’s author would? Who should own the copyright? What about the rights of the authors whose work trained the AI that created my book?

AI technology is developing rapidly, but it’s not clear how close are we to generative AI models regularly making final, release-ready products. People in the art scene have already raised concerns over how generative AI systems, particularly GPT and diffusion-based machine learning models, will disrupt their profession. But AI generated content could be a big deal in many other industries, from programmers to copywriters. So how long until all our proverbial gooses are metaphorically cooked?

To answer these questions, I spoke to three people who all work directly with AI but have varying takes on the technology:

Not your average broomThe Flippr makes traditional sweeping obsolete, with its two-in-one brush and roll functions.

As AI gets better at generating visual and written content, there’s a real risk that it could undermine creators whose livelihoods depend on their ability to generate content.

Some companies are already using cheap AI art instead of paying for the real thing. In December, science fiction and fantasy publisher Tor was put on blast after sharing the cover of an upcoming book that turned out to be AI art purchased off a stock image site. The work was uncredited, and even the person who touched up the image for the cover went unnamed.

Creators including Polish fantasy artist Greg Rutkowski have come out against AI art for fear their personal brands have been squashed by the proliferation of AI art specifically meant to imitate their work. Rutkowski has supported plans from the likes of the Concept Artists Association who want to lobby for updates to IP and data privacy laws.

Stability AI, the company behind the open source text-to-image diffusion model Stable Diffusion, has made a few concessions to artists over concerns that their work is being stolen or copied by AI. In the company’s release of Stable Diffusion 2, it created new parameters making it more difficult to create images based on celebrities, to craft porn, or to create art “in the style of” real-world artists. Some fans of the more open, open source model were none too happy with the changes. Stability AI also announced it was working with the company Spawning to allow artists to “opt-out” of having their work used in the training of Stable Diffusion 3, which will likely be released in 2023.

Though Stability AI has made some stated efforts to disable AI’s ability to explicitly copy work, there’s a genuine sense among small artists who depend on art and portrait commissions they’ll suffer as AI art generators become even more popular.

That’s not to say things are dire just yet. Mitchell said there are some artists who are seeing some benefits to current AI models, especially helping them speed up their work by creating a baseline they can work off of.

“Some of the specific artists who are speaking out about their work being stolen, I think, will also be artists that will potentially become even more valued as the actual artists, potentially even driving up their sales, or at least driving up the cost of their original pieces of work,” she said.

In November of 2022, Meta’s Galactica AI had to be pulled two days after release since researchers found it was propagating lies and misinformation.

“There’s such a drive and push to get things out and sort of ask for forgiveness later or not even ask for forgiveness,” Mitchell said. “The incentive for these safer structures is fear of bad PR, or wanting to get good PR, and what that incentivizes is the bare minimum.”

Emad Mosteque, the CEO of Stability AI, has been publicly resistant to, in his words, “all barriers from anyone being able to create anything they can imagine.” And even when Stability AI sets limits on output, there’s nothing that stops users from modifying the open source model and removing them. Mosteque, a former hedge fund manager, is bullish about pushing the technology as far as it will go.

But pushing too far too fast can expose a system’s internal biases. It’s almost a certainty that our current generative AI systems will grow more sophisticated and more lifelike, and the better a model is, the more it can be used to promote racism or hate speech.

Many of the companies in the space seem to understand how bad the optics can get. Meta’s BlenderBot language model seems to get very weird when talking about racism. Google has been hesitant to release its LaMDA chatbot to the public, in part, for this reason, but that only means the company is being beaten to the punch by smaller outfits that are being less careful.

“All generative models have harmful biases,” Solaiman said. “It’s the sort of garbage in, garbage out idiom. With improved quality of output comes increased risk of harm.”

Mitchell didn’t mince words regarding what it would take for somebody to put the brakes on the speed of AI development. “I imagine it’s going to take someone being seriously harmed —for example, a child drinking bleach and dying after they read misinformation generated by an AI— before there’s a public outcry about how these things can be harmful and problematic,” she said.

Copyright confusion is already causing legal battles. Last year New York-based artist Kris Kashtanova claimed they received the first copyright for a graphic novel that used AI art generator Midjourney as a baseline for the work. The problem is, it appears the copyright office did not understand the work had been generated using AI, even though it credits Midjourney on the cover. In October, Kashtanova said the copyright office notified them it wants to rescind the copyright. They have since appealed to the copyright office through Texas attorney Van Lindberg. The letter of appeal, which Linbergerg shared with Gizmodo, argued Kashtanova had used Midjourney as a “tool” like any other artist might use pencil or a brush.

Another legal battle pits software authors against AI: As detailed by Bloomberg Law, a class action lawsuit filed by two software developers states that an OpenAI program called Copilot, which is supposed to write code from scratch, is actually borrowing from and straight up copying copyrighted code, since it was is trained on open source software hosted on sites like GitHub.

More lawsuits are likely in 2023. Eric Bourdages, an artist for the game Dead by Daylight, has created a campaign on Twitter to proliferate and sell merchandise of Mickey Mouse and other copyrighted characters generated by AI. Bourdages wants to test just how far AI images can push their immunity to lawsuits, especially from the litigation-happy Disney corporation. “Legally there should be no recourse from Disney as according to the AI models TOS these images transcends copyright and the images are public domain,” he tweeted.

“Intellectual property, I think, is going to be a really big one moving forward, especially for imagery,” Solaiman said. “One of the big considerations that I think we’ll see in this openness debate is there’s a huge concentration of power, and who can train these models, not just because of the resources they have, but because of what data they have access to.”

While the act of scraping the web for all available images has routinely been found legal in the courts, Mitchell said this suit and future litigation will inevitably put the legal framework of fair use of copyrighted material to the test. While there has been some effort to watermark images to make them unusable by AI, “there hasn’t been enough to keep up with the progress of the technology.”

Meanwhile, we’re not likely to see new U.S. legislation put a damper on AI technology. The European Union’s AI Act, first proposed in 2021, recently passed a critical stage, though that bill focuses on using AI for social scoring or on biometric identification, not on generative AI.

Language generation is likely to get better and better, Mitchell said, and those systems will become much better at interpreting prompts to understand what users mean. In a year, people won’t need to collect and share prompts to use to get the best results, as the AI will simply be able to interpret context of users’ phrases.

This could also mean new models will facilitate larger and larger pools of dialogue within prompts. And inside a chat AI, users might be able to carry on larger, more detailed conversations. That means systems like Replika, a kind of AI-based chat meant to be a wellness aid, could become even more sophisticated and prolific.

Wahlforss said that there’s significant room for improvement of AI with even bigger training sets. GPT3 was trained on what Solaiman called an “unfathomable” 43 terabytes of data, but even that ludicrous amount of data could be dwarfed by upcoming projects. LAION 5B, for example, is a 250-terabyte dataset with 5.6 billion images scraped off the web.

But the advancements won’t stop there. RunwayML is already creating technology that can generate new images from empty space based on a single picture, so he said we can expect even more innovations from original AI generation. Source: Gizmodo

Previous

Next