Kickstarter shut down the campaign for AI porn group Unstable Diffusion amid changing guidelines
The group trying to monetize AI porn generation, Unstable Diffusion, raised more than $56,000 on Kickstarter from 867 backers. Now, as Kickstarter changes its thinking about what kind of AI-based projects it will allow, the crowdfunding platform has shut down Unstable Diffusion’s campaign. Since Kickstarter runs an all-or-nothing model and the campaign had not yet concluded, any money that Unstable Diffusion raised will be returned to the funders. In other words, Unstable Diffusion won’t see that $56,000, which more than doubled its initial $25,000 goal.
“Over the last several days, we’ve engaged our Community Advisory Council and we’ve read your feedback to us via our team and social media,” said CEO Everette Taylor in a blog post. “And one thing is clear: Kickstarter must, and will always be, on the side of creative work and the humans behind that work. We’re here to help creative work thrive.”
Kickstarter’s new approach to hosting AI projects is intentionally vague.
“This tech is really new, and we don’t have all the answers,” Taylor wrote. “The decisions we make now might not be the ones we make in the future, so we want this to be an ongoing conversation with all of you.”
Right now, the platform says it is considering how projects interface with copyrighted material, especially when artists’ work appears in an algorithm’s training data without consent. Kickstarter will also consider whether the project will “exploit a particular community or put anyone at risk of harm.”
In recent months, tools like OpenAI’s ChatGPT and Stability AI’s Stable Diffusion have been met with mainstream success, bringing conversations about the ethics of AI artwork into the forefront of public debate. If apps like Lensa AI can leverage the open source Stable Diffusion to instantly create artistic avatars that look like a professional artist’s work, how does that impact those same working artists?
Some artists took to Twitter to pressure Kickstarter into dropping the Unstable Diffusion project, citing concerns about how AI art generators can threaten artists’ careers.
Many cite the fate of Greg Rutkowski’s work as an example of what can go wrong. A living illustrator who has crafted detailed, high fantasy artwork for franchises like “Dungeons & Dragons,” Rutkowski’s name was one of Stable Diffusion‘s most popular search terms when it launched in September, allowing users to easily replicate his distinctive style. Rutkowski never consented to his artwork being used to train the algorithm, leading him to become a vocal advocate about how AI art generators impact working artists.
“With $25,000 in funding, we can afford to train the new model with 75 million high quality images consisting of ~25 million anime and cosplay images, ~25 million artistic images from Artstation/DeviantArt/Behance, and ~25 million photographic pictures,” Unstable Diffusion wrote in its Kickstarter.
Spawning, a set of AI tools designed to support artists, developed a website called Have I Been Trained, which lets artists see if their work appears in popular datasets and opt out. Per an April court case, there is legal precedent to defend the scraping of publicly accessible data.
Inherent problems in AI porn generation
Ethical questions about AI artwork get even murkier when considering projects like Unstable Diffusion, which center around the development of NSFW content.
Stable Diffusion uses a dataset of 2.3 billion images to train its text-to-image generator. But only an estimated 2.9% of the dataset contains NSFW material, giving the model little to go on when it comes to explicit content. That’s where Unstable Diffusion comes in. The project, which is part of Equilibrium AI, recruited volunteers from its Discord server to develop more robust porn datasets to fine-tune their algorithm, the same way you would upload more pictures of couches and chairs to a dataset if you wanted to make a furniture-generation AI.
But any AI generator is prone to fall victim to whatever biases the humans behind the algorithm have. Much of the porn that’s free and easily accessible online is developed for the male gaze, which means that’s likely what the AI will spit out, especially if those are the kinds of images that users are inputting into the dataset.
In its now-suspended Kickstarter, Unstable Diffusion said that it would work toward making an AI art model that can “better handle human anatomy, generate in diverse and controllable artistic styles, represent under-trained concepts like LGBTQ and races and genders more fairly.”
Plus, there’s no way of verifying whether much of the porn that’s freely available on the internet was made consensually (however, adult creators who use paid platforms like OnlyFans and ManyVids must verify their age and identity before using these services). Even then, if a model consents to appearing in porn, that doesn’t mean that they consent to their images being used to train an AI. While this technology can create stunningly realistic images, that also means that it can be weaponized to make nonconsensual deepfake pornography.
Currently, few laws around the world pertain to nonconsensual deepfaked porn. In the U.S., only Virginia and California have regulations restricting certain uses of faked and deepfaked pornographic media.
“One aspect that I’m particularly worried about is the disparate impact AI-generated porn has on women,” Ravit Dotan, VP of responsible AI at Mission Control, told TechCrunch last month. “For example, a previous AI-based app that can ‘undress’ people works only on women.”
Unstable Diffusion did not respond to request for comment at the time of publication.