Stable Diffusion image generator
Released: August 2022
Stable Diffusion is a free, open source AI-based system that generates images from text descriptions or prompts.
Stable Diffusion quickly became popular largely on account of its free availability, minimal usage restrictions, and the fact that it can run on ordinary consumer hardware.
But it is seen to suffer from a number of technical limitations, and to pose several actual and potential challenges and risks to organisations and individuals operating and using it:
Accuracy: The quality of images produced by the system degrade above 512×512-pixel resolution; it also has trouble generating images with human limbs and faces.
Safety: As TechCrunch notes, Stability AI bans lewd or sexual material, hateful or violent imagery, prompts containing copyrighted or trademarked material, and personal information such as phone numbers and Social Security numbers. But this has not stopped users generating an avalanche of abusive, violent, sexual and pornographic images.
Bias: Stable Diffusion was primarily trained on images with English descriptions, resulting in images that tend to reinforce western stereotypes and biases.
Dual/multi-use: The relative accessibility of the model, including the release of its model weights, together with its relative paucity of filters, means Stable Diffusion can be used for more or less any purpose, including nefarious ones such as spreading spam, propaganda, deepfakes, misinformation and disinformation.
Copyright: Stable Diffusion draws on copyrighted works of artists, without the owner’s consent or any form of attribution, prompting concerns about plagiarism and the misuse, abuse and loss of IP. Furthermore, the user is given full rights to images they create on Stable Diffusion, and is free to use them commercially, resulting in questions about fair use. Responding to pressure, Stability AI says artists will have the chance to opt out of the training for the next version of the software.
In January 2023, a group of lawyers filed a US class-action lawsuit against Stability AI, DeviantArt, and Midjourney on the basis that their use of Stable Diffusion illegally remixes the copyrighted works of millions of artists whose work was used as training data.
In the same month, Getty Images announced it will sue Stability AI for alleged copyright violation, on the basis that it believes Stability AI 'unlawfully copied and processed millions of images protected by copyright' to train its software.
Employment: Artists' loss of copyright and the ability of users to create images in their unique style has prompted increasing concern that artists, photographers, illustrators, cinematographers and others stand to lose their commercial viability and careers. A French game developer's release of a Stable Diffusion-based tool to generate images in the style of late Korean artist Kim Jung Gi with a simple text prompt sparked a furious response from artists who saw the tool as theft, and accused the developer of cultural appropriation.
Unlike DALL-E, Stability AI has made Stable Diffusion accessible courtesy of a public demo, a software beta called DreamStudio, a development notebook, and weights, model card and code.
The company also appears to have listened to feedback from users and artists as Stable Diffusion 2 has most NSFW content, celebrity images and famous artists’ names removed.
It has also made it easier to find out whose work is in the dataset, thereby making simpler for artists to opt-out.
Operator: Stability AI; Prisma Labs
Developer: Stability AI; CompVis LMU; Eleuther AI; RunwayML; LAION
Purpose: Generate images
Technology: NLP/text analysis; Computer vision; Text-to-image; Neural network; Deep learning
Issue: Safety; Bias/discrimination; Dual/multi-use; Mis/disinformation; Copyright; Employment
Transparency: Black box