Unable to load image

Lolsuit filed against Midjourney, DeviantArt, and Stable Diffusion for copyright infringement or something :marseysal:

https://old.reddit.com/r/aiwars/comments/10biogl/stable_diffusion_litigation

								

								

Orange site: https://news.ycombinator.com/item?id=34377910

Hello. This is Matthew Butterick. I'm a writer, designer, pro­gram­mer, and law­yer. In Novem­ber 2022, I teamed up with the amaz­ingly excel­lent class-action lit­i­ga­tors Joseph SaveriCadio Zir­poli, and Travis Man­fredi at the Joseph Saveri Law Firm to file a law­suit against GitHub Copi­lot for its "unprece­dented open-source soft­ware piracy". (That law­suit is still in progress.)

Since then, we've heard from peo­ple all over the world---espe­cially writ­ers, artists, pro­gram­mers, and other cre­ators---who are con­cerned about AI sys­tems being trained on vast amounts of copy­righted work with no con­sent, no credit, and no com­pen­sa­tion.

Today, we're tak­ing another step toward mak­ing AI fair & eth­i­cal for every­one. On behalf of three won­der­ful artist plain­tiffs---Sarah Ander­senKelly McK­er­nan, and Karla Ortiz---we've filed a class-action law­suit against Sta­bil­ity AIDeviantArt, and Mid­jour­ney for their use of Sta­ble Dif­fu­sion, a 21st-cen­tury col­lage tool that remixes the copy­righted works of mil­lions of artists whose work was used as train­ing data.

Join­ing as co-coun­sel are the ter­rific lit­i­ga­tors Brian Clark and Laura Mat­son of Lock­ridge Grindal Nauen P.L.L.P.

Today's fil­ings:

As a law­yer who is also a long­time mem­ber of the visual-arts com­mu­nity, it's an honor to stand up on behalf of fel­low artists and con­tinue this vital con­ver­sa­tion about how AI will coex­ist with human cul­ture and cre­ativ­ity.

The image-gen­er­a­tor com­pa­nies have made their views clear.\

Now they can hear from artists.

Sta­ble Dif­fu­sion is an arti­fi­cial intel­li­gence (AI) soft­ware prod­uct, released in August 2022 by a com­pany called Sta­bil­ity AI.

Sta­ble Dif­fu­sion con­tains unau­tho­rized copies of mil­lions---and pos­si­bly bil­lions---of copy­righted images. These copies were made with­out the knowl­edge or con­sent of the artists.

Even assum­ing nom­i­nal dam­ages of $1 per image, the value of this mis­ap­pro­pri­a­tion would be roughly $5 bil­lion. (For com­par­i­son, the largest art heist ever was the 1990 theft of 13 art­works from the Isabella Stew­art Gard­ner Museum, with a cur­rent esti­mated value of $500 mil­lion.)

Sta­ble Dif­fu­sion belongs to a cat­e­gory of AI sys­tems called gen­er­a­tive AI. These sys­tems are trained on a cer­tain kind of cre­ative work---for instance text, soft­ware code, or images---and then remix these works to derive (or "gen­er­ate") more works of the same kind.

Hav­ing copied the five bil­lion images---with­out the con­sent of the orig­i­nal artists---Sta­ble Dif­fu­sion relies on a math­e­mat­i­cal process called dif­fu­sion to store com­pressed copies of these train­ing images, which in turn are recom­bined to derive other images. It is, in short, a 21st-cen­tury col­lage tool.

These result­ing images may or may not out­wardly resem­ble the train­ing images. Nev­er­the­less, they are derived from copies of the train­ing images, and com­pete with them in the mar­ket­place. At min­i­mum, Sta­ble Dif­fu­sion's abil­ity to flood the mar­ket with an essen­tially unlim­ited num­ber of infring­ing images will inflict per­ma­nent dam­age on the mar­ket for art and artists.

Even Sta­bil­ity AI CEO Emad Mostaque has fore­cast that "[f]uture [AI] mod­els will be fully licensed". But Sta­ble Dif­fu­sion is not. It is a par­a­site that, if allowed to pro­lif­er­ate, will cause irrepara­ble harm to artists, now and in the future.

The prob­lem with dif­fu­sion


The dif­fu­sion tech­nique was invented in 2015 by AI researchers at Stan­ford Uni­ver­sity. The dia­gram below, taken from the Stan­ford team's research, illus­trates the two phases of the dif­fu­sion process using a spi­ral as the exam­ple train­ing image.

https://i.rdrama.net/images/1684135631929439.webp

The first phase in dif­fu­sion is to take an image and pro­gres­sively add more visual noise to it in a series of steps. (This process is depicted in the top row of the dia­gram.) At each step, the AI records how the addi­tion of noise changes the image. By the last step, the image has been "dif­fused" into essen­tially ran­dom noise.

The sec­ond phase is like the first, but in reverse. (This process is depicted in the bot­tom row of the dia­gram, which reads right to left.) Hav­ing recorded the steps that turn a cer­tain image into noise, the AI can run those steps back­wards. Start­ing with some ran­dom noise, the AI applies the steps in reverse. By remov­ing noise (or "denois­ing") the data, the AI will emit a copy of the orig­i­nal image.

In the dia­gram, the recon­structed spi­ral (in red) has some fuzzy parts in the lower half that the orig­i­nal spi­ral (in blue) does not. Though the red spi­ral is plainly a copy of the blue spi­ral, in com­puter terms it would be called a lossy copy, mean­ing some details are lost in trans­la­tion. This is true of numer­ous dig­i­tal data for­mats, includ­ing MP3 and JPEG, that also make highly com­pressed copies of dig­i­tal data by omit­ting small details.

In short, dif­fu­sion is a way for an AI pro­gram to fig­ure out how to recon­struct a copy of the train­ing data through denois­ing. Because this is so, in copy­right terms it's no dif­fer­ent from an MP3 or JPEG---a way of stor­ing a com­pressed copy of cer­tain dig­i­tal data.

Inter­po­lat­ing with latent images

In 2020, the dif­fu­sion tech­nique was improved by researchers at UC Berke­ley in two ways:

1. They showed how a dif­fu­sion model could store its train­ing images in a more com­pressed for­mat with­out impact­ing its abil­ity to recon­struct high-fidelity copies. These com­pressed copies of train­ing images are known as latent images.

2. They found that these latent images could be inter­po­lated---mean­ing, blended math­e­mat­i­cally---to pro­duce new deriv­a­tive images.

The dia­gram below, taken from the Berke­ley team's research, shows how this process works.

https://i.rdrama.net/images/16841356326009624.webp

The image in the red frame has been inter­po­lated from the two “Source” images pixel by pixel. It looks like two translu­cent face images stacked on top of each other, not a sin­gle con­vinc­ing face.

https://i.rdrama.net/images/16841356335785332.webp https://i.rdrama.net/images/16841356340051105.webp

The image in the green frame has been gen­er­ated dif­fer­ently. In that case, the two source images have been com­pressed into latent images. Once these latent images have been inter­po­lated, this newly inter­po­lated latent image has been recon­structed into pix­els using the denois­ing process. Com­pared to the pixel-by-pixel inter­po­la­tion, the advan­tage is appar­ent: the inter­po­la­tion based on latent images looks like a sin­gle con­vinc­ing human face, not an over­lay of two faces.

Despite the dif­fer­ence in results, in copy­right terms, these two modes of inter­po­la­tion are equiv­a­lent: they both gen­er­ate deriv­a­tive works by inter­po­lat­ing two source images.

Con­di­tion­ing with text prompts

In 2022, the dif­fu­sion tech­nique was fur­ther improved by researchers in Munich. These researchers fig­ured out how to shape the denois­ing process with extra infor­ma­tion. This process is called con­di­tion­ing. (One of these researchers, Robin Rom­bach, is now employed by Sta­bil­ity AI as a devel­oper of Sta­ble Dif­fu­sion.)

The most com­mon tool for con­di­tion­ing is short text descrip­tions, also known as text prompts, that describe ele­ments of the image, e.g.---"a dog wear­ing a base­ball cap while eat­ing ice cream". (Result shown at right.) This gave rise to the dom­i­nant inter­face of Sta­ble Dif­fu­sion and other AI image gen­er­a­tors: con­vert­ing a text prompt into an image.

The text-prompt inter­face serves another pur­pose, how­ever. It cre­ates a layer of mag­i­cal mis­di­rec­tion that makes it harder for users to coax out obvi­ous copies of the train­ing images (though not impos­si­ble). Nev­er­the­less, because all the visual infor­ma­tion in the sys­tem is derived from the copy­righted train­ing images, the images emit­ted---regard­less of out­ward appear­ance---are nec­es­sar­ily works derived from those train­ing images.

The defen­dants


Sta­bil­ity AI

Sta­bil­ity AI, founded by Emad Mostaque, is based in Lon­don.

Sta­bil­ity AI funded LAION, a Ger­man orga­ni­za­tion that is cre­at­ing ever-larger image datasets---with­out con­sent, credit, or com­pen­sa­tion to the orig­i­nal artists---for use by AI com­pa­nies.

Sta­bil­ity AI is the devel­oper of Sta­ble Dif­fu­sion. Sta­bil­ity AI trained Sta­ble Dif­fu­sion using the LAION dataset.

Sta­bil­ity AI also released Dream­Stu­dio, a paid app that pack­ages Sta­ble Dif­fu­sion in a web inter­face.

DeviantArt

DeviantArt was founded in 2000 and has long been one of the largest artist com­mu­ni­ties on the web.

As shown by Simon Willi­son and Andy Baio, thou­sands---and prob­a­bly closer to mil­lions---of images in LAION were copied from DeviantArt and used to train Sta­ble Dif­fu­sion.

Rather than stand up for its com­mu­nity of artists by pro­tect­ing them against AI train­ing, DeviantArt instead chose to release DreamUp, a paid app built around Sta­ble Dif­fu­sion. In turn, a flood of AI-gen­er­ated art has inun­dated DeviantArt, crowd­ing out human artists.

When con­fronted about the ethics and legal­ity of these maneu­vers dur­ing a live Q&A ses­sion in Novem­ber 2022, mem­bers of the DeviantArt man­age­ment team, includ­ing CEO Moti Levy, could not explain why they betrayed their artist com­mu­nity by embrac­ing Sta­ble Dif­fu­sion, while inten­tion­ally vio­lat­ing their own terms of ser­vice and pri­vacy pol­icy.

Mid­jour­ney

Mid­jour­ney was founded in 2021 by David Holz in San Fran­cisco. Mid­jour­ney offers a text-to-image gen­er­a­tor through Dis­cord and a web app.

Though hold­ing itself out as a "research lab", Mid­jour­ney has cul­ti­vated a large audi­ence of pay­ing cus­tomers who use Mid­jour­ney's image gen­er­a­tor pro­fes­sion­ally. Holz has said he wants Mid­jour­ney to be "focused toward mak­ing every­thing beau­ti­ful and artis­tic look­ing."

To that end, Holz has admit­ted that Mid­jour­ney is trained on "a big scrape of the inter­net". Though when asked about the ethics of mas­sive copy­ing of train­ing images, he said---

There are no laws specif­i­cally about that.

And when Holz was fur­ther asked about allow­ing artists to opt out of train­ing, he said---

We're look­ing at that. The chal­lenge now is find­ing out what the rules are.

We look for­ward to help­ing Mr. Holz find out about the many state and fed­eral laws that pro­tect artists and their work.

The plain­tiffs


Our plain­tiffs are won­der­ful, accom­plished artists who have stepped for­ward to rep­re­sent a class of thou­sands---pos­si­bly mil­lions---of fel­low artists affected by gen­er­a­tive AI.

https://i.rdrama.net/images/16841356346753697.webp

https://stablediffusionlitigation.com/

25
Jump in the discussion.

No email address required.

>Sarah Andersen

Proving once again that the artists who care the most about "muh copyright" are also the laziest and least talented.

:marseygoodnight::marseygoodnight::marseygoodnight:

Jump in the discussion.

No email address required.

AI turning racist at every opportunity makes much more sense now.

![](/images/16737099758760505.webp)

edit - LOL

>Fangs

>About The Book

>A New York Times bestselling love story between a vampire and a werewolf by the creator of the enormously popular Sarah's Scribbles comics.

Should call it 'Sarah's Derivatives'.

Jump in the discussion.

No email address required.



Now playing: Da Banana Bunch (DK64).mp3

Link copied to clipboard
Action successful!
Error, please refresh the page and try again.