this post was submitted on 21 Sep 2023
154 points (91.0% liked)

Technology

59174 readers
2125 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

On Wednesday, OpenAI announced DALL-E 3, the latest version of its AI image synthesis model that features full integration with ChatGPT. DALL-E 3 renders images by closely following complex descriptions and handling in-image text generation (such as labels and signs), which challenged earlier models. Currently in research preview, it will be available to ChatGPT Plus and Enterprise customers in early October.

Like its predecessor, DALLE-3 is a text-to-image generator that creates novel images based on written descriptions called prompts. Although OpenAI released no technical details about DALL-E 3, the AI model at the heart of previous versions of DALL-E was trained on millions of images created by human artists and photographers, some of them licensed from stock websites like Shutterstock. It's likely DALL-E 3 follows this same formula, but with new training techniques and more computational training time.

Judging by the samples provided by OpenAI on its promotional blog, DALL-E 3 appears to be a radically more capable image synthesis model than anything else available in terms of following prompts. While OpenAI's examples have been cherry-picked for their effectiveness, they appear to follow the prompt instructions faithfully and convincingly render objects with minimal deformations. Compared to DALL-E 2, OpenAI says that DALL-E 3 refines small details like hands more effectively, creating engaging images by default with "no hacks or prompt engineering required."

you are viewing a single comment's thread
view the rest of the comments
[–] lloram239@feddit.de 1 points 1 year ago* (last edited 1 year ago) (1 children)

And you quickly realize that when you generate things from similar prompts over and over the model gives you the same results but slightly adjusted.

That's quite true, however it's worth keeping in mind that this is largely not due to a limit in the model itself, but a limit how that model interfaces with the human. Text just isn't very good to get specific results, especially not when it lacks the incremental refinement that you can do in ChatGPT with follow up prompts.

On the other side if I take StableDiffusion with ControlNet, instead of just a text prompt, I can generate far more specific results, as I can feed other images and sketches into the generation.

but I think once the new toy factor wears off people will realize they aren’t as good as they seem.

Quite the opposite, there is a ton of hidden potential still left to uncover. We have barely even started to train them on video or 3D data, integration of image models with newer language models is also a work in progress and integration into old-school image manipulation tools has just began as well.

Worth keeping in mind that Dalle-1 isn't even three years old. We are basically still in the Atari2600 days of image generation.

Meanwhile Dalle-3 comes along and can produce this level of quality with a complete generic prompt: "A fan-art of Guardians of the Galaxy Vol. 3" on the first try.

I think the next “revolution” in art is going to be having human art as a selling point

The big problem for artists is that AI art drives the value of art down to zero. It'll be hard to convince anybody to pay hundred of dollars for something when AI can produce something similar in 30sec for free. Worse yet, AI can take any existing image and remix it. The whole idea of a singular static images feels quite restrictive once you played around with AI art for a while, as everything is just a few clicks away from being something different.

I think the idea of AI art as just generators for stock images doesn't capture the magnitude of the changes that are coming. We are straight up heading into Holodeck territory where you tell the computer what you want and you get it. The AI generators won't be a tool for the artists, but go right to the users. There won't be an static image that comes out the other end, the AI will be the medium of media consumption. Just like people today can flip through TikTok, future people will flip through a AI generated stream of content custom made for them.

Wanna play some 2D game with snow and ice? Tell the computer, a couple seconds waiting, and boom here it is. First try. Want lava instead? Done. How about an N64 game? How do you compete with that as a human when AI can pull that out of thin air in seconds?

[–] MysticKetchup@lemmy.world 1 points 1 year ago (1 children)

Nah that Guardians of the Galaxy art is exactly what I'm talking about. It makes basic mistakes even a child could point out and looks more long a knockoff. And refining it is just rolling the dice to get a better result, whereas an artist you can actually give feedback they can understand.

The game assets look a little better, but if you look carefully you'll notice that they don't tile correctly. It's 90% there but the last 10% is the hardest part and it's important especially for large projects and not just single static images. Not too mention they look generic as fuck, you're not going to get the next Hollow Knight or Darkest Dungeon with an amazing original style from AI, you're only going to get existing styles mashed together. The more specific the vision for the artstyle the harder it will be to generate it.

Also the idea of a Tiktok feed of AI generated content is exactly why I hate AI art. Sure, go ahead and use it to help existing artists generate cheap assets that would otherwise be random brush strokes. But replacing them? The idea that AI generated slop will have anything close to the quality and meaning of even cheap art is ridiculous. Why would anyone want that when they could have actual art made by real people, more of which exists today than anyone could go through in their entire life?

[–] lloram239@feddit.de 1 points 1 year ago* (last edited 1 year ago)

The idea that AI generated slop will have anything close to the quality and meaning of even cheap art is ridiculous.

You are missing the bigger picture: This took SECONDS, no effort on my part and it was a first try, using technology that was a little less than three years ago at this stage. I can generate new images on any topic I want, instantly. This stuff is already incredible today and is getting better rapidly.

Meanwhile here are examples of glorious human art:

Human art is full of mistakes. The best of the best human art has "quality and meaning", the average not really. Stuff like "Somehow, Palpatine returned" was written by humans. There is a lot of garbage that slips through, even in project that have so much money that there is really no excuse. I'll take a few additional AI generated fingers, that are trivial to fix, over that trash.

Here some of the box art recreated with AI, again zero effort, first try: https://imgur.com/a/kHcwv4j

And you can remix it at will: https://i.imgur.com/y38UPX6.jpg

Also the idea of a Tiktok feed of AI generated content is exactly why I hate AI art.

Netflix is already running personalized thumbnails, not with AI, but that's exactly the kind of stuff I expect AI to be used for real soon, if it isn't already in some capacity.

Why would anyone want that when they could have actual art made by real people, more of which exists today than anyone could go through in their entire life?

Nobody cares about who makes the art outside of some art historians. Every movie, TV show or game has dozens or even hundreds of people involved, you have no idea who was responsible for what or what was going on behind the scenes. All you see is the result and you either like it or you don't. "The Death of the Author" and all that.