Are you aware that this is how human artists work, too? They make countless works that aren't comparable to their top works. Even more so if you count the practice when they were just starting out. I think some people just refuse to believe that real art can be generated without humans and selectively look for things that confirm their pre-determined conclusion. Witnessing this always feels like witnessing someone before the wheel was invented, when things were rolled and logs, assuring everyone that technology will never beat manual labor because look at how cumbersome it is to work with logs.
There is a crucial difference: the human artist does this choosing themselves.
If DALL-E just spits out 1000 images and then a human goes through them and picks the best 2-3, and those are good - it's impressive, but the human was still a crucial part of the process. On the other hand, if DALL-E were to generate 10 billion images, and choose the best 2-3 itself and give those as output, and if at least one of those 2-3 would be consistently great, then DALL-E could be indeed considered to be creating (good) art.
> On the other hand, if DALL-E were to generate 10 billion images, and choose the best 2-3 itself and give those as output, and if at least one of those 2-3 would be consistently great, then DALL-E could be indeed considered to be creating (good) art.
It's worth noting that the OpenAI samples for DALL-E 1 used CLIP to rank generated samples, and got a big boost from that. For many model architectures, you can run them in reverse to do 'image -> caption', and 'score the caption' quality: if 'the caption is bad', that indicates your image was screwed up and low-quality (introduced by Cogview). DALL-E 2 doesn't use either approach, or finetuning on user choices like InstructGPT, and I dunno if OA is going to implement any of these, but there is a wide universe of techniques applicable here to improve quality and we should keep that in mind (https://www.gwern.net/Forking-Paths) if we are going to make any assertions more sweeping than "this specific model, at this very instant, with this particular interface, is only at this level of quality".
I disagree. The human artist's tastes at least partially originate in other people, both individuals and general societies/cultures, and oftentimes the artist directly incorporates feedback into future work. Are you aware that students in art school, music conservatories, etc constantly get feedback from instructors and peers? I reject your premise entirely unless you can give me an example of a human that created art without ever having been influenced as a human being by any other human being. Otherwise I believe it's just what I said before: concluding first that AI can't create art and finding reasons second.
Does DALL-E incorporate (or even receive) feedback about which of the pictures it generated were better? It of course does not, and it currently has no function to do so. IF it incorporated this feedback and changed its weights based on it, I would agree with you that the situation could be comparable.
Until then, my point remains: DALL-E is currently like an (extraordinarily good) hat that you can put words in and extract phrases out of. A human chooses what words to put in and which of the phrases they take out are better. Unlike pulling words out of a hat, the network has some criteria by which it produces phrases, but that's not enough to call it an artist.
This is not meant to minimize how good the achievement of this network is. The level of fidelity and even understanding of the prompts is extraordinary. But its purpose is not to be creative, it is to find a point on a hyperplane that matches the input it received. It is currently at the level of a tool - though there are potential advancements that could yet turn it into an artist in its own right.
We're basically just arguing semantics at this point. You define art/artist differently than I do. We won't make further progress with this discussion, but regardless it's great to see that technology is coming so far
Sure, it might. But until then, I wouldn't say it makes sense to consider the AI as "being the artist".
When DALL-E x.0 does that, and when it also generates similar quality from much higher-level prompts ("paint a sad picture", or "social commentary on BLM" or something like this, instead of a description of what the picture should show and in what style), then I for one will be in complete agreement that it's indeed an artist in itself.
Personally, I don't expect this to happen in the next few decades, as I don't think the current approaches are very promising for the type of intelligence that you would need to actually do this type of reasoning, but that remains to be seen, and I am fully confident that it will happen some day.