Setting aside the usual arguments on the anti- and pro-AI art debate and the nature of creativity itself, perhaps the negative reaction that the Redditor encountered is part of a sea change in opinion among many people that think corporate AI platforms are exploitive and extractive in nature because their datasets rely on copyrighted material without the original artists’ permission. And that’s without getting into AI’s negative drag on the environment.

  • barsoap
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    edit-2
    3 months ago

    Yeah and there are tons of angles and gestures for human subjects that AI just can’t figure out still.

    Actually less so because it can’t draw the stuff but because it doesn’t want to on its own, and there’s no way to ask it to do anything different with built-in tools, you have to bring your own.

    Say I ask you to draw a car. You’re probably going to do a profile or 3/4th view (is that the right terminology for car portraits?), possibly a head-on, you’re utterly unlikely to draw the car from the top, or from the perspective of a mechanic lying under it.

    Combine that tendency to draw cars from a limited set of perspectives because “that’s how you draw cars” with the inability of CLIP (the language model stable diffusion uses) to understand pretty much, well, anything (it’s not a LLM), and you’ll have no chance getting the model to draw the car from a non-standard perspective.

    Throw in some other kind of conditioning, though, like a depth map, doesn’t even need to be accurate it can be very rough, the information density equivalent of me gesturing the outline of a car and a camera, and suddenly all kinds of angles are possible. Probably not under the car as the model is unlikely to know much about it, but everything else should work just fine.

    SDXL can paint, say, a man in a tuxedo doing one-hand pullups while eating a sandwich with the other. Good luck prompting that only with text, though.