That doesn’t make much sense, in any digital art the computer is doing all the generation of the information. The human simply provides input as all these things are tools. For digital art the tool is something like Photoshop and the inputs are a series of tablet motion inputs, for 3D art the tool is Blender and the inputs are bit more complex but similar
If you make a humanoid robot that recognizes speech, and tell it to draw Mona Lisa, it doesn’t make you an artist. So why would it be any different with a web UI that understands text?
If you put your hand on theirs and you’re the one controlling what goes on the paper, you’re the artist.
And if you let the other person finish the drawing, and then you remove the 7 extra fingers the other person drew, you’re not the artist of that drawing, you just edited art made by someone else.
Now, if you were to redo a significant chunk of the drawing, you could perhaps argue that you are the artist and the other person assisted you by giving you a rough idea to work with. Now that I think about it, I think there might be a tag for the AI version of that?
It is not like these are just totally random outputs without any artistry behind them, even a single word prompt is enough to start guiding something in an artist direction. All these methods involve iterative refinement of ideas to reach some “goal” you have in your mind, they just use different methods of doing so.
A while ago I tried generating images with nothing but random verses of this song as the prompts. No quality prompts, negatives, poses, anything. Just random lyrics from that song. And it gave me some really good looking results that had next to nothing to do with the prompts.