Aside from the fact that your comment applies to photography as well, I think it’s fair to point out image generation can also be a complex pipeline instead of a simple prompt.
I use ComfyUI on my own hardware and frequently include steps for control net, depth maps, canny edge detection, segmentation, loras, and more. The text prompts, both positive and negative, are the least important parts in my workflow personally.
Hell sometimes I use my own photos as one of the dozens of inputs for the workflow, so in a sense photography was included.
You would think that but it isn’t always so clear. My college had a sizable chunk, if not a majority, of foreign Chinese students and they were extremely patriotic/nationalist.
But to be fair maybe those who never left China are even more patriotic, I wouldn’t know.