The entire training set isn’t used in each permutation. Your keywords are building the samples based on metadata tags tied back to the original images.
If you ask for “Iron Man in a cowboy hat”, the toolset will reach for some catalog of Iron Man images and some catalog of cowboy hat images and some catalog of person-in-cowboy-hat images, when looking for a basis of comparison as it renders the image.
These would be the images attributed to the output.
The entire training set isn’t used in each permutation. Your keywords are building the samples based on metadata tags tied back to the original images.
If you ask for “Iron Man in a cowboy hat”, the toolset will reach for some catalog of Iron Man images and some catalog of cowboy hat images and some catalog of person-in-cowboy-hat images, when looking for a basis of comparison as it renders the image.
These would be the images attributed to the output.