Alternate Subtraction Method, Faster #14
Replies: 4 comments
-
|
Beta Was this translation helpful? Give feedback.
-
Ha! Whoops, I was so focused on trying to do something involving the tendency for CLIP to label an image with a face as "a photo of a human face" with a higher score than "a photo of a human face" that I done went and did 2*enc1-enc2, shit. Back to the drawing board. |
Beta Was this translation helpful? Give feedback.
-
regarding preliminary text subtraction |
Beta Was this translation helpful? Give feedback.
-
just to ensure - i've tried direct subtraction method on a few meaningful sentences, and it predictably went totally aside of main topic. and just to make it clear - encoded embeddings are NOT losses, their summation/subtraction have different impact. |
Beta Was this translation helpful? Give feedback.
-
I was trying out ways of manipulating the encoded text and one that I tried was subtracting encoded text from the encoded text prompt. I tried four renders for each and they look about the same, except the one that changes the encoded text had less of the subtract prompt which suggests to me that it's more effective at subtracting a prompt. Also it ends up using just the one txt_enc rather than 2, and just the one cosine similarity.
Prompt: "a photo of a human face" and Negative: "a photo of a face"
Subtracting Subtract's txt_enc0 from text_enc resulted in these
Existing negative method what uses cosine similarity with the image and negative prompt for loss resulted in these
And for fun, using subtract to increase the difference between the two by txt_enc + (txt_enc - text_enc0) resulted in these
The encoded text and images seem to be explorable like latent space.
Beta Was this translation helpful? Give feedback.
All reactions