Currently trying to fine-tune SD1.5 to use patches from CLIP ViT/14 as the prompt rather than text tokens, using about 2k SD images I generated in October as the dataset. Should allow things like prompting with an image pair to get something in between. Eventual multi-modal (text + images as prompts) would be nice. #AIArt #StableDiffusion #machinelearning #AI
@nichg like this model here? https://huggingface.co/lambdalabs/sd-image-variations-diffusers
@GaggiX exactly like that I guess 😅
@nichg I don't think that 2k images are nearly enough for this job
@GaggiX interesting if that's the case, since Dreambooth seems to work so well with only 10-100. What if one just trained the projection layer and not the UNet?
@nichg with Dreambooth you're not trying to shift the entire conditioning distribution
@GaggiX Well anyhow, this probably saved me three months of training time, so thanks!
@nichg I don't understand what are you trying to do, I thought it was something like Midjourney when it's conditioned on two or more images
@GaggiX Broadly, if fine-tuning to new modalities is easy, you could have a multi-modality model where you choose what to drop in and what to remove. Use the ViT tokens except in an area to inpaint. Combine with audio because why not. Have some text be descriptive, other be seen as tags, one image for a mask, one for a depth map, segmentation map, and a texture reference, etc.
@nichg I still have no idea what you are trying to do but I guess good luck ahah
@nichg @GaggiX does the cross attention operation accept arbitrary length sequences? Sounds like a really cool idea.
Would be especially cool if you could do something like LoRA on the cross attention weights to separately fine-tune different conditioning modalities and then merge the ones you need at inference time.
@GaggiX Yeah but that means something different. Interpolating between two vectors gives you the concept halfway between the two. But if you use cross-attention over a token set and add additional tokens, its more like an 'and'.