Token dropout/limits #250
Replies: 1 comment
-
This is in fact a model limit. CLIP is limited to 75 tokens. Any solution you see using the exact same model is a hack. Only changing the text encoder truly solves this. The only model that isn’t limited to 75 is PixArt Sigma (it’s 120). |
Beta Was this translation helpful? Give feedback.
-
I have some captions with over 1000 tokens so that I be extremely specific in my prompts and get highly detailed results (same reason I am training at 1536-2048). I was wondering about this though: I know that stable diffusion is always trained on ~75 tokens, and its not a model limit but a training issue for prompts to not be higher quality with more descriptions.
there is the option for "keep tag count" which just doesnt shuffle the first n tags when shuffle is enabled (based on description), but does the tool do anything else automatically to the caption? does it drop the end of the prompt? Can it be made to where the prompt length is randomized, ie: sometimes 75 tokens, sometimes 150, 225, 300, etc for a similar reason to the resolution override?
Beta Was this translation helpful? Give feedback.
All reactions