exponentialml / comfyui_visualstyleprompting Goto Github PK
View Code? Open in Web Editor NEWComfyUI Version of "Visual Style Prompting with Swapping Self-Attention"
License: Apache License 2.0
ComfyUI Version of "Visual Style Prompting with Swapping Self-Attention"
License: Apache License 2.0
Ive been trying to generate images larger than 512 by 512 and realized that the method breaks if the reference image is 1024 by 1024, When the Image gets bigger than 752 x 1024 (not necessarily in that order or distribution of width and height) It stops working altogether and i have a feeling that the effect gets reduced the further away the reference image gets from 512 by 512. An easy way to test with a reference image of 512x512 and confirming that the output is capturing the style, then try with 752 x 1024 (captures the style somewhat strongly) then use 768 x 1024 and the style is gone pretty much completely finally a test with 1024 by 1024 confirms that the intended effect is not present at all.
Hi. I am looking forward to using VSP, the examples shown in the papers look awesome, and look indeed better than the 'competition'. However ... I can not get the expected outcome. I installed as described, via git clone, then opened the provided workflow, in which I chose model realisticvison4, loaded an image, wrote 'purple fur' in the style prompt and 'dog' in the pos prompt, expecting to get a purple fur dog. Please see the image I attached. Can you point me in the proper direction to get it working?
Greetings, Ruud.
Hi I am looking for a workflow with controlnet please.
The paper said the switching self-attention should only be applied to late upblocks of Unet to achieve the best results.
But this codes seem that the switching self-attention is applied to all stages.
I tested the updated plugin and several problems occurred:
I hope the author can notice
Is it possible to allow for longer positive prompts?
I kept getting the error:
RuntimeError: Sizes of tensors must match except in dimension 0. Expected size 77 but got size 154 for tensor number 1 in the list.
I had to keep reducing the length of my positive prompt until it was short enough to work successfully.
Hi! I am having the following error (it has worked fine before):
got prompt
Failed to validate prompt for output 15:
The input model is mutated in place, because the forward functions are replaced by monkey patching. This makes the input model unusable in alternative paths in the workflow, for example to denoise 15 steps with visual style applied and 5 steps without.
I tried to load the model twice like this for it to work, it takes 2x the memory as a result:
But it doesn't even work: comfyui ignores the second model and reuses the first one. A really dirty workaround, which has the same memory requirements as I intended above, is to add this at the start of the node fn:
model = model.clone()
model.model = copy.deepcopy(model.model)
As a rule of thumb, nodes should never mutate their input in place. This will lead to some very awkward situations and strange bugs.
If you really need to patch the forward functions and can't do it in any other way, I suggest deepcopying the shell of the model object and subobjects while reusing the weights at the leafs. Off the top of my head, you can do this by creating a second model on the meta device, and then loading the weights in from the existing model. Maybe there's an easier way.
im running an LCM model at a CFG of 1. I quickly realized that the tool does not work at all with a CFG of 1, but that makes sense due to things not being encoded fully at that value. That being said, I have noticed it works well on LCM with a CFG of 0.5. anything higher than 1 does a pretty lack luster result. 0.5 seems the best
After testing that a bit with different prompts, I found that the reference prompt does nothing at all to the output. This may be to the low CFG value, but I'm not sure
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.