Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 5 additions & 2 deletions prompt_enhancer_nodes.py
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ def __init__(
+ 1073741824
)

def forward(self, prompt, image_conditioning, max_resulting_tokens):
def forward(self, prompt, image_conditioning, max_resulting_tokens, seed=-1):
enhanced_prompt = generate_cinematic_prompt(
self.image_caption_model,
self.image_caption_processor,
Expand All @@ -47,6 +47,7 @@ def forward(self, prompt, image_conditioning, max_resulting_tokens):
prompt,
image_conditioning,
max_new_tokens=max_resulting_tokens,
seed=seed
)

return enhanced_prompt
Expand Down Expand Up @@ -172,6 +173,7 @@ def INPUT_TYPES(s):
},
"optional": {
"image_prompt": ("IMAGE",),
"seed": ("INT", {"default": 0, "min": 0, "max": 0xffffffffffffffff}),
},
}

Expand All @@ -192,6 +194,7 @@ def enhance(
prompt_enhancer: comfy.model_patcher.ModelPatcher,
image_prompt: torch.Tensor = None,
max_resulting_tokens=256,
seed=-1,
):
comfy.model_management.free_memory(
prompt_enhancer.memory_required([]),
Expand All @@ -204,5 +207,5 @@ def enhance(
permuted_image = image_prompt.permute(3, 0, 1, 2)[None, :]
image_conditioning = [(permuted_image, 0, 1.0)]

enhanced_prompt = model(prompt, image_conditioning, max_resulting_tokens)
enhanced_prompt = model(prompt, image_conditioning, max_resulting_tokens, seed)
return (enhanced_prompt[0],)
5 changes: 5 additions & 0 deletions prompt_enhancer_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,12 @@ def generate_cinematic_prompt(
prompt: Union[str, List[str]],
conditioning_items: Optional[List[Tuple[torch.Tensor, int, float]]] = None,
max_new_tokens: int = 256,
seed: int = -1,
) -> List[str]:
from transformers import set_seed
if seed > -1:
set_seed(seed)

prompts = [prompt] if isinstance(prompt, str) else prompt

if conditioning_items is None:
Expand Down