Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -60,12 +60,17 @@ def fill_image(prompt, image, model_selection, paste_back):
|
|
60 |
|
61 |
pipe = get_pipeline(model_selection)
|
62 |
|
|
|
63 |
(
|
64 |
prompt_embeds,
|
65 |
negative_prompt_embeds,
|
66 |
pooled_prompt_embeds,
|
67 |
negative_pooled_prompt_embeds,
|
68 |
) = pipe.encode_prompt(prompt, "cuda", True)
|
|
|
|
|
|
|
|
|
69 |
|
70 |
source = image["background"]
|
71 |
mask = image["layers"][0]
|
@@ -251,12 +256,17 @@ def infer(image, width, height, overlap_percentage, num_inference_steps, resize_
|
|
251 |
final_prompt = f"{prompt_input}, high quality, 4k"
|
252 |
|
253 |
pipe = get_pipeline(model_selection)
|
|
|
254 |
(
|
255 |
prompt_embeds,
|
256 |
negative_prompt_embeds,
|
257 |
pooled_prompt_embeds,
|
258 |
negative_pooled_prompt_embeds,
|
259 |
) = pipe.encode_prompt(final_prompt, "cuda", True)
|
|
|
|
|
|
|
|
|
260 |
|
261 |
for output_image in pipe(
|
262 |
prompt_embeds=prompt_embeds,
|
|
|
60 |
|
61 |
pipe = get_pipeline(model_selection)
|
62 |
|
63 |
+
# Encode the prompt and ensure the embeddings are in float16
|
64 |
(
|
65 |
prompt_embeds,
|
66 |
negative_prompt_embeds,
|
67 |
pooled_prompt_embeds,
|
68 |
negative_pooled_prompt_embeds,
|
69 |
) = pipe.encode_prompt(prompt, "cuda", True)
|
70 |
+
prompt_embeds = prompt_embeds.to(torch.float16)
|
71 |
+
negative_prompt_embeds = negative_prompt_embeds.to(torch.float16)
|
72 |
+
pooled_prompt_embeds = pooled_prompt_embeds.to(torch.float16)
|
73 |
+
negative_pooled_prompt_embeds = negative_pooled_prompt_embeds.to(torch.float16)
|
74 |
|
75 |
source = image["background"]
|
76 |
mask = image["layers"][0]
|
|
|
256 |
final_prompt = f"{prompt_input}, high quality, 4k"
|
257 |
|
258 |
pipe = get_pipeline(model_selection)
|
259 |
+
# Encode the prompt and ensure the embeddings are in float16
|
260 |
(
|
261 |
prompt_embeds,
|
262 |
negative_prompt_embeds,
|
263 |
pooled_prompt_embeds,
|
264 |
negative_pooled_prompt_embeds,
|
265 |
) = pipe.encode_prompt(final_prompt, "cuda", True)
|
266 |
+
prompt_embeds = prompt_embeds.to(torch.float16)
|
267 |
+
negative_prompt_embeds = negative_prompt_embeds.to(torch.float16)
|
268 |
+
pooled_prompt_embeds = pooled_prompt_embeds.to(torch.float16)
|
269 |
+
negative_pooled_prompt_embeds = negative_pooled_prompt_embeds.to(torch.float16)
|
270 |
|
271 |
for output_image in pipe(
|
272 |
prompt_embeds=prompt_embeds,
|