Menyu commited on
Commit
224e573
·
verified ·
1 Parent(s): 62031e4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -250,7 +250,7 @@ if torch.cuda.is_available():
250
  token = os.environ.get("HF_TOKEN") # 从环境变量读取令牌
251
  model_path = hf_hub_download(
252
  repo_id="Menyu/miaomiaoHarem_vPredDogma10", # 模型仓库名称(非完整URL)
253
- filename="miaomiaoHarem_vPredDogma10.safetensors",
254
  use_auth_token=token
255
  )
256
  pipe = StableDiffusionXLPipeline.from_single_file(
@@ -328,9 +328,9 @@ def infer(
328
  "height": height,
329
  "guidance_scale": guidance_scale,
330
  "num_inference_steps": num_inference_steps,
331
- "model": "miaomiaoHarem_vPredDogma10",
332
  "use_resolution_binning": use_resolution_binning,
333
- "PreUrl": "https://huggingface.co/spaces/Menyu/miaomiaoHaremDogma"
334
  }
335
  # Add metadata to the image
336
  image_with_metadata = add_metadata_to_image(image, metadata)
@@ -353,7 +353,7 @@ h1{text-align:center}
353
 
354
  with gr.Blocks(css=css) as demo:
355
  gr.Markdown("""# 梦羽的模型生成器
356
- ### 快速生成 MiaomiaoHarem vPred Dogma 模型的图片""")
357
  with gr.Group():
358
  with gr.Row():
359
  prompt = gr.Text(
 
250
  token = os.environ.get("HF_TOKEN") # 从环境变量读取令牌
251
  model_path = hf_hub_download(
252
  repo_id="Menyu/miaomiaoHarem_vPredDogma10", # 模型仓库名称(非完整URL)
253
+ filename="MiaoMiao Harem_V-Pred_dogma_1.1_FP16.safetensors",
254
  use_auth_token=token
255
  )
256
  pipe = StableDiffusionXLPipeline.from_single_file(
 
328
  "height": height,
329
  "guidance_scale": guidance_scale,
330
  "num_inference_steps": num_inference_steps,
331
+ "model": "miaomiaoHarem_vPredDogma11",
332
  "use_resolution_binning": use_resolution_binning,
333
+ "PreUrl": "https://huggingface.co/spaces/Menyu/miaomiaoHaremDogma11"
334
  }
335
  # Add metadata to the image
336
  image_with_metadata = add_metadata_to_image(image, metadata)
 
353
 
354
  with gr.Blocks(css=css) as demo:
355
  gr.Markdown("""# 梦羽的模型生成器
356
+ ### 快速生成 MiaomiaoHarem vPred Dogma 1.1 模型的图片""")
357
  with gr.Group():
358
  with gr.Row():
359
  prompt = gr.Text(