From 6fdfb51a49271fba40ebfe86b19d905389f20795 Mon Sep 17 00:00:00 2001 From: Karl Hudgell Date: Sat, 12 Apr 2025 10:17:24 +0100 Subject: [PATCH] support multiple models and random selection --- lib.py | 3 ++- user_config.cfg.sample | 7 ++++--- 2 files changed, 6 insertions(+), 4 deletions(-) diff --git a/lib.py b/lib.py index d532226..45299db 100644 --- a/lib.py +++ b/lib.py @@ -70,9 +70,10 @@ def rename_image() -> str | None: def create_prompt_on_openwebui(prompt: str) -> str: """Sends prompt to OpenWebui and returns the generated response.""" + model = random.choice(user_config["openwebui"]["models"].split(",")) response = litellm.completion( api_base=user_config["openwebui"]["base_url"], - model="openai/" + user_config["openwebui"]["model"], + model="openai/" + model, messages=[ { "role": "system", diff --git a/user_config.cfg.sample b/user_config.cfg.sample index 997e5d5..3005a98 100644 --- a/user_config.cfg.sample +++ b/user_config.cfg.sample @@ -6,12 +6,13 @@ port = 5000 [comfyui] comfyui_url = http://comfyui -models = zavychromaxl_v100.safetensors +models = zavychromaxl_v100.safetensors,ponyDiffusionV6XL_v6StartWithThisOne.safetensors output_dir = ./output/ -prompt = "Be explicit, only return the prompt and no other text, Generate a random detailed prompt for stable diffusion." +prompt = "Generate a random detailed prompt for stable diffusion." width = 1568 height = 672 [openwebui] base_url = https://openwebui -api_key = sk- \ No newline at end of file +api_key = sk- +models = llama3:latest,cogito:14b,gemma3:12b \ No newline at end of file