mirror of
https://github.com/karl0ss/ai_image_frame_server.git
synced 2025-08-05 09:58:28 +01:00
working openrouter and better create_image page formatting
This commit is contained in:
parent
76e33ea523
commit
bd1bb98160
1
.gitignore
vendored
1
.gitignore
vendored
@ -11,3 +11,4 @@ publish.sh
|
|||||||
test.py
|
test.py
|
||||||
.vscode/launch.json
|
.vscode/launch.json
|
||||||
favourites.json
|
favourites.json
|
||||||
|
.vscode/launch.json
|
||||||
|
2
.vscode/launch.json
vendored
2
.vscode/launch.json
vendored
@ -8,7 +8,7 @@
|
|||||||
"name": "Python Debugger: Current File",
|
"name": "Python Debugger: Current File",
|
||||||
"type": "debugpy",
|
"type": "debugpy",
|
||||||
"request": "launch",
|
"request": "launch",
|
||||||
"program": "${file}",
|
"program": "ai_frame_image_server.py",
|
||||||
"console": "integratedTerminal",
|
"console": "integratedTerminal",
|
||||||
"justMyCode": false,
|
"justMyCode": false,
|
||||||
"env": {"SECRET_KEY":"dkdkdk"}
|
"env": {"SECRET_KEY":"dkdkdk"}
|
||||||
|
@ -123,14 +123,14 @@ def select_model(model: str) -> tuple[str, str]:
|
|||||||
use_flux = json.loads(user_config["comfyui"].get("FLUX", "false").lower())
|
use_flux = json.loads(user_config["comfyui"].get("FLUX", "false").lower())
|
||||||
only_flux = json.loads(user_config["comfyui"].get("ONLY_FLUX", "false").lower())
|
only_flux = json.loads(user_config["comfyui"].get("ONLY_FLUX", "false").lower())
|
||||||
|
|
||||||
if model == "Random":
|
if model == "Random Image Model":
|
||||||
selected_workflow = "FLUX" if (use_flux and (only_flux or random.choice([True, False]))) else "SDXL"
|
selected_workflow = "FLUX" if (use_flux and (only_flux or random.choice([True, False]))) else "SDXL"
|
||||||
elif "flux" in model.lower():
|
elif "flux" in model.lower():
|
||||||
selected_workflow = "FLUX"
|
selected_workflow = "FLUX"
|
||||||
else:
|
else:
|
||||||
selected_workflow = "SDXL"
|
selected_workflow = "SDXL"
|
||||||
|
|
||||||
if model == "Random":
|
if model == "Random Image Model":
|
||||||
if selected_workflow == "FLUX":
|
if selected_workflow == "FLUX":
|
||||||
valid_models = user_config["comfyui:flux"]["models"].split(",")
|
valid_models = user_config["comfyui:flux"]["models"].split(",")
|
||||||
else: # SDXL
|
else: # SDXL
|
||||||
@ -141,7 +141,7 @@ def select_model(model: str) -> tuple[str, str]:
|
|||||||
return selected_workflow, model
|
return selected_workflow, model
|
||||||
|
|
||||||
|
|
||||||
def create_image(prompt: str | None = None, model: str = "Random") -> None:
|
def create_image(prompt: str | None = None, model: str = "Random Image Model") -> None:
|
||||||
"""Generate an image with a chosen workflow (Random, FLUX*, or SDXL*)."""
|
"""Generate an image with a chosen workflow (Random, FLUX*, or SDXL*)."""
|
||||||
|
|
||||||
if prompt is None:
|
if prompt is None:
|
||||||
|
@ -4,6 +4,7 @@ import logging
|
|||||||
import sys
|
import sys
|
||||||
import time
|
import time
|
||||||
import os
|
import os
|
||||||
|
import random
|
||||||
from PIL import Image
|
from PIL import Image
|
||||||
import nest_asyncio
|
import nest_asyncio
|
||||||
import json
|
import json
|
||||||
@ -130,6 +131,14 @@ def load_openrouter_models_from_config():
|
|||||||
models = config["openrouter"]["models"].split(",")
|
models = config["openrouter"]["models"].split(",")
|
||||||
return sorted([model.strip() for model in models if model.strip()], key=str.lower)
|
return sorted([model.strip() for model in models if model.strip()], key=str.lower)
|
||||||
return []
|
return []
|
||||||
|
|
||||||
|
def load_openwebui_models_from_config():
|
||||||
|
config = load_config()
|
||||||
|
if "openwebui" in config and "models" in config["openwebui"]:
|
||||||
|
models = config["openwebui"]["models"].split(",")
|
||||||
|
return sorted([model.strip() for model in models if model.strip()], key=str.lower)
|
||||||
|
return []
|
||||||
|
|
||||||
def load_prompt_models_from_config():
|
def load_prompt_models_from_config():
|
||||||
"""Load and return a list of available prompt generation models (both OpenWebUI and OpenRouter)."""
|
"""Load and return a list of available prompt generation models (both OpenWebUI and OpenRouter)."""
|
||||||
config = load_config()
|
config = load_config()
|
||||||
|
@ -13,7 +13,7 @@ LOG_FILE = "./prompts_log.jsonl"
|
|||||||
user_config = load_config()
|
user_config = load_config()
|
||||||
output_folder = user_config["comfyui"]["output_dir"]
|
output_folder = user_config["comfyui"]["output_dir"]
|
||||||
|
|
||||||
def create_prompt_on_openwebui(prompt: str, topic: str = "random") -> str:
|
def create_prompt_on_openwebui(prompt: str, topic: str = "random", model: str = None) -> str:
|
||||||
"""Sends prompt to OpenWebui and returns the generated response."""
|
"""Sends prompt to OpenWebui and returns the generated response."""
|
||||||
topic_instruction = ""
|
topic_instruction = ""
|
||||||
selected_topic = ""
|
selected_topic = ""
|
||||||
@ -40,6 +40,11 @@ def create_prompt_on_openwebui(prompt: str, topic: str = "random") -> str:
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
if model:
|
||||||
|
# Use the specified model
|
||||||
|
model = model
|
||||||
|
else:
|
||||||
|
# Select a random model
|
||||||
model = random.choice(user_config["openwebui"]["models"].split(","))
|
model = random.choice(user_config["openwebui"]["models"].split(","))
|
||||||
response = litellm.completion(
|
response = litellm.completion(
|
||||||
api_base=user_config["openwebui"]["base_url"],
|
api_base=user_config["openwebui"]["base_url"],
|
||||||
|
@ -2,7 +2,7 @@ from flask import Blueprint, request, render_template, redirect, url_for, sessio
|
|||||||
import threading
|
import threading
|
||||||
from libs.comfyui import create_image, select_model, get_available_models
|
from libs.comfyui import create_image, select_model, get_available_models
|
||||||
from libs.ollama import create_prompt_on_openwebui
|
from libs.ollama import create_prompt_on_openwebui
|
||||||
from libs.generic import load_models_from_config, load_topics_from_config, load_openrouter_models_from_config, create_prompt_with_random_model
|
from libs.generic import load_models_from_config, load_topics_from_config, load_openrouter_models_from_config, load_openwebui_models_from_config, create_prompt_with_random_model
|
||||||
import os
|
import os
|
||||||
|
|
||||||
bp = Blueprint("create_routes", __name__)
|
bp = Blueprint("create_routes", __name__)
|
||||||
@ -12,10 +12,24 @@ user_config = None # will be set in init_app
|
|||||||
def create():
|
def create():
|
||||||
if request.method == "POST":
|
if request.method == "POST":
|
||||||
prompt = request.form.get("prompt")
|
prompt = request.form.get("prompt")
|
||||||
selected_workflow, model = select_model(request.form.get("model") or "Random")
|
image_model = request.form.get("model") or "Random Image Model"
|
||||||
|
selected_workflow, model = select_model(image_model)
|
||||||
topic = request.form.get("topic")
|
topic = request.form.get("topic")
|
||||||
|
|
||||||
if not prompt:
|
if not prompt:
|
||||||
|
# Get the prompt model from the form data
|
||||||
|
prompt_model = request.form.get("prompt_model") or ""
|
||||||
|
if prompt_model and prompt_model != "Random Prompt Model":
|
||||||
|
# Use the specified prompt model
|
||||||
|
service, service_model = prompt_model.split(":", 1) if ":" in prompt_model else (prompt_model, "")
|
||||||
|
if service == "openwebui":
|
||||||
|
from libs.ollama import create_prompt_on_openwebui
|
||||||
|
prompt = create_prompt_on_openwebui(user_config["comfyui"]["prompt"], topic, service_model)
|
||||||
|
elif service == "openrouter":
|
||||||
|
from libs.openrouter import create_prompt_on_openrouter
|
||||||
|
prompt = create_prompt_on_openrouter(user_config["comfyui"]["prompt"], topic, service_model)
|
||||||
|
else:
|
||||||
|
# Use a random prompt model
|
||||||
prompt = create_prompt_with_random_model(user_config["comfyui"]["prompt"], topic)
|
prompt = create_prompt_with_random_model(user_config["comfyui"]["prompt"], topic)
|
||||||
|
|
||||||
threading.Thread(target=lambda: create_image(prompt, model)).start()
|
threading.Thread(target=lambda: create_image(prompt, model)).start()
|
||||||
@ -23,16 +37,24 @@ def create():
|
|||||||
|
|
||||||
# Load all models (SDXL and FLUX only)
|
# Load all models (SDXL and FLUX only)
|
||||||
sdxl_models, flux_models = load_models_from_config()
|
sdxl_models, flux_models = load_models_from_config()
|
||||||
|
openwebui_models = load_openwebui_models_from_config()
|
||||||
|
openrouter_models = load_openrouter_models_from_config()
|
||||||
|
|
||||||
return render_template("create_image.html",
|
return render_template("create_image.html",
|
||||||
sdxl_models=sdxl_models,
|
sdxl_models=sdxl_models,
|
||||||
flux_models=flux_models,
|
flux_models=flux_models,
|
||||||
|
openwebui_models=openwebui_models,
|
||||||
|
openrouter_models=openrouter_models,
|
||||||
topics=load_topics_from_config())
|
topics=load_topics_from_config())
|
||||||
|
|
||||||
@bp.route("/image_queued")
|
@bp.route("/image_queued")
|
||||||
def image_queued():
|
def image_queued():
|
||||||
prompt = request.args.get("prompt", "No prompt provided.")
|
prompt = request.args.get("prompt", "No prompt provided.")
|
||||||
model = request.args.get("model", "No model selected.").split(".")[0]
|
model = request.args.get("model", "No model selected.")
|
||||||
|
if model == "Random Image Model":
|
||||||
|
model = "Random"
|
||||||
|
else:
|
||||||
|
model = model.split(".")[0]
|
||||||
return render_template("image_queued.html", prompt=prompt, model=model)
|
return render_template("image_queued.html", prompt=prompt, model=model)
|
||||||
|
|
||||||
@bp.route("/create_image", methods=["GET"])
|
@bp.route("/create_image", methods=["GET"])
|
||||||
@ -42,10 +64,14 @@ def create_image_page():
|
|||||||
|
|
||||||
# Load all models (SDXL and FLUX only)
|
# Load all models (SDXL and FLUX only)
|
||||||
sdxl_models, flux_models = load_models_from_config()
|
sdxl_models, flux_models = load_models_from_config()
|
||||||
|
openwebui_models = load_openwebui_models_from_config()
|
||||||
|
openrouter_models = load_openrouter_models_from_config()
|
||||||
|
|
||||||
return render_template("create_image.html",
|
return render_template("create_image.html",
|
||||||
sdxl_models=sdxl_models,
|
sdxl_models=sdxl_models,
|
||||||
flux_models=flux_models,
|
flux_models=flux_models,
|
||||||
|
openwebui_models=openwebui_models,
|
||||||
|
openrouter_models=openrouter_models,
|
||||||
topics=load_topics_from_config())
|
topics=load_topics_from_config())
|
||||||
|
|
||||||
|
|
||||||
|
@ -33,6 +33,28 @@
|
|||||||
align-items: center;
|
align-items: center;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.model-selection {
|
||||||
|
display: flex;
|
||||||
|
flex-wrap: wrap;
|
||||||
|
gap: 20px;
|
||||||
|
justify-content: center;
|
||||||
|
margin: 20px 0;
|
||||||
|
width: 100%;
|
||||||
|
max-width: 800px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.model-group {
|
||||||
|
display: flex;
|
||||||
|
flex-direction: column;
|
||||||
|
align-items: flex-start;
|
||||||
|
gap: 5px;
|
||||||
|
}
|
||||||
|
|
||||||
|
.model-group label {
|
||||||
|
font-weight: bold;
|
||||||
|
color: #ddd;
|
||||||
|
}
|
||||||
|
|
||||||
button,
|
button,
|
||||||
select {
|
select {
|
||||||
background: #333;
|
background: #333;
|
||||||
@ -43,6 +65,7 @@
|
|||||||
font-size: 16px;
|
font-size: 16px;
|
||||||
cursor: pointer;
|
cursor: pointer;
|
||||||
transition: background 0.3s;
|
transition: background 0.3s;
|
||||||
|
min-width: 150px;
|
||||||
}
|
}
|
||||||
|
|
||||||
button:hover,
|
button:hover,
|
||||||
@ -90,6 +113,15 @@
|
|||||||
width: 100%;
|
width: 100%;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
.model-selection {
|
||||||
|
flex-direction: column;
|
||||||
|
align-items: stretch;
|
||||||
|
}
|
||||||
|
|
||||||
|
.model-group {
|
||||||
|
align-items: stretch;
|
||||||
|
}
|
||||||
|
|
||||||
button,
|
button,
|
||||||
select {
|
select {
|
||||||
width: 100%;
|
width: 100%;
|
||||||
@ -109,13 +141,15 @@
|
|||||||
|
|
||||||
<div class="button-group">
|
<div class="button-group">
|
||||||
<button onclick="showSpinner(); location.href='/'">Back</button>
|
<button onclick="showSpinner(); location.href='/'">Back</button>
|
||||||
|
|
||||||
<button onclick="sendPrompt()">Send Prompt</button>
|
<button onclick="sendPrompt()">Send Prompt</button>
|
||||||
|
|
||||||
<button onclick="randomPrompt()">Random Prompt</button>
|
<button onclick="randomPrompt()">Random Prompt</button>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="model-selection">
|
||||||
|
<div class="model-group">
|
||||||
|
<label for="model-select">Image Model:</label>
|
||||||
<select id="model-select">
|
<select id="model-select">
|
||||||
<option value="" selected>Random</option>
|
<option value="" selected>Random Image Model</option>
|
||||||
{% if flux_models %}
|
{% if flux_models %}
|
||||||
<optgroup label="FLUX">
|
<optgroup label="FLUX">
|
||||||
{% for m in flux_models %}
|
{% for m in flux_models %}
|
||||||
@ -131,7 +165,31 @@
|
|||||||
</optgroup>
|
</optgroup>
|
||||||
{% endif %}
|
{% endif %}
|
||||||
</select>
|
</select>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="model-group">
|
||||||
|
<label for="prompt-model-select">Prompt Model:</label>
|
||||||
|
<select id="prompt-model-select">
|
||||||
|
<option value="" selected>Random Prompt Model</option>
|
||||||
|
{% if openwebui_models %}
|
||||||
|
<optgroup label="OpenWebUI">
|
||||||
|
{% for m in openwebui_models %}
|
||||||
|
<option value="openwebui:{{ m }}">{{ m }}</option>
|
||||||
|
{% endfor %}
|
||||||
|
</optgroup>
|
||||||
|
{% endif %}
|
||||||
|
{% if openrouter_models %}
|
||||||
|
<optgroup label="OpenRouter">
|
||||||
|
{% for m in openrouter_models %}
|
||||||
|
<option value="openrouter:{{ m }}">{{ m }}</option>
|
||||||
|
{% endfor %}
|
||||||
|
</optgroup>
|
||||||
|
{% endif %}
|
||||||
|
</select>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="model-group">
|
||||||
|
<label for="topic-select">Topic:</label>
|
||||||
<select id="topic-select">
|
<select id="topic-select">
|
||||||
<option value="">No Topic</option>
|
<option value="">No Topic</option>
|
||||||
<option value="random">Random</option>
|
<option value="random">Random</option>
|
||||||
@ -141,6 +199,7 @@
|
|||||||
{% endfor %}
|
{% endfor %}
|
||||||
</optgroup>
|
</optgroup>
|
||||||
</select>
|
</select>
|
||||||
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
<div id="spinner-overlay">
|
<div id="spinner-overlay">
|
||||||
@ -158,10 +217,12 @@
|
|||||||
showSpinner();
|
showSpinner();
|
||||||
const prompt = document.getElementById('prompt-box').value;
|
const prompt = document.getElementById('prompt-box').value;
|
||||||
const model = document.getElementById('model-select').value;
|
const model = document.getElementById('model-select').value;
|
||||||
|
const promptModel = document.getElementById('prompt-model-select').value;
|
||||||
|
|
||||||
const formData = new URLSearchParams();
|
const formData = new URLSearchParams();
|
||||||
formData.append('prompt', prompt);
|
formData.append('prompt', prompt);
|
||||||
formData.append('model', model);
|
formData.append('model', model);
|
||||||
|
formData.append('prompt_model', promptModel);
|
||||||
|
|
||||||
fetch('/create', {
|
fetch('/create', {
|
||||||
method: 'POST',
|
method: 'POST',
|
||||||
@ -180,10 +241,12 @@
|
|||||||
function randomPrompt() {
|
function randomPrompt() {
|
||||||
showSpinner();
|
showSpinner();
|
||||||
const model = document.getElementById('model-select').value;
|
const model = document.getElementById('model-select').value;
|
||||||
|
const promptModel = document.getElementById('prompt-model-select').value;
|
||||||
const topic = document.getElementById('topic-select').value;
|
const topic = document.getElementById('topic-select').value;
|
||||||
|
|
||||||
const formData = new URLSearchParams();
|
const formData = new URLSearchParams();
|
||||||
formData.append('model', model);
|
formData.append('model', model);
|
||||||
|
formData.append('prompt_model', promptModel);
|
||||||
formData.append('topic', topic);
|
formData.append('topic', topic);
|
||||||
|
|
||||||
fetch('/create', {
|
fetch('/create', {
|
||||||
|
Loading…
x
Reference in New Issue
Block a user