William Gill c07e962cae Image tools: migrate to OWUI 0.9.0 async model accessors
Open WebUI 0.9.0 made every model-class accessor (Users.get_user_by_id,
Chats.get_chat_by_id, Files.get_file_by_id, …) a coroutine. Both tools
were still calling them synchronously, so the calls returned coroutines
instead of model objects; the first downstream attribute access threw,
the bare `except Exception: return False` swallowed it, and uploads
silently fell through to the data-URI fallback. The data-URI markdown
rendered during streaming but didn't survive post-stream commit, which
looked like "image flashes in, then disappears."

Add await to the six call sites; promote `_read_file_dict` to async
since it now contains an await; restore `_push_image_to_chat` to the
canonical `files` event so the file-attachment chrome (thumbnail +
download) comes back.

This supersedes commit d034700, which mis-diagnosed the symptom as a
virtualization regression and switched to a `message`-event markdown
workaround. The workaround didn't help (same flash-and-vanish) because
the upload pre-check still failed for the same async-migration reason
and the data-URI fallback path still ran.

smart_image_gen.py 0.7.9 -> 0.7.10
smart_image_pipe.py 0.1.1 -> 0.1.2

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-26 06:16:02 -05:00

comfyui-nvidia

ComfyUI image-generation backend, NVIDIA-accelerated, fronted by Open WebUI for multi-user chat and image generation/editing.

Built from the official ComfyUI manual install for NVIDIA — no third-party base image. CI publishes the image to git.anomalous.dev/alphacentri/comfyui-nvidia on every v* tag (see .gitea/workflows/release.yml).

Repository layout

Path What
Dockerfile ComfyUI on NVIDIA, manual-install pattern
workflows/ txt2img + img2img workflow JSONs and node mappings
deployments/ai-stack/ The deployment — compose, Caddyfile, env, model preseed
.gitea/workflows/ Release pipeline (build & push image on tag)

Deploy

The full stack — Caddy + Ollama + ComfyUI + Open WebUI (+ optional Anubis) — lives under deployments/ai-stack/. Bring-up steps, host prerequisites, Open WebUI workflow wiring, and gotchas are in deployments/ai-stack/README.md.

Replaces

This repo supersedes the previous figment + segment + Forge stack. ComfyUI's node graph covers everything those services provided (txt2img, img2img, inpaint, mask generation via SAM/GroundingDINO custom nodes), and Open WebUI talks to it natively.

Description
No description provided
Readme 553 KiB
Languages
Python 84%
Shell 11.7%
Dockerfile 4.3%