William Gill f26dfbee02 smart_image_gen v0.7.2: chat-DB fallback + diagnostic 'no image' msg
If __messages__ doesn't include the assistant's prior file attachments
(which is what the screenshot is showing), the new fallback queries
the chat by id via Chats.get_chat_by_id and walks every persisted
message for files. Open WebUI's socket handler always upserts files
onto the assistant message via {'files': files} so this path is
authoritative.

The 'No image found' return now includes diagnostic counts —
__files__, __messages__, messages_with_files, chat_id_present,
openwebui_runtime — so subsequent failures actually show what the
tool saw instead of being opaque.

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-19 15:07:55 -05:00

comfyui-nvidia

ComfyUI image-generation backend, NVIDIA-accelerated, fronted by Open WebUI for multi-user chat and image generation/editing.

Built from the official ComfyUI manual install for NVIDIA — no third-party base image. CI publishes the image to git.anomalous.dev/alphacentri/comfyui-nvidia on every v* tag (see .gitea/workflows/release.yml).

Repository layout

Path What
Dockerfile ComfyUI on NVIDIA, manual-install pattern
workflows/ txt2img + img2img workflow JSONs and node mappings
deployments/ai-stack/ The deployment — compose, Caddyfile, env, model preseed
.gitea/workflows/ Release pipeline (build & push image on tag)

Deploy

The full stack — Caddy + Ollama + ComfyUI + Open WebUI (+ optional Anubis) — lives under deployments/ai-stack/. Bring-up steps, host prerequisites, Open WebUI workflow wiring, and gotchas are in deployments/ai-stack/README.md.

Replaces

This repo supersedes the previous figment + segment + Forge stack. ComfyUI's node graph covers everything those services provided (txt2img, img2img, inpaint, mask generation via SAM/GroundingDINO custom nodes), and Open WebUI talks to it natively.

Description
No description provided
Readme 553 KiB
Languages
Python 84%
Shell 11.7%
Dockerfile 4.3%