The Tool now exposes two methods the LLM picks between based on whether the user attached an image: generate_image — txt2img (existing, unchanged behavior) edit_image — img2img on the most recently attached image edit_image extracts the source image from __messages__ (base64 data URIs in image_url content blocks) or __files__ (local path or URL), uploads to ComfyUI's /upload/image, runs an img2img workflow at the caller-specified denoise (default 0.7), and returns the edited result. Same per-style routing / sampler / CFG / prefix logic as generation. Refactored the submit-and-poll loop into _submit_and_fetch shared by both methods. Image extraction is defensive — tries messages first, then files (path then URL), returns a clear "no image attached" message rather than silently generating from scratch. Image Studio system prompt rewritten to teach the LLM when to call edit_image vs generate_image and how to pick denoise. Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
comfyui-nvidia
ComfyUI image-generation backend, NVIDIA-accelerated, fronted by Open WebUI for multi-user chat and image generation/editing.
Built from the official ComfyUI manual install for
NVIDIA — no
third-party base image. CI publishes the image to
git.anomalous.dev/alphacentri/comfyui-nvidia on every v* tag (see
.gitea/workflows/release.yml).
Repository layout
| Path | What |
|---|---|
Dockerfile |
ComfyUI on NVIDIA, manual-install pattern |
workflows/ |
txt2img + img2img workflow JSONs and node mappings |
deployments/ai-stack/ |
The deployment — compose, Caddyfile, env, model preseed |
.gitea/workflows/ |
Release pipeline (build & push image on tag) |
Deploy
The full stack — Caddy + Ollama + ComfyUI + Open WebUI (+ optional
Anubis) — lives under deployments/ai-stack/.
Bring-up steps, host prerequisites, Open WebUI workflow wiring, and
gotchas are in deployments/ai-stack/README.md.
Replaces
This repo supersedes the previous figment + segment + Forge stack. ComfyUI's node graph covers everything those services provided (txt2img, img2img, inpaint, mask generation via SAM/GroundingDINO custom nodes), and Open WebUI talks to it natively.