274 questions
-1
votes
0
answers
45
views
How to use Stable Diffusion WebUI with ControlNet for pose-to-image generation (img2img)? [closed]
I'm trying to generate images using Stable Diffusion WebUI with ControlNet (pose) via img2img mode, and I want the output image to follow the pose of a reference skeleton image.
Here’s what I have so ...
-3
votes
0
answers
26
views
T2I Adapter with SDXL produces only black images when conditioning on COCO-WholeBody skeletons (no error, loss decreasing) [closed]
I'm trying to fine-tune a T2I Adapter (full_adapter_xl) on COCO-WholeBody skeleton images, using Hugging Face Diffusers 0.33.0.dev0 + Stable Diffusion XL (stabilityai/stable-diffusion-xl-base-1.0).
...
0
votes
0
answers
18
views
Stable Diffusion v1.4 PTQ on both weight and activation
I'm currently working on quantizing the Stable Diffusion v1.4 checkpoint without relying on external libraries such as torch.quantization or other quantization toolkits. I’m exploring two scenarios:
...
-1
votes
1
answer
52
views
Deconstructiong the Stable Diffusion 3.5 pipeline
I am trying to deconstruct the SD3.5 (specifically 3.5 medium) pipeline in order to have a controlled process over the denoising steps. I can't do callbacks because I need to modify the latent ...
0
votes
0
answers
125
views
where does ollama get the parameters to run the model with automatic1111
I tried to start image generation via chat in OpenWebUI on AUTOMATIC1111 (Schedule type: Кarras, Sampling method: DPM++ 2M, CFG Scale: 5.5, W:768, H:768, Sampling Steps: 20), and after ~20 seconds I ...
0
votes
1
answer
148
views
Running Stable Diffusion locally
I've been trying to follow the instructions here to run StableDiffusion locally, but the code appears to just hang.
I've cloned the repo, and installed dependencies:
$ git show --stat
commit ...
1
vote
2
answers
131
views
How can I resolve "RuntimeError: The size of tensor a (64) must match the size of tensor b (6) at non-singleton dimension 1"?
I am training a diffusion model for an image-to-image translation task. Image size is 128 by 128, and batch size is 8 (due to memory limitations).
This is the model I am using:
class ResidualBlock(nn....
0
votes
0
answers
43
views
Stable Diffusion - small dataset
I am trying to generate a bigger dataset of sattelite images from Google Earth Engine or Sentinel-2 data depicting aggregated litter.
For this purpose I have found that I can use a GAN and stable ...
0
votes
0
answers
46
views
OutOfMemoryError: CUDA out of memory (StableDiffusionInpaintPipeline)
The error explains:
Tried to allocate 8.00 GiB. GPU 0 has a total
capacity of 39.56 GiB of which 3.94 GiB is free. Process 3162 has
35.61 GiB memory in use.
Runtime: A-100, High RAM (best option to ...
0
votes
1
answer
335
views
How to show dedicated progress bar in each tab in a Gradio app?
I am developing an image generation Gradio app that uses multiple models like SD3.5, Flux, and others to generate images from a given prompt.
The app has 7 tabs, each corresponding to a specific model....
1
vote
0
answers
43
views
error mat1 and mat2 shapes cannot be multiplied (4x256 and 768x1280) while fine tuning stable video diffusion
I want to fine tune a stable diffusion model with my dataset. Getting error on line output = lora_unet
model used: stable-video-diffusion-img2vid-xt-1-1
Tried some code fixes given by chatgpt but no ...
0
votes
1
answer
429
views
How to find corresponding node in a comfyUI custom node?
I'm trying to write a comfyUI custom node, when dealing with the frontend part, in the callback async nodeCreated(node), console.log(node.id) just print -1. Or to avoid XYZ problem, how am I supposed ...
0
votes
0
answers
125
views
Diffusers pipeline Instant ID with Ipadapter
I want to use an implementation of InstantID with Ipadapter using Diffusers library.
So far I got :
import diffusers
from diffusers.utils import load_image
from diffusers.models import ControlNetModel
...
0
votes
0
answers
73
views
Differences in no of ResNet blocks in up blocks and no of channels for Unet2D model of diffusers
I have been reading about Unets and Stable diffusion and want to train one. I understand the original architecture for unets and how its channels, height and width evolve over down blocks and up ...
0
votes
0
answers
104
views
How to run Stable Diffusion 3.5 Large on 4090?
I have 4090 and this code:
login(HF_TOKEN)
self.pipe = StableDiffusion3Pipeline.from_pretrained("stabilityai/stable-diffusion-3.5-large", torch_dtype=torch.bfloat16)
self.pipe = ...