this post was submitted on 01 Aug 2023
20 points (100.0% liked)

Stable Diffusion

4301 readers
5 users here now

Discuss matters related to our favourite AI Art generation technology

Also see

Other communities

founded 1 year ago
MODERATORS
 

What are the hardware requirements to run SDXL?

In particular, how much VRAM is required?

This is assuming A1111 and not using --lowvram or --medvram.

all 14 comments
sorted by: hot top controversial new old
[–] amenotef@lemmy.world 6 points 1 year ago* (last edited 1 year ago) (2 children)

I don't have A1111 but in ComfyUI using a shared workflow that does base and then refiner, SDXL 0.9 was using 12GB of VRAM and 22GB of ram in Ubuntu for me. Doing images of 1024x1024~ GPU: AMD RX 6800

[–] Scew@lemmy.world 5 points 1 year ago

Also using Comfy. Have been able to get away with 6GB of VRAM doing 1024x1024 and it took a bit longer but I've done a couple of 1024x2048's and they're coming out good :3

[–] xenla@programming.dev 4 points 1 year ago (1 children)

Also have a 6800xt, 32gb ram. SDXL 1.0 running with A1111, but I can only generate images using --medvram. This is on windows admittedly.

[–] amenotef@lemmy.world 1 points 1 year ago

Yeah in windows you don't have ROCm so it sucks.

[–] Vivarevo@sopuli.xyz 5 points 1 year ago* (last edited 1 year ago) (1 children)

3070 8gb vram, 16gb ram

In confy About 16-17sec to do everything at 1024x1024 with 20steps and 5 refiner.

A1111 About the same, using batch of 4 and then using batch img2img refining. Just more clicks without extensions etc as getting the same it/s between confy and a1111 - - medvram

[–] Thanks4Nothing@lemm.ee 3 points 1 year ago (1 children)

I am using a 3070 8gb FTW, and 32gb ddr5....I have yet to get SDXL to even generate without an error yet. I assumed it was my hardware.

[–] altima_neo@lemmy.zip 1 points 1 year ago (1 children)

Nah, auto1111 seriously struggles with SDXL for me. But comfy manages to do it without issue.

[–] Thanks4Nothing@lemm.ee 1 points 1 year ago (1 children)

I tried ComyfyUI last night - but, unless I am just missing something, I couln't get past the workflow screen. Thanks for the tip, I will keep tinkering with things.

I tried InvokeAI, but was having the same problem. I got the safetensor directly from Stability AI's hugging face page - so I have to be doing something wrong...3 different UIs and I couldn't get any to work. I am losing my technical aptitude :)

[–] altima_neo@lemmy.zip 1 points 1 year ago

The best thing with comfy is you can drag a PNG image that was generated by it and it will replicate the workflow nodes. They have some images on their GitHub for examples

[–] InterSynth@lemmy.dbzer0.com 2 points 1 year ago

I have a 6600 XT (so 8GBs of VRAM) and had no luck with A1111 or Vladmandic's port, they would crash. ComfyUI worked with no fiddling.

[–] FactorSD@lemmy.dbzer0.com 2 points 1 year ago

It's hard to give precise figures, because there's always tricks to getting a little more or less but from my (admittedly limited) testing SDXL is significantly more demanding, and 10+GB of VRAM is probably going to be the minimum to run it. I don't remember exactly what I was doing but I run on an RTX A4500 card, and I managed to max out the 20GB of VRAM just with one SDXL process, where I can normally run a LORA training and 512x768 size images at the same time.

[–] ehsanrt@lemmy.dbzer0.com 2 points 1 year ago

and im tryin to make SDXL work on my 1660ti laptop lol , comfyui runs it like 1:30 min for each pic , A1111 can’t even load the vae , however yesterday i saw and update on hugging face page of sd that they chaned to 0.9 vae for sdxl1 , seems like there was an issue with their provided 1.0 vae

[–] altima_neo@lemmy.zip 1 points 1 year ago* (last edited 1 year ago) (1 children)

I can run it on my 3080 10 gig card, but Its ridiculously slow. I HAVE to use --medvram or I get out of memory errors and NaN errors. And I mean ridiculously slow. Loading the model takes a few minutes. Generating an image requires me to minimize the browser window, or stable diffusion just stalls. Switching to the refiner isnt even an option because it takes so long to switch between models.

This is on a 5930K, 32 GB Ram, 3080 10G trying to generate 1024x1024 images.

However with comfyUI, it runs just fine, PC doesnt struggle, and it generates the images in about 40 seconds at 50 steps base, 10 refiner.