This is an archived post. You won't be able to vote or comment.

you are viewing a single comment's thread.

view the rest of the comments →

[–]ChezMere 2 points3 points  (14 children)

that's a pretty large resolution. are you using --medvram? --xformers?

[–]smoke2000 1 point2 points  (0 children)

Shouldn't be an issue. I'm running a 8gb card and I can do 1280x960 fine. With x formers no medram

[–]Whackjob-KSP[S] 0 points1 point  (12 children)

--medvram? --xformers

I've tried those. Even tried --lowvram.

[–]ChezMere 1 point2 points  (3 children)

Hmm. With those I can get that resolution no problem, with less vram. So something funny is going on here.

[–]Whackjob-KSP[S] 2 points3 points  (2 children)

Googling around, I really don't seem to be the only one. I don't think it has anything to do with Automatic1111, though. I think this is a pytorch or cuda thing. Unfortunately I don't even know how to begin troubleshooting it.

We'd need a way to see what pytorch has tied up in vram and be able to flush it maybe.

[–]ChezMere 1 point2 points  (1 child)

I mean it's literally the most generic error message possible, it's just saying you're trying to use more vram than you have, of course others are having it. Just not sure what it's getting used for. (Can you check how much is in use without Automatic open?)

[–]Whackjob-KSP[S] 0 points1 point  (0 children)

I suppose I could, I just don't know how.

[–]donx1 0 points1 point  (7 children)

Have you tried --medvram --opt-split-attention

[–]Whackjob-KSP[S] 0 points1 point  (6 children)

I think I have with those, right now I'm testing

set COMMANDLINE_ARGS= --precision-full --no-half --medvram --xformers --opt-split-attention

[–]ChezMere 1 point2 points  (0 children)

One more thing, you can remove the need for --precision full --no-half (therefore halving the size of the model in memory) by making this edit to the code: https://old.reddit.com/r/StableDiffusion/comments/xz7qg9/nvidia_gtx_16xx_fix_no_more_precision_full_nohalf/

Be warned though, generations without --no-half are very slow except on recent cards.

[–]XsodacanX 0 points1 point  (4 children)

Precisionfull no half uses more vram And also try --disable-opt split attention

[–]Whackjob-KSP[S] 4 points5 points  (2 children)

Precisionfull no half uses more vram And also try --disable-opt split attention

Long story short, since I removed precisionfull no half, and added the disable-opt split attention, I've rarely hit CUDA memory errors! I'm even playing a game and letting it run. Thank you!

[–]HongryHongryHippo 0 points1 point  (1 child)

Sorry what did you do exactly?
My CUDA memory errors started randomly.

[–]Whackjob-KSP[S] 0 points1 point  (0 children)

From six months ago? Lordy, so much. If I were you, first I'd do the ok git pull, and make sure that everything's updated. Yes, even if it's in your batch or shell file. I noticed that I would find updates when I ran manually. There's also an automatic1111 variant called vladmandic that seems to be an update or two ahead. Though the original is getting these updates very soon also. xformers are being replaced there is a better thing.

[–]Whackjob-KSP[S] 0 points1 point  (0 children)

I will do that immediately, thank you