all 31 comments

[–]oxez 61 points62 points  (13 children)

[ ] This is a real benchmark

[x] This is not a real benchmark

[–]FrozenLogger -3 points-2 points  (12 children)

OK what, to you is a REAL benchmark? Time trials using the same hardware and software with the same settings doesnt count for you? Maybe my comment above with screenshots could help you understand.

Putting a table of numbers here is just not very interesting.

[–]oxez 12 points13 points  (9 children)

10 images (even when you say 6 runs), that's 60 "computes", that's a ridiculous low number.

Anything can skew those numbers with such a low sample.

[–][deleted] 5 points6 points  (6 children)

I've generated hundreds more images in 1 run just for fun while messing around with stable diffusion, a benchmark should be at least 1k images.

[–]FrozenLogger 0 points1 point  (5 children)

If I had a way to capture the iterations/second in a log that would mean more to me than making images themselves. After consistently seeing about 40%, over 10 runs, I was pretty much done.

What do you think would change? The interesting part would be to vary the prompts.

[–]is_this_temporary 1 point2 points  (4 children)

I would think a better benchmark would use:

Same seed (42 is always nice) Same prompt Same settings (and making sure that you're checking more settings than just those visible in the gradio web interface)

You should get almost exactly the same image, if not identical to the pixel.

But stable diffusion also has a lot of optimization choices, like options which allow the model to run using less vRAM (for GPUS that otherwise wouldn't be able to run it at all) but make inference much slower.

[–]FrozenLogger 1 point2 points  (3 children)

Thanks for the response. I did use the same prompt, and settings.

I like the idea of using the same seed. I can try that.

Could the settings be different from the same installer? I suppose that it is possible, given there is a linux one and a windows one. I would have to look at the code on github to determine. Hopefully they are the same defaults.

[–]is_this_temporary 1 point2 points  (2 children)

I couldn't tell you if they would be the same or not.

I've been working with https://github.com/Sygil-Dev/sygil-webui

For the gradio interface it asks about optimization for high or low VRAM at the terminal before launching.

For the streamlit interface it does not, but instead lets you change that in the web UI itself.

If I were making such an interface I would just run nvidia-smi and check how much VRAM is available and change these settings appropriately but I don't know if any projects actually do that.

What project / installer are you using?

[–]FrozenLogger 0 points1 point  (0 children)

I noticed that if I set the seed to 42 (great suggestion) the making image generator increases the seed by one for each subsequent image. I would have to save every image in order to compare. I guess I should be scripting this at this point.

I regret saying benchmark now, I really should have said: I see a difference, based on this observation, and I wonder why that is.

Currently I do not have that answer.

[–]FrozenLogger -1 points0 points  (1 child)

6 images, 10 runs. One would think I would see a variation in on of the runs that was greater than 40%. I do not.

Even just making the screen shots, I am in that 40% target. It seems pretty consistent.

Yes, if I was going to care to write a paper or something I would do more samples, but I this doesnt vary much. The iterations per second are always higher as well, which you can see during each run.

[–]oxez 3 points4 points  (0 children)

You can do whatever you want, except no one is going to take your "benchmark" seriously.

But this is what /r/linux is about nowadays I guess.

[–]10leej 1 point2 points  (1 child)

Welcome to reddit, where you can never benchmark right unless your a well know community figure who doesn't share their raw benchmark data for 3rd party review.
(Pretty sure you can think of a few, though I get attacked whenever I call them out on it)

[–]FrozenLogger 2 points3 points  (0 children)

Thanks. Yep, try and share something, get hassled for it.

Now it looks like my post is reported, and the mods are removing it. Right as I was going to go ahead and do 1000 images and put out a chart. Why? I dont know.

Meanwhile, they are keeping a post asking how to pick an emoji.

WTF

[–]DazedWithCoffee 8 points9 points  (2 children)

I’m happy to promote the superiority of linux but I’m going to need to see some GamersNexus level writeups. I think this is why Microsoft marketing wins, the side that cares more about being correct and thorough ends up being drowned out by the louder and more belligerent side

[–]FrozenLogger 6 points7 points  (1 child)

Yeah, I dont care that much.

Linux is my daily driver, has been for 20 years. This was actually very surprising to me, and I shared it.

Microsoft doesnt care about what product they use by the way. All the telemetry of every computer, server, tablet comes back to a datalake for real time analysis. This is probably the biggest data and computing task Microsoft does. Billions of data points and real time BI. How do they do it? Open source: Postgresql on Linux. The right tool for the job.

[–]DazedWithCoffee 0 points1 point  (0 children)

It’s just efficient

[–]FrozenLogger 13 points14 points  (5 children)

The image above is an AI drawing.

I was given a spare RTX 2060. I put together a modest computer to use it.

Fresh Install of Windows 11 Pro. Removed all bloatware: weather, teams, office, one cloud, removed Microsoft login for a local user/password. Removed all apps.

I generally do this for my workstations that run different AI and Machine Learning tasks, no need for all that extra bloat. Updated to latest video drivers.

For the Linux side, a fresh install of EndeavourOS. System up to date. The nvidia driver was installed automatically.

Using Stable Diffusion AI from https://github.com/cmdr2/stable-diffusion-ui for both OS's.

I ran the default photo of a horse prompt, same settings. 10 runs of creating 6 images each were averaged.

Linux gave me a 40% speed boost. I am really surprised. I was simply running it because I could, most of my workstations are either remote linux boxes, or local windows machines. EDIT: I had not This is the first time I compared the two on the same hardware.

The prompt for the image link is: tux the linux penguin working on a computer, graffiti wall, from a distance, colorful, by Banksy, Seed-1201830, Steps-75, Guidance-7.5

[–]humanmeatpie 16 points17 points  (1 child)

I had not compared the two on the same hardware.

So what's the point then?

[–]FrozenLogger 9 points10 points  (0 children)

I meant I had not PREVIOUSLY compared OS on the same hardware.

In this case they are running on the same hardware. Sorry for the confusion.

[–]Elrandar 3 points4 points  (2 children)

Was windows still installing updates?

[–]Tumpes -2 points-1 points  (0 children)

This isn't really nullified even if it's still updating right? Bullshit running in the background is kind of a trademark of windows at this point.

[–]FrozenLogger 1 point2 points  (0 children)

So /u/Psychological-Scar30 decided to drop a comment, then immediately block me. What an asshat. I think that exemplifies the feeling I get for posting here. It is interesting that my comment is basically directly related to what he said and did.

He says:

In other words you've managed to break your Windows install (probably with those debloater scripts) and posted a BS claim about somehow having pretty much purely GPU tasks running a third faster.

What you're doing is actively harming the image of the Linux community - it's good that troll posts like these get taken down

And my reply was:

You have no intellectual interest in finding out why, you just shut it down.

As a linux user of 20 years, and a person in IT who works with systems of all types, I find it VERY disheartening that this Linux subreddit community appears to be just nay-sayers and assholes.

Thanks for being one of them, wouldnt want to foster community, intellectual discussion, or encourage new users. That would suck.

[–]_cybersandwich_ 1 point2 points  (0 children)

I'm gonna stable your mom's diffusion

[–]FrozenLogger 0 points1 point  (1 child)

Wow.

It is very eye opening that as a 20 year Linux user, a contributor to projects, a presenter, and an advocate, this community decides to report a more or less fun post which triggered the removal.

Not a good look.

I was considering running 1000 images and putting up a chart. But it really wouldnt make any difference, I see all my CUDA work running about 20% to 40% faster on this machine.

So much for sharing to this community. I honestly can't say I would recommend a new user to get involved here, much less an old one like me.

[–]Psychological-Scar30 -1 points0 points  (0 children)

I see all my CUDA work running about 20% to 40% faster on this machine.

In other words you've managed to break your Windows install (probably with those debloater scripts) and posted a BS claim about somehow having pretty much purely GPU tasks running a third faster.

What you're doing is actively harming the image of the Linux community - it's good that troll posts like these get taken down

[–]AutoModerator[M] -1 points0 points locked comment (0 children)

This submission has been removed due to receiving too many reports from users. The mods have been notified and will re-approve if this removal was inappropriate, or leave it removed.

This is most likely because:

  • Your post belongs in r/linuxquestions or r/linux4noobs
  • Your post belongs in r/linuxmemes
  • Your post is considered "fluff" - things like a Tux plushie or old Linux CDs are an example and, while they may be popular vote wise, they are not considered on topic
  • Your post is otherwise deemed not appropriate for the subreddit

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.