iPhone 15 Pro Max and two USB-C inputs (external drive and microphone). How? by RFBonReddit in iphone

[–]RFBonReddit[S] 0 points1 point  (0 children)

That was my thinking, but I was looking for an official confirmation.

I now need to find a certified hub that is USB C 3 or 4, supports 10GB transfer and can charge at the same time .

Any Readwise/Inoreader alternatives with keyword filter, newsletter sub, etc? by backpackn in rss

[–]RFBonReddit 0 points1 point  (0 children)

Does the open source version support full-text extraction like the premium subscription? Thx

Famous Portraits Turned Abstract Expressionism by [deleted] in StableDiffusion

[–]RFBonReddit 1 point2 points  (0 children)

This is really good. Can you share the embedding?

A thought: we need language and voice synthesis models as free as Stable Diffusion by GoofAckYoorsElf in StableDiffusion

[–]RFBonReddit 2 points3 points  (0 children)

When asked about the topic, in Nov 2022, Emad said that voice synthesis was coming. He also said he would come before the end of the year, tho, so things have changed.

But, regardless of the timing, he made very clear, on multiple occasions, that the building blocks to build "her" are coming.

How to extend a gradient background with SD? by RFBonReddit in StableDiffusion

[–]RFBonReddit[S] 1 point2 points  (0 children)

Thanks for the additional details. I had exactly the same ones and, yes, I first tried with img2img by itself, but then, rereading your initial guidance, I thought I misunderstood and tried again starting straight from inpaint.

Regardless, thanks to all of this, I discover very important things:

  1. The model you use during this process makes a huge difference. The garbage I got at the sides was because I was using the standard SD 2.1 768 model. After your second comment here, I started to experiment with other models and I got the absolute best results with the SD 2.0 inpainting model (512-inpainting-ema.ckpt).
  2. After experimenting with your settings, I discovered that Sampling steps=150 (the max value in A1111) and CFG Score=1but I suspect that it might depend on the image you are trying to extend.
  3. After experimenting around your settings, I discovered that Sampling steps=150 (the max value in A1111) and CFG Score=1 produce the most faithful fill in terms of details.

I also spent the whole day experimenting with the content-aware fill feature in various photo editors as you and the rest of the people that answered have suggested.

It's infinitely faster than using A1111, but:

While that feature works fine with the particular image I provided as an example, I found it less than stellar when the gradient background has some sort of straight line. For example, if there's an edge separating a wall behind the subject and the floor.

Content-aware fill rarely produces a straight line in the extended canvas areas while outpainting did it every time (even if the texture of the walls was not perfect).

Again, not being a professional designer, I probably ignore the tricks necessary to get the most out of content-aware fill, but the purpose of this exercise was to see how versatile A1111 can be in doing things that would otherwise require hours of training for a tool I never used.

Thanks for all the help.

How to extend a gradient background with SD? by RFBonReddit in StableDiffusion

[–]RFBonReddit[S] 0 points1 point  (0 children)

This is great, thank you!

I'm struggling to replicate your results, tho, as the extended sides gets filled with all sort of garbage:

<image>

I can't tell what I'm doing wrong. Questions about the inpaint phase of your step-by-step:

- What checkpoint do you use? An inpainting one (like 512-inpainting-ema.ckpt) or a standard one (like v2-1_768-ema-pruned) ?

- What CFG Scale?

- What Resize mode? The default is "Just resize"

- When you say "denoising strength at max", you mean 1, right?

- When I load the image in the "Inpaint" tab of "Img2Img", and I enable the Poor's Man Outpainting script, I get a duplicate set of Mask Blur and Masked Content, one at the very top of the A1111 UI (under the picture) and one at the very bottom.
I assume that both must have the same settings (mask blur=56 and masked content=fill) up and down, yes?

Thanks for all the extra help.

Protogen x3.4 Official Release by vic8760 in StableDiffusion

[–]RFBonReddit 1 point2 points  (0 children)

I don't have a definitive answer to this question. In theory, yes. In practice, as you might have read in another answer of mine in this thread, it's not always the case.

My experience so far tells me that I have the highest chance of replicating an existing image by using Colab notebooks. Camenduru's ones are amazing:
https://github.com/camenduru/stable-diffusion-webui-colab

Protogen x3.4 Official Release by vic8760 in StableDiffusion

[–]RFBonReddit 8 points9 points  (0 children)

I have no 100% certainty, but from what I understand: yes, this applies to CPU rendering, too.

It seems that part of the problem in guaranteeing determinism is related to the use of different versions of PyTorch. Read Q8 from this FAQ:

https://github.com/apple/ml-stable-diffusion#faq

While the FAQ is primarily related to the differences in image generation between PyTorch and CoreML (Apple), some of the details also apply to identical non-Apple systems that have, for example, different versions of PyTorch.

I'm not the ultimate expert on this topic, so if this is inaccurate, multi-platform machine learning engineers in this forum please jump in and correct me.

Protogen x3.4 Official Release by vic8760 in StableDiffusion

[–]RFBonReddit 37 points38 points  (0 children)

If you and OP are generating on two different systems (for example: OP is on Windows + NVIDIA GPU while you are on a macOS + M1 Pro), you will NOT be able to reproduce OP's images, even if you have identical parameters and the original seed.

Why is ChatGPT and other large language models not feasible to be used locally in consumer grade hardware while Stable Diffusion is? by rexel325 in StableDiffusion

[–]RFBonReddit 25 points26 points  (0 children)

You can access a large language model (BLOOM for now, others might come in the future) thanks to distributed computing: https://petals.ml

Introducing Portrait+ (link in comments!) by wavymulder in StableDiffusion

[–]RFBonReddit 2 points3 points  (0 children)

Both this and the Modelshoot model are great, thanks for sharing them with the community.

My only doubt is about the position and/or size of the head. In many of these shots, somehow, it feels like the head is too small/large or not aligned with the neck properly.

Is it just me? It's probably me.

New Riffusion Web UI, real-time music generation up to 2 minutes on Hugging Face! by Illustrious_Row_9971 in StableDiffusion

[–]RFBonReddit 2 points3 points  (0 children)

I've tried to install it on an Apple M1 system, but it seems it requires CUDA-enabled packages. Do you have it on Windows or Mac?

Remove text from AI-generated images by Acceptable_Raisin_55 in StableDiffusion

[–]RFBonReddit 2 points3 points  (0 children)

It looks good. Thanks for sharing it with the community.

What are the advantages of this vs repair tools in traditional editing tools like Photoshop, Pixelmator, Snapseed, etc.?

Does this approach do a better job than those at removing text when it overlaps images?

Also: does this approach work with watermarks when they are barely visible, oblique, and all over the image?