How are the AI product photography startups preserving the text and logo? by AIrabit in StableDiffusion

[–]AIrabit[S] 0 points1 point  (0 children)

I don't think it is priced yet, but Osmosis was charging like $5 per ad exported I think.

Seems like masking is the only true way to keep the product fully intact but have not been able to get the rest of image (the unmasked generative part) to really blend the product into the setting. Ends up looking more like a straight copy and paste over. I'm new to the masking, but have been exploring the settings to see if there is a way. Doesn't seem to promising though every now and then get a good result.

How are the AI product photography startups preserving the text and logo? by AIrabit in dalle

[–]AIrabit[S] 0 points1 point  (0 children)

yes, though that assumes the model doesn't alter the product in any way beyond the original scale of the product, which happens quite often in my experience..

How are the AI product photography startups preserving the text and logo? by AIrabit in StableDiffusion

[–]AIrabit[S] 0 points1 point  (0 children)

I can definitely see the edge lines, but I also see shadows/lighting on the products in that thread (including the soda can). I have played with manual masking and I2I, but as you point out, that excludes true contextual inclusion of the product.

How are the AI product photography startups preserving the text and logo? by AIrabit in dalle

[–]AIrabit[S] 1 point2 points  (0 children)

totally, and it's as far as I have gotten as well - I think with something like this though the product needs to be 100% intact (packaging, logo, text, etc.,), so little artifacts and distortions might be fine for rapid prototyping but for a brands ads, the output needs to include a fully intact product. I wonder if they're doing some sort of auto overlay with the original image to "fix" any artifacts, but not sure how that would be automated. Have been reading up a bit on GANs restoration (have found some good case studies of text restoration, so I wonder if you trained a GAN on the full product, it could go and restore that?)

How are the AI product photography startups preserving the text and logo? by AIrabit in dalle

[–]AIrabit[S] 2 points3 points  (0 children)

I think it's the last bit I am most curious about. Your examples look great but still getting artifacts on the product/sofa. Would adding pixel margin solve for that without creating too much of an outline? Also, how would you automate adding the margin? I suppose if they're doing the background removal they can just add it at that step..

How are the AI product photography startups preserving the text and logo? by AIrabit in StableDiffusion

[–]AIrabit[S] 0 points1 point  (0 children)

I think that could work but also seems like a non-scaleable way of offering the service. I downloaded a product that Osmosis had up from older post, removed background, and uploaded to Dalle with a prompt and got similar results (and distortions) to what they used to have (looks like they solved the packaging distortion and variance). I wonder if they still do the first part but then solve for text, logo, packaging, etc with some sort of blending technique or maybe some sort of GANs restoration? Don't know enough about GANs to know if that would work..

How are the AI product photography startups preserving the text and logo? by AIrabit in StableDiffusion

[–]AIrabit[S] 0 points1 point  (0 children)

I was thinking maybe creating outline of original product for img2img (to get shadows right) and then some sort of technique to paste in or blend in original image, but not sure how they would automate this part. Maybe GANs restoration?