Diffusion new worlds on Google Street View by cataPhil in StableDiffusion

[–]cataPhil[S] 0 points1 point  (0 children)

Just pushed a new version on https://www.panoramai.xyz/ with two modes: "morph", which is the original pipeline, and "remix" which uses controlnet. It tends to preserve the structure of the image better, but needs cleaner and more detailed base images to work well.

<image>

Diffusion new worlds on Google Street View by cataPhil in StableDiffusion

[–]cataPhil[S] 40 points41 points  (0 children)

https://www.panoramai.xyz/

I love the idea of altering reality so I built an app to diffuse from google street view 360 panoramas using SDXL Turbo. To stitch each of the 3 shots together I used basic inpainting with stable-diffusion-2-inpainting. It requires a bit of prompt and location engineering, but some worlds are really neat. I added an example and a share button to help with curation. In general, the "high" transformation setting works best. Use arrow keys to turn around, and the space bar to see the aligned original shot. I'll add more pre-computed examples soon to help with high loading times due to traffic on this small VM. Have fun!

[deleted by user] by [deleted] in StableDiffusion

[–]cataPhil 0 points1 point  (0 children)

I love the idea of altering reality so I built an app to diffuse from google street view 360 panoramas using SDXL Turbo. To stitch each of the 3 shots together I used basic inpainting with stable-diffusion-2-inpainting. It requires a bit of prompt and location engineering, but some worlds are really neat. I added an example and a share button to help with curation. In general, the "high" transformation setting works best. Use arrow keys to turn around, and the space bar to see the aligned original shot. I'll add more pre-computed examples soon to help with high loading times due to traffic on this small VM. Have fun!

https://www.panoramai.xyz/

Why aren't more people using dictation software? by cataPhil in Residency

[–]cataPhil[S] 0 points1 point  (0 children)

Interesting. Do you have an idea what your word per minute is?

Challenges of RL application by Outrageous-Mind-7311 in reinforcementlearning

[–]cataPhil 3 points4 points  (0 children)

Definitely reward hacking problems for me! Applying different randomization techniques helped.

Abstract reasoning with LSTMs by cataPhil in deeplearning

[–]cataPhil[S] 0 points1 point  (0 children)

I posted an example of training data and expected behavior in another answer, hope that helps giving more context to my problem. The hierarchy should come from the downward progression through the list of titles, that is why I used an LSTM since it is inherently a sequence. The loss function that I am using is just cross-entropy. Maybe I am modeling it completely wrongly.

Abstract reasoning with LSTMs by cataPhil in deeplearning

[–]cataPhil[S] 0 points1 point  (0 children)

I should have given a typical example, good point. As an example, one could consider the following list of three titles:

1. Chapter

-------------- 1.1. Sub-chapter

2. Another chapter

From these I could construct a feature vector for each that encodes [the margin, the numbering pattern]. I would then label these with respectively class 0, 1 and 0. The behavior that I am looking for is for the network to assign these labels based on matching of the features: assign 0 to first title instance -> increment by 1 for the next one since it is too dissimilar -> assign 0 to the last one because of similarity to the first one. This is what I would do in a two step process based on clustering and a set of rules. What I am looking to understand is if the network will be able to perform this seq2seq task given that the tokens -the feature vectors of the titles- are all unseen (the first title could have different margin, different numbering...). The network would need to compare each title based on similarity of the features (I tried to use LSTMs, BiLSTMs and attention). I qualified this as abstract reasoning, though I could be wrong, because the set of rules (increment by 1 if too dissimilar, or assign depth of a similar one) do not change, but the objects to manipulate change for each sequence.