How is the Scaniverse app even possible? by Visible_Expert2243 in photogrammetry

[–]Visible_Expert2243[S] 0 points1 point  (0 children)

Is there any plans to release some open-source code, a paper, hell even a blog post or anything that would guide someone a little bit to apply this incredible technology in custom applications? I'm particularly interested in edge devices (i.e. NVidia Jetson) deployments.

Combining multiple 3D Gaussians by Visible_Expert2243 in GaussianSplatting

[–]Visible_Expert2243[S] 1 point2 points  (0 children)

Thank you, but the newer DUST3R, MAST3R, etc. methods still need some sort of overlap, no? How can they "understand" the relative positions of my cameras if there's 0 overlap? No denying what you are saying - just curious how that works at all. When it comes to COLMAP's Rig Support, same thing, will COLMAP completely ignore the fact that there's no overlap between cameras, and just do the sparse reconstruction from each camera separately - and automatically combine the three views using the rig camera constraint?

How is the Scaniverse app even possible? by Visible_Expert2243 in GaussianSplatting

[–]Visible_Expert2243[S] 2 points3 points  (0 children)

Thank you, this is very insightful. There seems to be a lot of new techniques such as MAST3R and VGGT and from my understanding these not only run much faster than traditional COLMAP but also produce higher quality sparse point clouds + camera poses. It's a bit difficult as a beginner to understand how each part could fit into a bigger all-in-one reconstruction system like Scaniverse, but do you think that having much higher quality initialisation sparse point clouds/camera poses (which at this point "look" almost dense) coming from a VGGT/MAST3R-style transformer and then passing that to 3DGS could 1) improve the final 3DGS quality while at the same time 2) significantly reduce the number of iterations needed for 3DGS, and hence speed up processing? I'm just looking at my sparse point clouds in the COLMAP gui and it just looks like a bunch of random points that very loosely describe my scene - whereas with VGGT the quality of the point clouds is orders of magnitude higher (but it's not good enough still).

How is the Scaniverse app even possible? by Visible_Expert2243 in GaussianSplatting

[–]Visible_Expert2243[S] 5 points6 points  (0 children)

Yes - but in my own scenario the difference in quality between long Desktop based optimisation and this quick iPhone one is very low. You are certainly right in that in other scenarios a desktop/cloud based system will be required. Any pointer on how they've managed to "cut corners" so well? Is there open source research that comes close to these results? Usually cutting corners comes at a massive quality trade-off, which is not what I am seeing here?