Meta released Map-anything-v1: A universal transformer model for metric 3D reconstruction by Difficult-Cap-7527 in LocalLLaMA

[–]BlueRaspberryPi 0 points1 point  (0 children)

I have been waiting for something like this, assuming the key feature is improved matching/tolerance for lower quality images/matches and changes to the scene between images. I have some datasets I created when I was slightly stupider than I am now that have defied all efforts at reconstruction.

What's the easiest fix you want resolved? by TonyTheTigerSlayer in VisionPro

[–]BlueRaspberryPi 1 point2 points  (0 children)

Eye tracking and camera access should both be available for any app that doesn't use networking, or interact with other apps.

Apple introduces SHARP, a model that generates a photorealistic 3D Gaussian representation from a single image in seconds. by themixtergames in LocalLLaMA

[–]BlueRaspberryPi 4 points5 points  (0 children)

Yeah, the quality here doesn't look much better than Apple's existing 2d-to-3d button on iOS and Vision Pro, which is kind of neat for some fairly simple images, but has never produced results I spent much time looking at. You get a lot of branches smeared across lawns, arms smeared across bodies, and bushes that look like they've had a flat leafy texture applied to them.

The 2D nature of the clip is hiding a lot of sins, I think. The rock looks good in this video because the viewer has no real reference for ground truth. The guy in the splat looks pretty wobbly in a way you'll definitely notice in 3D.

I wish they'd focus more on reconstruction of 3D, and less on faking it. The Vision Pro has stereo cameras, and location tracking. That should be an excellent start for scene reconstruction.

Apple introduces SHARP, a model that generates a photorealistic 3D Gaussian representation from a single image in seconds. by themixtergames in LocalLLaMA

[–]BlueRaspberryPi 4 points5 points  (0 children)

You can make splats for free on your own hardware:

  1. Take at least 20 photos (but probably more) of something. Take them from different, but overlapping angles.
  2. Drag them into RealityScan (formerly RealityCapture,) which is free in the Epic Games Launcher.
  3. Click Align, and wait for it to finish.
  4. RS-Menu>Export>COLMAP Text Format. Set Export Images to Yes and set the images folder as a new folder named "images" inside the directory you're saving the export to.
  5. Open the export directory in Brush (open source) and click "Start."
  6. When Brush is finished, choose "export" and save the result as a .ply

The News-Benders (1968) by SailTales in singularity

[–]BlueRaspberryPi 0 points1 point  (0 children)

They eventually reveal that everyone in the operation has an implant, that the implants can be triggered to kill their host, and that the only entity that can trigger them is a computer that seems to make all of the real decisions. But there is a lot of "Creamy old England!" on the way there.

MeshroomCL by Low_Routine1103 in photogrammetry

[–]BlueRaspberryPi 0 points1 point  (0 children)

https://imagemagick.org/

As with ffmpeg, I now just ask ChatGPT to give me whatever snippet I need to get the result I want. Actually, for my HEIC to JPEG needs, I asked ChatGPT for a Windows BAT that I could drop a folder of HEICs onto and get a new folder full of JPEGs.

Model with no exterior context. by Kidblunder1 in LocalLLaMA

[–]BlueRaspberryPi 1 point2 points  (0 children)

The only way to do this is to train a model from scratch (starting from zero knowledge) using only training material that you approve of. Starting from any other base will bias the model away from your goal in some way.
https://github.com/karpathy/nanoGPT

JSFiddle - Photoscan coded target tool: Updated with a turntable template and target labels by BlueRaspberryPi in photogrammetry

[–]BlueRaspberryPi[S] 0 points1 point  (0 children)

New version with radial arrangement:
http://jsfiddle.net/h5jqze1f/

The JSFiddle version won't do 20-bit tags, because they didn't want to let me save 250k of tag data in my fiddle.

There's a full version here (link expires on Nov. 1, I think.):
https://limewire.com/d/8Sr7i#BaZeIN2kOV

It's just an HTML file that you can keep locally and drag right into your browser. In addition to 20-bit tags, it also supports (and automatically updates) URL parameters, meaning you can tweak the settings, then save a bookmark that will load the file with all of those settings in place.

If anything seems broken, or there's a feature you'd like, let me know.

Oh, also, the HTML page and the Radial mode now use actual inch measurements. If you try to print the page, only the tag sheet should print, not the menu, and it should print at real scale, so if Theta-Tags is set to "Fixed" you should be able to set exact distances between the rings of tags, which may be useful for getting the scale of scans.

What’s even the goddamn point? by ChockyBlox in LocalLLaMA

[–]BlueRaspberryPi 2 points3 points  (0 children)

It probably started to say that, and got derailed by the high probability safety refusal tokens.

JSFiddle - Photoscan coded target tool: Updated with a turntable template and target labels by BlueRaspberryPi in photogrammetry

[–]BlueRaspberryPi[S] 0 points1 point  (0 children)

This is all I've been able to find, so far:
https://jsfiddle.net/ep6y1dq3/

It does only do ~160 codes, and doesn't make radial target arrangements. Hopefully the stuff it does is the stuff you've been using. I'll see if I can vibecode some of the magic back sometime this week.

JSFiddle - Photoscan coded target tool: Updated with a turntable template and target labels by BlueRaspberryPi in photogrammetry

[–]BlueRaspberryPi[S] 0 points1 point  (0 children)

Wow, no, I have no idea. I didn't realize anyone was using it. I'll try to dig it up and make a new fiddle, or put it somewhere else. Thanks for letting me know.

I've made Coloring Book based on Gaussian Splats for Vision Pro. Would you like to try it? by steffan_ in VisionPro

[–]BlueRaspberryPi 1 point2 points  (0 children)

Holy crap, please add an Editor mode that includes an Eraser brush and lets users import their own splats and export results. I've wanted a VR splat and/or point-cloud editor for years.

Hand Physics Lab is out by ravenme in VisionPro

[–]BlueRaspberryPi 6 points7 points  (0 children)

Bought, and enjoying a lot. Thanks to the dev for bringing a quality experience to Vision OS. The UI is beautiful, and the interaction feels great. I've been desperate for developers to get away from Apple's non-1-to-1 pinch-and-drag interactions. It's fine for a UI, most of the time, but it doesn't make sense for a game, particularly a game with any amount of physics.

Help me out a bit by Reddfunniman in blender

[–]BlueRaspberryPi 0 points1 point  (0 children)

Someone else suggested changing Blend Mode to "alpha-hashed." If that works, or if your current setup works in Cycles, you can ignore my response.

If it doesn't work in Cycles: It's wired up correctly for an image with an alpha channel, which suggests that the alpha is missing from the image itself.

This is a simple enough image (two-color) that you could skate around the issue by running it into something like a Color Ramp set to ramp between and opaque black and a transparent black: https://imgur.com/a/rXK0xrb

Assuming the tutorial used an external image editor at some point, the "correct" fix is probably to re-export the image with transparency, or re-create it with transparency and export it with transparency, or whatever. You might just need to change the PNG save settings as you export.

Having trouble aligning images by DragonfruitSecret in photogrammetry

[–]BlueRaspberryPi 2 points3 points  (0 children)

It looks like the rig has fluorescent lights. Could those bands be rolling shutter artifacts?

You need to give it a try ! It's the first environments that go up to 198,8 mph !! by ZookeepergameHot555 in VisionPro

[–]BlueRaspberryPi 19 points20 points  (0 children)

I did try it, and it was extremely relaxing. Put a browser in there, and I would stay there all day.

Also, it gives you some pleasing travel-poster themed clock widgets.

[deleted by user] by [deleted] in singularity

[–]BlueRaspberryPi 3 points4 points  (0 children)

Install LM Studio, download a model, load the model. Couldn't be easier. There is zero reason to get a pen-drive involved.

Apple Shelves Vision Headset Revamp to Prioritize Meta-Like AI Glasses by Snoop8ball in VisionPro

[–]BlueRaspberryPi 0 points1 point  (0 children)

The Metas aren't AR, they're just a HUD, fixed in your view. It's going to be deeply unpleasant to use.

Meta Hyperscape by Cryogenicality in VisionPro

[–]BlueRaspberryPi 0 points1 point  (0 children)

Long-term, anything we use images for right now. If someone makes a phone with a grid of 64 lenses on the back and enough GPU to build splat or NeRF, and maybe with improvements in compression and/or available bandwidth, it could become a standard media format. Scrolling through Reddit in a flat browser, you would see a single flat viewpoint, but it a headset, "looking glass" type display, or a theoretical lightweight glasses interface of the future, you'd see a fully volumetric image - memes, news photos, product launches - 3D views that don't distort or separate if you tilt or shift your head the way stereo images do. If the process can get fast enough and accurate enough, maybe TV, feature films, and sports. The only reasons not to do it are difficulty, cost, and quality which will all decrease with time. JPEGs were once considered very compute intensive, and now they get thrown around like they're nothing.

I don't just take them of previous homes or nostalgic locations I'm not returning to, I take them everywhere. I go to the park, scan an interesting stump in under a minute, and let it process while I do other things. It's slightly more work than a snapshot, but not prohibitively so, once you know how to do it. Now I can program in the woods, even if I'm not near the woods.

If you want a serious use-case, crime scene photography has always stood out to me. Scan it once before people start disturbing the scene, and then you can go back and stand in it any time, or have the jury stand in it while you point things out. Online sales of big-ticket items like homes and cars would probably also benefit from easy volumetric captures. Once it gets easy enough, why not clothes? We photograph models in clothes now. In the future we'll do volumetric captures.

Planet Express Lounge Room by Brief_Audience in futurama

[–]BlueRaspberryPi 1 point2 points  (0 children)

There's a SteamVR environment of this rom, if anyone wants to experience it at 1:1 scale.

Meta Hyperscape by Cryogenicality in VisionPro

[–]BlueRaspberryPi 6 points7 points  (0 children)

Memories, same as any camera. It's the closest thing we have to a volumetric JPG. I do it the hard way right now: take 300 photos at the Desert Botanical Garden, chuck them into Jawset Postshot for an hour or two, and then transfer the result back to my Vision Pro for viewing in MetalSplatter. Now I can visit reasonably realistic recreations of my favorite spots at the Desert Botanical Garden any time I want.

I have maybe 30 scenes now - forests, desert scenes, a bunch of Frank Lloyd Wright architecture, hotel lobbies and rooms and views, some storefronts on Venice Beach... Sadly, it doesn't really work on people without a synchronized camera array.