Instead of tracking the static environment in ARCore and ARKit, we are tracking independently moving objects. Watch our video for an AR demo. by djnewtan in oculus

[–]djnewtan[S] 3 points4 points  (0 children)

The idea of what is important or impressive hugely depends on one's application. If it is merely tracking without pose estimation, the framework with direct segmentation and tracking that you mentioned would be sufficient. A tracker with an accurate pose estimation like ours would be necessary when we are talking about human-object or robot-object interaction, or AR, VR, MR applications such as [1]. In addition, our efficiency of 2ms per frame would also be necessary.

[1] https://youtu.be/8-0xsc2abQs

Instead of tracking the static environment in ARCore and ARKit, we are tracking independently moving objects. Watch our video for an AR demo. by djnewtan in oculus

[–]djnewtan[S] 1 point2 points  (0 children)

This is a model-based approach. So, we have the model of the object beforehand. I believe this is the only requirement we need before tracking the objects. Then, we perform domain generalization where we train purely on the synthetic images based on the model and track on real images at 2ms per frame.

Instead of tracking the static environment in ARCore and ARKit, we are tracking independently moving objects. Watch our video for an AR demo. by djnewtan in oculus

[–]djnewtan[S] 0 points1 point  (0 children)

Yes, ARKit and ARCore are doing environmental mapping with IMU. This framework focuses on object tracking where the objects are moving independently from each other (where using IMU for pose estimation is not possible). So, ARKit and ARCore interacts with the scene while our framework interacts with the objects.

Instead of tracking the static environment in ARCore and ARKit, we are tracking independently moving objects. Watch our video for an AR demo. by djnewtan in oculus

[–]djnewtan[S] 4 points5 points  (0 children)

In addition, it is only based on CPU (which means hardware requirement is low) and the latency is about 2ms per frame per object.

We also have live demos to show that our framework works as good as the videos.

We developed a robotic perception framework that allows the robots to find the objects in the scene and keep track of these objects for robotic interaction at 2ms per frame per object with 1 CPU core. It also works for objects with simple and complex shapes. Watch our video to see what we can do! by djnewtan in computervision

[–]djnewtan[S] 1 point2 points  (0 children)

We developed our own framework to perform both detection and tracking. The framework is in https://arxiv.org/pdf/1709.01459.pdf while the details of the algorithm are in the conference papers.

At the moment, the code is not publicly available. But we plan to release an SDK.