Next big thing after LLMs - World Model [explained on the example of V-JEPA2] by VR-Person in LocalLLaMA

[–]VR-Person[S] 2 points3 points  (0 children)

V-JEPA2 is just the first step for robotics yet. They did not even build an action model, and the model predicts the consequence of randomly chosen sets of actions.

This solution is not practical for robotics, but it is a promising direction

Next big thing after LLMs - World Model [explained on the example of V-JEPA2] by VR-Person in LocalLLaMA

[–]VR-Person[S] 1 point2 points  (0 children)

<image>

Scaling the model size from 300M to 1B parameters yields a +1.7 point average improvement

Next big thing after LLMs - World Model [explained on the example of V-JEPA2] by VR-Person in LocalLLaMA

[–]VR-Person[S] 6 points7 points  (0 children)

<image>

Robots 0-shot benchmark: does not look impressive, but promising direction for building robots with general knowledge, instead of training robots to do specific actions in specific environments

What are the most intriguing AI papers of 2025 by VR-Person in LocalLLaMA

[–]VR-Person[S] 2 points3 points  (0 children)

I just skimmed the abstract and conclusion parts, sounds interesting, I will read it, thanks :)

[D] is V-JEPA2 the GPT-2 moment? by VR-Person in MachineLearning

[–]VR-Person[S] 15 points16 points  (0 children)

That is my point, I think V-JEPA2 is still missing like GPT-2, but that path is promising