Connect

V-JEPA 2: Meta’s New Self-Supervised World Model for Advanced AI Capabilities

Jack Carter

Translate this article

Updated:
June 13, 2025

Meta has introduced V-JEPA 2, a Video Joint Embedding Predictive Architecture that marks a significant step forward in AI world modeling. Trained on video data, this self-supervised foundation model achieves impressive visual understanding and predictive capabilities, enabling applications like zero-shot robot control in unfamiliar environments. Here’s a closer look at what V-JEPA 2 offers and its potential impact.


Understanding and Anticipating the Physical World

V-JEPA 2 is designed to understand, predict, and plan within physical environments with minimal supervision. By leveraging a world model, it interprets physical reality, anticipates future events, and devises efficient strategies. The model excels in motion understanding and, when paired with language modeling, delivers strong visual reasoning capabilities. Its ability to predict how the world evolves sets a new benchmark in action anticipation based on contextual cues.


Enabling Zero-Shot Robot Control

One of V-JEPA 2’s standout features is its application in robotics. Trained on just 62 hours of robot data from the Droid dataset, the model can be deployed on a robot arm to perform tasks like reaching, grasping, and pick-and-place in new environments. By using goal images to specify tasks, V-JEPA 2 enables task-agnostic planning without requiring extensive robot data or task-specific demonstrations, making it highly adaptable.


A Two-Phase Training Approach

V-JEPA 2’s architecture relies on a two-phase training process. First, its encoder and predictor are pre-trained through self-supervised learning on abundant natural video data, building a foundation for understanding and predicting physical world dynamics. Then, fine-tuning on a small amount of robot data allows the model to plan efficiently without needing large-scale expert demonstrations, which are often challenging to collect.


Meta’s Vision for World Models

Meta sees world models like V-JEPA 2 as a way to enable AI to reason and plan as intuitively as humans. This ambition drives their research, aiming to tackle one of AI’s grand scientific challenges. By releasing V-JEPA 2 to the research community, Meta invites collaboration to explore new applications and build on this work.



Resources and Community Engagement

Meta encourages researchers and developers to explore V-JEPA 2 through various resources:

  1. The AI at Meta blog provides insights into the model’s development.
  2. The research paper details its technical foundations. https://arxiv.org/abs/2506.09985
  3. V-JEPA 2 is available on Hugging Face, along with options to download both V-JEPA 2 and its predecessor, V-JEPA 1.


Meta’s release of V-JEPA 2 underscores their commitment to advancing AI through open collaboration, inviting the community to push the boundaries of what world models can achieve.


For more details, visit Meta’s AI resources or download V-JEPA 2 to explore its capabilities firsthand using this link

Artificial Intelligence

About the Author

Jack Carter

Jack Carter is an AI Correspondent from United States of America.

Subscribe to Newsletter

Enter your email address to register to our newsletter subscription!

Contact

+1 336-825-0330

Connect