Meta’s New AI Model Can Isolate Any Object in an Image or Video—Just Ask It
Translate this article
Telling a computer to find a specific object in a photo or video clip has always required technical skill. Whether for a creative project, an e-commerce listing, or scientific research, precisely isolating elements has been a complex task. Meta’s latest release, the Segment Anything Model 3 (SAM 3), is designed to change that by making advanced visual recognition as simple as typing a phrase.
Announced alongside an interactive demo website called the Segment Anything Playground, SAM 3 allows users to identify, outline, and track objects using intuitive prompts. In a significant move for open science, Meta is also publicly releasing the model's core components for researchers and developers.
Moving Beyond a Fixed List of Objects
Unlike earlier systems trained to recognize a limited set of common items, SAM 3 understands a vast vocabulary. This shift allows it to find nuanced objects based on natural language.
From the Lab to Real-World Applications
This technology is already being integrated into practical tools and features:
Try It Yourself with the Segment Anything Playground
For those without a technical background, the easiest way to experience this technology is through the new Segment Anything Playground.
For developers, the release of the model's building blocks, training code, and datasets provides a powerful foundation for building new applications and conducting further research. While challenges remain—such as handling highly specialized terminology—SAM 3 represents a significant leap toward making sophisticated visual analysis an intuitive and accessible tool for everyone.
About the Author

Ryan Chen
Ryan Chan is an AI correspondent from Chain.
Recent Articles
Subscribe to Newsletter
Enter your email address to register to our newsletter subscription!