What is Segment Anything 3 (SAM 3)?
Published Nov 19, 2025 • 10 min read

Released on November 19th, 2025, Segment Anything 3 (SAM 3) is a zero-shot image segmentation model that “detects, segments, and tracks objects in images and videos based on concept prompts.” This model was developed by Meta as the third model in the Segment Anything series.

Unlike its previous SAM models (Segment Anything and Segment Anything 2), you can provide SAM 3 with the prompt “shipping container” and it will generate precise segmentation masks for all shipping containers in an image. SAM 3 generates segmentation masks that correspond to the location of the objects found with a text prompt.

In this guide, we are going to talk about what SAM 3 is, what you can do with SAM 3, and how you can use the model today.

Just looking for links? We have the most important SAM 3 links here:

Eager to try SAM 3? Drag and drop an image into our interactive playground below with your own text prompt:

Without further ado, let’s get started!

What is Segment Anything 3?

SAM 3 expands the Segment Anything series beyond visual prompts to Promptable Concept Segmentation (PCS) — the task of identifying every instance of a concept in an image or video from a short noun phrase (e.g., “yellow school bus”) or image exemplar.

Unlike SAM 1 and 2, which predicted a single object per prompt, SAM 3 performs open-vocabulary instance detection, returning unique masks and IDs for all matching objects simultaneously. SAM 3 transforms SAM from a geometric segmentation tool into a concept-level vision foundation model.

How does Segment Anything 3 work?

SAM 3 draws precise segmentation masks around the contours of objects that have been identified by a prompt. For example, you can ask SAM 3 to identify a “striped cat” and it will find the striped cats, or you could click on a cat in an image and calculate a segmentation mask that corresponds with the specific cat on which you have clicked:

SAM 3 works across images and videos. If you provide a video as an input to SAM 3, the model will track all identified or selected instances of objects through frames in the video. If you provide an image as an input, SAM 3 will run on the input image.

SAM 3 runs at ~30 ms per image on an H200 GPU, handling 100+ objects, but at ~840 M parameters (≈3.4 GB) it remains a server-scale model. For object detection on the edge, you can use SAM-3 to label data and use the labeled data to train a smaller, faster, supervised object detection model. The smaller model will not have text prompting abilities, but would allow you to detect exactly what objects you have trained the model to find in real time.

You can run SAM 3 without fine-tuning. You can also fine-tune SAM 3 to improve its performance on domain-specific tasks, which we discuss later in this article.

SAM 3 is licensed under a custom license developed by Meta. The model is available for public use subject to the terms and conditions provided on the model weights download page.

SAM 3 Task Types

SAM 3 unifies several segmentation paradigms — single-image, video, interactive refinement, and concept-driven detection — under a single model backbone. It can operate in Promptable Concept Segmentation (PCS) mode (text or exemplars), Promptable Visual Segmentation (PVS) mode (points, boxes, masks), or interactive hybrid mode that blends both.

This unified interface means a single SAM 3 checkpoint can handle everything from dataset labeling to live tracking and multimodal reasoning.

SAM 3 Point and Click

As before in SAM 1 and SAM 2, you can prompt SAM 3 with visual positive and negative clicks.

SAM 3 keeps the same interactive refinement loop — positive clicks to add, negative to remove regions — but these can now be applied to concept-wide detections or to individual masklets within videos.

Internally, each click is encoded as a positional token and fused into the prompt embedding space before being re-decoded into refined masks. This design allows incremental correction without restarting inference, enabling fluid dataset annotation and fine-tuning workflows.

Providing point and click positive and negative feedback points for SAM 3

SAM 3 Video Processing

SAM 3 deploys the same video memory layer used in SAM 2 where predictions from prior frames are used to inform predictions in new frames.

SAM 3’s tracker inherits the transformer-based memory bank from SAM 2 but introduces tighter coupling between the detector and tracker through the shared Perception Encoder. On each frame, the detector finds new concept instances, while the tracker propagates “masklets” — temporal object segments — from previous frames via self- and cross-attention.

A matching and update stage merges propagated and newly detected masks, ensuring consistency under occlusion or re-appearance.
The tracker also employs temporal disambiguation strategies such as detection-score decay and re-prompting, preventing drift and maintaining identity integrity across clips.

Together, this streaming memory design enables real-time concept tracking across complex video domains.

SAM 3 Examples

Below, we show a few of the images and text prompts that we have run through SAM 3 to demonstrate the powerful zero-shot segmentation capabilities of the model. 

Solar Panels

We gave SAM-3 the prompt “solar panels” and an image that contained several solar panels. The model was able to both successfully identify all solar panels as well as draw precise masks around each, while being careful not to highlight occluded areas on the solar panel:

Vehicles on a Street

We provided SAM-3 with the prompt “vehicle” and a photo of a street with several cars and motorbikes. SAM-3 was able to successfully identify all vehicles:

Boxes in a Warehouse

We provided SAM-3 with a photo of dozens of boxes in a warehouse and gave the model the prompt “box”. SAM-3 correctly identified the location of almost every box in the foreground of the image, without fine-tuning:

This demonstrates SAM-3’s proficiency at identifying a large number of objects in a single image.

Aerial Imagery

We asked SAM-3 to identify all trailers and stall numbers in an aerial image of an intermodal yard. SAM-3 was able to successfully identify all instances of the objects we prompted it to identify:

SAM 3 Demo

If you want to experiment with SAM 3 on your own input data and prompts, you can use Roboflow Playground, a free, web-based playground for experimenting with models. Roboflow Playground lets you upload an image and see the segmentation masks returned by the model.

We also have a widget below that you can drag-and-drop an image into to try SAM-3:

Meta has prepared a web-based tool that is available on the Segment Anything 3 website.

How to Use SAM 3

There are several ways that you can use SAM 3.

Use a SAM 3 API

With Roboflow, you can deploy your own SAM 3 API. You can use this API to make requests to SAM 3 without having to worry about managing your own infrastructure.

To get started, you can click “Fork Workflow” on the widget below:

This will create an endpoint that can run SAM 3. Here is an example of what the code snippet will look like:

# 1. Import the library
from inference_sdk import InferenceHTTPClient

# 2. Connect to your workflow
client = InferenceHTTPClient(
    api_url="https://serverless.roboflow.com",
    api_key="ZBGiII0DVVkK7LWLjlYE"
)

# 3. Run your workflow on an image
result = client.run_workflow(
    workspace_name="roboflow-jvuqo",
    workflow_id="sam3",
    images={
        "image": "YOUR_IMAGE.jpg" # Path to your image file
    },
    parameters={
        "prompt": "PARAMETER_VALUE"
    },
    use_cache=True # Speeds up repeated requests
)

# 4. Get your results
print(result)

Fine-Tune SAM 3

If you notice that SAM 3 doesn’t perform well on a task in a specific domain, you may want to experiment with fine-tuning SAM 3. This will help you achieve better performance on tasks for which the model may not have originally been trained.

To fine-tune SAM 3, you will need:

  1. An annotated image segmentation dataset focused on the task that you want to fine-tune SAM 3 to solve, and;
  2. A place to train your model.

To start fine-tuning the model, you can upload your instance segmentation dataset to Roboflow – in any common format, like COCO JSON – generate a dataset version, then start a SAM 3 fine-tuning job.

As you fine-tune SAM-3 models in Roboflow, you can measure your model performance with each epoch. You can then try your models in a Workflow and deploy them to the cloud or the edge.

Use SAM 3 to Train a Smaller Model

SAM 3 is a computationally expensive model. You need a large GPU to be able to run the model. With that said, you can use the capabilities of SAM 3 to label data for use in training a smaller model that is fine-tuned for your specific task.

All you need are a few images or a short video to get started. You can then use the text prompting capabilities of SAM 3 in the Roboflow interface to label your data. Roboflow uses the fine-tuned SAM 3 model to train an RF-DETR model.

Models can be deployed in the cloud using Roboflow Workflows with serverless API infrastructure, on dedicated cloud infrastructure, or on your own hardware.

SAM 3 Data Engine

In order to train SAM 3 the Meta team took on an enormous data annotation and review process through their SAM 3 data engine.

To power open-vocabulary learning, Meta built a four-phase data engine that combines humans, SAM models, and fine-tuned LLMs in a feedback loop.
Phases 1–3 focused on images, progressively increasing automation, while Phase 4 extended to videos.

The SAM 3 data engine. [Source]

The resulting SA-Co dataset (“Segment Anything with Concepts”) includes ≈ 5.2 M high-quality images and 52.5 K videos with > 4 M unique noun phrases and ≈ 1.4 B masks — making it the largest concept-segmentation corpus to date.
Each iteration used AI annotators to propose candidate noun phrases and AI verifiers (fine-tuned Llama 3.2) to assess mask quality and exhaustivity. Human effort was then concentrated on failure cases — a process that doubled throughput compared to human-only pipelines.

SA-Co dataset ontology introduced in SAM 3

The SA-Co ontology spans 22 million entities across 17 top-level and 72 sub-categories, bringing fine-grained coverage from common objects to long-tail concepts.

SAM 3 Architecture Overview

SAM 3 builds on architecture from SAM 1, SAM 2, and Perception Encoder.

The SAM 3 architecture is a dual encoder-decoder transformer comprising a DETR-style detector and a SAM 2-inspired tracker that share a unified Perception Encoder (PE). The PE aligns visual features from the image encoder with text and exemplar embeddings from the prompt encoder, creating a joint embedding space for vision-language fusion.

To handle open-vocabulary concepts, the detector adds a global presence head that first determines if the target concept exists in the scene before localizing its instances. This decoupling of recognition (what) and localization (where) significantly improves accuracy on unseen concepts and hard negatives, which are prevalent in natural imagery.

The result is a zero-shot segment-anything engine capable of detecting all matching objects with minimal user input.

SAM 3 architecture inherits memory from SAM 2 and builds a new encoder and detection engine

Training occurs in four stages: (1) Perception Encoder pre-training, (2) detector pre-training on synthetic and HQ data, (3) detector fine-tuning on SA-Co HQ, and (4) tracker training with a frozen backbone to stabilize temporal learning.

Promptable Composable Segmentation: A New Vision Task 

SAM and SAM 2 were both capable of identifying objects with a specific prompt. SAM 3, however, introduces text prompting, and a new task type to evaluate the efficacy of segmentation models that accept text prompts as inputs: Promptable Concept Segmentation (PCS).

Part of the contributions made in the SAM 3 paper were both the introduction of PCS as a task type as well as a “scalable data engine” that was used to produce a dataset with splits for training and benchmarking. This is discussed in Section 5, Segment Anything with Concepts (SA-Co) Dataset in the SAM 3 paper.

Meta benchmarked SAM 3 across both instance segmentation, box detection, and semantic segmentation, and reported the following results:

SAM 3 achieves state-of-the-art performance across most of the benchmarks run, including on LVIS and SA-Co for instance segmentation.

Measuring Few-Shot Adaptation

In addition to benchmarking for instance and semantic segmentation and box detection, Meta evaluated the few-shot adaptation capabilities of SAM 3. The Meta research team used the ODinW13 and Roboflow100-VL benchmarks for this purpose.

When reporting on the results on the Roboflow100-VL benchmark, Meta found that “While SAM 3 underperforms the current state-of-the-art Liu et al. (2023) in zero-shot evaluation, it surpasses leading methods Liu et al. (2023); Chen et al. (2024a) in both few-shot and full fine-tuning scenarios.” Their research team concluded this result demonstrated SAM 3’s “strong visual generalisation capabilities when provided with task-specific training data.”

Conclusion

Released on November 19th, 2025 by Meta, SAM 3 is a zero-shot image and video segmentation model. You can prompt SAM 3 with “concept prompts” which are text prompts (i.e. “shipping container”) or points on an image that correspond to the object you want to identify.

You can use SAM 3 out-of-the-box with text or point prompts. You can also use the model to label data for use in training a smaller, supervised model that specialises in the particular task you are working on. Furthermore, you can fine-tune SAM 3 on a custom dataset, ideal if you want to adapt SAM 3 itself to a particular visual problem you want to solve.

To get started with SAM 3, you can use Roboflow Playground to experiment for free.

Cite this Post

Use the following entry to cite this post in your research:

James Gallagher, Jacob Solawetz. (Nov 19, 2025). What is Segment Anything 3 (SAM 3)?. Roboflow Blog: https://blog.roboflow.com/what-is-sam3/

Stay Connected
Get the Latest in Computer Vision First
Unsubscribe at any time. Review our Privacy Policy.

Written by

James Gallagher
James is a technical writer at Roboflow, with experience writing documentation on how to train and use state-of-the-art computer vision models.
Jacob Solawetz
Vision language research @ Roboflow