Back to stories
Models

Meta Releases SAM 3.1 — Open-Source Image Segmentation Coming to Instagram

Michael Ouroumis2 min read
Meta Releases SAM 3.1 — Open-Source Image Segmentation Coming to Instagram

Meta dropped a new version of one of the most useful open-source computer vision models in the field.

SAM 3.1 — the latest iteration of Meta's Segment Anything Model — is now available on Hugging Face and GitHub, with inference code, finetuning code, downloadable checkpoints, and example notebooks. The release also confirmed that SAM 3 technology is coming to Instagram Edits and Vibes in the Meta AI app.

What SAM Does

Segment Anything is exactly what it sounds like. You point at something — with a click, a box, text, or another visual prompt — and SAM precisely outlines and isolates it. In images or video.

This is the kind of technology that makes "remove the background" buttons work well. It's what powers "select the person" in photo editors. SAM is the open-source version that anyone can build with, study, or improve.

What's New in 3.1

The headline addition in 3.1 is finetuning support. Previous releases included inference code — you could run the model — but 3.1 adds the training infrastructure to adapt it to new domains. Medical imaging, satellite imagery, industrial inspection, autonomous vehicles: any field with specialized segmentation needs can now finetune SAM 3 on domain-specific data.

That matters because "segment anything" works well for general consumer images, but specialized applications need a model that understands their specific objects, lighting conditions, and contexts. Finetuning support turns SAM from a useful tool into a foundation model that specialists can actually build on.

The Consumer Application

Meta confirmed that SAM 3 is headed to Instagram Edits and Vibes — the company's consumer video and photo editing surfaces. For most users, that means better object selection, cleaner cutouts, and more accurate background editing in Instagram's native tools.

It's also a significant distribution play: SAM's capabilities will reach billions of users through apps they're already using, not just developers who build with the API.

Why Open-Source Matters Here

Meta has been consistently open-sourcing its computer vision and AI research. SAM, Llama, and related models have become foundational pieces of the AI ecosystem precisely because they're available to build with, study, and improve.

SAM 3.1's finetuning release extends that pattern. Instead of keeping the best capabilities inside Meta products, the company publishes the tools that let the broader research and developer community push the technology further — and build applications Meta's own teams haven't imagined yet.

How AI Actually Works — Free Book on FreeLibrary

A free book that explains the AI concepts behind the headlines — no jargon, just clarity.

More in Models

Anthropic's Next Model Is Called 'Mythos' — Leaked via Security Lapse in Its Own CMS
Models

Anthropic's Next Model Is Called 'Mythos' — Leaked via Security Lapse in Its Own CMS

A misconfigured content management system exposed nearly 3,000 unpublished Anthropic assets, including a draft blog post revealing 'Claude Mythos' — a new model tier Anthropic describes as a 'step change' that surpasses all previous Claude models.

1 day ago3 min read
Anthropic Ships Claude Computer Use: Claude Can Now Control Your Mac While You Do Something Else
Models

Anthropic Ships Claude Computer Use: Claude Can Now Control Your Mac While You Do Something Else

Anthropic has launched Claude Computer Use as a research preview — Claude can take over your keyboard, mouse, browser, and apps via the Dispatch iOS app, completing tasks autonomously on your desktop.

1 day ago3 min read
Google Launches Gemini 3.1 Flash Live and Search Live Globally — Point Your Camera and Talk
Models

Google Launches Gemini 3.1 Flash Live and Search Live Globally — Point Your Camera and Talk

Google has launched Gemini 3.1 Flash Live, its highest-quality voice model yet, powering a global rollout of Search Live across 200+ countries. Users can point their phone camera at anything and have a real-time conversation about it.

1 day ago3 min read