×
Meta’s SAM 2.1 brings complex video editing to Instagram creators
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Meta’s Segment Anything Model (SAM) 2.1 has rapidly transitioned from research project to practical application, now powering the innovative Cutouts feature in Instagram’s new Edits app. This technology enables creators to perform sophisticated video editing tasks previously reserved for desktop applications, demonstrating how advanced AI research can evolve into consumer-facing features that empower digital creativity.

The big picture: Meta has successfully deployed its open-source segmentation model SAM 2.1 into Instagram’s Edits app, allowing mobile creators to perform complex video editing through the Cutouts feature.

  • The feature was used hundreds of thousands of times within 24 hours of the app’s launch, showing strong user adoption.
  • This implementation represents a rapid transition from research to practical application, with less than a year between SAM 2’s research demo and its integration into a consumer product.

How it works: Cutouts uses an object detection pipeline that can automatically suggest objects in video frames or allow manual selection through interactive clicking.

  • Once an object is selected, SAM 2.1 predicts a high-quality mask defining the object’s boundary in the selected frame.
  • Users can track the object throughout the video, with SAM 2.1 automatically generating consistent masks across all frames.

Key improvements: The engineering team made significant performance enhancements to make the technology viable for mobile use.

  • Model throughput was increased by 1.8x, making the feature more responsive.
  • End-to-end first frame preview latency was reduced by 3x on NVIDIA H100 GPUs, creating a smoother user experience.

What’s next: Meta is already developing SAM 3, which will expand capabilities to automatically detect, segment, and track objects in both images and videos.

  • The next-generation model will introduce open vocabulary text prompts alongside click prompts, making the technology more intuitive to use.
  • This advancement could further democratize sophisticated video editing capabilities for mobile creators.
How Meta Segment Anything Model enables Cutouts in the Instagram Edits app

Recent News

Monster cats? Gnarly Minions? AI-generated cartoon gore floods YouTube

AI tools are being misused to create violent, grotesque cartoons featuring children's characters that have garnered millions of views on the platform, echoing the 2017 Elsagate controversy.

Scanning the horizon: Meta expands data collection from Ray-Bans to enhance AI

Meta's smart glasses now automatically record and store voice commands, giving users fewer privacy options as the company harvests more data for AI development.

AI guardian Anubis thwarts massive DDoS attack on websites

Specialized filtering tool successfully mitigates 35,000-IP DDoS attack on open-source project without resorting to traditional IP blocking methods.