398 results for “topic:segment-anything”
Ultralytics YOLO 🚀
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
Segment Anything in High Quality [NeurIPS 2023]
A Python package for segmenting geospatial data with the Segment Anything Model (SAM)
Segment Anything for Stable Diffusion WebUI
Efficient vision foundation models for high-resolution generation and perception.
Effortless AI-assisted data labeling with AI support from YOLO, Segment Anything (SAM+SAM2/2.1+SAM3), MobileSAM!!
InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)
An open-source project dedicated to tracking and segmenting any objects in videos, either automatically or interactively. The primary algorithms utilized include the Segment Anything Model (SAM) for key-frame segmentation and Associating Objects with Transformers (AOT) for efficient tracking and propagation purposes.
Images to inference with no labeling (use foundation models to train supervised models).
Labeling tool with SAM(segment anything model),supports SAM, SAM2, SAM3, sam-hq, MobileSAM EdgeSAM etc.交互式半自动图像标注工具
General AI methods for Anything: AnyObject, AnyGeneration, AnyModel, AnyTask, AnyX
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
Tracking and collecting papers/projects/others related to Segment Anything.
Must-have resource for anyone who wants to experiment with and build on the OpenAI vision API 🔥
Segment-Anything + 3D. Let's lift anything to 3D.
Open Source Generative Process Automation (i.e. Generative RPA). AI-First Process Automation with Large ([Language (LLMs) / Action (LAMs) / Multimodal (LMMs)] / Visual Language (VLMs)) Models
Adapting Meta AI's Segment Anything to Downstream Tasks with Adapters and Prompts
收集 CVPR 最新的成果,包括论文、代码和demo视频等,欢迎大家推荐!Collect the latest CVPR (Conference on Computer Vision and Pattern Recognition) results, including papers, code, and demo videos, etc., and welcome recommendations from everyone!
Inpaint Anything extension performs stable diffusion inpainting on a browser UI using masks from Segment Anything.
[CVPR2024 Highlight]GLEE: General Object Foundation Model for Images and Videos at Scale
Wunjo CE: Face Swap, Lip Sync, Control Remove Objects & Text & Background, Restyling, Audio Separator, Clone Voice, Video Generation. Open Source, Local & Free.
Official PyTorch implementation of "EdgeSAM: Prompt-In-the-Loop Distillation for On-Device Deployment of SAM"
Based on GroundingDino and SAM, use semantic strings to segment any element in an image. The comfyui version of sd-webui-segment-anything.
SAM-PT: Extending SAM to zero-shot video segmentation with point-based tracking.
Segment Anything in 3D with NeRFs (NeurIPS 2023 & IJCV 2025)
MetaSeg: Packaged version of the Segment Anything repository
The official implementation of Segment Any 3D GAussians (AAAI-25)
Medical SAM 2: Segment 3D Medical Images Via Segment Anything Model 2
[CVPR 2025] Official PyTorch implementation of "EdgeTAM: On-Device Track Anything Model"