# EVF-SAM
**Repository Path**: xiguaner/evf-sam
## Basic Information
- **Project Name**: EVF-SAM
- **Description**: 基于evf-sam来搞参考遥感图像分割
- **Primary Language**: Unknown
- **License**: Apache-2.0
- **Default Branch**: master
- **Homepage**: None
- **GVP Project**: No
## Statistics
- **Stars**: 1
- **Forks**: 0
- **Created**: 2024-10-20
- **Last Updated**: 2026-01-06
## Categories & Tags
**Categories**: Uncategorized
**Tags**: None
## README
📷 EVF-SAM
Early Vision-Language Fusion for Text-Prompted Segment Anything Model
[Yuxuan Zhang](https://github.com/CoderZhangYx)
1,\*, [Tianheng Cheng](https://scholar.google.com/citations?user=PH8rJHYAAAAJ&hl=zh-CN)
1,\*, Lei Liu
2, Heng Liu
2, Longjin Ran
2, Xiaoxin Chen
2, [Wenyu Liu](http://eic.hust.edu.cn/professor/liuwenyu)
1, [Xinggang Wang](https://xwcv.github.io/)
1,📧
1 Huazhong University of Science and Technology,
2 vivo AI Lab
(\* equal contribution, 📧 corresponding author)
[](https://arxiv.org/abs/2406.20076)
[](https://huggingface.co/YxZhang/)
[](https://huggingface.co/spaces/wondervictor/evf-sam)
[](https://huggingface.co/spaces/wondervictor/evf-sam2)
[](https://colab.research.google.com/github/hustvl/EVF-SAM/blob/main/inference_image.ipynb)
## News
We have expanded our EVF-SAM to powerful [SAM-2](https://github.com/facebookresearch/segment-anything-2). Besides improvements on image prediction, our new model also performs well on video prediction (powered by SAM-2). Only at the expense of a simple image training process on RES datasets, we find our EVF-SAM has zero-shot video text-prompted capability. Try our code!
## Highlight