pull/139/head
MaureenZOU 2023-10-04 16:59:23 -05:00
parent 82ccd5e55a
commit b0c4787775
1 changed files with 6 additions and 6 deletions

View File

@ -44,12 +44,6 @@ git clone git@github.com:UX-Decoder/Segment-Everything-Everywhere-All-At-Once.gi
<img src="https://user-images.githubusercontent.com/11957155/233526415-a0a44963-19a3-4e56-965a-afaa598e6127.gif" width="400" />
</p>
## :bookmark_tabs: Catelog
- [x] SEEM Demo
- [x] Inference and Installation Code
- [x] Evaluation Code
- [x] Training Code
:fire: **Related projects:**
* [FocalNet](https://github.com/microsoft/FocalNet) and [DaViT](https://github.com/dingmyu/davit) : We used FocalNet and DaViT as the vision backbones.
@ -63,6 +57,12 @@ git clone git@github.com:UX-Decoder/Segment-Everything-Everywhere-All-At-Once.gi
* [X-GPT](https://github.com/microsoft/X-Decoder/tree/xgpt) : Conversational Visual Agent supported by X-Decoder.
* [LLaVA](https://github.com/haotian-liu/LLaVA) : Large Language and Vision Assistant.
## :bookmark_tabs: Catelog
- [x] SEEM Demo
- [x] Inference and Installation Code
- [x] Evaluation Code
- [x] Training Code
## :bulb: Highlights
Inspired by the appealing universal interface in LLMs, we are advocating a universal, interactive multi-modal interface for any type of segmentation with **ONE SINGLE MODEL**. We emphasize **4** important features of **SEEM** below.
1. **Versatility**: work with various types of prompts, for example, clicks, boxes, polygons, scribbles, texts, and referring image;