Update README.md

pull/22/head
Xueyan Zou 2023-04-23 10:50:13 -05:00 committed by GitHub
parent ef010a0d0b
commit 3311d6df73
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
1 changed files with 13 additions and 11 deletions

View File

@ -4,6 +4,19 @@ We introduce **SEEM** that can **S**egment **E**verything **E**verywhere with **
:grapes:\[[Read our arXiv Paper](https://arxiv.org/pdf/2304.06718.pdf)\]   :apple:\[[Try Hugging Face Demo](https://huggingface.co/spaces/xdecoder/SEEM)\]
:fire: **Related projects:**
* [FocalNet](https://github.com/microsoft/FocalNet) : Focal Modulation Networks; **We used FocalNet as the vision backbone**.
* [UniCL](https://github.com/microsoft/UniCL) : Unified Contrasative Learning; **We used this technique for image-text contrastive larning**.
* [X-Decoder](https://github.com/microsoft/X-Decoder) : Generic decoder that can do multiple tasks with one model only**We built SEEM based on X-Decoder**.
:fire: **Other projects you may find interesting:**
* [OpenSeed](https://github.com/IDEA-Research/OpenSeeD) : Strong open-set segmentation methods.
* [Grounding SAM](https://github.com/IDEA-Research/Grounded-Segment-Anything) : Combining Grounding DINO and Segment Anythin.
* [LLaVA](https://github.com/haotian-liu/LLaVA) : Large Language and Vision Assistant.
## :rocket: Updates
* Try our [Video Demo (Beta)](https://083193edbb48f717.gradio.app) on referring video object segmentation.
<p float="left">
@ -11,17 +24,6 @@ We introduce **SEEM** that can **S**egment **E**verything **E**verywhere with **
<img src="https://user-images.githubusercontent.com/11957155/233526415-a0a44963-19a3-4e56-965a-afaa598e6127.gif" width="400" />
</p>
:fire: **Related projects:**
* [X-Decoder](https://github.com/microsoft/X-Decoder) : Generic decoder that can do multiple tasks with one model only**We built SEEM based on X-Decoder**.
* [FocalNet](https://github.com/microsoft/FocalNet) : Focal Modulation Networks; **We used FocalNet as the vision backbone**.
* [UniCL](https://github.com/microsoft/UniCL) : Unified Contrasative Learning; **We used this technique for image-text contrastive larning**.
:fire: **Other projects you may find interesting:**
* [OpenSeed](https://github.com/IDEA-Research/OpenSeeD) : Strong open-set segmentation methods.
* [Grounding SAM](https://github.com/IDEA-Research/Grounded-Segment-Anything) : Combining Grounding DINO and Segment Anythin.
* [LLaVA](https://github.com/haotian-liu/LLaVA) : Large Language and Vision Assistant.
## :bulb: Highlights