Update README.md
parent
ef010a0d0b
commit
3311d6df73
24
README.md
24
README.md
|
@ -4,6 +4,19 @@ We introduce **SEEM** that can **S**egment **E**verything **E**verywhere with **
|
|||
|
||||
:grapes:\[[Read our arXiv Paper](https://arxiv.org/pdf/2304.06718.pdf)\] :apple:\[[Try Hugging Face Demo](https://huggingface.co/spaces/xdecoder/SEEM)\]
|
||||
|
||||
:fire: **Related projects:**
|
||||
|
||||
* [FocalNet](https://github.com/microsoft/FocalNet) : Focal Modulation Networks; **We used FocalNet as the vision backbone**.
|
||||
* [UniCL](https://github.com/microsoft/UniCL) : Unified Contrasative Learning; **We used this technique for image-text contrastive larning**.
|
||||
* [X-Decoder](https://github.com/microsoft/X-Decoder) : Generic decoder that can do multiple tasks with one model only;**We built SEEM based on X-Decoder**.
|
||||
|
||||
:fire: **Other projects you may find interesting:**
|
||||
|
||||
* [OpenSeed](https://github.com/IDEA-Research/OpenSeeD) : Strong open-set segmentation methods.
|
||||
* [Grounding SAM](https://github.com/IDEA-Research/Grounded-Segment-Anything) : Combining Grounding DINO and Segment Anythin.
|
||||
* [LLaVA](https://github.com/haotian-liu/LLaVA) : Large Language and Vision Assistant.
|
||||
|
||||
|
||||
## :rocket: Updates
|
||||
* Try our [Video Demo (Beta)](https://083193edbb48f717.gradio.app) on referring video object segmentation.
|
||||
<p float="left">
|
||||
|
@ -11,17 +24,6 @@ We introduce **SEEM** that can **S**egment **E**verything **E**verywhere with **
|
|||
<img src="https://user-images.githubusercontent.com/11957155/233526415-a0a44963-19a3-4e56-965a-afaa598e6127.gif" width="400" />
|
||||
</p>
|
||||
|
||||
:fire: **Related projects:**
|
||||
|
||||
* [X-Decoder](https://github.com/microsoft/X-Decoder) : Generic decoder that can do multiple tasks with one model only;**We built SEEM based on X-Decoder**.
|
||||
* [FocalNet](https://github.com/microsoft/FocalNet) : Focal Modulation Networks; **We used FocalNet as the vision backbone**.
|
||||
* [UniCL](https://github.com/microsoft/UniCL) : Unified Contrasative Learning; **We used this technique for image-text contrastive larning**.
|
||||
|
||||
:fire: **Other projects you may find interesting:**
|
||||
|
||||
* [OpenSeed](https://github.com/IDEA-Research/OpenSeeD) : Strong open-set segmentation methods.
|
||||
* [Grounding SAM](https://github.com/IDEA-Research/Grounded-Segment-Anything) : Combining Grounding DINO and Segment Anythin.
|
||||
* [LLaVA](https://github.com/haotian-liu/LLaVA) : Large Language and Vision Assistant.
|
||||
|
||||
## :bulb: Highlights
|
||||
|
||||
|
|
Loading…
Reference in New Issue