From b0c4787775f973c50a85a650410e37b093c560e4 Mon Sep 17 00:00:00 2001 From: MaureenZOU Date: Wed, 4 Oct 2023 16:59:23 -0500 Subject: [PATCH] update --- README.md | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-) diff --git a/README.md b/README.md index b4fcc89..e4b50a7 100644 --- a/README.md +++ b/README.md @@ -44,12 +44,6 @@ git clone git@github.com:UX-Decoder/Segment-Everything-Everywhere-All-At-Once.gi

-## :bookmark_tabs: Catelog -- [x] SEEM Demo -- [x] Inference and Installation Code -- [x] Evaluation Code -- [x] Training Code - :fire: **Related projects:** * [FocalNet](https://github.com/microsoft/FocalNet) and [DaViT](https://github.com/dingmyu/davit) : We used FocalNet and DaViT as the vision backbones. @@ -63,6 +57,12 @@ git clone git@github.com:UX-Decoder/Segment-Everything-Everywhere-All-At-Once.gi * [X-GPT](https://github.com/microsoft/X-Decoder/tree/xgpt) : Conversational Visual Agent supported by X-Decoder. * [LLaVA](https://github.com/haotian-liu/LLaVA) : Large Language and Vision Assistant. +## :bookmark_tabs: Catelog +- [x] SEEM Demo +- [x] Inference and Installation Code +- [x] Evaluation Code +- [x] Training Code + ## :bulb: Highlights Inspired by the appealing universal interface in LLMs, we are advocating a universal, interactive multi-modal interface for any type of segmentation with **ONE SINGLE MODEL**. We emphasize **4** important features of **SEEM** below. 1. **Versatility**: work with various types of prompts, for example, clicks, boxes, polygons, scribbles, texts, and referring image;