Learning Open-World Object Proposals without Learning to Classify
 
 
Go to file
Dahun Kim 7636de2df1 minor fix 2021-08-31 20:48:51 +09:00
configs minor fix 2021-08-31 20:48:51 +09:00
docker first commit 2021-08-30 01:36:15 +09:00
docs first commit 2021-08-30 01:36:15 +09:00
images add figures 2021-08-30 12:21:16 +09:00
mmdet minor fix 2021-08-31 20:48:51 +09:00
requirements first commit 2021-08-30 01:36:15 +09:00
resources first commit 2021-08-30 01:36:15 +09:00
tests first commit 2021-08-30 01:36:15 +09:00
tools first commit 2021-08-30 12:07:34 +09:00
.gitignore first commit 2021-08-30 12:07:34 +09:00
LICENSE first commit 2021-08-30 01:36:15 +09:00
README.md readme 2021-08-31 19:18:31 +09:00
pytest.ini first commit 2021-08-30 01:36:15 +09:00
requirements.txt first commit 2021-08-30 01:36:15 +09:00
setup.cfg first commit 2021-08-30 01:36:15 +09:00
setup.py first commit 2021-08-30 01:36:15 +09:00

README.md

Learning Open-World Object Proposals without Learning to Classify

Pytorch implementation for "Learning Open-World Object Proposals without Learning to Classify" (arXiv 2021)

Dahun Kim, Tsung-Yi Lin, Anelia Angelova, In So Kweon, and Weicheng Kuo.

@article{kim2021oln,
  title={Learning Open-World Object Proposals without Learning to Classify},
  author={Kim, Dahun and Lin, Tsung-Yi and Angelova, Anelia and Kweon, In So and Kuo, Weicheng},
  journal={arXiv preprint arXiv:2108.06753},
  year={2021}
}

Introduction

Humans can recognize novel objects in this image despite having never seen them before. “Is it possible to learn open-world (novel) object proposals?” In this paper we propose Object Localization Network (OLN) that learns localization cues instead of foreground vs background classification. Only trained on COCO, OLN is able to propose many novel objects (top) missed by Mask R-CNN (bottom) on an out-of-sample frame in an ego-centric video.


Disclaimer

This repo is tested under Python 3.7, PyTorch 1.7.0, Cuda 11.0, and mmcv==1.2.5.

Installation

This repo is built based on mmdetection.

You can use following commands to create conda env with related dependencies.

conda create -n oln python=3.7 -y
conda activate oln
conda install pytorch=1.7.0 torchvision cudatoolkit=11.0 -c pytorch -y
pip install mmcv-full
pip install -r requirements.txt
pip install -v -e . 

Please also refer to get_started.md for more details of installation.

Prepare datasets

COCO dataset is available from official websites. It is recommended to download and extract the dataset somewhere outside the project directory and symlink the dataset root to $OLN/data as below.

object_localization_network
├── mmdet
├── tools
├── configs
├── data
│   ├── coco
│   │   ├── annotations
│   │   ├── train2017
│   │   ├── val2017
│   │   ├── test2017

Testing

Our trained models are available for download here. Place it under trained_weights/latest.pth and run the following commands to test OLN on COCO dataset.

# Multi-GPU distributed testing
bash tools/dist_test_bbox.sh configs/oln_box/oln_box.py \
trained_weights/latest.pth ${NUM_GPUS}
# OR
python tools/test.py configs/oln_box/oln_box.py work_dirs/oln_box/latest.pth --eval bbox

Training

# Multi-GPU distributed training
bash tools/dist_train.sh configs/oln_box/oln_box.py ${NUM_GPUS}

Contact

If you have any questions regarding the repo, please contact Dahun Kim (mcahny01@gmail.com) or create an issue.