Coobiw
ed5924b6fe
[Feature] Implement of RAM with a gradio interface. ( #1802 )
...
* [CodeCamp2023-584]Support DINO self-supervised learning in project (#1756 )
* feat: impelemt DINO
* chore: delete debug code
* chore: impplement pre-commit
* fix: fix imported package
* chore: pre-commit check
* [CodeCamp2023-340] New Version of config Adapting MobileNet Algorithm (#1774 )
* add new config adapting MobileNetV2,V3
* add base model config for mobile net v3, modified all training configs of mobile net v3 inherit from the base model config
* removed directory _base_/models/mobilenet_v3
* [Feature] Implement of Zero-Shot CLIP Classifier (#1737 )
* zero-shot CLIP
* modify zero-shot clip config
* add in1k_sub_prompt(8 prompts) for improvement
* add some annotations doc
* clip base class & clip_zs sub-class
* some modifications of details after review
* convert into and use mmpretrain-vit
* modify names of some files and directories
* ram init commit
* [Fix] Fix pipeline bug in image retrieval inferencer
* [CodeCamp2023-341] 多模态数据集文档补充-COCO Retrieval
* Update OFA to compat with latest huggingface.
* Update train.py to compat with new config
* Bump version to v1.1.0
* Update __init__.py
---------
Co-authored-by: LALBJ <40877073+LALBJ@users.noreply.github.com>
Co-authored-by: DE009 <57087096+DE009@users.noreply.github.com>
Co-authored-by: mzr1996 <mzr1996@163.com>
Co-authored-by: 飞飞 <102729089+ASHORE1225@users.noreply.github.com>
2023-10-25 16:23:45 +08:00
fanqiNO1
7cbfb36c14
[Refactor] Fix spelling ( #1681 )
...
* [Refactor] Fix spelling
* [Refactor] Fix spelling
* [Refactor] Fix spelling
* [Refactor] Fix spelling
2023-07-05 11:07:43 +08:00
Ma Zerun
6847d20d57
[Feature] Support multiple multi-modal algorithms and inferencers. ( #1561 )
...
* [Feat] Migrate blip caption to mmpretrain. (#50 )
* Migrate blip caption to mmpretrain
* minor fix
* support train
* [Feature] Support OFA caption task. (#51 )
* [Feature] Support OFA caption task.
* Remove duplicated files.
* [Feature] Support OFA vqa task. (#58 )
* [Feature] Support OFA vqa task.
* Fix lint.
* [Feat] Add BLIP retrieval to mmpretrain. (#55 )
* init
* minor fix for train
* fix according to comments
* refactor
* Update Blip retrieval. (#62 )
* [Feature] Support OFA visual grounding task. (#59 )
* [Feature] Support OFA visual grounding task.
* minor add TODO
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feat] Add flamingos coco caption and vqa. (#60 )
* first init
* init flamingo coco
* add vqa
* minor fix
* remove unnecessary modules
* Update config
* Use `ApplyToList`.
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature]: BLIP2 coco retrieval (#53 )
* [Feature]: Add blip2 retriever
* [Feature]: Add blip2 all modules
* [Feature]: Refine model
* [Feature]: x1
* [Feature]: Runnable coco ret
* [Feature]: Runnable version
* [Feature]: Fix lint
* [Fix]: Fix lint
* [Feature]: Use 364 img size
* [Feature]: Refactor blip2
* [Fix]: Fix lint
* refactor files
* minor fix
* minor fix
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* Remove
* fix blip caption inputs (#68 )
* [Feat] Add BLIP NLVR support. (#67 )
* first init
* init flamingo coco
* add vqa
* add nlvr
* refactor nlvr
* minor fix
* minor fix
* Update dataset
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature]: BLIP2 Caption (#70 )
* [Feature]: Add language model
* [Feature]: blip2 caption forward
* [Feature]: Reproduce the results
* [Feature]: Refactor caption
* refine config
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feat] Migrate BLIP VQA to mmpretrain (#69 )
* reformat
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* refactor code
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* Update RefCOCO dataset
* [Fix] fix lint
* [Feature] Implement inference APIs for multi-modal tasks. (#65 )
* [Feature] Implement inference APIs for multi-modal tasks.
* [Project] Add gradio demo.
* [Improve] Update requirements
* Update flamingo
* Update blip
* Add NLVR inferencer
* Update flamingo
* Update hugging face model register
* Update ofa vqa
* Update BLIP-vqa (#71 )
* Update blip-vqa docstring (#72 )
* Refine flamingo docstring (#73 )
* [Feature]: BLIP2 VQA (#61 )
* [Feature]: VQA forward
* [Feature]: Reproduce accuracy
* [Fix]: Fix lint
* [Fix]: Add blank line
* minor fix
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feature]: BLIP2 docstring (#74 )
* [Feature]: Add caption docstring
* [Feature]: Add docstring to blip2 vqa
* [Feature]: Add docstring to retrieval
* Update BLIP-2 metafile and README (#75 )
* [Feature]: Add readme and docstring
* Update blip2 results
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature] BLIP Visual Grounding on MMPretrain Branch (#66 )
* blip grounding merge with mmpretrain
* remove commit
* blip grounding test and inference api
* refcoco dataset
* refcoco dataset refine config
* rebasing
* gitignore
* rebasing
* minor edit
* minor edit
* Update blip-vqa docstring (#72 )
* rebasing
* Revert "minor edit"
This reverts commit 639cec757c215e654625ed0979319e60f0be9044.
* blip grounding final
* precommit
* refine config
* refine config
* Update blip visual grounding
---------
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: mzr1996 <mzr1996@163.com>
* Update visual grounding metric
* Update OFA docstring, README and metafiles. (#76 )
* [Docs] Update installation docs and gradio demo docs. (#77 )
* Update OFA name
* Update Visual Grounding Visualizer
* Integrate accelerate support
* Fix imports.
* Fix timm backbone
* Update imports
* Update README
* Update circle ci
* Update flamingo config
* Add gradio demo README
* [Feature]: Add scienceqa (#1571 )
* [Feature]: Add scienceqa
* [Feature]: Change param name
* Update docs
* Update video
---------
Co-authored-by: Hubert <42952108+yingfhu@users.noreply.github.com>
Co-authored-by: yingfhu <yingfhu@gmail.com>
Co-authored-by: Yuan Liu <30762564+YuanLiuuuuuu@users.noreply.github.com>
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: Rongjie Li <limo97@163.com>
2023-05-19 16:50:04 +08:00
Yixiao Fang
1ee9bbe050
[Docs] Update links ( #1457 )
...
* update links
* update readtherdocs
* update
* update
* fix lint
* update
* update
* update
* update cov branch
* update
* update
* update
2023-04-06 20:58:52 +08:00
Ma Zerun
6cedce234e
[Refactor] Update dev scripts to be compatible with selfsup tasks. ( #1412 )
...
* [Refactor] Update dev scripts to be compatible with selfsup tasks.
* Fix some missing fields in config files.
* Set maximum number of gpus for local training.
* Update README files
* Update according to comments.
2023-03-20 14:30:57 +08:00
Ma Zerun
4f5b38f225
[Refactor] Update almost tools and add unit tests for these tools. ( #1418 )
...
* [Refactor] Update almost tools and add unit tests for these tools.
* Fix Windows UT.
2023-03-17 10:50:51 +08:00
Ma Zerun
dda3d6565b
[Docs] Update generate_readme.py and readme files. ( #1388 )
...
* Update generate_readme.py and readme files.
* Update reamde
* Update docs
* update metafile
* update simmim readme
* update
* update mae
* fix lint
* update mocov2
* update readme pic
* fix lint
* Fix mmcls download links.
* Fix Chinese docs.
* Decrease readthedocs requirements.
---------
Co-authored-by: fangyixiao18 <fangyx18@hotmail.com>
2023-03-02 13:29:07 +08:00
Rongjie Li
5b266d9e7c
[Feature] Add clip backbone. ( #1258 )
...
* clip backbone added
* passed precommit
* update readme
* update according to PR review
* add missing file
* add unittest
* refine metafile
* refine metafile and readme for readdocs
* refine metafile
* refine metafile
* Update metafile
Co-authored-by: mzr1996 <mzr1996@163.com>
2022-12-22 16:33:57 +08:00