fanqiNO1
|
7cbfb36c14
|
[Refactor] Fix spelling (#1681)
* [Refactor] Fix spelling
* [Refactor] Fix spelling
* [Refactor] Fix spelling
* [Refactor] Fix spelling
|
2023-07-05 11:07:43 +08:00 |
Yixiao Fang
|
d4a6dfa00a
|
Add benchmark options (#1654)
* update dev_scripts
* update metafile
* update multimodal floating range
* fix lint
* update
* update
* fix lint
* Update metric map
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
|
2023-06-20 14:18:57 +08:00 |
Yiqin Wang 王逸钦
|
6d7fe91a98
|
[Feature] Support Flickr30k Retrieval dataset (#1625)
* format
* remove abs path
* init add flickr30k caption
* remove abs dir
* update blip readme
* add convert sscripts
* minor
* minor
|
2023-06-19 15:15:03 +08:00 |
Yiqin Wang 王逸钦
|
bb415b91be
|
[Feature] Support OCR-VQA dataset (#1621)
* support ocrvqa dataset
* minor
* remove abs path
* refine README
|
2023-06-13 10:28:45 +08:00 |
Yiqin Wang 王逸钦
|
dbfb84ccbd
|
[Feature] Support OK-VQA dataset (#1615)
* add okvqa
* refine README
|
2023-06-08 16:57:18 +08:00 |
Wangbo Zhao(黑色枷锁)
|
3a277ee9e6
|
[Feature] support TextVQA dataset (#1596)
* [Support] Suport TextVQA dataset
* add folder structure
* fix readme
|
2023-06-02 11:50:38 +08:00 |
Wangbo Zhao(黑色枷锁)
|
a779c8c5a7
|
[Feature] Support NoCap dataset based on BLIP. (#1582)
* [Feature] Support nocaps dataset
* precommit
* Use official coco format
* add nocp readme
* fix readme
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
|
2023-05-23 18:06:43 +08:00 |
Ma Zerun
|
6847d20d57
|
[Feature] Support multiple multi-modal algorithms and inferencers. (#1561)
* [Feat] Migrate blip caption to mmpretrain. (#50)
* Migrate blip caption to mmpretrain
* minor fix
* support train
* [Feature] Support OFA caption task. (#51)
* [Feature] Support OFA caption task.
* Remove duplicated files.
* [Feature] Support OFA vqa task. (#58)
* [Feature] Support OFA vqa task.
* Fix lint.
* [Feat] Add BLIP retrieval to mmpretrain. (#55)
* init
* minor fix for train
* fix according to comments
* refactor
* Update Blip retrieval. (#62)
* [Feature] Support OFA visual grounding task. (#59)
* [Feature] Support OFA visual grounding task.
* minor add TODO
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feat] Add flamingos coco caption and vqa. (#60)
* first init
* init flamingo coco
* add vqa
* minor fix
* remove unnecessary modules
* Update config
* Use `ApplyToList`.
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature]: BLIP2 coco retrieval (#53)
* [Feature]: Add blip2 retriever
* [Feature]: Add blip2 all modules
* [Feature]: Refine model
* [Feature]: x1
* [Feature]: Runnable coco ret
* [Feature]: Runnable version
* [Feature]: Fix lint
* [Fix]: Fix lint
* [Feature]: Use 364 img size
* [Feature]: Refactor blip2
* [Fix]: Fix lint
* refactor files
* minor fix
* minor fix
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* Remove
* fix blip caption inputs (#68)
* [Feat] Add BLIP NLVR support. (#67)
* first init
* init flamingo coco
* add vqa
* add nlvr
* refactor nlvr
* minor fix
* minor fix
* Update dataset
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature]: BLIP2 Caption (#70)
* [Feature]: Add language model
* [Feature]: blip2 caption forward
* [Feature]: Reproduce the results
* [Feature]: Refactor caption
* refine config
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feat] Migrate BLIP VQA to mmpretrain (#69)
* reformat
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* refactor code
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* Update RefCOCO dataset
* [Fix] fix lint
* [Feature] Implement inference APIs for multi-modal tasks. (#65)
* [Feature] Implement inference APIs for multi-modal tasks.
* [Project] Add gradio demo.
* [Improve] Update requirements
* Update flamingo
* Update blip
* Add NLVR inferencer
* Update flamingo
* Update hugging face model register
* Update ofa vqa
* Update BLIP-vqa (#71)
* Update blip-vqa docstring (#72)
* Refine flamingo docstring (#73)
* [Feature]: BLIP2 VQA (#61)
* [Feature]: VQA forward
* [Feature]: Reproduce accuracy
* [Fix]: Fix lint
* [Fix]: Add blank line
* minor fix
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feature]: BLIP2 docstring (#74)
* [Feature]: Add caption docstring
* [Feature]: Add docstring to blip2 vqa
* [Feature]: Add docstring to retrieval
* Update BLIP-2 metafile and README (#75)
* [Feature]: Add readme and docstring
* Update blip2 results
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature] BLIP Visual Grounding on MMPretrain Branch (#66)
* blip grounding merge with mmpretrain
* remove commit
* blip grounding test and inference api
* refcoco dataset
* refcoco dataset refine config
* rebasing
* gitignore
* rebasing
* minor edit
* minor edit
* Update blip-vqa docstring (#72)
* rebasing
* Revert "minor edit"
This reverts commit 639cec757c215e654625ed0979319e60f0be9044.
* blip grounding final
* precommit
* refine config
* refine config
* Update blip visual grounding
---------
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: mzr1996 <mzr1996@163.com>
* Update visual grounding metric
* Update OFA docstring, README and metafiles. (#76)
* [Docs] Update installation docs and gradio demo docs. (#77)
* Update OFA name
* Update Visual Grounding Visualizer
* Integrate accelerate support
* Fix imports.
* Fix timm backbone
* Update imports
* Update README
* Update circle ci
* Update flamingo config
* Add gradio demo README
* [Feature]: Add scienceqa (#1571)
* [Feature]: Add scienceqa
* [Feature]: Change param name
* Update docs
* Update video
---------
Co-authored-by: Hubert <42952108+yingfhu@users.noreply.github.com>
Co-authored-by: yingfhu <yingfhu@gmail.com>
Co-authored-by: Yuan Liu <30762564+YuanLiuuuuuu@users.noreply.github.com>
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: Rongjie Li <limo97@163.com>
|
2023-05-19 16:50:04 +08:00 |