Commit Graph

16 Commits (d35c778a6f6c5c85cf56ad8aba2892fff97861b4)

Author SHA1 Message Date
mzr1996 d35c778a6f Merge remote-tracking branch 'origin/main' into dev 2023-10-12 17:19:27 +08:00
hmtbgc c0766519b1
[Feature] Add minigpt4 gradio demo and training script. (#1758)
* Add minigpt4 gradio demo

* update minigpt4 demo

* update minigpt4 demo (inference with float16)

* update minigpt4 and some dependent files

* add minigpt4 dataset for training

* add training script for minigpt4

* restore files deleted by mistake

* fix an error

* remove useless modification

* provide command line arguments for minigpt4 gradio demo and update some comments

* update code

* Update minigpt-4 readme

---------

Co-authored-by: mzr1996 <mzr1996@163.com>
2023-10-12 10:36:17 +08:00
mzr1996 b0a792eb08 Update OFA to compat with latest huggingface. 2023-10-11 11:13:40 +08:00
Coobiw bb59c9ad82
[Feature] Implement of Zero-Shot CLIP Classifier (#1737)
* zero-shot CLIP

* modify zero-shot clip config

* add in1k_sub_prompt(8 prompts) for improvement

* add some annotations doc

* clip base class & clip_zs sub-class

* some modifications of details after review

* convert into and use mmpretrain-vit

* modify names of some files and directories
2023-09-04 10:30:28 +08:00
fanqiNO1 29d706248c
[Enhancement] Support training of BLIP2 (#1700)
* [Fix] Fix BEiT pre_norm

* [Enhancement] Support BLIP2 training

* [Fix] Fix quoted strings

* [Fix] Fix init_weights

* [Fix] Fix with_cls_token

* [Fix] Fix tokenizer

* [Fix] Fix quoted strings

* [Fix] Fix predict

* [Fix] Cancel changing BEiT

* [Fix] Add loading hook

* [Fix] Reformat with yapf

* [Fix] Fix prompt

* [Fix] Fix typo
2023-08-10 11:15:38 +08:00
fangyixiao18 58a2243d99 Merge branch 'main' into dev 2023-07-28 15:35:55 +08:00
Yixiao Fang 1f99279657
[Fix] Fix dict update in minigpt4. (#1709) 2023-07-28 15:30:30 +08:00
Yixiao Fang b1cd05caf2
[Enhance] Set 'is_init' in some multimodal methods (#1718)
* update is_init of multimodal

* Update minigpt4.py

---------

Co-authored-by: Ma Zerun <mzr1996@163.com>
2023-07-28 15:28:07 +08:00
mzr1996 60d780f99e Fix docs 2023-07-20 10:21:15 +08:00
fanqiNO1 465b6bdeec
[Refactor] Fix spelling (#1689) 2023-07-13 15:38:58 +08:00
Ma Zerun 7d850dfadd
[Improve] Update Otter and LLaVA docs and config. (#1653) 2023-06-19 20:16:13 +08:00
Ma Zerun bfd49b0d52
[Feature] Support LLaVA (#1652) 2023-06-17 16:05:52 +08:00
Ma Zerun e69bace03f
[Feature] Support otter (#1651)
* [Feature] Support Otter

* Update docs
2023-06-17 16:03:21 +08:00
Yixiao Fang 9d3fc43073
[Feature] Support MiniGPT-4 (#1642)
* support inference of MiniGPT-4

* refine codes

* update metafile, readme and docs

* fix typo

* fix lint

* add ckpt load hook
2023-06-16 22:50:34 +08:00
Yixiao Fang 1e478462b8
[Feature] Support Chinese CLIP. (#1576)
* support cn-clip

* update README

* Update progress bar

* update order of category

* fix lint

* update

* update readme and metafile

* update

* update docstring

* refactor tokenizer

* fix lint

* Update README and progress bar

---------

Co-authored-by: mzr1996 <mzr1996@163.com>
2023-05-22 15:46:13 +08:00
Ma Zerun 6847d20d57
[Feature] Support multiple multi-modal algorithms and inferencers. (#1561)
* [Feat] Migrate blip caption to mmpretrain. (#50)

* Migrate blip caption to mmpretrain

* minor fix

* support train

* [Feature] Support OFA caption task. (#51)

* [Feature] Support OFA caption task.

* Remove duplicated files.

* [Feature] Support OFA vqa task. (#58)

* [Feature] Support OFA vqa task.

* Fix lint.

* [Feat] Add BLIP retrieval to mmpretrain. (#55)

* init

* minor fix for train

* fix according to comments

* refactor

* Update Blip retrieval. (#62)

* [Feature] Support OFA visual grounding task. (#59)

* [Feature] Support OFA visual grounding task.

* minor add TODO

---------

Co-authored-by: yingfhu <yingfhu@gmail.com>

* [Feat] Add flamingos coco caption and vqa. (#60)

* first init

* init flamingo coco

* add vqa

* minor fix

* remove unnecessary modules

* Update config

* Use `ApplyToList`.

---------

Co-authored-by: mzr1996 <mzr1996@163.com>

* [Feature]: BLIP2 coco retrieval  (#53)

* [Feature]: Add blip2 retriever

* [Feature]: Add blip2 all modules

* [Feature]: Refine model

* [Feature]: x1

* [Feature]: Runnable coco ret

* [Feature]: Runnable version

* [Feature]: Fix lint

* [Fix]: Fix lint

* [Feature]: Use 364 img size

* [Feature]: Refactor blip2

* [Fix]: Fix lint

* refactor files

* minor fix

* minor fix

---------

Co-authored-by: yingfhu <yingfhu@gmail.com>

* Remove

* fix blip caption inputs (#68)

* [Feat] Add BLIP NLVR support. (#67)

* first init

* init flamingo coco

* add vqa

* add nlvr

* refactor nlvr

* minor fix

* minor fix

* Update dataset

---------

Co-authored-by: mzr1996 <mzr1996@163.com>

* [Feature]: BLIP2 Caption (#70)

* [Feature]: Add language model

* [Feature]: blip2 caption forward

* [Feature]: Reproduce the results

* [Feature]: Refactor caption

* refine config

---------

Co-authored-by: yingfhu <yingfhu@gmail.com>

* [Feat] Migrate BLIP VQA to mmpretrain (#69)

* reformat

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* change

* refactor code

---------

Co-authored-by: yingfhu <yingfhu@gmail.com>

* Update RefCOCO dataset

* [Fix] fix lint

* [Feature] Implement inference APIs for multi-modal tasks. (#65)

* [Feature] Implement inference APIs for multi-modal tasks.

* [Project] Add gradio demo.

* [Improve] Update requirements

* Update flamingo

* Update blip

* Add NLVR inferencer

* Update flamingo

* Update hugging face model register

* Update ofa vqa

* Update BLIP-vqa (#71)

* Update blip-vqa docstring (#72)

* Refine flamingo docstring (#73)

* [Feature]: BLIP2 VQA (#61)

* [Feature]: VQA forward

* [Feature]: Reproduce accuracy

* [Fix]: Fix lint

* [Fix]: Add blank line

* minor fix

---------

Co-authored-by: yingfhu <yingfhu@gmail.com>

* [Feature]: BLIP2 docstring (#74)

* [Feature]: Add caption docstring

* [Feature]: Add docstring to blip2 vqa

* [Feature]: Add docstring to retrieval

* Update BLIP-2 metafile and README (#75)

* [Feature]: Add readme and docstring

* Update blip2 results

---------

Co-authored-by: mzr1996 <mzr1996@163.com>

* [Feature] BLIP Visual Grounding on MMPretrain Branch (#66)

* blip grounding merge with mmpretrain

* remove commit

* blip grounding test and inference api

* refcoco dataset

* refcoco dataset refine config

* rebasing

* gitignore

* rebasing

* minor edit

* minor edit

* Update blip-vqa docstring (#72)

* rebasing

* Revert "minor edit"

This reverts commit 639cec757c215e654625ed0979319e60f0be9044.

* blip grounding final

* precommit

* refine config

* refine config

* Update blip visual grounding

---------

Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: mzr1996 <mzr1996@163.com>

* Update visual grounding metric

* Update OFA docstring, README and metafiles. (#76)

* [Docs] Update installation docs and gradio demo docs. (#77)

* Update OFA name

* Update Visual Grounding Visualizer

* Integrate accelerate support

* Fix imports.

* Fix timm backbone

* Update imports

* Update README

* Update circle ci

* Update flamingo config

* Add gradio demo README

* [Feature]: Add scienceqa (#1571)

* [Feature]: Add scienceqa

* [Feature]: Change param name

* Update docs

* Update video

---------

Co-authored-by: Hubert <42952108+yingfhu@users.noreply.github.com>
Co-authored-by: yingfhu <yingfhu@gmail.com>
Co-authored-by: Yuan Liu <30762564+YuanLiuuuuuu@users.noreply.github.com>
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: Rongjie Li <limo97@163.com>
2023-05-19 16:50:04 +08:00