mzr1996
d35c778a6f
Merge remote-tracking branch 'origin/main' into dev
2023-10-12 17:19:27 +08:00
hmtbgc
c0766519b1
[Feature] Add minigpt4 gradio demo and training script. ( #1758 )
...
* Add minigpt4 gradio demo
* update minigpt4 demo
* update minigpt4 demo (inference with float16)
* update minigpt4 and some dependent files
* add minigpt4 dataset for training
* add training script for minigpt4
* restore files deleted by mistake
* fix an error
* remove useless modification
* provide command line arguments for minigpt4 gradio demo and update some comments
* update code
* Update minigpt-4 readme
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
2023-10-12 10:36:17 +08:00
mzr1996
b0a792eb08
Update OFA to compat with latest huggingface.
2023-10-11 11:13:40 +08:00
Coobiw
bb59c9ad82
[Feature] Implement of Zero-Shot CLIP Classifier ( #1737 )
...
* zero-shot CLIP
* modify zero-shot clip config
* add in1k_sub_prompt(8 prompts) for improvement
* add some annotations doc
* clip base class & clip_zs sub-class
* some modifications of details after review
* convert into and use mmpretrain-vit
* modify names of some files and directories
2023-09-04 10:30:28 +08:00
fanqiNO1
29d706248c
[Enhancement] Support training of BLIP2 ( #1700 )
...
* [Fix] Fix BEiT pre_norm
* [Enhancement] Support BLIP2 training
* [Fix] Fix quoted strings
* [Fix] Fix init_weights
* [Fix] Fix with_cls_token
* [Fix] Fix tokenizer
* [Fix] Fix quoted strings
* [Fix] Fix predict
* [Fix] Cancel changing BEiT
* [Fix] Add loading hook
* [Fix] Reformat with yapf
* [Fix] Fix prompt
* [Fix] Fix typo
2023-08-10 11:15:38 +08:00
fangyixiao18
58a2243d99
Merge branch 'main' into dev
2023-07-28 15:35:55 +08:00
Yixiao Fang
1f99279657
[Fix] Fix dict update in minigpt4. ( #1709 )
2023-07-28 15:30:30 +08:00
Yixiao Fang
b1cd05caf2
[Enhance] Set 'is_init' in some multimodal methods ( #1718 )
...
* update is_init of multimodal
* Update minigpt4.py
---------
Co-authored-by: Ma Zerun <mzr1996@163.com>
2023-07-28 15:28:07 +08:00
mzr1996
60d780f99e
Fix docs
2023-07-20 10:21:15 +08:00
fanqiNO1
465b6bdeec
[Refactor] Fix spelling ( #1689 )
2023-07-13 15:38:58 +08:00
Ma Zerun
7d850dfadd
[Improve] Update Otter and LLaVA docs and config. ( #1653 )
2023-06-19 20:16:13 +08:00
Ma Zerun
bfd49b0d52
[Feature] Support LLaVA ( #1652 )
2023-06-17 16:05:52 +08:00
Ma Zerun
e69bace03f
[Feature] Support otter ( #1651 )
...
* [Feature] Support Otter
* Update docs
2023-06-17 16:03:21 +08:00
Yixiao Fang
9d3fc43073
[Feature] Support MiniGPT-4 ( #1642 )
...
* support inference of MiniGPT-4
* refine codes
* update metafile, readme and docs
* fix typo
* fix lint
* add ckpt load hook
2023-06-16 22:50:34 +08:00
Yixiao Fang
1e478462b8
[Feature] Support Chinese CLIP. ( #1576 )
...
* support cn-clip
* update README
* Update progress bar
* update order of category
* fix lint
* update
* update readme and metafile
* update
* update docstring
* refactor tokenizer
* fix lint
* Update README and progress bar
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
2023-05-22 15:46:13 +08:00
Ma Zerun
6847d20d57
[Feature] Support multiple multi-modal algorithms and inferencers. ( #1561 )
...
* [Feat] Migrate blip caption to mmpretrain. (#50 )
* Migrate blip caption to mmpretrain
* minor fix
* support train
* [Feature] Support OFA caption task. (#51 )
* [Feature] Support OFA caption task.
* Remove duplicated files.
* [Feature] Support OFA vqa task. (#58 )
* [Feature] Support OFA vqa task.
* Fix lint.
* [Feat] Add BLIP retrieval to mmpretrain. (#55 )
* init
* minor fix for train
* fix according to comments
* refactor
* Update Blip retrieval. (#62 )
* [Feature] Support OFA visual grounding task. (#59 )
* [Feature] Support OFA visual grounding task.
* minor add TODO
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feat] Add flamingos coco caption and vqa. (#60 )
* first init
* init flamingo coco
* add vqa
* minor fix
* remove unnecessary modules
* Update config
* Use `ApplyToList`.
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature]: BLIP2 coco retrieval (#53 )
* [Feature]: Add blip2 retriever
* [Feature]: Add blip2 all modules
* [Feature]: Refine model
* [Feature]: x1
* [Feature]: Runnable coco ret
* [Feature]: Runnable version
* [Feature]: Fix lint
* [Fix]: Fix lint
* [Feature]: Use 364 img size
* [Feature]: Refactor blip2
* [Fix]: Fix lint
* refactor files
* minor fix
* minor fix
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* Remove
* fix blip caption inputs (#68 )
* [Feat] Add BLIP NLVR support. (#67 )
* first init
* init flamingo coco
* add vqa
* add nlvr
* refactor nlvr
* minor fix
* minor fix
* Update dataset
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature]: BLIP2 Caption (#70 )
* [Feature]: Add language model
* [Feature]: blip2 caption forward
* [Feature]: Reproduce the results
* [Feature]: Refactor caption
* refine config
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feat] Migrate BLIP VQA to mmpretrain (#69 )
* reformat
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* change
* refactor code
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* Update RefCOCO dataset
* [Fix] fix lint
* [Feature] Implement inference APIs for multi-modal tasks. (#65 )
* [Feature] Implement inference APIs for multi-modal tasks.
* [Project] Add gradio demo.
* [Improve] Update requirements
* Update flamingo
* Update blip
* Add NLVR inferencer
* Update flamingo
* Update hugging face model register
* Update ofa vqa
* Update BLIP-vqa (#71 )
* Update blip-vqa docstring (#72 )
* Refine flamingo docstring (#73 )
* [Feature]: BLIP2 VQA (#61 )
* [Feature]: VQA forward
* [Feature]: Reproduce accuracy
* [Fix]: Fix lint
* [Fix]: Add blank line
* minor fix
---------
Co-authored-by: yingfhu <yingfhu@gmail.com>
* [Feature]: BLIP2 docstring (#74 )
* [Feature]: Add caption docstring
* [Feature]: Add docstring to blip2 vqa
* [Feature]: Add docstring to retrieval
* Update BLIP-2 metafile and README (#75 )
* [Feature]: Add readme and docstring
* Update blip2 results
---------
Co-authored-by: mzr1996 <mzr1996@163.com>
* [Feature] BLIP Visual Grounding on MMPretrain Branch (#66 )
* blip grounding merge with mmpretrain
* remove commit
* blip grounding test and inference api
* refcoco dataset
* refcoco dataset refine config
* rebasing
* gitignore
* rebasing
* minor edit
* minor edit
* Update blip-vqa docstring (#72 )
* rebasing
* Revert "minor edit"
This reverts commit 639cec757c215e654625ed0979319e60f0be9044.
* blip grounding final
* precommit
* refine config
* refine config
* Update blip visual grounding
---------
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: mzr1996 <mzr1996@163.com>
* Update visual grounding metric
* Update OFA docstring, README and metafiles. (#76 )
* [Docs] Update installation docs and gradio demo docs. (#77 )
* Update OFA name
* Update Visual Grounding Visualizer
* Integrate accelerate support
* Fix imports.
* Fix timm backbone
* Update imports
* Update README
* Update circle ci
* Update flamingo config
* Add gradio demo README
* [Feature]: Add scienceqa (#1571 )
* [Feature]: Add scienceqa
* [Feature]: Change param name
* Update docs
* Update video
---------
Co-authored-by: Hubert <42952108+yingfhu@users.noreply.github.com>
Co-authored-by: yingfhu <yingfhu@gmail.com>
Co-authored-by: Yuan Liu <30762564+YuanLiuuuuuu@users.noreply.github.com>
Co-authored-by: Yiqin Wang 王逸钦 <wyq1217@outlook.com>
Co-authored-by: Rongjie Li <limo97@163.com>
2023-05-19 16:50:04 +08:00