Commit Graph

24 Commits (866a196d1937e267cca515ac942fb539df595505)

Author SHA1 Message Date
liaoxingyu 866a196d19 add more datasets support 2020-09-01 16:13:12 +08:00
liaoxingyu 57ebf75f83 fix bug about changing cfg values 2020-08-10 10:29:24 +08:00
liaoxingyu d516d59fb5 fix autoaug total_iter bug #219
max_epoch does not convert to max_iter when needing to build dataloader first
add extra convert in `build_reid_train_loader`
2020-08-09 19:43:01 +08:00
liaoxingyu d1c20cbe50 fix pre-train model bugs
fix bugs locks when downloading pre-train model
2020-08-04 15:56:36 +08:00
liaoxingyu 65169b40bd support ddp testing 2020-07-30 20:15:28 +08:00
liaoxingyu 3f35eb449d minor update 2020-07-14 11:58:06 +08:00
liaoxingyu b7a2b1c21a support DDP sampler and augment 2020-07-06 16:55:23 +08:00
liaoxingyu 8879db3fba update training instruction
Summary: update dataset configuration and training instruction
2020-06-16 19:43:36 +08:00
liaoxingyu ecc2b1a790 update naive sampler
Summary: update naive sampler which will introduce unbalanced sampling
2020-06-15 20:50:25 +08:00
liaoxingyu 84c733fa85 fix: remove prefetcher, put normalizer in model
1. remove messy data prefetcher which will cause  confusion
2. put normliazer in model to accelerate training via GPU computing
2020-05-25 23:39:11 +08:00
liaoxingyu fd90555e19 feat: add multi-dataset joint training
new feature that can support joint training, and find some bugs in funtion combine_all of datasets/bases.py that assume person id in dataset has been relabeld from 0 to num_class.
Another bug appears in msmt17 which trainset and testset person id both begin from 0, and we should change testset id from num_class of trainset.
2020-05-18 20:06:04 +08:00
liaoxingyu 4be4cacb73 fix: add a simple way to reset data prefetcher when resume training
use data prefetcher build-in reset function to reload it rather than
redefining a new data prefetcher, otherwise it will introduce other
problems in eval-only mode.
2020-05-09 11:58:27 +08:00
liaoxingyu 6d96529d4c fix(data): fix resume training bug
fix dataset pid dictionary loading bug when resume training,
data prefetcher will pre-load a batch of data, this will lead to
misalignment of old pid dict and updated pid dict.
We can address this problem by redefine a prefetcher in resume_or_load
2020-05-05 23:20:42 +08:00
liaoxingyu d27729c5bb refactor(preciseBN): add preciseBN datasets show 2020-04-29 21:05:53 +08:00
liaoxingyu 9684500a57 chagne arch
1. change dataset show to trainset show and testset show seperately
2. add cls layer to easily plug in circle loss and arcface
2020-04-19 12:54:01 +08:00
liaoxingyu be9faa5605 update focal loss
update dataset info display
update seperate lr
update adaptive label smooth regularization
2020-04-17 13:46:10 +08:00
L1aoXingyu b020c7f0ae Fix data prefetcher minor bug 2020-02-27 12:16:57 +08:00
L1aoXingyu 12957f66aa Change architecture:
1. delete redundant preprocess
2. add data prefetcher to accelerate data loading
3. fix minor bug of triplet sampler when only one image for one id
2020-02-18 21:01:23 +08:00
L1aoXingyu 327d74ffbb Update strong baseline result
Change data sampler
2020-02-13 00:19:15 +08:00
L1aoXingyu a2f69d0537 Update StrongBaseline results for market1501 and dukemtmc 2020-02-11 22:38:40 +08:00
L1aoXingyu 8a9c0ccfad Finish first version for fastreid 2020-02-10 22:13:04 +08:00
L1aoXingyu db6ed12b14 Update sampler code 2020-02-10 07:38:56 +08:00
liaoxingyu 71950d2c09 1. Fix evaluation code
2. Finish multi-dataset evaluation
3. Decouple image preprocess and output postprocess with model forward for DataParallel training
4. Finish build backbone registry
5. Fix dataset sampler
2020-01-21 20:24:26 +08:00
liaoxingyu b761b656f3 Finish basic training loop and evaluation results 2020-01-20 21:33:37 +08:00