Commit Graph

194 Commits (a00291fec2c1f4ad89095361c5574b89e94d42a3)

Author SHA1 Message Date
HydrogenSulfate f2982e5e47 update code 2022-04-22 12:01:05 +08:00
HydrogenSulfate 17fd1bc2c0 refine code 2022-04-22 12:00:03 +08:00
HydrogenSulfate aa26a8c1d8 fix bug for static graph 2022-04-21 16:31:28 +08:00
HydrogenSulfate daf7eea23d fix code 2022-04-21 15:43:53 +08:00
HydrogenSulfate dfd7749828 refine hard code 2022-04-21 14:37:00 +08:00
HydrogenSulfate 41e1a86caf add center loss 2022-04-21 00:17:54 +08:00
Wei Shengyu e6feb68bb8
Merge pull request #1824 from TingquanGao/dev/spt_amp_eval
fix: fp32 eval by default when enable amp
2022-04-20 14:40:46 +08:00
HydrogenSulfate 3a1276d315 train_loss_func only used in train mode 2022-04-19 19:54:48 +08:00
HydrogenSulfate 24abea151a support for multi optimizer case 2022-04-19 14:26:42 +08:00
gaotingquan 83ed5195c3
fix: set use_fp16_test to True when AMP O2 is enabled 2022-04-18 06:14:43 +00:00
weishengyu 1789da6422 fix bug 2022-04-18 11:26:32 +08:00
gaotingquan a35cdd2aec
uncommit: sync bn is too slow to use and convert_sync_batchnorm() is not effective for BatchNorm 2022-04-14 08:19:39 +00:00
gaotingquan 13d5e59051
fix: convert bn to sync_bn
the running_mean and running_var of bn would not be synchronized in dist,
so which leads to bug that eval loss in training is inconsistent with eval only.
2022-04-14 07:36:39 +00:00
gaotingquan efde56ffc6
fix: only fp16 evaluation is supported when ampO2 is enabled 2022-04-13 12:14:14 +00:00
gaotingquan 474c918b27
fix: fix bug of batch_size statistics error 2022-04-13 09:19:30 +00:00
gaotingquan c46189bad0
fix: fix bug about calc loss in dist 2022-04-12 06:56:44 +00:00
HydrogenSulfate af90cd7c59 update center loss config and related code 2022-04-12 13:07:53 +08:00
weishengyu 9de22673df dbg 2022-04-08 14:29:03 +08:00
gaotingquan b761325faa fix: fp32 eval by default when enable amp
If you want to eval by fp16 when enable amp, please set Amp.use_fp16_test=True, False by default.
2022-04-02 19:22:10 +08:00
dongshuilong a944603da0 fix log twice bug 2022-03-30 08:31:35 +00:00
huangqipeng b62b98d79f feat: support mlu device and amp of mlu 2022-03-14 15:48:26 +08:00
littletomatodonkey f68c098a4a fix train acc log 2022-03-09 19:58:36 +08:00
WangChen0902 7595ba6d70
add AFD (#1683)
* add AFD
2022-02-28 19:11:50 +08:00
dongshuilong dc6281a6d4 add benchmark for tipc 2022-02-10 08:25:52 +00:00
Tingquan Gao 42134cd8dd fix: raise warning when using Global.class_num 2022-01-25 15:06:36 +08:00
Tingquan Gao bb6581d21b refactor: raise warning when gpu numbers is not 4 2022-01-25 15:06:36 +08:00
Tingquan Gao 8f0bd5b582 fix: fix vdl makedir 2022-01-25 15:06:36 +08:00
gaotingquan 10c93c55d1 fix: enable amp only in training 2022-01-25 11:58:07 +08:00
gaotingquan 7040ce8314 refactor: change params to be consistent with amp 2022-01-25 11:58:07 +08:00
zhangbo9674 cd039a7b37 add save_dtype 2022-01-10 18:19:03 +08:00
zhangbo9674 d437bb0a7e use fp32 to eval 2022-01-10 18:19:03 +08:00
zhangbo9674 bb19c1f7a6 fix eval bug 2022-01-10 18:19:03 +08:00
zhangbo9674 b2956c1b41 refine code 2022-01-10 18:19:03 +08:00
zhangbo9674 205592a3e3 fix amp with distribute bug 2022-01-10 18:19:03 +08:00
littletomatodonkey aea712cc87
add dist of rec model (#1574)
* add distillation loss func and rec distillation
2022-01-05 19:25:36 +08:00
gaotingquan 6e13ff3068 fix: use hasattr() to check if collate_fn is in dataloader
fix bug caused by PR #1596
2021-12-30 16:35:05 +08:00
gaotingquan 7da2a997e9 fix: save latest model every epoch 2021-12-27 22:04:26 +08:00
gaotingquan 5d53e9f152 fix: raise warning when setting batch_transform_ops and TopkAcc 2021-12-24 21:22:52 +08:00
zhangbo9674 28061f537c refine optimizer init logice 2021-12-21 06:28:13 +00:00
zhangbo9674 b54ee04491 Accelerate dynamic graph amp training 2021-12-20 06:36:56 +00:00
gaotingquan 7732a69f1b fix: fix key error in distillation 2021-12-16 18:21:08 +08:00
weishengyu 534037d145 dbg 2021-12-10 11:14:14 +08:00
weishengyu ad1a2fd137 move slim into arch 2021-12-09 20:08:57 +08:00
weishengyu 7c6567cc6b dbg 2021-12-09 18:08:16 +08:00
weishengyu 6c5d1ebc28 add pruner and quanter for theseus 2021-12-09 14:51:40 +08:00
cuicheng01 a3b54f15d5 fix export quant_model 2021-12-03 07:39:31 +00:00
dongshuilong f7ccc874e2 fix dali distributed eval bug 2021-11-16 11:09:21 +08:00
stephon 7a17f72fc2 fix seed=0 bug 2021-11-01 06:16:24 +00:00
Walter a5d0e37b02
Merge pull request #1341 from RainFrost1/googlenet_bug
fig goooglenet distributed eval bug
2021-10-29 10:28:37 +08:00
gaotingquan ed459a2a16 refactor: adapt to static graph in deprecating MixCELoss 2021-10-27 19:47:43 +08:00
gaotingquan ba2dd01a13 refactor: deprecate MixCELoss 2021-10-27 19:47:43 +08:00
dongshuilong 278f6d8050 fig goooglenet distributed eval bug 2021-10-26 11:56:30 +00:00
ronnywang a0eb34a642
Add npu supporting (#1324) 2021-10-22 11:02:29 +08:00
Walter cc00a51af7
Merge pull request #1322 from RainFrost1/train_without_eval
fix train without eval bug
2021-10-21 15:35:05 +08:00
dongshuilong fd6f1ad2ca fix clas distributed eval bug 2021-10-21 03:47:03 +00:00
dongshuilong 06382519cf fix train without eval bug 2021-10-21 02:39:27 +00:00
dongshuilong c93d638f4c fix clas distributed eval bug 2021-10-20 11:22:37 +00:00
Walter e4bb18766a
Merge pull request #1245 from RainFrost1/benchmark
add version 1 for benchmark
2021-10-19 10:18:16 +08:00
dongshuilong 0e20e60f6b fix slim load pretrained model bug 2021-10-18 10:07:14 +00:00
Wei Shengyu 3c148cb56a
update return res method and fix amp bug (#1309)
* fix amp bug; update return res method

* support index for list object for return res
2021-10-15 18:25:50 +08:00
dongshuilong c2a9f8830e update profiler 2021-10-13 09:30:11 +00:00
dongshuilong 272bc9481d Merge branch 'develop' of https://github.com/PaddlePaddle/PaddleClas into benchmark 2021-10-13 09:26:28 +00:00
cuicheng01 fe6f614680 Update multilabel 2021-09-27 03:00:37 +00:00
cuicheng01 a90881c99f
Merge branch 'develop' into add_multilabel 2021-09-26 22:47:20 +08:00
cuicheng01 af9aae730e add multilabel feature 2021-09-26 07:05:13 +00:00
weishengyu af25e25640 modify format 2021-09-23 11:22:25 +08:00
dongshuilong 9f2ab06ec2 add profiler 2021-09-22 08:29:44 +00:00
cuicheng01 ba9b708acb
Merge pull request #1174 from TingquanGao/dev/add_adamw
feat: add AdamW
2021-09-17 18:53:18 +08:00
Wei Shengyu d98b881603
Merge pull request #1208 from weisy11/fix_bug_of_theseus
dbg theseus
2021-09-15 14:04:51 +08:00
weishengyu 600067f4fc dbg 2021-09-15 11:35:49 +08:00
dongshuilong 94c9900ece add seed 2021-09-14 04:06:37 +00:00
Walter 74622af482
Merge pull request #1093 from RainFrost1/slim
添加slim功能
2021-09-13 19:11:43 +08:00
Walter ce39aea97f
Merge pull request #1187 from RainFrost1/retrieval_dataloader
fix issues when gallery == query dataset
2021-09-03 14:33:23 +08:00
dongshuilong 50f2547005 Merge branch 'develop' of https://github.com/PaddlePaddle/PaddleClas into slim 2021-09-03 03:19:14 +00:00
dongshuilong d49657ad08 fix rec forward bug 2021-09-02 07:42:22 +00:00
gaotingquan 079434dc5f feat: add AdamW 2021-09-01 08:07:48 +00:00
dongshuilong 68c1b089e8 update slim for new trainer 2021-08-27 09:32:37 +00:00
dongshuilong 1a1eb3a1ae fix issues when gallery == query dataset 2021-08-27 06:25:27 +00:00
Bin Lu ec5e07da99
Merge pull request #1166 from Intsigstephon/develop
add Deephash method: DLBHC
2021-08-26 20:11:36 +08:00
Bin Lu d388d69a63
Update retrieval.py 2021-08-26 19:38:40 +08:00
dongshuilong c9be0fe00a fix loss_msg bugs 2021-08-26 07:13:11 +00:00
Bin Lu 52663edf60
Update retrieval.py 2021-08-25 13:57:24 +08:00
dongshuilong 5bf74beb46 update according comments 2021-08-24 07:07:17 +00:00
dongshuilong 15f6f58139 refactor trainer v2 2021-08-24 03:02:55 +00:00
dongshuilong ebde0e13cb refactor trainer 2021-08-22 15:10:23 +00:00
weishengyu b4678fd3fa revert trainer 2021-08-10 11:27:56 +08:00
weishengyu f41b09ef7f dbg 2021-08-08 17:28:57 +08:00
weishengyu ef13f8c640 modify code 2021-08-08 15:04:26 +08:00
weishengyu e232a97a84 dbg 2021-08-08 02:10:17 +08:00
weishengyu 91af66dccd replace self.model() with self.forward 2021-08-08 02:05:26 +08:00
weishengyu 6c164c9e2b dbg 2021-08-08 02:02:42 +08:00
weishengyu 6ebe7f09ed add return_inter flag 2021-08-08 01:52:18 +08:00
weishengyu 9790cc517f add return_dict to trainer 2021-08-08 01:49:26 +08:00
zhiqiu bb1376db45 add flags setting 2021-07-16 13:26:03 +00:00
Walter 73004f78f5
add fp16 amp training and dali (#993)
* add dygraph amp train
2021-07-14 22:59:23 +08:00
Wei Shengyu 5cd7a05d68
Merge pull request #1034 from cuicheng01/develop
support load pretrain from url
2021-07-13 17:35:04 +08:00
cuicheng01 592f5b10f5 support batch mix 2021-07-13 06:38:01 +00:00
littletomatodonkey 4af9f51088
fix windows training (#1038)
* fix windows training

* fix typo
2021-07-13 13:57:24 +08:00
cuicheng01 0c63f64bec Update trainer.py 2021-07-12 08:06:05 +00:00
cuicheng01 f7329ddf74 support load pretrain from url 2021-07-12 07:45:44 +00:00