Commit Graph

14 Commits (c890ceb9303468cf47553da8764d7febabd9df68)

Author SHA1 Message Date
Hugo Touvron f512394620
Merge pull request #118 from developer0hye/patch-1
save best checkpoint
2021-11-21 15:24:12 +01:00
Yonghye Kwon fdafaf0b8c
save best checkpoint 2021-10-01 17:12:25 +09:00
Maxwell Aladago 2f76ec64fd BUG Fix: ensure the right loss function is selected when mixtup or cutmix is enabeled 2021-05-02 10:27:25 -04:00
Maxwell Aladago 686507a6a5 BUG Fix: ensure the right loss function is selected when mixtup or cutmix is enabeled 2021-05-02 10:09:58 -04:00
Francisco Massa ab5715372d
Add deit_base_distilled_patch16_384 (#55) 2021-01-18 12:19:47 +01:00
Francisco Massa 1f961b060a
Fix no-weight-decay in optimizer (#52) 2021-01-16 11:53:29 +01:00
Francisco Massa adc7f1e849
Add loss_scaler to checkpoints (#49)
Ensure bit-wise reproducibility when rescheduling jobs
2021-01-15 10:52:41 +01:00
Francisco Massa a8e90967a3
Add option to finetune on larger resolution (#43)
* Add option for finetuning a model

* Fixes

* Keep model in eval mode during finetuning

* Only skip head weights if size mismatch

* Remove finetune-epochs

Might not be needed

* Raise error if distillation + finetune are enabled
2021-01-15 10:13:52 +01:00
Francisco Massa 8eae3269da
Add Knowledge-Distillation (#42)
* Add knowledge distillation

* Bugfix

* Bugfix

* Make names more readable and use single torch.cat call

* Remove criterion.train() in engine

The teacher should stay in eval mode

* Change default argument for teacher-model

* Return the average of classifiers during inference

* Cleanup unused code

* Add docstring for DistillationLoss

* Remove warnings from newer PyTorch

Also uses more stable variant, instead of using softmax + log, use directly log_softmax
2021-01-13 14:19:23 +01:00
Changlin Li b06fcba6bf
Support parallelized evaluation (#24)
* support parallelized evaluation

* remove shuffle arg of loader val, add sampler val in non-dist branch

* replace timm eval sampler with torch sampler

* add logger synchronizing to support parallelized evaluation

* add command line argument dist-eval and warning
2021-01-08 11:05:39 +01:00
Francisco Massa 0c4b8f60bd Change LICENSE to Apache 2.0 2021-01-08 10:51:58 +01:00
Francisco Massa dcd888df20
Update default data-path (#25)
Points to the new valid location
2021-01-07 22:18:54 +01:00
Zhiyuan Chen 0282d2a175
Remove drop block args (#23)
timm's Vision Transformer does not support drop block
2021-01-07 17:28:41 +01:00
Francisco Massa 1d38fa4c37 Initial commit 2020-12-23 10:47:58 -08:00