11 Commits

Author SHA1 Message Date
brian1009
43d74ab8b8 [Refactor, Feature Distillation] refactor models, clean distillation token and add feature distillation 2023-06-17 02:51:32 +08:00
410011max
2ba43bbca1 update main & test config 2023-06-16 21:49:10 +08:00
brian1009
afadc46e64 [NAS] Modify the infrasturcture for nas training 2023-03-12 17:52:14 +08:00
brian1009
ce08e408dd Add the basic implementation of the sparsity factory 2023-02-23 15:12:28 +08:00
Hugo Touvron
4da66916c7
Add files via upload 2022-05-08 20:06:44 +02:00
Francisco Massa
a8e90967a3
Add option to finetune on larger resolution (#43)
* Add option for finetuning a model

* Fixes

* Keep model in eval mode during finetuning

* Only skip head weights if size mismatch

* Remove finetune-epochs

Might not be needed

* Raise error if distillation + finetune are enabled
2021-01-15 10:13:52 +01:00
Francisco Massa
8eae3269da
Add Knowledge-Distillation (#42)
* Add knowledge distillation

* Bugfix

* Bugfix

* Make names more readable and use single torch.cat call

* Remove criterion.train() in engine

The teacher should stay in eval mode

* Change default argument for teacher-model

* Return the average of classifiers during inference

* Cleanup unused code

* Add docstring for DistillationLoss

* Remove warnings from newer PyTorch

Also uses more stable variant, instead of using softmax + log, use directly log_softmax
2021-01-13 14:19:23 +01:00
Changlin Li
b06fcba6bf
Support parallelized evaluation (#24)
* support parallelized evaluation

* remove shuffle arg of loader val, add sampler val in non-dist branch

* replace timm eval sampler with torch sampler

* add logger synchronizing to support parallelized evaluation

* add command line argument dist-eval and warning
2021-01-08 11:05:39 +01:00
Francisco Massa
0c4b8f60bd Change LICENSE to Apache 2.0 2021-01-08 10:51:58 +01:00
sanjaydatasciencedojo
4e91d2588f
Remove unused libraries (#9) 2020-12-27 21:43:54 +01:00
Francisco Massa
1d38fa4c37 Initial commit 2020-12-23 10:47:58 -08:00