mirror of
https://github.com/open-mmlab/mmselfsup.git
synced 2025-06-03 14:59:38 +08:00
update
This commit is contained in:
parent
4cd6eb6cba
commit
d2284b373f
@ -248,9 +248,12 @@ def _non_dist_train(model,
|
|||||||
seed=cfg.seed,
|
seed=cfg.seed,
|
||||||
drop_last=getattr(cfg.data, 'drop_last', False)) for ds in dataset
|
drop_last=getattr(cfg.data, 'drop_last', False)) for ds in dataset
|
||||||
]
|
]
|
||||||
|
|
||||||
|
if 'use_fp16' in cfg and cfg.use_fp16 == True:
|
||||||
|
raise NotImplementedError('apex do not support non_dist_trian!')
|
||||||
# put model on gpus
|
# put model on gpus
|
||||||
model = MMDataParallel(model, device_ids=range(cfg.gpus)).cuda()
|
model = MMDataParallel(model, device_ids=range(cfg.gpus)).cuda()
|
||||||
|
|
||||||
# build runner
|
# build runner
|
||||||
optimizer = build_optimizer(model, cfg.optimizer)
|
optimizer = build_optimizer(model, cfg.optimizer)
|
||||||
runner = Runner(
|
runner = Runner(
|
||||||
|
Loading…
x
Reference in New Issue
Block a user