mirror of
https://github.com/huggingface/pytorch-image-models.git
synced 2025-06-03 15:01:08 +08:00
Fix bug introduced in non layer_decay weight_decay application. Remove debug print, fix arg desc.
This commit is contained in:
parent
372ad5fa0d
commit
0557c8257d
@ -660,7 +660,6 @@ def group_with_matcher(
|
||||
for k in sorted(filter(lambda x: x is not None, grouping.keys())):
|
||||
if lid < 0 or k[-1] != MATCH_PREV_GROUP[0]:
|
||||
lid += 1
|
||||
print(lid, k, grouping[k])
|
||||
layer_id_to_param[lid].extend(grouping[k])
|
||||
|
||||
if reverse:
|
||||
|
@ -44,7 +44,7 @@ def param_groups_weight_decay(
|
||||
if not param.requires_grad:
|
||||
continue
|
||||
|
||||
if param.ndim or name.endswith(".bias") or name in no_weight_decay_list:
|
||||
if param.ndim <= 1 or name.endswith(".bias") or name in no_weight_decay_list:
|
||||
no_decay.append(param)
|
||||
else:
|
||||
decay.append(param)
|
||||
|
2
train.py
2
train.py
@ -140,7 +140,7 @@ parser.add_argument('--clip-grad', type=float, default=None, metavar='NORM',
|
||||
parser.add_argument('--clip-mode', type=str, default='norm',
|
||||
help='Gradient clipping mode. One of ("norm", "value", "agc")')
|
||||
parser.add_argument('--layer-decay', type=float, default=None,
|
||||
help='weight decay (default: None)')
|
||||
help='layer-wise learning rate decay (default: None)')
|
||||
|
||||
# Learning rate schedule parameters
|
||||
parser.add_argument('--sched', default='cosine', type=str, metavar='SCHEDULER',
|
||||
|
Loading…
x
Reference in New Issue
Block a user