mirror of
https://github.com/huggingface/pytorch-image-models.git
synced 2025-06-03 15:01:08 +08:00
Fix adopt descriptions
This commit is contained in:
parent
ce42cc4846
commit
61305cc26a
@ -432,17 +432,17 @@ def _register_adam_variants(registry: OptimizerRegistry) -> None:
|
|||||||
OptimInfo(
|
OptimInfo(
|
||||||
name='adafactorbv',
|
name='adafactorbv',
|
||||||
opt_class=AdafactorBigVision,
|
opt_class=AdafactorBigVision,
|
||||||
description='Big Vision variant of Adafactor with factored gradients, half precision momentum.',
|
description='Big Vision variant of Adafactor with factored gradients, half precision momentum',
|
||||||
),
|
),
|
||||||
OptimInfo(
|
OptimInfo(
|
||||||
name='adopt',
|
name='adopt',
|
||||||
opt_class=Adopt,
|
opt_class=Adopt,
|
||||||
description='Memory-efficient implementation of Adam with factored gradients',
|
description='Modified Adam that can converge with any β2 with the optimal rate',
|
||||||
),
|
),
|
||||||
OptimInfo(
|
OptimInfo(
|
||||||
name='adoptw',
|
name='adoptw',
|
||||||
opt_class=Adopt,
|
opt_class=Adopt,
|
||||||
description='Memory-efficient implementation of Adam with factored gradients',
|
description='Modified AdamW (decoupled decay) that can converge with any β2 with the optimal rate',
|
||||||
defaults={'decoupled': True}
|
defaults={'decoupled': True}
|
||||||
),
|
),
|
||||||
]
|
]
|
||||||
|
@ -51,6 +51,10 @@ def _get_value(x):
|
|||||||
|
|
||||||
|
|
||||||
class Adopt(Optimizer):
|
class Adopt(Optimizer):
|
||||||
|
"""
|
||||||
|
ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate: https://arxiv.org/abs/2411.02853
|
||||||
|
|
||||||
|
"""
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
params,
|
params,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user