Commit Graph

16 Commits (28061f537cca8a93d7e82d0ad0cd16391ebd461d)

Author SHA1 Message Date
zhangbo9674 28061f537c refine optimizer init logice 2021-12-21 06:28:13 +00:00
zhangbo9674 b54ee04491 Accelerate dynamic graph amp training 2021-12-20 06:36:56 +00:00
gaotingquan 7dcb2d4fd0 fix: raise exception
raise exception about using no_weight_decay of AdamW in static graph
2021-09-30 10:48:36 +00:00
gaotingquan c7aeec28e2 fix: support static graph 2021-09-30 06:57:17 +00:00
gaotingquan 079434dc5f feat: add AdamW 2021-09-01 08:07:48 +00:00
littletomatodonkey 0bc5b54d19
fix opt (#1048) 2021-07-15 15:19:41 +08:00
littletomatodonkey 9d9cd3719e
add static training (#1037)
* add static training

* fix typo

* add se fp16

* rm note

* fix loader

* fix cfg
2021-07-15 10:30:07 +08:00
littletomatodonkey 487c797230
fix optimizer builder (#751) 2021-05-28 10:39:51 +08:00
huangxu96 4e43ec6995 new usage of amp training. (#564)
* new usage of amp training.

* change the usage of amp and pure fp16 training.

* modified code as reviews
2021-02-26 09:25:54 +00:00
littletomatodonkey 6a5f4626d7
add static running in dygraph (#399)
* add static running in dygraph
2020-11-18 09:48:56 +08:00
littletomatodonkey 7ef474169d
polish api to Paddle2.0-rc (#346)
polish codes and docs to adapt Paddle2.0-rc
2020-10-30 00:20:48 +08:00
littletomatodonkey 4773a2177e fix opt 2020-09-17 13:24:33 +00:00
littletomatodonkey b8a7d186d7 fix optimizer and regularizer 2020-09-15 09:43:19 +00:00
littletomatodonkey a43aac3250 fix optimizer 2020-09-03 08:39:14 +00:00
WuHaobo ed2f71ca68 dygraph optimizer 2020-06-07 15:39:26 +08:00
WuHaobo 9f39da8859 Init PaddleClas 2020-04-09 02:16:30 +08:00