liaoxingyu 3d1bae9f13 fix triplet loss backward propagation on multi-gpu training (#82)
Summary: fix torch.distributed.all_gather has no gradient when performing all_gather operation on the provided tensors, instead using `GatherLayer`.
2020-09-28 17:16:51 +08:00
..
2020-09-23 14:41:44 +08:00
2020-09-23 19:45:13 +08:00
2020-03-25 10:58:26 +08:00