diff --git a/timm/layers/mlp.py b/timm/layers/mlp.py index a68ae4bc..11d9eeca 100644 --- a/timm/layers/mlp.py +++ b/timm/layers/mlp.py @@ -130,8 +130,6 @@ class SwiGLU(nn.Module): self.fc2 = nn.Linear(hidden_features, out_features, bias=bias[1]) self.drop2 = nn.Dropout(drop_probs[1]) - self.drop = nn.Dropout(drop) - def init_weights(self): # override init of fc1 w/ gate portion set to weight near zero, bias=1 nn.init.ones_(self.fc1_g.bias)