from typing import Dict, Sequence import torch import torch.nn as nn import torch.nn.functional as F from mmcv.cnn import build_activation_layer, build_norm_layer from mmcv.runner import BaseModule, ModuleList from ..builder import HEADS from .cls_head import ClsHead class LinearBlock(BaseModule): def __init__(self, in_channels, out_channels, dropout_rate=0., norm_cfg=None, act_cfg=None, init_cfg=None): super().__init__(init_cfg=init_cfg) self.fc = nn.Linear(in_channels, out_channels) self.norm = None self.act = None self.dropout = None if norm_cfg is not None: self.norm = build_norm_layer(norm_cfg, out_channels)[1] if act_cfg is not None: self.act = build_activation_layer(act_cfg) if dropout_rate > 0: self.dropout = nn.Dropout(p=dropout_rate) def forward(self, x): x = self.fc(x) if self.norm is not None: x = self.norm(x) if self.act is not None: x = self.act(x) if self.dropout is not None: x = self.dropout(x) return x @HEADS.register_module() class StackedLinearClsHead(ClsHead): """Classifier head with several hidden fc layer and a output fc layer. Args: num_classes (int): Number of categories excluding the background category. in_channels (int): Number of channels in the input feature map. mid_channels (Sequence): Number of channels in the hidden fc layers. dropout_rate (float): Dropout rate after each hidden fc layer, except the last layer. Defaults to 0. norm_cfg (dict, optional): Config dict of normalization layer after each hidden fc layer, except the last layer. Defaults to None. act_cfg (dict, optional): Config dict of activation function after each hidden layer, except the last layer. Defaults to use "ReLU". """ def __init__(self, num_classes: int, in_channels: int, mid_channels: Sequence, dropout_rate: float = 0., norm_cfg: Dict = None, act_cfg: Dict = dict(type='ReLU'), **kwargs): super(StackedLinearClsHead, self).__init__(**kwargs) assert num_classes > 0, \ f'`num_classes` of StackedLinearClsHead must be a positive ' \ f'integer, got {num_classes} instead.' self.num_classes = num_classes self.in_channels = in_channels assert isinstance(mid_channels, Sequence), \ f'`mid_channels` of StackedLinearClsHead should be a sequence, ' \ f'instead of {type(mid_channels)}' self.mid_channels = mid_channels self.dropout_rate = dropout_rate self.norm_cfg = norm_cfg self.act_cfg = act_cfg self._init_layers() def _init_layers(self): self.layers = ModuleList( init_cfg=dict( type='Normal', layer='Linear', mean=0., std=0.01, bias=0.)) in_channels = self.in_channels for hidden_channels in self.mid_channels: self.layers.append( LinearBlock( in_channels, hidden_channels, dropout_rate=self.dropout_rate, norm_cfg=self.norm_cfg, act_cfg=self.act_cfg)) in_channels = hidden_channels self.layers.append( LinearBlock( self.mid_channels[-1], self.num_classes, dropout_rate=0., norm_cfg=None, act_cfg=None)) def init_weights(self): self.layers.init_weights() def simple_test(self, img): """Test without augmentation.""" cls_score = img for layer in self.layers: cls_score = layer(cls_score) if isinstance(cls_score, list): cls_score = sum(cls_score) / float(len(cls_score)) pred = F.softmax(cls_score, dim=1) if cls_score is not None else None if torch.onnx.is_in_onnx_export(): return pred pred = list(pred.detach().cpu().numpy()) return pred def forward_train(self, x, gt_label): cls_score = x for layer in self.layers: cls_score = layer(cls_score) losses = self.loss(cls_score, gt_label) return losses