| ''' |
| MobileNetv1 in PyTorch. |
| |
| 论文: "MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications" |
| 参考: https://arxiv.org/abs/1704.04861 |
| |
| 主要特点: |
| 1. 使用深度可分离卷积(Depthwise Separable Convolution)减少参数量和计算量 |
| 2. 引入宽度乘子(Width Multiplier)和分辨率乘子(Resolution Multiplier)进一步压缩模型 |
| 3. 适用于移动设备和嵌入式设备的轻量级CNN架构 |
| ''' |
|
|
| import torch |
| import torch.nn as nn |
|
|
|
|
| class Block(nn.Module): |
| '''深度可分离卷积块 (Depthwise Separable Convolution Block) |
| |
| 包含: |
| 1. 深度卷积(Depthwise Conv): 对每个通道单独进行空间卷积 |
| 2. 逐点卷积(Pointwise Conv): 1x1卷积实现通道混合 |
| |
| Args: |
| in_channels: 输入通道数 |
| out_channels: 输出通道数 |
| stride: 卷积步长 |
| ''' |
| def __init__(self, in_channels, out_channels, stride=1): |
| super(Block, self).__init__() |
| |
| |
| self.conv1 = nn.Conv2d( |
| in_channels=in_channels, |
| out_channels=in_channels, |
| kernel_size=3, |
| stride=stride, |
| padding=1, |
| groups=in_channels, |
| bias=False |
| ) |
| self.bn1 = nn.BatchNorm2d(in_channels) |
| self.relu1 = nn.ReLU(inplace=True) |
|
|
| |
| self.conv2 = nn.Conv2d( |
| in_channels=in_channels, |
| out_channels=out_channels, |
| kernel_size=1, |
| stride=1, |
| padding=0, |
| bias=False |
| ) |
| self.bn2 = nn.BatchNorm2d(out_channels) |
| self.relu2 = nn.ReLU(inplace=True) |
|
|
| def forward(self, x): |
| |
| x = self.conv1(x) |
| x = self.bn1(x) |
| x = self.relu1(x) |
| |
| |
| x = self.conv2(x) |
| x = self.bn2(x) |
| x = self.relu2(x) |
| return x |
|
|
|
|
| class MobileNet(nn.Module): |
| '''MobileNet v1网络 |
| |
| Args: |
| num_classes: 分类数量 |
| alpha: 宽度乘子,用于控制网络宽度(默认1.0) |
| beta: 分辨率乘子,用于控制输入分辨率(默认1.0) |
| init_weights: 是否初始化权重 |
| ''' |
| |
| cfg = [64, (128,2), 128, (256,2), 256, (512,2), |
| 512, 512, 512, 512, 512, (1024,2), 1024] |
|
|
| def __init__(self, num_classes=10, alpha=1.0, beta=1.0, init_weights=True): |
| super(MobileNet, self).__init__() |
|
|
| |
| self.conv1 = nn.Sequential( |
| nn.Conv2d(3, 32, kernel_size=3, stride=1, bias=False), |
| nn.BatchNorm2d(32), |
| nn.ReLU(inplace=True) |
| ) |
|
|
| |
| self.layers = self._make_layers(in_channels=32) |
|
|
| |
| self.avg = nn.AdaptiveAvgPool2d(1) |
| self.linear = nn.Linear(1024, num_classes) |
|
|
| |
| if init_weights: |
| self._initialize_weights() |
|
|
| def _make_layers(self, in_channels): |
| '''构建深度可分离卷积层 |
| |
| Args: |
| in_channels: 输入通道数 |
| ''' |
| layers = [] |
| for x in self.cfg: |
| out_channels = x if isinstance(x, int) else x[0] |
| stride = 1 if isinstance(x, int) else x[1] |
| layers.append(Block(in_channels, out_channels, stride)) |
| in_channels = out_channels |
| return nn.Sequential(*layers) |
|
|
| def forward(self, x): |
| |
| x = self.conv1(x) |
| |
| |
| x = self.layers(x) |
| |
| |
| x = self.avg(x) |
| x = x.view(x.size(0), -1) |
| x = self.linear(x) |
| return x |
|
|
| def feature(self, x): |
| |
| x = self.conv1(x) |
| |
| |
| x = self.layers(x) |
| x = self.avg(x) |
| return x |
|
|
| def prediction(self, x): |
| x = x.view(x.size(0), -1) |
| x = self.linear(x) |
| return x |
| |
| def _initialize_weights(self): |
| '''初始化模型权重''' |
| for m in self.modules(): |
| if isinstance(m, nn.Conv2d): |
| |
| nn.init.kaiming_normal_(m.weight, mode='fan_out') |
| if m.bias is not None: |
| nn.init.zeros_(m.bias) |
| elif isinstance(m, nn.BatchNorm2d): |
| |
| nn.init.ones_(m.weight) |
| nn.init.zeros_(m.bias) |
| elif isinstance(m, nn.Linear): |
| |
| nn.init.normal_(m.weight, 0, 0.01) |
| nn.init.zeros_(m.bias) |
|
|
|
|
| def test(): |
| """测试函数""" |
| net = MobileNet() |
| x = torch.randn(2, 3, 32, 32) |
| y = net(x) |
| print(y.size()) |
| |
| |
| from torchinfo import summary |
| device = 'cuda' if torch.cuda.is_available() else 'cpu' |
| net = net.to(device) |
| summary(net, (2, 3, 32, 32)) |
|
|
| if __name__ == '__main__': |
| test() |