cifar

您所在的位置:网站首页 resnet34代码 cifar

cifar

2023-07-01 04:07| 来源: 网络整理| 查看: 265

须知:

同前文

代码变化部分如下图,其余部分与上文一同CIFAR-10基础优化一(加入标准化和激活函数)_百炼成丹的博客-CSDN博客

优化思路:

再加入多层卷积层

卷积层层间和全连接层层间加入dropout

优化网络顺序结构

图像增广

网络结构: class Model(nn.Module): def __init__(self): super(Model, self).__init__() self.model1 = nn.Sequential( nn.Conv2d(3, 32, 5, padding=2), nn.BatchNorm2d(32), nn.Dropout2d(0.25), nn.ReLU(), nn.Conv2d(32, 32, 5, padding=2), nn.BatchNorm2d(32), nn.MaxPool2d(2)) # nn.ReLU(), self.model2 = nn.Sequential( nn.Conv2d(32, 64, 5, padding=2), nn.BatchNorm2d(64), nn.Dropout2d(0.25), nn.ReLU(), nn.Conv2d(32, 64, 5, padding=2), nn.BatchNorm2d(64), nn.MaxPool2d(2)) # nn.ReLU(), self.model3 = nn.Sequential( nn.Conv2d(32, 128, 5, padding=2), nn.BatchNorm2d(128), nn.Dropout2d(0.3), nn.ReLU(), nn.Conv2d(32, 128, 5, padding=2), nn.BatchNorm2d(128), nn.MaxPool2d(2)) # nn.ReLU(), self.model4 = nn.Sequential( nn.Conv2d(32, 256, 5, padding=2), nn.BatchNorm2d(256), nn.Dropout2d(0.3), nn.ReLU(), nn.Conv2d(32, 256, 5, padding=2), nn.BatchNorm2d(256), nn.MaxPool2d(2)) # nn.ReLU(), self.model5 = nn.Sequential( nn.Conv2d(256, 512, 5, padding=2), nn.BatchNorm2d(512), nn.Dropout2d(0.35), nn.ReLU(), nn.Conv2d(512, 512, 5, padding=2), nn.BatchNorm2d(512), nn.MaxPool2d(2)) # nn.ReLU(), self.linear1 = nn.Sequential( nn.Flatten(), nn.Linear(512, 128), nn.BatchNorm1d(128), nn.ReLU(), nn.Dropout(0.5), nn.Linear(128, 10)) 图像增广:torchvision.transforms.ColorJitter(0.5), torchvision.transforms.RandomHorizontalFlip() train_data = torchvision.datasets.CIFAR10("../dataset", train=True, transform=torchvision.transforms.Compose( [torchvision.transforms.ColorJitter(0.5), torchvision.transforms.RandomHorizontalFlip(), torchvision.transforms.ToTensor()]))

效果:

 测试集准确率峰值88.2%

 约64epoch后无增幅

表现效果进一步提高

但时间不理想,约40分钟上下

下一步,加入残差网络



【本文地址】


今日新闻


推荐新闻


CopyRight 2018-2019 办公设备维修网 版权所有 豫ICP备15022753号-3