期刊论文详细信息
Applied Sciences
CDTNet: Improved Image Classification Method Using Standard, Dilated and Transposed Convolutions
Yuepeng Zhou1  Huiyou Chang1  Yonghe Lu2  Xili Lu3 
[1] School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou 510006, China;School of Information Management, Sun Yat-sen University, Guangzhou 510006, China;School of Information and Engineering, Shaoguan University, Shaoguan 512005, China;
关键词: CDTNet;    dilated convolution;    transposed convolution;    feature fusion;    receptive field;   
DOI  :  10.3390/app12125984
来源: DOAJ
【 摘 要 】

Convolutional neural networks (CNNs) have achieved great success in image classification tasks. In the process of a convolutional operation, a larger input area can capture more context information. Stacking several convolutional layers can enlarge the receptive field, but this increases the parameters. Most CNN models use pooling layers to extract important features, but the pooling operations cause information loss. Transposed convolution can increase the spatial size of the feature maps to recover the lost low-resolution information. In this study, we used two branches with different dilated rates to obtain different size features. The dilated convolution can capture richer information, and the outputs from the two channels are concatenated together as input for the next block. The small size feature maps of the top blocks are transposed to increase the spatial size of the feature maps to recover low-resolution prediction maps. We evaluated the model on three image classification benchmark datasets (CIFAR-10, SVHN, and FMNIST) with four state-of-the-art models, namely, VGG16, VGG19, ResNeXt, and DenseNet. The experimental results show that CDTNet achieved lower loss, higher accuracy, and faster convergence speed in the training and test stages. The average test accuracy of CDTNet increased by 54.81% at most on SVHN with VGG19 and by 1.28% at least on FMNIST with VGG16, which proves that CDTNet has better performance and strong generalization abilities, as well as fewer parameters.

【 授权许可】

Unknown   

  文献评价指标  
  下载次数:0次 浏览次数:0次