【AI达人特训营第三期】时频图分类项目

P粉084495128
发布: 2025-07-31 10:04:33
原创
789人浏览过
本项目基于ResNet50,将其BottleneckBlock模块中conv2特征层的标准卷积替换为DCN可变性卷积,构建ResNet50-DCN模型,用于科大讯飞24类语音时序图谱分类。使用2143条训练样本、429条验证样本训练,90轮后验证集最高准确率79%,未深入调优,有提升空间,还包含数据处理、模型训练及推理等过程。

☞☞☞AI 智能聊天, 问答助手, AI 智能搜索, 免费无限量使用 DeepSeek R1 模型☜☜☜

【ai达人特训营第三期】时频图分类项目 - php中文网

基于ResNet50的音频时序图分类任务


本项目是基于 resnet50 对其 BottleneckBlock 模块中 conv2 特征层的标准卷积替换为 DCN(Deformable ConvNets V1) 可变性卷积,从而提升模型的泛化性。本项目只是实现了 resnet50-dcn 模型搭建与跑通,分类在验证集准确率最高79%,没有进行更多的调优,有提升可能性。

数据集是科大讯飞的语音时序图谱(已脱敏处理),24种不同种类,2000多条训练样本,1020张测试图片。 表述不当的地方,还请多多包涵指正。

模型介绍

Resnet50

基础残差块

【AI达人特训营第三期】时频图分类项目 - php中文网        

resnet50

resnet50 是由 resnet34 将其基础残差块上图左替换为图右而得到。

【AI达人特训营第三期】时频图分类项目 - php中文网        
resnet34结构图 

机器视觉深度学习算法,即使很深,也能保持初心,还易于训练,得益于残差思想,弃其糟粕,取其精华,最终使自己更强大。实现方法以及原论文都在这里(飞桨API)。

注意:在计算交叉熵损失过程中使用softmax,故而resnet50 网络中没有softmax层,所以 output 不是 0~1 概率值。

DCN概述

背景

在计算机视觉领域,同一物体在不同场景,角度中未知的几何变换是检测/识别的一大挑战,通常的做法:

  • 通过数据增广,尽可能多的获得不同场景下的样本以增强模型的适应能力。

  • 设置一些针对几何变换不变的特征或者算法。

两种方法都有缺陷,第一种方法因为样本的局限性显然模型的泛化能力比较低,无法泛化到一般场景中,第二种方法则因为手工设计的不变特征和算法对于过于复杂的变换是很难的而无法设计。所以作者提出了Deformable Conv(可变形卷积)和 Deformable Pooling(可变形池化)来解决这个问题。 这里上传两张图片来形象展示一下,具体可参考其他大佬文章(参阅)。

【AI达人特训营第三期】时频图分类项目 - php中文网        

在上面这张图里面,左边传统的卷积显然没有提取到完整绵羊的特征,而右边的可变形卷积则提取到了完整的不规则绵羊的特征。

【AI达人特训营第三期】时频图分类项目 - php中文网        

(a) 所示的正常卷积规律的采样 9 个点(绿点),(b)(c)(d) 为可变形卷积,在正常的采样坐标上加上一个位移量(蓝色箭头),其中 (c)(d) 作为 (b) 的特殊情况,展示了可变形卷积可以作为尺度变换,比例变换和旋转变换等特殊情况。

怪兽AI数字人
怪兽AI数字人

数字人短视频创作,数字人直播,实时驱动数字人

怪兽AI数字人 44
查看详情 怪兽AI数字人
【AI达人特训营第三期】时频图分类项目 - php中文网        

offset偏置卷积 kernel_size = [kw, kh], 形状[N, 2 * kw * kh, W, H],对原卷积进行x,y两个方向偏置,所以他的输出通道 2 * kw * kh。最终输出是图左边的亮绿色原始卷积核变为松散开来的蓝色卷积核(有点像空洞),然后再以新卷积进行计算。(个人理解:offset得到的是浮点型位置坐标,是由那一点的周围四个整型坐标加权得到,即一个点坐标是由 x,y 两个方向的偏置确定,可以看作用一组正交向量表示一个任意向量, 每个kernel总共有 kw * kh 个点,那就需要 2 * kw * kh个输出,总共有N个 kernel,所以形状为 [N, 2 * kw * kh, W, H]。)

可形变卷积 v1(mask = None)计算公式如下:

y(p)=k=1Kwkx(p+pk+Δpk)y(p)=∑k=1Kwk∗x(p+pk+Δpk)

其中

pkpk是标准卷积的第k个位置的坐标(固定不变的)

ΔpkΔpk 和 ΔmkΔmk 分别为第 k 个位置的可学习偏移和调制标量。在 deformable conv v1 中 ΔmkΔmk 为 1(参考链接)。

In [1]
# 解压数据!unzip -q -d /home/aistudio /home/aistudio/data/data155487/data.zip
登录后复制
   
In [45]
# 导入依赖模块import paddleimport paddle.vision.transforms as Tfrom PIL import Imageimport timeimport os 
import numpy as npimport matplotlib.pyplot as pltimport pandas as pd# 导入自定义模型import models


paddle.device.set_device("gpu:0") # 开启GPU训练ROOT_DIR = '/home/aistudio/data'
登录后复制
   

数据处理

读取并查看数据

In [3]
data = pd.read_csv('data/train.csv')# 图片张数print('图片的张数:', data.shape[0])# 时序图的类别数classes_nums = data['label'].nunique()print('时序图的类别数:', classes_nums)# 查看标签列表classes = data['label']print('第一张图片标签', classes[0])# 图片路径列表img_paths = data['image']
img_path = os.path.join('data', data['image'][0])print('第一张图片路径:', img_path)
登录后复制
       
图片的张数: 2143
时序图的类别数: 24
第一张图片标签 12
第一张图片路径: data/train_0.jpg
登录后复制
       
In [4]
# 切分train和valid数据train_data = data[:int(classes.shape[0]*0.8)]
valid_data = data[int(classes.shape[0]*0.8):]
登录后复制
   
In [ ]
valid_data['label']
登录后复制
   
In [54]
# 读取图片方法def load_image(img_path):
    img = Image.open(img_path).convert(mode='RGB') # 将图片转成RGB模式三通道模式
    # img = img.resize(img.size) 
    img = img.resize((224,224))    # return np.array(img).astype('float32')
    return np.array(img)
登录后复制
   

训练集、验证集加载方法

In [56]
# 创建数据集class MyDataset(paddle.io.Dataset):
    def __init__(self, data, mode='train', valid_start=None):
        super(MyDataset,self).__init__(),
        self.data = data
        self.mode = mode        if self.mode == 'train':
            self.transform = T.Compose([                      
                T.BrightnessTransform(0.4), # 亮度调节(要求像素值为int类型)
                T.ContrastTransform(0.4),   # 对比度调节
                T.HueTransform(0.4),        # 色调
                T.RandomErasing(),          # 随机修剪
                T.Normalize(mean=[123.675, 116.28, 103.53],
                    std=[58.395, 57.12, 57.375], data_format='HWC'), # 标准化
                T.Transpose()             # 数据格式转换,Transpose默认参数(2,0,1)
            ])        if self.mode == 'valid':
            self.transform = T.Compose([
                T.Normalize(mean=[123.675, 116.28, 103.53],
                    std=[58.395, 57.12, 57.375], data_format='HWC'),
                T.Transpose()
            ])
        self.valid_start = valid_start    def __getitem__(self, index):
        # 验证集索引处理
        if self.mode == 'valid':
            index = index + self.valid_start        # 获取标签
        label = self.data['label'][index]        # 获取路径
        img_path = os.path.join(ROOT_DIR, 'train', self.data['image'][index])        # 读取图片内容
        image =  load_image(img_path)        # 图形增广
        image = self.transform(image)        
        return image.astype('float32'), label    def __len__(self):
        return len(self.data)
登录后复制
   
In [57]
# 创建train,valid数据集train_dataset = MyDataset(train_data, mode='train')
valid_dataset = MyDataset(valid_data, mode='valid', valid_start=int(classes.shape[0]*0.8))# 创建train,valid生成器,加速数据读取,设置batch size以及打乱训练数据的顺序train_loader = paddle.io.DataLoader(train_dataset, shuffle = True, batch_size = 32, drop_last=True)
valid_loader = paddle.io.DataLoader(valid_dataset, shuffle = False, batch_size = 10, drop_last=False)
登录后复制
   

测试集加载方法

In [58]
# 获取测试数据所有图片名称test_img_names = os.listdir('/home/aistudio/data/test')print(len(test_img_names))
登录后复制
       
1020
登录后复制
       
In [59]
# 按照数字顺序重新排序图片名称列表test_img_names_order = sorted(test_img_names, key=lambda x:int(x.split('.')[0].split('_')[-1]))# test_img_names_order
登录后复制
   
In [60]
# 创建测试数据集class TestDataset(paddle.io.Dataset):
    def __init__(self, data, mode='test'):
        super(TestDataset,self).__init__(),
        self.data = data
        self.mode = mode
        self.transform = T.Compose([                      
            T.Normalize(mean=[123.675, 116.28, 103.53],
                    std=[58.395, 57.12, 57.375], data_format='HWC'
            ), # 标准化
            T.Transpose() # 数据格式转换,Transpose默认参数(2,0,1)
        ])    def __getitem__(self, index):
        # 获取文件名
        img_name = self.data[index]        # 获取路径
        img_path = os.path.join(ROOT_DIR, 'test', self.data[index])        # 读取图片内容
        image =  load_image(img_path)        # 图形增广
        image = self.transform(image)        
        return image.astype('float32'), img_name    def __len__(self):
        return len(self.data)
登录后复制
   
In [61]
# 测试集加载test_dataset = TestDataset(data=test_img_names_order,mode='test')
登录后复制
   
In [62]
# 测试测试数据集,输出测试图片名称test_dataset[1][1]
登录后复制
       
'test_1.jpg'
登录后复制
               

模型组建

In [ ]
# 加载 ResNet50-DCN 模型,并预训练resnet_net = models.resnet50(pretrained=True, num_classes=24)
登录后复制
   

模型训练

In [11]
# 高层API封装模型model = paddle.Model(resnet_net)# 创建优化器,配置学习率optim = paddle.optimizer.Adam(learning_rate= 1e-4,parameters = model.parameters())# 训练配置model.prepare(
    optim,
    loss=paddle.nn.CrossEntropyLoss(),
    metrics=paddle.metric.Accuracy()
)

callback1 = paddle.callbacks.VisualDL(log_dir='visualdl_log_dir') # 可视化训练过程,提供给VisualDL监控机制# 开启训练model.fit(train_loader, valid_loader, epochs=90, eval_freq=2, verbose=1, callbacks=[callback1])
登录后复制
       
The loss value printed in the log is the current step, and the metric is the average value of previous steps.
Epoch 1/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 3.1927 - acc: 0.0419 - 503ms/step          
Eval begin...
step 43/43 [==============================] - loss: 3.1498 - acc: 0.0466 - 111ms/step          
Eval samples: 429
Epoch 2/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 3.2429 - acc: 0.0383 - 458ms/step          
Epoch 3/90
step 53/53 [==============================] - loss: 3.2695 - acc: 0.0448 - 441ms/step          
Eval begin...
step 43/43 [==============================] - loss: 3.3218 - acc: 0.0303 - 122ms/step           
Eval samples: 429
Epoch 4/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 3.1471 - acc: 0.0531 - 464ms/step          
Epoch 5/90
step 53/53 [==============================] - loss: 3.1612 - acc: 0.0513 - 466ms/step          
Eval begin...
step 43/43 [==============================] - loss: 3.1134 - acc: 0.0723 - 105ms/step          
Eval samples: 429
Epoch 6/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 3.1925 - acc: 0.0643 - 462ms/step          
Epoch 7/90
step 53/53 [==============================] - loss: 3.1686 - acc: 0.0660 - 463ms/step          
Eval begin...
step 43/43 [==============================] - loss: 2.5680 - acc: 0.0723 - 108ms/step          
Eval samples: 429
Epoch 8/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 3.1395 - acc: 0.0743 - 449ms/step          
Epoch 9/90
step 53/53 [==============================] - loss: 3.0175 - acc: 0.0784 - 485ms/step          
Eval begin...
step 43/43 [==============================] - loss: 2.7788 - acc: 0.0956 - 107ms/step          
Eval samples: 429
Epoch 10/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 2.9871 - acc: 0.0949 - 463ms/step          
Epoch 11/90
step 53/53 [==============================] - loss: 2.8698 - acc: 0.1138 - 454ms/step          
Eval begin...
step 43/43 [==============================] - loss: 2.8364 - acc: 0.0979 - 127ms/step          
Eval samples: 429
Epoch 12/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 2.8052 - acc: 0.1221 - 455ms/step          
Epoch 13/90
step 53/53 [==============================] - loss: 2.8459 - acc: 0.1616 - 459ms/step          
Eval begin...
step 43/43 [==============================] - loss: 2.6274 - acc: 0.1259 - 105ms/step          
Eval samples: 429
Epoch 14/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 2.5441 - acc: 0.1887 - 468ms/step          
Epoch 15/90
step 53/53 [==============================] - loss: 2.3795 - acc: 0.2246 - 467ms/step          
Eval begin...
step 43/43 [==============================] - loss: 1.9842 - acc: 0.2774 - 127ms/step          
Eval samples: 429
Epoch 16/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 2.3704 - acc: 0.2942 - 525ms/step          
Epoch 17/90
step 53/53 [==============================] - loss: 2.2671 - acc: 0.3290 - 486ms/step          
Eval begin...
step 43/43 [==============================] - loss: 1.4051 - acc: 0.3287 - 109ms/step          
Eval samples: 429
Epoch 18/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 1.9189 - acc: 0.3809 - 478ms/step          
Epoch 19/90
step 53/53 [==============================] - loss: 1.8451 - acc: 0.4381 - 488ms/step          
Eval begin...
step 43/43 [==============================] - loss: 1.2951 - acc: 0.4056 - 116ms/step          
Eval samples: 429
Epoch 20/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 1.6947 - acc: 0.5018 - 459ms/step          
Epoch 21/90
step 53/53 [==============================] - loss: 1.5430 - acc: 0.5389 - 470ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.9600 - acc: 0.5618 - 113ms/step          
Eval samples: 429
Epoch 22/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 1.3332 - acc: 0.6197 - 485ms/step          
Epoch 23/90
step 53/53 [==============================] - loss: 1.1644 - acc: 0.6268 - 447ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.6927 - acc: 0.6037 - 108ms/step          
Eval samples: 429
Epoch 24/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 1.0518 - acc: 0.6946 - 448ms/step          
Epoch 25/90
step 53/53 [==============================] - loss: 0.7262 - acc: 0.7329 - 472ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.5536 - acc: 0.6993 - 104ms/step          
Eval samples: 429
Epoch 26/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.6846 - acc: 0.7406 - 447ms/step          
Epoch 27/90
step 53/53 [==============================] - loss: 0.6977 - acc: 0.7765 - 479ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.5581 - acc: 0.7179 - 112ms/step          
Eval samples: 429
Epoch 28/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.5008 - acc: 0.8154 - 497ms/step          
Epoch 29/90
step 53/53 [==============================] - loss: 0.6462 - acc: 0.8438 - 458ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2971 - acc: 0.7086 - 102ms/step          
Eval samples: 429
Epoch 30/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.6442 - acc: 0.8526 - 466ms/step          
Epoch 31/90
step 53/53 [==============================] - loss: 0.4941 - acc: 0.8667 - 445ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2377 - acc: 0.7506 - 152ms/step          
Eval samples: 429
Epoch 32/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.3167 - acc: 0.9021 - 434ms/step          
Epoch 33/90
step 53/53 [==============================] - loss: 0.1696 - acc: 0.9116 - 466ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1650 - acc: 0.7646 - 109ms/step          
Eval samples: 429
Epoch 34/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1703 - acc: 0.9074 - 467ms/step          
Epoch 35/90
step 53/53 [==============================] - loss: 0.1415 - acc: 0.9216 - 465ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.3412 - acc: 0.7739 - 120ms/step          
Eval samples: 429
Epoch 36/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2013 - acc: 0.9304 - 439ms/step          
Epoch 37/90
step 53/53 [==============================] - loss: 0.1673 - acc: 0.9287 - 456ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1123 - acc: 0.7669 - 103ms/step          
Eval samples: 429
Epoch 38/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2036 - acc: 0.9499 - 475ms/step          
Epoch 39/90
step 53/53 [==============================] - loss: 0.1945 - acc: 0.9351 - 465ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.8336 - acc: 0.7529 - 115ms/step          
Eval samples: 429
Epoch 40/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1870 - acc: 0.9404 - 475ms/step          
Epoch 41/90
step 53/53 [==============================] - loss: 0.1277 - acc: 0.9611 - 451ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1676 - acc: 0.7669 - 116ms/step          
Eval samples: 429
Epoch 42/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2235 - acc: 0.9493 - 443ms/step          
Epoch 43/90
step 53/53 [==============================] - loss: 0.1583 - acc: 0.9452 - 472ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.6762 - acc: 0.7692 - 108ms/step          
Eval samples: 429
Epoch 44/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2066 - acc: 0.9469 - 468ms/step          
Epoch 45/90
step 53/53 [==============================] - loss: 0.0378 - acc: 0.9399 - 455ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2462 - acc: 0.7646 - 113ms/step          
Eval samples: 429
Epoch 46/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.3767 - acc: 0.9564 - 486ms/step          
Epoch 47/90
step 53/53 [==============================] - loss: 0.1808 - acc: 0.9593 - 448ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1106 - acc: 0.7646 - 121ms/step          
Eval samples: 429
Epoch 48/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1386 - acc: 0.9546 - 435ms/step          
Epoch 49/90
step 53/53 [==============================] - loss: 0.2941 - acc: 0.9640 - 494ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1883 - acc: 0.7599 - 116ms/step          
Eval samples: 429
Epoch 50/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1425 - acc: 0.9617 - 457ms/step          
Epoch 51/90
step 53/53 [==============================] - loss: 0.0866 - acc: 0.9617 - 481ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2950 - acc: 0.7622 - 105ms/step          
Eval samples: 429
Epoch 52/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1512 - acc: 0.9629 - 469ms/step          
Epoch 53/90
step 53/53 [==============================] - loss: 0.1060 - acc: 0.9752 - 446ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1060 - acc: 0.7949 - 111ms/step          
Eval samples: 429
Epoch 54/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0280 - acc: 0.9758 - 492ms/step          
Epoch 55/90
step 53/53 [==============================] - loss: 0.0724 - acc: 0.9752 - 466ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.5074 - acc: 0.7949 - 111ms/step          
Eval samples: 429
Epoch 56/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0205 - acc: 0.9723 - 474ms/step          
Epoch 57/90
step 53/53 [==============================] - loss: 0.0429 - acc: 0.9705 - 466ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2588 - acc: 0.7762 - 113ms/step          
Eval samples: 429
Epoch 58/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1691 - acc: 0.9723 - 467ms/step          
Epoch 59/90
step 53/53 [==============================] - loss: 0.1695 - acc: 0.9605 - 453ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.3465 - acc: 0.7576 - 112ms/step          
Eval samples: 429
Epoch 60/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2683 - acc: 0.9658 - 445ms/step          
Epoch 61/90
step 53/53 [==============================] - loss: 0.0321 - acc: 0.9634 - 447ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.4822 - acc: 0.7576 - 105ms/step          
Eval samples: 429
Epoch 62/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0861 - acc: 0.9664 - 486ms/step          
Epoch 63/90
step 53/53 [==============================] - loss: 0.0670 - acc: 0.9717 - 451ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1761 - acc: 0.7506 - 118ms/step          
Eval samples: 429
Epoch 64/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0327 - acc: 0.9758 - 464ms/step          
Epoch 65/90
step 53/53 [==============================] - loss: 0.0834 - acc: 0.9735 - 448ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2367 - acc: 0.7436 - 103ms/step          
Eval samples: 429
Epoch 66/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0353 - acc: 0.9782 - 462ms/step          
Epoch 67/90
step 53/53 [==============================] - loss: 0.2589 - acc: 0.9746 - 464ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.4783 - acc: 0.7832 - 111ms/step          
Eval samples: 429
Epoch 68/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0448 - acc: 0.9841 - 466ms/step          
Epoch 69/90
step 53/53 [==============================] - loss: 0.1902 - acc: 0.9776 - 457ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.1164 - acc: 0.7622 - 104ms/step          
Eval samples: 429
Epoch 70/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2499 - acc: 0.9746 - 465ms/step          
Epoch 71/90
step 53/53 [==============================] - loss: 0.1311 - acc: 0.9817 - 463ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2597 - acc: 0.7786 - 114ms/step          
Eval samples: 429
Epoch 72/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0990 - acc: 0.9752 - 446ms/step          
Epoch 73/90
step 53/53 [==============================] - loss: 0.0124 - acc: 0.9705 - 480ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.4437 - acc: 0.7413 - 120ms/step          
Eval samples: 429
Epoch 74/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1568 - acc: 0.9717 - 458ms/step          
Epoch 75/90
step 53/53 [==============================] - loss: 0.1583 - acc: 0.9711 - 464ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2247 - acc: 0.7599 - 112ms/step          
Eval samples: 429
Epoch 76/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1303 - acc: 0.9782 - 454ms/step          
Epoch 77/90
step 53/53 [==============================] - loss: 0.0870 - acc: 0.9717 - 454ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2439 - acc: 0.7622 - 115ms/step          
Eval samples: 429
Epoch 78/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1416 - acc: 0.9735 - 486ms/step          
Epoch 79/90
step 53/53 [==============================] - loss: 0.0401 - acc: 0.9794 - 455ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.5282 - acc: 0.7692 - 110ms/step          
Eval samples: 429
Epoch 80/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.0090 - acc: 0.9794 - 502ms/step          
Epoch 81/90
step 53/53 [==============================] - loss: 0.0869 - acc: 0.9699 - 458ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.4070 - acc: 0.7506 - 105ms/step          
Eval samples: 429
Epoch 82/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2644 - acc: 0.9658 - 445ms/step          
Epoch 83/90
step 53/53 [==============================] - loss: 0.0142 - acc: 0.9770 - 489ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.4563 - acc: 0.7436 - 109ms/step          
Eval samples: 429
Epoch 84/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1400 - acc: 0.9870 - 467ms/step          
Epoch 85/90
step 53/53 [==============================] - loss: 0.0161 - acc: 0.9705 - 458ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.4269 - acc: 0.7622 - 110ms/step          
Eval samples: 429
Epoch 86/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1201 - acc: 0.9853 - 479ms/step          
Epoch 87/90
step 53/53 [==============================] - loss: 0.0189 - acc: 0.9805 - 454ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.9262 - acc: 0.7506 - 131ms/step          
Eval samples: 429
Epoch 88/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.1661 - acc: 0.9770 - 472ms/step          
Epoch 89/90
step 53/53 [==============================] - loss: 0.1646 - acc: 0.9752 - 463ms/step          
Eval begin...
step 43/43 [==============================] - loss: 0.2383 - acc: 0.7786 - 111ms/step          
Eval samples: 429
Epoch 90/90
登录后复制
       
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/nn/layer/norm.py:712: UserWarning: When training, we now always track global mean and variance.
  "When training, we now always track global mean and variance."
登录后复制
       
step 53/53 [==============================] - loss: 0.2614 - acc: 0.9776 - 448ms/step
登录后复制
       

保存模型

In [ ]
model.save('checkpoint/resnet50-dcn')
登录后复制
   

模型推理

In [27]
# 创建模型resnet50_dcn = models.resnet50(pretrained=False, num_classes=24)# 封装模型model = paddle.Model(resnet50_dcn)# 加载参数model.load("checkpoint/resnet50-dcn.pdparams")# 加载模型配置model.prepare()
登录后复制
       
W0416 11:30:43.951839 11026 gpu_resources.cc:61] Please NOTE: device: 0, GPU Compute Capability: 7.0, Driver API Version: 11.2, Runtime API Version: 11.2
W0416 11:30:43.956612 11026 gpu_resources.cc:91] device: 0, cuDNN Version: 8.2.
登录后复制
       
In [28]
# 用 predict 在验证集上对模型进行预测test_result = model.predict(valid_dataset)# 由于模型是单一输出,test_result的形状为[1, 429, 24],429是测试数据集的数据量。#   这里打印第一个数据的结果,这个数组表示每个数字的预测得分# print(len(test_result))# print(test_result[0][0])
登录后复制
       
Predict begin...
step 429/429 [==============================] - 24ms/step          
Predict samples: 429
登录后复制
       
In [92]
len(valid_dataset)
登录后复制
       
429
登录后复制
               
In [127]
# 验证集准确率# 转换字典可以提高计算效率# 预测标签pred_labels = dict((i,pred_label.argmax()) for i, pred_label in enumerate(test_result[0]))# 实际标签labels = dict((i, label) for i, label in enumerate(valid_data['label']))

Acc = 0for i in range(len(pred_labels)):    if pred_labels[i] == labels[i]:
        Acc += 1print("Acc:", Acc / len(pred_labels))
登录后复制
       
Acc: 0.7668997668997669
登录后复制
       

效果展示

将预测结果与真实标签进行比较,通过一致性来判断模型的观点的正误。

In [16]
# 从验证集中取出一张图片,sample_n表示第几张,小于429sample_n = 219img, label = valid_dataset[sample_n]# 打印推理结果,这里的argmax函数用于取出预测值最高的一个的下标,作为预测标签pred_label = test_result[0][sample_n].argmax()print('true label: {}, pred label: {}'.format(label, pred_label))# 使用matplotlib库,可视化图片plt.imshow(img[0])
登录后复制
       
true label: 9, pred label: 9
登录后复制
       
<matplotlib.image.AxesImage at 0x7f6c64137150>
登录后复制
               
<Figure size 640x480 with 1 Axes>
登录后复制
               

测试结果及保存文件

In [142]
# 通过模型预测结果predicts = model.predict(test_data= test_dataset)
登录后复制
       
Predict begin...
step 1020/1020 [==============================] - 24ms/step          
Predict samples: 1020
登录后复制
       
In [195]
# predicts数据形状[1,1020,1,24],外层是 list[tuple[nparray]],获取得分最大的索引results = np.array(predicts).reshape((1020,24)).argmax(axis=1)
登录后复制
   
In [198]
# 预测结果转列表,每个元素是一个类别pre_labels = list(results)

pre_results = zip(test_img_names_order, pre_labels)
pre_results = list(pre_results)# 将结果保存为csv文件df = pd.DataFrame(pre_results, columns=['image','label'])
df.to_csv('submit.csv',index=False)
登录后复制
   

以上就是【AI达人特训营第三期】时频图分类项目的详细内容,更多请关注php中文网其它相关文章!

最佳 Windows 性能的顶级免费优化软件
最佳 Windows 性能的顶级免费优化软件

每个人都需要一台速度更快、更稳定的 PC。随着时间的推移,垃圾文件、旧注册表数据和不必要的后台进程会占用资源并降低性能。幸运的是,许多工具可以让 Windows 保持平稳运行。

下载
来源:php中文网
本文内容由网友自发贡献,版权归原作者所有,本站不承担相应法律责任。如您发现有涉嫌抄袭侵权的内容,请联系admin@php.cn
最新问题
开源免费商场系统广告
热门教程
更多>
最新下载
更多>
网站特效
网站源码
网站素材
前端模板
关于我们 免责申明 举报中心 意见反馈 讲师合作 广告合作 最新更新 English
php中文网:公益在线php培训,帮助PHP学习者快速成长!
关注服务号 技术交流群
PHP中文网订阅号
每天精选资源文章推送
PHP中文网APP
随时随地碎片化学习

Copyright 2014-2025 https://www.php.cn/ All Rights Reserved | php.cn | 湘ICP备2023035733号