首页
学习
活动
专区
工具
TVP
发布
精选内容/技术社群/优惠产品,尽在小程序
立即前往
您找到你想要的搜索结果了吗?
是的
没有找到

手写体数字识别

Extracting MNIST_data/train-images-idx3-ubyte.gz Extracting MNIST_data/train-labels-idx1-ubyte.gz Extracting MNIST_data/t10k-images-idx3-ubyte.gz Extracting MNIST_data/t10k-labels-idx1-ubyte.gz Train Epoch: 01 Loss= 53.529132843 Accuracy= 0.8624 Train Epoch: 02 Loss= 33.633792877 Accuracy= 0.9002 Train Epoch: 03 Loss= 26.070878983 Accuracy= 0.9166 Train Epoch: 04 Loss= 21.878053665 Accuracy= 0.9256 Train Epoch: 05 Loss= 19.345390320 Accuracy= 0.9310 Train Epoch: 06 Loss= 18.134420395 Accuracy= 0.9348 Train Epoch: 07 Loss= 15.690796852 Accuracy= 0.9414 Train Epoch: 08 Loss= 15.634956360 Accuracy= 0.9372 Train Epoch: 09 Loss= 15.122309685 Accuracy= 0.9414 Train Epoch: 10 Loss= 14.535149574 Accuracy= 0.9426 Train Epoch: 11 Loss= 14.188427925 Accuracy= 0.9450 Train Epoch: 12 Loss= 14.709759712 Accuracy= 0.9426 Train Epoch: 13 Loss= 13.977644920 Accuracy= 0.9448 Train Epoch: 14 Loss= 13.594002724 Accuracy= 0.9466 Train Epoch: 15 Loss= 12.868132591 Accuracy= 0.9492 Train Epoch: 16 Loss= 12.838119507 Accuracy= 0.9506 Train Epoch: 17 Loss= 12.817976952 Accuracy= 0.9496 Train Epoch: 18 Loss= 12.890332222 Accuracy= 0.9506 Train Epoch: 19 Loss= 12.724534988 Accuracy= 0.9502 Train Epoch: 20 Loss= 13.171916008 Accuracy= 0.9494 Train Epoch: 21 Loss= 12.193360329 Accuracy= 0.9558 Train Epoch: 22 Loss= 11.771809578 Accuracy= 0.9516 Train Epoch: 23 Loss= 12.657453537 Accuracy= 0.9532 Train Epoch: 24 Loss= 12.012898445 Accuracy= 0.9552 Train Epoch: 25 Loss= 12.073326111 Accuracy= 0.9542 Train Epoch: 26 Loss= 12.455985069 Accuracy= 0.9556 Train Epoch: 27 Loss= 11.321227074 Accuracy= 0.9564 Train Epoch: 28 Loss= 12.093022346 Accuracy= 0.9568 Train Epoch: 29 Loss= 11.713661194 Accuracy= 0.9580 Train Epoch: 30 Loss= 11.451450348 Accuracy= 0.9588 Train Finished takes: 76.92 Starting another session for prediction

02

手把手带你Transformer图像分类

使用Transformer来提升模型的性能 最近几年,Transformer体系结构已成为自然语言处理任务的实际标准, 但其在计算机视觉中的应用还受到限制。在视觉上,注意力要么与卷积网络结合使用, 要么用于替换卷积网络的某些组件,同时将其整体结构保持在适当的位置。2020年10月22日,谷歌人工智能研究院发表一篇题为“An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale”的文章。文章将图像切割成一个个图像块,组成序列化的数据输入Transformer执行图像分类任务。当对大量数据进行预训练并将其传输到多个中型或小型图像识别数据集(如ImageNet、CIFAR-100、VTAB等)时,与目前的卷积网络相比,Vision Transformer(ViT)获得了出色的结果,同时所需的计算资源也大大减少。 这里我们以ViT我模型,实现对数据CiFar10的分类工作,模型性能得到进一步的提升。

01
领券