Python 为什么卷积网络使用每64幅图像进行训练?

Python 为什么卷积网络使用每64幅图像进行训练?,python,python-3.x,tensorflow,tflearn,Python,Python 3.x,Tensorflow,Tflearn,我正在查找Python 3.5+TensorFlow+TFLearn的代码: # -*- coding: utf-8 -*- """ Convolutional Neural Network for MNIST dataset classification task. References: Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. "Gradient-based learning applied to document

我正在查找Python 3.5+TensorFlow+TFLearn的代码:

# -*- coding: utf-8 -*-

""" Convolutional Neural Network for MNIST dataset classification task.

References:
    Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. "Gradient-based
    learning applied to document recognition." Proceedings of the IEEE,
    86(11):2278-2324, November 1998.

Links:
    [MNIST Dataset] http://yann.lecun.com/exdb/mnist/

"""

from __future__ import division, print_function, absolute_import

import tflearn
from tflearn.layers.core import input_data, dropout, fully_connected
from tflearn.layers.conv import conv_2d, max_pool_2d
from tflearn.layers.normalization import local_response_normalization
from tflearn.layers.estimator import regression

# Data loading and preprocessing
import tflearn.datasets.mnist as mnist
X, Y, testX, testY = mnist.load_data(one_hot=True)
X = X.reshape([-1, 28, 28, 1])
testX = testX.reshape([-1, 28, 28, 1])

# Building convolutional network
network = input_data(shape=[None, 28, 28, 1], name='input')
network = conv_2d(network, 32, 3, activation='relu', regularizer="L2")
network = max_pool_2d(network, 2)
network = local_response_normalization(network)
network = conv_2d(network, 64, 3, activation='relu', regularizer="L2")
network = max_pool_2d(network, 2)
network = local_response_normalization(network)
network = fully_connected(network, 128, activation='tanh')
network = dropout(network, 0.8)
network = fully_connected(network, 256, activation='tanh')
network = dropout(network, 0.8)
network = fully_connected(network, 10, activation='softmax')
network = regression(network, optimizer='adam', learning_rate=0.01,
                     loss='categorical_crossentropy', name='target')

# Training
model = tflearn.DNN(network, tensorboard_verbose=0)
model.fit({'input': X}, {'target': Y}, n_epoch=20,
           validation_set=({'input': testX}, {'target': testY}),
           snapshot_step=100, show_metric=True, run_id='convnet_mnist')
好的,它起作用了。但在学习过程中,它只使用了来自集合的每64幅图像。为什么会这样? 如果我有一个小集合,并且希望网络使用每个第一个图像,我应该怎么做

培训信息示例

Training Step: 1  | time: 2.416s
| Adam | epoch: 001 | loss: 0.00000 -- iter: 064/55000
Training Step: 2  | total loss: 0.24470 | time: 4.621s
| Adam | epoch: 001 | loss: 0.24470 -- iter: 128/55000
Training Step: 3  | total loss: 0.10852 | time: 6.876s
| Adam | epoch: 001 | loss: 0.10852 -- iter: 192/55000
Training Step: 4  | total loss: 0.20421 | time: 9.129s
| Adam | epoch: 001 | loss: 0.20421 -- iter: 256/55000

它不仅使用每一个第64个图像,还加载了64个图像的批处理。这就是为什么国际热核实验堆每次增加64幅,因为它在每个训练步骤中处理了64幅图像。 查看回归层的文档,在这里您可以设置批量大小

我自己做的。 此参数由模型中的批量大小调节。适合。默认值是64。 因此,要使用每个第一个图像,您需要用下一种方法重写最后一个字符串:

model.fit({'input': X}, {'target': Y}, n_epoch=20,
           validation_set=({'input': testX}, {'target': testY}),
           snapshot_step=100, batch_size=1, 
           show_metric=True,run_id='convnet_mnist')

是的,你说得对,但我已经设法自己做了。不过,非常感谢。它使用每个图像进行训练,即使批量大小为64。它只是没有为每一个打印一行。那么为什么批大小为64的训练步骤比批大小为1的训练步骤少呢?因为每一步训练64个图像。