VGG19网络的Keras实现有26层。怎么用?
VGG-19网络有25层,如图所示。但是如果我检查Keras实现中的层数,它会显示26层。怎么做VGG19网络的Keras实现有26层。怎么用?,keras,vgg-net,Keras,Vgg Net,VGG-19网络有25层,如图所示。但是如果我检查Keras实现中的层数,它会显示26层。怎么做 model = VGG19() len(model.layers) 输出 26 如果您感到困惑,可以使用model.summary()直接打印出VGG19的结构。它显示一个层input\u 1(InputLayer)作为输入层 _________________________________________________________________ Layer (type)
model = VGG19()
len(model.layers)
输出
26
如果您感到困惑,可以使用
model.summary()
直接打印出VGG19
的结构。它显示一个层input\u 1(InputLayer)
作为输入层
_________________________________________________________________
Layer (type) Output Shape Param #
=================================================================
input_1 (InputLayer) (None, 224, 224, 3) 0
_________________________________________________________________
block1_conv1 (Conv2D) (None, 224, 224, 64) 1792
_________________________________________________________________
block1_conv2 (Conv2D) (None, 224, 224, 64) 36928
_________________________________________________________________
block1_pool (MaxPooling2D) (None, 112, 112, 64) 0
_________________________________________________________________
block2_conv1 (Conv2D) (None, 112, 112, 128) 73856
_________________________________________________________________
block2_conv2 (Conv2D) (None, 112, 112, 128) 147584
_________________________________________________________________
block2_pool (MaxPooling2D) (None, 56, 56, 128) 0
_________________________________________________________________
block3_conv1 (Conv2D) (None, 56, 56, 256) 295168
_________________________________________________________________
block3_conv2 (Conv2D) (None, 56, 56, 256) 590080
_________________________________________________________________
block3_conv3 (Conv2D) (None, 56, 56, 256) 590080
_________________________________________________________________
block3_conv4 (Conv2D) (None, 56, 56, 256) 590080
_________________________________________________________________
block3_pool (MaxPooling2D) (None, 28, 28, 256) 0
_________________________________________________________________
block4_conv1 (Conv2D) (None, 28, 28, 512) 1180160
_________________________________________________________________
block4_conv2 (Conv2D) (None, 28, 28, 512) 2359808
_________________________________________________________________
block4_conv3 (Conv2D) (None, 28, 28, 512) 2359808
_________________________________________________________________
block4_conv4 (Conv2D) (None, 28, 28, 512) 2359808
_________________________________________________________________
block4_pool (MaxPooling2D) (None, 14, 14, 512) 0
_________________________________________________________________
block5_conv1 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_conv2 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_conv3 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_conv4 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_pool (MaxPooling2D) (None, 7, 7, 512) 0
_________________________________________________________________
flatten (Flatten) (None, 25088) 0
_________________________________________________________________
fc1 (Dense) (None, 4096) 102764544
_________________________________________________________________
fc2 (Dense) (None, 4096) 16781312
_________________________________________________________________
predictions (Dense) (None, 1000) 4097000
=================================================================
Total params: 143,667,240
Trainable params: 143,667,240
Non-trainable params: 0
_________________________________________________________________
如果要从第一个FC层获取输出,应使用model.layers[23]
而不是22
。实际上,您可以直接打印出形状,并将其与model.summary()
的输出进行比较
此外,您可以使用层名'fc1'
直接获得第一个FC层
print(model.get_layer('fc1').output.shape)
(?, 4096)
如果您感到困惑,可以使用
model.summary()
直接打印出VGG19
的结构。它显示一个层input\u 1(InputLayer)
作为输入层
_________________________________________________________________
Layer (type) Output Shape Param #
=================================================================
input_1 (InputLayer) (None, 224, 224, 3) 0
_________________________________________________________________
block1_conv1 (Conv2D) (None, 224, 224, 64) 1792
_________________________________________________________________
block1_conv2 (Conv2D) (None, 224, 224, 64) 36928
_________________________________________________________________
block1_pool (MaxPooling2D) (None, 112, 112, 64) 0
_________________________________________________________________
block2_conv1 (Conv2D) (None, 112, 112, 128) 73856
_________________________________________________________________
block2_conv2 (Conv2D) (None, 112, 112, 128) 147584
_________________________________________________________________
block2_pool (MaxPooling2D) (None, 56, 56, 128) 0
_________________________________________________________________
block3_conv1 (Conv2D) (None, 56, 56, 256) 295168
_________________________________________________________________
block3_conv2 (Conv2D) (None, 56, 56, 256) 590080
_________________________________________________________________
block3_conv3 (Conv2D) (None, 56, 56, 256) 590080
_________________________________________________________________
block3_conv4 (Conv2D) (None, 56, 56, 256) 590080
_________________________________________________________________
block3_pool (MaxPooling2D) (None, 28, 28, 256) 0
_________________________________________________________________
block4_conv1 (Conv2D) (None, 28, 28, 512) 1180160
_________________________________________________________________
block4_conv2 (Conv2D) (None, 28, 28, 512) 2359808
_________________________________________________________________
block4_conv3 (Conv2D) (None, 28, 28, 512) 2359808
_________________________________________________________________
block4_conv4 (Conv2D) (None, 28, 28, 512) 2359808
_________________________________________________________________
block4_pool (MaxPooling2D) (None, 14, 14, 512) 0
_________________________________________________________________
block5_conv1 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_conv2 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_conv3 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_conv4 (Conv2D) (None, 14, 14, 512) 2359808
_________________________________________________________________
block5_pool (MaxPooling2D) (None, 7, 7, 512) 0
_________________________________________________________________
flatten (Flatten) (None, 25088) 0
_________________________________________________________________
fc1 (Dense) (None, 4096) 102764544
_________________________________________________________________
fc2 (Dense) (None, 4096) 16781312
_________________________________________________________________
predictions (Dense) (None, 1000) 4097000
=================================================================
Total params: 143,667,240
Trainable params: 143,667,240
Non-trainable params: 0
_________________________________________________________________
如果要从第一个FC层获取输出,应使用model.layers[23]
而不是22
。实际上,您可以直接打印出形状,并将其与model.summary()
的输出进行比较
此外,您可以使用层名'fc1'
直接获得第一个FC层
print(model.get_layer('fc1').output.shape)
(?, 4096)
VGG-19中的
19
指的是具有可学习权重的层。如果打印模型摘要,您将获得以下信息
这里有
7个
层,它们没有任何可学习的权重。这是一个InputLayer
,五个MaxPooling2D
层和一个Flatten
层。这就是如何获得26
层(19+1+5+1)
VGG-19中的19
指具有可学习权重的层。如果打印模型摘要,您将获得以下信息
这里有
7个
层,它们没有任何可学习的权重。这是一个InputLayer
,五个MaxPooling2D
层和一个Flatten
层。这就是如何获得26
层(19+1+5+1)
model.summary()
将显示一个层input\u 1(InputLayer)
作为输入层。因此,如果我必须从第一个FC层获取输出,我应该执行model.layers[23]
而不是22
?您不能列出这些层吗?通常会自动添加一个“输出”层。model.summary()
将显示一个层input\u 1(InputLayer)
作为输入层。因此,如果我必须从第一个FC层获得输出,我应该做model.layers[23]
而不是22
?您不能列出这些层吗?通常会自动添加一个“输出”层。