Warning: file_get_contents(/data/phpspider/zhask/data//catemap/2/python/298.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181

Warning: file_get_contents(/data/phpspider/zhask/data//catemap/9/opencv/3.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
Python yolo but-nan和nan的培训模式_Python_Opencv_Image Processing_Artificial Intelligence_Yolo - Fatal编程技术网

Python yolo but-nan和nan的培训模式

Python yolo but-nan和nan的培训模式,python,opencv,image-processing,artificial-intelligence,yolo,Python,Opencv,Image Processing,Artificial Intelligence,Yolo,我想在yolo上训练模特儿,但一步后给nan和-nan 我有300张不同大小的图片(几乎600*600) 一类用于图像检测。之前,我对100张图像给出了很好的结果(%75检测精度) 但我想给出最好的结果 tiny_yolo.cfg 我分割了80-20个列车和测试数据 我用这个 请帮帮我 我忘了在黑暗中配置makefile。 我使用谷歌Colab,首先必须定义使用GPU和CUDNN [net] batch=64 subdivisions=8 width=416 height=416 chann

我想在yolo上训练模特儿,但一步后给nan和-nan 我有300张不同大小的图片(几乎600*600) 一类用于图像检测。之前,我对100张图像给出了很好的结果(%75检测精度) 但我想给出最好的结果

tiny_yolo.cfg

我分割了80-20个列车和测试数据 我用这个


请帮帮我

我忘了在黑暗中配置makefile。 我使用谷歌Colab,首先必须定义使用GPU和CUDNN

[net]
batch=64
subdivisions=8
width=416
height=416
channels=3
momentum=0.9
decay=0.0005
angle=0
saturation = 1.5
exposure = 1.5
hue=.1

learning_rate=0.001
max_batches = 120000
policy=steps
steps=-1,100,80000,100000
scales=.1,10,.1,.1

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=2

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

[maxpool]
size=2
stride=1

[convolutional]
batch_normalize=1
filters=30
size=3
stride=1
pad=1
activation=leaky

###########

[convolutional]
batch_normalize=1
size=3
stride=1
pad=1
filters=30
activation=leaky

[convolutional]
size=1
stride=1
pad=1
filters=30
activation=linear

[region]
anchors = 0.738768,0.874946,  2.42204,2.65704,  4.30971,7.04493,  10.246,4.59428,  12.6868,11.8741
bias_match=1
classes=1
coords=4
num=5
softmax=1
jitter=.2
rescore=1
small_object=1

object_scale=5
noobject_scale=1
class_scale=1
coord_scale=1

absolute=1
thresh = .6
random=1