caffe搭建以及初步学习--win7-vs2013-gtx650tiboost-cuda8.0-cifar10训练和测试-2-完整解决方案cifar10_full_solver.prototxt

首先总结前一节的内容。

简单的讲,就是训练并测试了快速解决方案。

转换数据格式:

convert_cifar_data.exe data/cifar10 examples/cifar10 lmdb

计算平均值

compute_image_mean.exe -backend=lmdb examples/cifar10/cifar10_train_lmdb examples/cifar10/mean.binaryproto

训练网络--快速解决方案

caffe train --solver=examples/cifar10/cifar10_quick_solver.prototxt

继续训练网络--再多迭代1000次

caffe train --solver=examples/cifar10/cifar10_quick_solver_lr1.prototxt --snapshot=examples/cifar10/cifar10_quick_iter_4000.solverstate

测试模型的准确率

caffe test -model examples/cifar10/cifar10_quick_train_test.prototxt -weights examples/cifar10/cifar10_quick_iter_5000.caffemodel.h5 -iterations 100

识别一只猫

classification.exe examples/cifar10/cifar10_quick.prototxt examples/cifar10/cifar10_quick_iter_5000.caffemodel.h5 examples/cifar10/mean.binaryproto data/cifar10/synset_words.txt examples/images/cat.jpg

识别人鱼自行车

classification.exe examples/cifar10/cifar10_quick.prototxt examples/cifar10/cifar10_quick_iter_5000.caffemodel.h5 examples/cifar10/mean.binaryproto data/cifar10/synset_words.txt examples/images/fish-bike

===========================================华丽的分割线=============================================

现在开始 尝试 完整解决方案 cifar10_full_solver.prototxt

1. 训练

caffe train --solver=examples/cifar10/cifar10_full_solver.prototxt

原来以为半个小时能跑完的,结果发现,40分钟只跑了10000次迭代。总数是6万。所以,跑了很久。

好在最后还是跑完了。以后事先要先预估一下计算量才好,不然,跑了也是白跑。

2.测试

测试:做10次迭代的结果

caffe test -model examples/cifar10/cifar10_full_train_test.prototxt -weights examples/cifar10

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$ caffe test -model examples/cifar10/cifar10_full_train_test.prototxt -weights examples/cifar10/cifar10_full_iter_60000.caffemodel.h5 -iterations 10

------------------------------------------------------------------------------

I0704 12:26:43.571471 5465 caffe.cpp:284] Use CPU.

I0704 12:26:45.302641 5465 net.cpp:294] The NetState phase (1) differed from the phase (0) specified by a rule in layer cifar

I0704 12:26:45.302878 5465 net.cpp:51] Initializing net from parameters:

name: "CIFAR10_full"

state {

phase: TEST

level: 0

stage: ""

}

layer {

name: "cifar"

type: "Data"

top: "data"

top: "label"

include {

phase: TEST

}

transform_param {

mean_file: "examples/cifar10/mean.binaryproto"

}

data_param {

source: "examples/cifar10/cifar10_test_lmdb"

batch_size: 100

backend: LMDB

}

}

layer {

name: "conv1"

type: "Convolution"

bottom: "data"

top: "conv1"

param {

lr_mult: 1

}

param {

lr_mult: 2

}

convolution_param {

num_output: 32

pad: 2

kernel_size: 5

stride: 1

weight_filler {

type: "gaussian"

std: 0.0001

}

bias_filler {

type: "constant"

}

}

}

layer {

name: "pool1"

type: "Pooling"

bottom: "conv1"

top: "pool1"

pooling_param {

pool: MAX

kernel_size: 3

stride: 2

}

}

layer {

name: "relu1"

type: "ReLU"

bottom: "pool1"

top: "pool1"

}

layer {

name: "norm1"

type: "LRN"

bottom: "pool1"

top: "norm1"

lrn_param {

local_size: 3

alpha: 5e-05

beta: 0.75

norm_region: WITHIN_CHANNEL

}

}

layer {

name: "conv2"

type: "Convolution"

bottom: "norm1"

top: "conv2"

param {

lr_mult: 1

}

param {

lr_mult: 2

}

convolution_param {

num_output: 32

pad: 2

kernel_size: 5

stride: 1

weight_filler {

type: "gaussian"

std: 0.01

}

bias_filler {

type: "constant"

}

}

}

layer {

name: "relu2"

type: "ReLU"

bottom: "conv2"

top: "conv2"

}

layer {

name: "pool2"

type: "Pooling"

bottom: "conv2"

top: "pool2"

pooling_param {

pool: AVE

kernel_size: 3

stride: 2

}

}

layer {

name: "norm2"

type: "LRN"

bottom: "pool2"

top: "norm2"

lrn_param {

local_size: 3

alpha: 5e-05

beta: 0.75

norm_region: WITHIN_CHANNEL

}

}

layer {

name: "conv3"

type: "Convolution"

bottom: "norm2"

top: "conv3"

convolution_param {

num_output: 64

pad: 2

kernel_size: 5

stride: 1

weight_filler {

type: "gaussian"

std: 0.01

}

bias_filler {

type: "constant"

}

}

}

layer {

name: "relu3"

type: "ReLU"

bottom: "conv3"

top: "conv3"

}

layer {

name: "pool3"

type: "Pooling"

bottom: "conv3"

top: "pool3"

pooling_param {

pool: AVE

kernel_size: 3

stride: 2

}

}

layer {

name: "ip1"

type: "InnerProduct"

bottom: "pool3"

top: "ip1"

param {

lr_mult: 1

decay_mult: 250

}

param {

lr_mult: 2

decay_mult: 0

}

inner_product_param {

num_output: 10

weight_filler {

type: "gaussian"

std: 0.01

}

bias_filler {

type: "constant"

}

}

}

layer {

name: "accuracy"

type: "Accuracy"

bottom: "ip1"

bottom: "label"

top: "accuracy"

include {

phase: TEST

}

}

layer {

name: "loss"

type: "SoftmaxWithLoss"

bottom: "ip1"

bottom: "label"

top: "loss"

}

I0704 12:26:45.329610 5465 layer_factory.hpp:77] Creating layer cifar

I0704 12:26:45.329771 5465 db_lmdb.cpp:35] Opened lmdb examples/cifar10/cifar10_test_lmdb

I0704 12:26:45.329805 5465 net.cpp:84] Creating Layer cifar

I0704 12:26:45.329823 5465 net.cpp:380] cifar -> data

I0704 12:26:45.329852 5465 net.cpp:380] cifar -> label

I0704 12:26:45.329874 5465 data_transformer.cpp:25] Loading mean file from: examples/cifar10/mean.binaryproto

I0704 12:26:45.329970 5465 data_layer.cpp:45] output data size: 100,3,32,32

I0704 12:26:45.342496 5465 net.cpp:122] Setting up cifar

I0704 12:26:45.342577 5465 net.cpp:129] Top shape: 100 3 32 32 (307200)

I0704 12:26:45.342600 5465 net.cpp:129] Top shape: 100 (100)

I0704 12:26:45.342612 5465 net.cpp:137] Memory required for data: 1229200

I0704 12:26:45.342634 5465 layer_factory.hpp:77] Creating layer label_cifar_1_split

I0704 12:26:45.352715 5465 net.cpp:84] Creating Layer label_cifar_1_split

I0704 12:26:45.352736 5465 net.cpp:406] label_cifar_1_split <- label

I0704 12:26:45.352809 5465 net.cpp:380] label_cifar_1_split -> label_cifar_1_split_0

I0704 12:26:45.352833 5465 net.cpp:380] label_cifar_1_split -> label_cifar_1_split_1

I0704 12:26:45.352854 5465 net.cpp:122] Setting up label_cifar_1_split

I0704 12:26:45.352866 5465 net.cpp:129] Top shape: 100 (100)

I0704 12:26:45.352890 5465 net.cpp:129] Top shape: 100 (100)

I0704 12:26:45.352898 5465 net.cpp:137] Memory required for data: 1230000

I0704 12:26:45.352908 5465 layer_factory.hpp:77] Creating layer conv1

I0704 12:26:45.352936 5465 net.cpp:84] Creating Layer conv1

I0704 12:26:45.352947 5465 net.cpp:406] conv1 <- data

I0704 12:26:45.352963 5465 net.cpp:380] conv1 -> conv1

I0704 12:26:47.343575 5465 net.cpp:122] Setting up conv1

I0704 12:26:47.343633 5465 net.cpp:129] Top shape: 100 32 32 32 (3276800)

I0704 12:26:47.343646 5465 net.cpp:137] Memory required for data: 14337200

I0704 12:26:47.343694 5465 layer_factory.hpp:77] Creating layer pool1

I0704 12:26:47.343719 5465 net.cpp:84] Creating Layer pool1

I0704 12:26:47.343730 5465 net.cpp:406] pool1 <- conv1

I0704 12:26:47.343745 5465 net.cpp:380] pool1 -> pool1

I0704 12:26:47.343776 5465 net.cpp:122] Setting up pool1

I0704 12:26:47.343789 5465 net.cpp:129] Top shape: 100 32 16 16 (819200)

I0704 12:26:47.343799 5465 net.cpp:137] Memory required for data: 17614000

I0704 12:26:47.343809 5465 layer_factory.hpp:77] Creating layer relu1

I0704 12:26:47.343827 5465 net.cpp:84] Creating Layer relu1

I0704 12:26:47.343838 5465 net.cpp:406] relu1 <- pool1

I0704 12:26:47.343852 5465 net.cpp:367] relu1 -> pool1 (in-place)

I0704 12:26:47.344049 5465 net.cpp:122] Setting up relu1

I0704 12:26:47.344064 5465 net.cpp:129] Top shape: 100 32 16 16 (819200)

I0704 12:26:47.344074 5465 net.cpp:137] Memory required for data: 20890800

I0704 12:26:47.344084 5465 layer_factory.hpp:77] Creating layer norm1

I0704 12:26:47.344103 5465 net.cpp:84] Creating Layer norm1

I0704 12:26:47.344115 5465 net.cpp:406] norm1 <- pool1

I0704 12:26:47.344127 5465 net.cpp:380] norm1 -> norm1

I0704 12:26:47.344894 5465 net.cpp:122] Setting up norm1

I0704 12:26:47.344913 5465 net.cpp:129] Top shape: 100 32 16 16 (819200)

I0704 12:26:47.344923 5465 net.cpp:137] Memory required for data: 24167600

I0704 12:26:47.344933 5465 layer_factory.hpp:77] Creating layer conv2

I0704 12:26:47.344955 5465 net.cpp:84] Creating Layer conv2

I0704 12:26:47.344966 5465 net.cpp:406] conv2 <- norm1

I0704 12:26:47.344981 5465 net.cpp:380] conv2 -> conv2

I0704 12:26:47.346536 5465 net.cpp:122] Setting up conv2

I0704 12:26:47.346554 5465 net.cpp:129] Top shape: 100 32 16 16 (819200)

I0704 12:26:47.346565 5465 net.cpp:137] Memory required for data: 27444400

I0704 12:26:47.346583 5465 layer_factory.hpp:77] Creating layer relu2

I0704 12:26:47.346596 5465 net.cpp:84] Creating Layer relu2

I0704 12:26:47.346607 5465 net.cpp:406] relu2 <- conv2

I0704 12:26:47.346621 5465 net.cpp:367] relu2 -> conv2 (in-place)

I0704 12:26:47.346993 5465 net.cpp:122] Setting up relu2

I0704 12:26:47.347010 5465 net.cpp:129] Top shape: 100 32 16 16 (819200)

I0704 12:26:47.347021 5465 net.cpp:137] Memory required for data: 30721200

I0704 12:26:47.347031 5465 layer_factory.hpp:77] Creating layer pool2

I0704 12:26:47.347048 5465 net.cpp:84] Creating Layer pool2

I0704 12:26:47.347059 5465 net.cpp:406] pool2 <- conv2

I0704 12:26:47.347071 5465 net.cpp:380] pool2 -> pool2

I0704 12:26:47.347445 5465 net.cpp:122] Setting up pool2

I0704 12:26:47.347462 5465 net.cpp:129] Top shape: 100 32 8 8 (204800)

I0704 12:26:47.347472 5465 net.cpp:137] Memory required for data: 31540400

I0704 12:26:47.347482 5465 layer_factory.hpp:77] Creating layer norm2

I0704 12:26:47.347497 5465 net.cpp:84] Creating Layer norm2

I0704 12:26:47.347507 5465 net.cpp:406] norm2 <- pool2

I0704 12:26:47.347522 5465 net.cpp:380] norm2 -> norm2

I0704 12:26:47.348095 5465 net.cpp:122] Setting up norm2

I0704 12:26:47.348112 5465 net.cpp:129] Top shape: 100 32 8 8 (204800)

I0704 12:26:47.348122 5465 net.cpp:137] Memory required for data: 32359600

I0704 12:26:47.348132 5465 layer_factory.hpp:77] Creating layer conv3

I0704 12:26:47.348186 5465 net.cpp:84] Creating Layer conv3

I0704 12:26:47.348197 5465 net.cpp:406] conv3 <- norm2

I0704 12:26:47.348212 5465 net.cpp:380] conv3 -> conv3

I0704 12:26:47.358871 5465 net.cpp:122] Setting up conv3

I0704 12:26:47.358929 5465 net.cpp:129] Top shape: 100 64 8 8 (409600)

I0704 12:26:47.358940 5465 net.cpp:137] Memory required for data: 33998000

I0704 12:26:47.358973 5465 layer_factory.hpp:77] Creating layer relu3

I0704 12:26:47.358996 5465 net.cpp:84] Creating Layer relu3

I0704 12:26:47.359007 5465 net.cpp:406] relu3 <- conv3

I0704 12:26:47.359025 5465 net.cpp:367] relu3 -> conv3 (in-place)

I0704 12:26:47.359354 5465 net.cpp:122] Setting up relu3

I0704 12:26:47.359377 5465 net.cpp:129] Top shape: 100 64 8 8 (409600)

I0704 12:26:47.359387 5465 net.cpp:137] Memory required for data: 35636400

I0704 12:26:47.359397 5465 layer_factory.hpp:77] Creating layer pool3

I0704 12:26:47.359411 5465 net.cpp:84] Creating Layer pool3

I0704 12:26:47.359422 5465 net.cpp:406] pool3 <- conv3

I0704 12:26:47.359439 5465 net.cpp:380] pool3 -> pool3

I0704 12:26:47.359905 5465 net.cpp:122] Setting up pool3

I0704 12:26:47.359925 5465 net.cpp:129] Top shape: 100 64 4 4 (102400)

I0704 12:26:47.359935 5465 net.cpp:137] Memory required for data: 36046000

I0704 12:26:47.359944 5465 layer_factory.hpp:77] Creating layer ip1

I0704 12:26:47.359964 5465 net.cpp:84] Creating Layer ip1

I0704 12:26:47.359975 5465 net.cpp:406] ip1 <- pool3

I0704 12:26:47.359992 5465 net.cpp:380] ip1 -> ip1

I0704 12:26:47.360213 5465 net.cpp:122] Setting up ip1

I0704 12:26:47.360225 5465 net.cpp:129] Top shape: 100 10 (1000)

I0704 12:26:47.360234 5465 net.cpp:137] Memory required for data: 36050000

I0704 12:26:47.360249 5465 layer_factory.hpp:77] Creating layer ip1_ip1_0_split

I0704 12:26:47.360265 5465 net.cpp:84] Creating Layer ip1_ip1_0_split

I0704 12:26:47.360276 5465 net.cpp:406] ip1_ip1_0_split <- ip1

I0704 12:26:47.360291 5465 net.cpp:380] ip1_ip1_0_split -> ip1_ip1_0_split_0

I0704 12:26:47.360307 5465 net.cpp:380] ip1_ip1_0_split -> ip1_ip1_0_split_1

I0704 12:26:47.360324 5465 net.cpp:122] Setting up ip1_ip1_0_split

I0704 12:26:47.360337 5465 net.cpp:129] Top shape: 100 10 (1000)

I0704 12:26:47.360347 5465 net.cpp:129] Top shape: 100 10 (1000)

I0704 12:26:47.360355 5465 net.cpp:137] Memory required for data: 36058000

I0704 12:26:47.360365 5465 layer_factory.hpp:77] Creating layer accuracy

I0704 12:26:47.360388 5465 net.cpp:84] Creating Layer accuracy

I0704 12:26:47.360399 5465 net.cpp:406] accuracy <- ip1_ip1_0_split_0

I0704 12:26:47.360410 5465 net.cpp:406] accuracy <- label_cifar_1_split_0

I0704 12:26:47.360426 5465 net.cpp:380] accuracy -> accuracy

I0704 12:26:47.360442 5465 net.cpp:122] Setting up accuracy

I0704 12:26:47.360455 5465 net.cpp:129] Top shape: (1)

I0704 12:26:47.360463 5465 net.cpp:137] Memory required for data: 36058004

I0704 12:26:47.360472 5465 layer_factory.hpp:77] Creating layer loss

I0704 12:26:47.360492 5465 net.cpp:84] Creating Layer loss

I0704 12:26:47.360503 5465 net.cpp:406] loss <- ip1_ip1_0_split_1

I0704 12:26:47.360514 5465 net.cpp:406] loss <- label_cifar_1_split_1

I0704 12:26:47.360528 5465 net.cpp:380] loss -> loss

I0704 12:26:47.360553 5465 layer_factory.hpp:77] Creating layer loss

I0704 12:26:47.360777 5465 net.cpp:122] Setting up loss

I0704 12:26:47.360795 5465 net.cpp:129] Top shape: (1)

I0704 12:26:47.360805 5465 net.cpp:132] with loss weight 1

I0704 12:26:47.360834 5465 net.cpp:137] Memory required for data: 36058008

I0704 12:26:47.360846 5465 net.cpp:198] loss needs backward computation.

I0704 12:26:47.360860 5465 net.cpp:200] accuracy does not need backward computation.

I0704 12:26:47.360872 5465 net.cpp:198] ip1_ip1_0_split needs backward computation.

I0704 12:26:47.360882 5465 net.cpp:198] ip1 needs backward computation.

I0704 12:26:47.360891 5465 net.cpp:198] pool3 needs backward computation.

I0704 12:26:47.360901 5465 net.cpp:198] relu3 needs backward computation.

I0704 12:26:47.360911 5465 net.cpp:198] conv3 needs backward computation.

I0704 12:26:47.360954 5465 net.cpp:198] norm2 needs backward computation.

I0704 12:26:47.360965 5465 net.cpp:198] pool2 needs backward computation.

I0704 12:26:47.360975 5465 net.cpp:198] relu2 needs backward computation.

I0704 12:26:47.360985 5465 net.cpp:198] conv2 needs backward computation.

I0704 12:26:47.360996 5465 net.cpp:198] norm1 needs backward computation.

I0704 12:26:47.361006 5465 net.cpp:198] relu1 needs backward computation.

I0704 12:26:47.361016 5465 net.cpp:198] pool1 needs backward computation.

I0704 12:26:47.361026 5465 net.cpp:198] conv1 needs backward computation.

I0704 12:26:47.361037 5465 net.cpp:200] label_cifar_1_split does not need backward computation.

I0704 12:26:47.361047 5465 net.cpp:200] cifar does not need backward computation.

I0704 12:26:47.361057 5465 net.cpp:242] This network produces output accuracy

I0704 12:26:47.361068 5465 net.cpp:242] This network produces output loss

I0704 12:26:47.361099 5465 net.cpp:255] Network initialization done.

I0704 12:26:47.440634 5465 hdf5.cpp:32] Datatype class: H5T_FLOAT

I0704 12:26:47.442189 5465 caffe.cpp:290] Running for 10 iterations.

I0704 12:26:48.122941 5465 caffe.cpp:313] Batch 0, accuracy = 0.81

I0704 12:26:48.123001 5465 caffe.cpp:313] Batch 0, loss = 0.669872

I0704 12:26:48.722308 5465 caffe.cpp:313] Batch 1, accuracy = 0.81

I0704 12:26:48.722373 5465 caffe.cpp:313] Batch 1, loss = 0.600901

I0704 12:26:49.322628 5465 caffe.cpp:313] Batch 2, accuracy = 0.72

I0704 12:26:49.322688 5465 caffe.cpp:313] Batch 2, loss = 0.712268

I0704 12:26:49.923393 5465 caffe.cpp:313] Batch 3, accuracy = 0.79

I0704 12:26:49.923454 5465 caffe.cpp:313] Batch 3, loss = 0.630273

I0704 12:26:50.522907 5465 caffe.cpp:313] Batch 4, accuracy = 0.8

I0704 12:26:50.522970 5465 caffe.cpp:313] Batch 4, loss = 0.519629

I0704 12:26:51.123852 5465 caffe.cpp:313] Batch 5, accuracy = 0.84

I0704 12:26:51.123913 5465 caffe.cpp:313] Batch 5, loss = 0.398108

I0704 12:26:51.724124 5465 caffe.cpp:313] Batch 6, accuracy = 0.73

I0704 12:26:51.724195 5465 caffe.cpp:313] Batch 6, loss = 0.712703

I0704 12:26:52.333739 5465 caffe.cpp:313] Batch 7, accuracy = 0.76

I0704 12:26:52.333806 5465 caffe.cpp:313] Batch 7, loss = 0.711946

I0704 12:26:52.934726 5465 caffe.cpp:313] Batch 8, accuracy = 0.8

I0704 12:26:52.934793 5465 caffe.cpp:313] Batch 8, loss = 0.67638

I0704 12:26:53.534425 5465 caffe.cpp:313] Batch 9, accuracy = 0.73

I0704 12:26:53.534492 5465 caffe.cpp:313] Batch 9, loss = 0.703718

I0704 12:26:53.534503 5465 caffe.cpp:318] Loss: 0.63358

I0704 12:26:53.534533 5465 caffe.cpp:330] accuracy = 0.779

I0704 12:26:53.534556 5465 caffe.cpp:330] loss = 0.63358 (* 1 = 0.63358 loss)

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$

-------------------------------------------------------------------------------------------------------------------------------------------------------

测试:做100次迭代的结果

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$ caffe test -model examples/cifar10/cifar10_full_train_test.prototxt -weights examples/cifar10/cifar10_full_iter_60000.caffemodel.h5 -iterations 100

I0704 13:12:44.414038 6622 caffe.cpp:284] Use CPU.

I0704 13:12:44.681273 6622 net.cpp:294] The NetState phase (1) differed from the phase (0) specified by a rule in layer cifar

I0704 13:12:44.681501 6622 net.cpp:51] Initializing net from parameters:

..............................

..............................

I0704 13:12:44.998669 6622 caffe.cpp:290] Running for 100 iterations.

I0704 13:12:45.619782 6622 caffe.cpp:313] Batch 0, accuracy = 0.81

I0704 13:12:45.619843 6622 caffe.cpp:313] Batch 0, loss = 0.669872

I0704 13:12:46.217348 6622 caffe.cpp:313] Batch 1, accuracy = 0.81

I0704 13:12:46.217408 6622 caffe.cpp:313] Batch 1, loss = 0.600901

I0704 13:12:46.818264 6622 caffe.cpp:313] Batch 2, accuracy = 0.72

I0704 13:12:46.818323 6622 caffe.cpp:313] Batch 2, loss = 0.712268

I0704 13:12:47.415946 6622 caffe.cpp:313] Batch 3, accuracy = 0.79

I0704 13:12:47.416007 6622 caffe.cpp:313] Batch 3, loss = 0.630273

I0704 13:12:48.015729 6622 caffe.cpp:313] Batch 4, accuracy = 0.8

I0704 13:12:48.015792 6622 caffe.cpp:313] Batch 4, loss = 0.519629

I0704 13:12:48.621882 6622 caffe.cpp:313] Batch 5, accuracy = 0.84

I0704 13:12:48.621939 6622 caffe.cpp:313] Batch 5, loss = 0.398108

I0704 13:12:49.227767 6622 caffe.cpp:313] Batch 6, accuracy = 0.73

I0704 13:12:49.227831 6622 caffe.cpp:313] Batch 6, loss = 0.712703

I0704 13:12:49.833628 6622 caffe.cpp:313] Batch 7, accuracy = 0.76

I0704 13:12:49.833700 6622 caffe.cpp:313] Batch 7, loss = 0.711946

I0704 13:12:50.431000 6622 caffe.cpp:313] Batch 8, accuracy = 0.8

I0704 13:12:50.431063 6622 caffe.cpp:313] Batch 8, loss = 0.67638

I0704 13:12:51.036695 6622 caffe.cpp:313] Batch 9, accuracy = 0.73

I0704 13:12:51.036754 6622 caffe.cpp:313] Batch 9, loss = 0.703718

I0704 13:12:51.654356 6622 caffe.cpp:313] Batch 10, accuracy = 0.81

I0704 13:12:51.654417 6622 caffe.cpp:313] Batch 10, loss = 0.628615

I0704 13:12:52.271215 6622 caffe.cpp:313] Batch 11, accuracy = 0.77

I0704 13:12:52.271272 6622 caffe.cpp:313] Batch 11, loss = 0.640812

I0704 13:12:52.871871 6622 caffe.cpp:313] Batch 12, accuracy = 0.84

I0704 13:12:52.871930 6622 caffe.cpp:313] Batch 12, loss = 0.48857

I0704 13:12:53.500851 6622 caffe.cpp:313] Batch 13, accuracy = 0.82

I0704 13:12:53.500921 6622 caffe.cpp:313] Batch 13, loss = 0.545125

I0704 13:12:54.106851 6622 caffe.cpp:313] Batch 14, accuracy = 0.83

I0704 13:12:54.106916 6622 caffe.cpp:313] Batch 14, loss = 0.454256

I0704 13:12:54.712241 6622 caffe.cpp:313] Batch 15, accuracy = 0.8

I0704 13:12:54.712308 6622 caffe.cpp:313] Batch 15, loss = 0.604123

I0704 13:12:55.317904 6622 caffe.cpp:313] Batch 16, accuracy = 0.82

I0704 13:12:55.317970 6622 caffe.cpp:313] Batch 16, loss = 0.602975

I0704 13:12:55.915468 6622 caffe.cpp:313] Batch 17, accuracy = 0.78

I0704 13:12:55.915534 6622 caffe.cpp:313] Batch 17, loss = 0.653725

I0704 13:12:56.521464 6622 caffe.cpp:313] Batch 18, accuracy = 0.74

I0704 13:12:56.521531 6622 caffe.cpp:313] Batch 18, loss = 0.775862

I0704 13:12:57.127270 6622 caffe.cpp:313] Batch 19, accuracy = 0.7

I0704 13:12:57.127341 6622 caffe.cpp:313] Batch 19, loss = 0.927896

I0704 13:12:57.724392 6622 caffe.cpp:313] Batch 20, accuracy = 0.75

I0704 13:12:57.724458 6622 caffe.cpp:313] Batch 20, loss = 0.631937

I0704 13:12:58.330080 6622 caffe.cpp:313] Batch 21, accuracy = 0.75

I0704 13:12:58.330200 6622 caffe.cpp:313] Batch 21, loss = 0.689163

I0704 13:12:58.927534 6622 caffe.cpp:313] Batch 22, accuracy = 0.79

I0704 13:12:58.927600 6622 caffe.cpp:313] Batch 22, loss = 0.718018

I0704 13:12:59.533542 6622 caffe.cpp:313] Batch 23, accuracy = 0.76

I0704 13:12:59.533630 6622 caffe.cpp:313] Batch 23, loss = 0.772289

I0704 13:13:00.131110 6622 caffe.cpp:313] Batch 24, accuracy = 0.78

I0704 13:13:00.131175 6622 caffe.cpp:313] Batch 24, loss = 0.616908

I0704 13:13:00.727895 6622 caffe.cpp:313] Batch 25, accuracy = 0.69

I0704 13:13:00.727962 6622 caffe.cpp:313] Batch 25, loss = 0.921352

I0704 13:13:01.324116 6622 caffe.cpp:313] Batch 26, accuracy = 0.86

I0704 13:13:01.324188 6622 caffe.cpp:313] Batch 26, loss = 0.43752

I0704 13:13:01.920688 6622 caffe.cpp:313] Batch 27, accuracy = 0.76

I0704 13:13:01.920760 6622 caffe.cpp:313] Batch 27, loss = 0.694387

I0704 13:13:02.518180 6622 caffe.cpp:313] Batch 28, accuracy = 0.88

I0704 13:13:02.518250 6622 caffe.cpp:313] Batch 28, loss = 0.503546

I0704 13:13:03.123641 6622 caffe.cpp:313] Batch 29, accuracy = 0.76

I0704 13:13:03.123706 6622 caffe.cpp:313] Batch 29, loss = 0.675348

I0704 13:13:03.729601 6622 caffe.cpp:313] Batch 30, accuracy = 0.75

I0704 13:13:03.729679 6622 caffe.cpp:313] Batch 30, loss = 0.641321

I0704 13:13:04.326668 6622 caffe.cpp:313] Batch 31, accuracy = 0.79

I0704 13:13:04.326732 6622 caffe.cpp:313] Batch 31, loss = 0.668134

I0704 13:13:04.932490 6622 caffe.cpp:313] Batch 32, accuracy = 0.78

I0704 13:13:04.932554 6622 caffe.cpp:313] Batch 32, loss = 0.587462

I0704 13:13:05.529435 6622 caffe.cpp:313] Batch 33, accuracy = 0.69

I0704 13:13:05.529508 6622 caffe.cpp:313] Batch 33, loss = 0.809499

I0704 13:13:06.125875 6622 caffe.cpp:313] Batch 34, accuracy = 0.72

I0704 13:13:06.125944 6622 caffe.cpp:313] Batch 34, loss = 0.841917

I0704 13:13:06.722806 6622 caffe.cpp:313] Batch 35, accuracy = 0.81

I0704 13:13:06.722877 6622 caffe.cpp:313] Batch 35, loss = 0.65033

I0704 13:13:07.328688 6622 caffe.cpp:313] Batch 36, accuracy = 0.74

I0704 13:13:07.328760 6622 caffe.cpp:313] Batch 36, loss = 0.73518

I0704 13:13:07.926215 6622 caffe.cpp:313] Batch 37, accuracy = 0.77

I0704 13:13:07.926282 6622 caffe.cpp:313] Batch 37, loss = 0.626204

I0704 13:13:08.531821 6622 caffe.cpp:313] Batch 38, accuracy = 0.84

I0704 13:13:08.531885 6622 caffe.cpp:313] Batch 38, loss = 0.50705

I0704 13:13:09.128522 6622 caffe.cpp:313] Batch 39, accuracy = 0.86

I0704 13:13:09.128587 6622 caffe.cpp:313] Batch 39, loss = 0.45618

I0704 13:13:09.725127 6622 caffe.cpp:313] Batch 40, accuracy = 0.82

I0704 13:13:09.725200 6622 caffe.cpp:313] Batch 40, loss = 0.594011

I0704 13:13:10.321892 6622 caffe.cpp:313] Batch 41, accuracy = 0.83

I0704 13:13:10.321964 6622 caffe.cpp:313] Batch 41, loss = 0.673196

I0704 13:13:10.918488 6622 caffe.cpp:313] Batch 42, accuracy = 0.85

I0704 13:13:10.918555 6622 caffe.cpp:313] Batch 42, loss = 0.45519

I0704 13:13:11.524237 6622 caffe.cpp:313] Batch 43, accuracy = 0.79

I0704 13:13:11.524307 6622 caffe.cpp:313] Batch 43, loss = 0.628985

I0704 13:13:12.129947 6622 caffe.cpp:313] Batch 44, accuracy = 0.78

I0704 13:13:12.130018 6622 caffe.cpp:313] Batch 44, loss = 0.704623

I0704 13:13:12.735232 6622 caffe.cpp:313] Batch 45, accuracy = 0.71

I0704 13:13:12.735304 6622 caffe.cpp:313] Batch 45, loss = 0.695174

I0704 13:13:13.340771 6622 caffe.cpp:313] Batch 46, accuracy = 0.84

I0704 13:13:13.340842 6622 caffe.cpp:313] Batch 46, loss = 0.552928

I0704 13:13:13.947043 6622 caffe.cpp:313] Batch 47, accuracy = 0.76

I0704 13:13:13.947109 6622 caffe.cpp:313] Batch 47, loss = 0.651739

I0704 13:13:14.553036 6622 caffe.cpp:313] Batch 48, accuracy = 0.82

I0704 13:13:14.553324 6622 caffe.cpp:313] Batch 48, loss = 0.441534

I0704 13:13:15.180112 6622 caffe.cpp:313] Batch 49, accuracy = 0.76

I0704 13:13:15.180176 6622 caffe.cpp:313] Batch 49, loss = 0.729064

I0704 13:13:15.779243 6622 caffe.cpp:313] Batch 50, accuracy = 0.8

I0704 13:13:15.779307 6622 caffe.cpp:313] Batch 50, loss = 0.584773

I0704 13:13:16.378615 6622 caffe.cpp:313] Batch 51, accuracy = 0.79

I0704 13:13:16.378679 6622 caffe.cpp:313] Batch 51, loss = 0.541237

I0704 13:13:16.978050 6622 caffe.cpp:313] Batch 52, accuracy = 0.79

I0704 13:13:16.978111 6622 caffe.cpp:313] Batch 52, loss = 0.656132

I0704 13:13:17.577879 6622 caffe.cpp:313] Batch 53, accuracy = 0.79

I0704 13:13:17.577941 6622 caffe.cpp:313] Batch 53, loss = 0.568454

I0704 13:13:18.177491 6622 caffe.cpp:313] Batch 54, accuracy = 0.72

I0704 13:13:18.177562 6622 caffe.cpp:313] Batch 54, loss = 0.797033

I0704 13:13:18.776235 6622 caffe.cpp:313] Batch 55, accuracy = 0.82

I0704 13:13:18.776298 6622 caffe.cpp:313] Batch 55, loss = 0.679914

I0704 13:13:19.375723 6622 caffe.cpp:313] Batch 56, accuracy = 0.76

I0704 13:13:19.375792 6622 caffe.cpp:313] Batch 56, loss = 0.681724

I0704 13:13:19.974643 6622 caffe.cpp:313] Batch 57, accuracy = 0.86

I0704 13:13:19.974711 6622 caffe.cpp:313] Batch 57, loss = 0.460299

I0704 13:13:20.574430 6622 caffe.cpp:313] Batch 58, accuracy = 0.78

I0704 13:13:20.574501 6622 caffe.cpp:313] Batch 58, loss = 0.694127

I0704 13:13:21.174055 6622 caffe.cpp:313] Batch 59, accuracy = 0.74

I0704 13:13:21.174118 6622 caffe.cpp:313] Batch 59, loss = 0.748216

I0704 13:13:21.804786 6622 caffe.cpp:313] Batch 60, accuracy = 0.79

I0704 13:13:21.804847 6622 caffe.cpp:313] Batch 60, loss = 0.566229

I0704 13:13:22.410190 6622 caffe.cpp:313] Batch 61, accuracy = 0.8

I0704 13:13:22.410251 6622 caffe.cpp:313] Batch 61, loss = 0.53662

I0704 13:13:23.011648 6622 caffe.cpp:313] Batch 62, accuracy = 0.77

I0704 13:13:23.011713 6622 caffe.cpp:313] Batch 62, loss = 0.630629

I0704 13:13:23.618448 6622 caffe.cpp:313] Batch 63, accuracy = 0.8

I0704 13:13:23.618508 6622 caffe.cpp:313] Batch 63, loss = 0.553974

I0704 13:13:24.215498 6622 caffe.cpp:313] Batch 64, accuracy = 0.82

I0704 13:13:24.215561 6622 caffe.cpp:313] Batch 64, loss = 0.640087

I0704 13:13:24.812703 6622 caffe.cpp:313] Batch 65, accuracy = 0.78

I0704 13:13:24.812767 6622 caffe.cpp:313] Batch 65, loss = 0.720206

I0704 13:13:25.417578 6622 caffe.cpp:313] Batch 66, accuracy = 0.85

I0704 13:13:25.417637 6622 caffe.cpp:313] Batch 66, loss = 0.469227

I0704 13:13:26.015293 6622 caffe.cpp:313] Batch 67, accuracy = 0.74

I0704 13:13:26.015355 6622 caffe.cpp:313] Batch 67, loss = 0.662439

I0704 13:13:26.621019 6622 caffe.cpp:313] Batch 68, accuracy = 0.73

I0704 13:13:26.621078 6622 caffe.cpp:313] Batch 68, loss = 0.832033

I0704 13:13:27.226202 6622 caffe.cpp:313] Batch 69, accuracy = 0.74

I0704 13:13:27.226261 6622 caffe.cpp:313] Batch 69, loss = 0.7851

I0704 13:13:27.831529 6622 caffe.cpp:313] Batch 70, accuracy = 0.79

I0704 13:13:27.831593 6622 caffe.cpp:313] Batch 70, loss = 0.594752

I0704 13:13:28.437043 6622 caffe.cpp:313] Batch 71, accuracy = 0.81

I0704 13:13:28.437101 6622 caffe.cpp:313] Batch 71, loss = 0.600966

I0704 13:13:29.034024 6622 caffe.cpp:313] Batch 72, accuracy = 0.89

I0704 13:13:29.034085 6622 caffe.cpp:313] Batch 72, loss = 0.544593

I0704 13:13:29.639663 6622 caffe.cpp:313] Batch 73, accuracy = 0.81

I0704 13:13:29.639719 6622 caffe.cpp:313] Batch 73, loss = 0.578454

I0704 13:13:30.245527 6622 caffe.cpp:313] Batch 74, accuracy = 0.71

I0704 13:13:30.245589 6622 caffe.cpp:313] Batch 74, loss = 0.85487

I0704 13:13:30.851264 6622 caffe.cpp:313] Batch 75, accuracy = 0.75

I0704 13:13:30.851325 6622 caffe.cpp:313] Batch 75, loss = 0.703642

I0704 13:13:31.448464 6622 caffe.cpp:313] Batch 76, accuracy = 0.77

I0704 13:13:31.448526 6622 caffe.cpp:313] Batch 76, loss = 0.702493

I0704 13:13:32.053625 6622 caffe.cpp:313] Batch 77, accuracy = 0.77

I0704 13:13:32.053694 6622 caffe.cpp:313] Batch 77, loss = 0.616773

I0704 13:13:32.659436 6622 caffe.cpp:313] Batch 78, accuracy = 0.77

I0704 13:13:32.659543 6622 caffe.cpp:313] Batch 78, loss = 0.624216

I0704 13:13:33.256774 6622 caffe.cpp:313] Batch 79, accuracy = 0.76

I0704 13:13:33.256835 6622 caffe.cpp:313] Batch 79, loss = 0.702435

I0704 13:13:33.862572 6622 caffe.cpp:313] Batch 80, accuracy = 0.83

I0704 13:13:33.862634 6622 caffe.cpp:313] Batch 80, loss = 0.479035

I0704 13:13:34.460429 6622 caffe.cpp:313] Batch 81, accuracy = 0.77

I0704 13:13:34.460491 6622 caffe.cpp:313] Batch 81, loss = 0.679997

I0704 13:13:35.066623 6622 caffe.cpp:313] Batch 82, accuracy = 0.76

I0704 13:13:35.066686 6622 caffe.cpp:313] Batch 82, loss = 0.668821

I0704 13:13:35.672404 6622 caffe.cpp:313] Batch 83, accuracy = 0.79

I0704 13:13:35.672462 6622 caffe.cpp:313] Batch 83, loss = 0.59493

I0704 13:13:36.277426 6622 caffe.cpp:313] Batch 84, accuracy = 0.77

I0704 13:13:36.277488 6622 caffe.cpp:313] Batch 84, loss = 0.636169

I0704 13:13:36.883759 6622 caffe.cpp:313] Batch 85, accuracy = 0.79

I0704 13:13:36.883818 6622 caffe.cpp:313] Batch 85, loss = 0.655568

I0704 13:13:37.489516 6622 caffe.cpp:313] Batch 86, accuracy = 0.79

I0704 13:13:37.489574 6622 caffe.cpp:313] Batch 86, loss = 0.637793

I0704 13:13:38.086877 6622 caffe.cpp:313] Batch 87, accuracy = 0.84

I0704 13:13:38.086935 6622 caffe.cpp:313] Batch 87, loss = 0.604765

I0704 13:13:38.691869 6622 caffe.cpp:313] Batch 88, accuracy = 0.81

I0704 13:13:38.691931 6622 caffe.cpp:313] Batch 88, loss = 0.525659

I0704 13:13:39.297114 6622 caffe.cpp:313] Batch 89, accuracy = 0.76

I0704 13:13:39.297176 6622 caffe.cpp:313] Batch 89, loss = 0.657071

I0704 13:13:39.902731 6622 caffe.cpp:313] Batch 90, accuracy = 0.81

I0704 13:13:39.902787 6622 caffe.cpp:313] Batch 90, loss = 0.5901

I0704 13:13:40.499342 6622 caffe.cpp:313] Batch 91, accuracy = 0.85

I0704 13:13:40.499403 6622 caffe.cpp:313] Batch 91, loss = 0.433673

I0704 13:13:41.105139 6622 caffe.cpp:313] Batch 92, accuracy = 0.76

I0704 13:13:41.105197 6622 caffe.cpp:313] Batch 92, loss = 0.68505

I0704 13:13:41.711143 6622 caffe.cpp:313] Batch 93, accuracy = 0.84

I0704 13:13:41.711205 6622 caffe.cpp:313] Batch 93, loss = 0.526908

I0704 13:13:42.317083 6622 caffe.cpp:313] Batch 94, accuracy = 0.81

I0704 13:13:42.317149 6622 caffe.cpp:313] Batch 94, loss = 0.54016

I0704 13:13:42.923281 6622 caffe.cpp:313] Batch 95, accuracy = 0.81

I0704 13:13:42.923348 6622 caffe.cpp:313] Batch 95, loss = 0.568361

I0704 13:13:42.924571 6628 data_layer.cpp:73] Restarting data prefetching from start.

I0704 13:13:43.529124 6622 caffe.cpp:313] Batch 96, accuracy = 0.82

I0704 13:13:43.529189 6622 caffe.cpp:313] Batch 96, loss = 0.455237

I0704 13:13:44.135169 6622 caffe.cpp:313] Batch 97, accuracy = 0.78

I0704 13:13:44.135228 6622 caffe.cpp:313] Batch 97, loss = 0.753922

I0704 13:13:44.740659 6622 caffe.cpp:313] Batch 98, accuracy = 0.73

I0704 13:13:44.740911 6622 caffe.cpp:313] Batch 98, loss = 0.709454

I0704 13:13:45.338565 6622 caffe.cpp:313] Batch 99, accuracy = 0.78

I0704 13:13:45.338631 6622 caffe.cpp:313] Batch 99, loss = 0.737646

I0704 13:13:45.338644 6622 caffe.cpp:318] Loss: 0.634399

I0704 13:13:45.338677 6622 caffe.cpp:330] accuracy = 0.7859

I0704 13:13:45.338701 6622 caffe.cpp:330] loss = 0.634399 (* 1 = 0.634399 loss)

识别猫:

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$ classification examples/cifar10/cifar10_full.prototxt examples/cifar10/cifar10_full_iter_60000.caffemodel.h5 examples/cifar10/mean.binaryproto data/cifar10/synset_words.txt examples/images/cat.jpg

---------- Prediction for examples/images/cat.jpg ----------

0.7481 - " deer "

0.1352 - " bird "

0.0476 - " cat "

0.0162 - " frog "

0.0143 - " horse "

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$

灰色猫:

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$ classification examples/cifar10/cifar10_full.prototxt examples/cifar10/cifar10_full_iter_60000.caffemodel.h5 examples/cifar10/mean.binaryproto data/cifar10/synset_words.txt examples/images/cat_gray.jpg

---------- Prediction for examples/images/cat_gray.jpg ----------

0.2487 - " bird "

0.2476 - " horse "

0.1985 - " dog "

0.1277 - " cat "

0.0853 - " deer "

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$

人鱼自行车:

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$ classification examples/cifar10/cifar10_full.prototxt examples/cifar10/cifar10_full_iter_60000.caffemodel.h5 examples/cifar10/mean.binaryproto data/cifar10/synset_words.txt examples/images/fish-bike.jpg

---------- Prediction for examples/images/fish-bike.jpg ----------

0.6517 - " horse "

0.1291 - " truck "

0.0530 - " deer "

0.0441 - " cat "

0.0435 - " frog "

seag@seag-G41MT-S2PT:~/wsCaffe/caffe$

看起来效果很糟糕的。