Upload 9 files
5cde30c
|
name: "CaffeNet" |
|
input: "data" |
|
input_dim: 10 |
|
input_dim: 3 |
|
input_dim: 227 |
|
input_dim: 227 |
|
layers { |
|
name: "conv1" |
|
type: CONVOLUTION |
|
bottom: "data" |
|
top: "conv1" |
|
convolution_param { |
|
num_output: 96 |
|
kernel_size: 7 |
|
stride: 4 |
|
} |
|
} |
|
layers { |
|
name: "relu1" |
|
type: RELU |
|
bottom: "conv1" |
|
top: "conv1" |
|
} |
|
layers { |
|
name: "pool1" |
|
type: POOLING |
|
bottom: "conv1" |
|
top: "pool1" |
|
pooling_param { |
|
pool: MAX |
|
kernel_size: 3 |
|
stride: 2 |
|
} |
|
} |
|
layers { |
|
name: "norm1" |
|
type: LRN |
|
bottom: "pool1" |
|
top: "norm1" |
|
lrn_param { |
|
local_size: 5 |
|
alpha: 0.0001 |
|
beta: 0.75 |
|
} |
|
} |
|
layers { |
|
name: "conv2" |
|
type: CONVOLUTION |
|
bottom: "norm1" |
|
top: "conv2" |
|
convolution_param { |
|
num_output: 256 |
|
pad: 2 |
|
kernel_size: 5 |
|
} |
|
} |
|
layers { |
|
name: "relu2" |
|
type: RELU |
|
bottom: "conv2" |
|
top: "conv2" |
|
} |
|
layers { |
|
name: "pool2" |
|
type: POOLING |
|
bottom: "conv2" |
|
top: "pool2" |
|
pooling_param { |
|
pool: MAX |
|
kernel_size: 3 |
|
stride: 2 |
|
} |
|
} |
|
layers { |
|
name: "norm2" |
|
type: LRN |
|
bottom: "pool2" |
|
top: "norm2" |
|
lrn_param { |
|
local_size: 5 |
|
alpha: 0.0001 |
|
beta: 0.75 |
|
} |
|
} |
|
layers { |
|
name: "conv3" |
|
type: CONVOLUTION |
|
bottom: "norm2" |
|
top: "conv3" |
|
convolution_param { |
|
num_output: 384 |
|
pad: 1 |
|
kernel_size: 3 |
|
} |
|
} |
|
layers{ |
|
name: "relu3" |
|
type: RELU |
|
bottom: "conv3" |
|
top: "conv3" |
|
} |
|
layers { |
|
name: "pool5" |
|
type: POOLING |
|
bottom: "conv3" |
|
top: "pool5" |
|
pooling_param { |
|
pool: MAX |
|
kernel_size: 3 |
|
stride: 2 |
|
} |
|
} |
|
layers { |
|
name: "fc6" |
|
type: INNER_PRODUCT |
|
bottom: "pool5" |
|
top: "fc6" |
|
inner_product_param { |
|
num_output: 512 |
|
} |
|
} |
|
layers { |
|
name: "relu6" |
|
type: RELU |
|
bottom: "fc6" |
|
top: "fc6" |
|
} |
|
layers { |
|
name: "drop6" |
|
type: DROPOUT |
|
bottom: "fc6" |
|
top: "fc6" |
|
dropout_param { |
|
dropout_ratio: 0.5 |
|
} |
|
} |
|
layers { |
|
name: "fc7" |
|
type: INNER_PRODUCT |
|
bottom: "fc6" |
|
top: "fc7" |
|
inner_product_param { |
|
num_output: 512 |
|
} |
|
} |
|
layers { |
|
name: "relu7" |
|
type: RELU |
|
bottom: "fc7" |
|
top: "fc7" |
|
} |
|
layers { |
|
name: "drop7" |
|
type: DROPOUT |
|
bottom: "fc7" |
|
top: "fc7" |
|
dropout_param { |
|
dropout_ratio: 0.5 |
|
} |
|
} |
|
layers { |
|
name: "fc8" |
|
type: INNER_PRODUCT |
|
bottom: "fc7" |
|
top: "fc8" |
|
inner_product_param { |
|
num_output: 2 |
|
} |
|
} |
|
layers { |
|
name: "prob" |
|
type: SOFTMAX |
|
bottom: "fc8" |
|
top: "prob" |
|
} |