Comments (4)
Thanks.
If you are discussing the problem in general. Make sure that you use the correct prototxt to load the snapshot model, which means all the layer names in testing prototxt should be the same as that in your training prototxt. Otherwise, it will load the default values.
Also, did you try to make the learning rate higher for the crf layer? It should give you quite big different parameters.
e.g.
layers {
name: "inference1"
type: MULTI_STAGE_MEANFIELD
bottom: "unary"
bottom: "Q0"
bottom: "data"
top: "pred"
blobs_lr: 10000
blobs_lr: 10000
blobs_lr:1000 #new parameter
multi_stage_meanfield_param {
num_iterations: 10
compatibility_mode: POTTS
threshold: 2
theta_alpha: 160
theta_beta: 3
theta_gamma: 3
spatial_filter_weight: 3
bilateral_filter_weight: 5
}
}
from crfasrnn.
@bittnt
I am training the net with your prototxt file, where I just modified the data layer in order to use my data. It looks like this:
`name: 'TVG_CRF_RNN_SEG'
force_backward: true
layers { top: "data" top: "label_argmax" name: "train_data" type: HDF5_DATA include { phase: TRAIN }
hdf5_data_param { source: "caffe/model/train_data.txt" batch_size: 1 } }
layers { top: "data" top: "label_argmax" name: "val_data" type: HDF5_DATA include { phase: TEST }
hdf5_data_param { source: "caffe/model/val_data.txt" batch_size: 1 } }
layers { bottom: 'data' top: 'conv1_1' name: 'conv1_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 64 pad: 100 kernel_size: 3 } }
layers { bottom: 'conv1_1' top: 'conv1_1' name: 'relu1_1' type: RELU }
layers { bottom: 'conv1_1' top: 'conv1_2' name: 'conv1_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 64 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv1_2' top: 'conv1_2' name: 'relu1_2' type: RELU }
layers { name: 'pool1' bottom: 'conv1_2' top: 'pool1' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { name: 'conv2_1' bottom: 'pool1' top: 'conv2_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 128 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv2_1' top: 'conv2_1' name: 'relu2_1' type: RELU }
layers { bottom: 'conv2_1' top: 'conv2_2' name: 'conv2_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 128 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv2_2' top: 'conv2_2' name: 'relu2_2' type: RELU }
layers { bottom: 'conv2_2' top: 'pool2' name: 'pool2' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool2' top: 'conv3_1' name: 'conv3_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 256 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv3_1' top: 'conv3_1' name: 'relu3_1' type: RELU }
layers { bottom: 'conv3_1' top: 'conv3_2' name: 'conv3_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 256 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv3_2' top: 'conv3_2' name: 'relu3_2' type: RELU }
layers { bottom: 'conv3_2' top: 'conv3_3' name: 'conv3_3' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 256 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv3_3' top: 'conv3_3' name: 'relu3_3' type: RELU }
layers { bottom: 'conv3_3' top: 'pool3' name: 'pool3' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool3' top: 'conv4_1' name: 'conv4_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv4_1' top: 'conv4_1' name: 'relu4_1' type: RELU }
layers { bottom: 'conv4_1' top: 'conv4_2' name: 'conv4_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv4_2' top: 'conv4_2' name: 'relu4_2' type: RELU }
layers { bottom: 'conv4_2' top: 'conv4_3' name: 'conv4_3' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv4_3' top: 'conv4_3' name: 'relu4_3' type: RELU }
layers { bottom: 'conv4_3' top: 'pool4' name: 'pool4' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool4' top: 'conv5_1' name: 'conv5_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv5_1' top: 'conv5_1' name: 'relu5_1' type: RELU }
layers { bottom: 'conv5_1' top: 'conv5_2' name: 'conv5_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv5_2' top: 'conv5_2' name: 'relu5_2' type: RELU }
layers { bottom: 'conv5_2' top: 'conv5_3' name: 'conv5_3' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv5_3' top: 'conv5_3' name: 'relu5_3' type: RELU }
layers { bottom: 'conv5_3' top: 'pool5' name: 'pool5' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool5' top: 'fc6' name: 'fc6' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE kernel_size: 7 num_output: 4096 } }
layers { bottom: 'fc6' top: 'fc6' name: 'relu6' type: RELU }
layers { bottom: 'fc6' top: 'fc6' name: 'drop6' type: DROPOUT
dropout_param { dropout_ratio: 0.5 } }
layers { bottom: 'fc6' top: 'fc7' name: 'fc7' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE kernel_size: 1 num_output: 4096 } }
layers { bottom: 'fc7' top: 'fc7' name: 'relu7' type: RELU }
layers { bottom: 'fc7' top: 'fc7' name: 'drop7' type: DROPOUT
dropout_param { dropout_ratio: 0.5 } }
layers { name: 'score-fr' type: CONVOLUTION bottom: 'fc7' top: 'score'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 3 kernel_size: 1 } }
layers { type: DECONVOLUTION name: 'score2' bottom: 'score' top: 'score2'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { kernel_size: 4 stride: 2 num_output: 3 } }
layers { name: 'score-pool4' type: CONVOLUTION bottom: 'pool4' top: 'score-pool4'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 3 kernel_size: 1 } }
layers { type: CROP name: 'crop' bottom: 'score-pool4' bottom: 'score2'
top: 'score-pool4c' }
layers { type: ELTWISE name: 'fuse' bottom: 'score2' bottom: 'score-pool4c'
top: 'score-fused'
eltwise_param { operation: SUM } }
layers { type: DECONVOLUTION name: 'score4' bottom: 'score-fused'
top: 'score4'
blobs_lr: 1 weight_decay: 1
convolution_param { bias_term: false kernel_size: 4 stride: 2 num_output: 3 } }
layers { name: 'score-pool3' type: CONVOLUTION bottom: 'pool3' top: 'score-pool3'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 3 kernel_size: 1 } }
layers { type: CROP name: 'crop' bottom: 'score-pool3' bottom: 'score4'
top: 'score-pool3c' }
layers { type: ELTWISE name: 'fuse' bottom: 'score4' bottom: 'score-pool3c'
top: 'score-final'
eltwise_param { operation: SUM } }
layers { type: DECONVOLUTION name: 'upsample'
bottom: 'score-final' top: 'bigscore'
blobs_lr: 0
convolution_param { bias_term: false num_output: 3 kernel_size: 16 stride: 8 } }
layers { type: CROP name: 'crop' bottom: 'bigscore' bottom: 'data' top: 'coarse' }
layers { type: SPLIT name: 'splitting'
bottom: 'coarse' top: 'unary' top: 'Q0'
}
layers {
name: "inference1"
type: MULTI_STAGE_MEANFIELD
bottom: "unary"
bottom: "Q0"
bottom: "data"
top: "pred"
blobs_lr: 10
blobs_lr: 10
blobs_lr: 10 #new parameter
multi_stage_meanfield_param {
num_iterations: 10
compatibility_mode: POTTS
threshold: 2
theta_alpha: 160
theta_beta: 3
theta_gamma: 3
spatial_filter_weight: 3
bilateral_filter_weight: 5
}
}
layers { bottom: "pred" bottom: "label_argmax" top: "loss" name: "loss" type: SOFTMAX_LOSS }
}`
After the snapshot is saved during training, I load this prototxt together with the caffemodel file in python to check the parameters. The weights have the initial zero values.
I've tried to train the net with smaller and also bigger values for crf's learning rate and also withou the crf itself. The outcome is always the same as described above..
from crfasrnn.
I am not sure if your script is correct. But could you make sure that you
got the correct data and label in both training and testing phase? Ideally,
for semantic image segmentation problem, you should have input image and
its corresponding label map at the same resolution.
Check out Jon long & Evan's scripts:
https://gist.github.com/shelhamer/80667189b218ad570e82
name: "FCN"
layer {
name: "data"
type: "Data"
top: "data"
include {
phase: TRAIN
}
transform_param {
mean_value: 104.00699
mean_value: 116.66877
mean_value: 122.67892
}
data_param {
source: "../../data/pascal-context/pascal-context-train-lmdb"
batch_size: 1
backend: LMDB
}
}
layer {
name: "label"
type: "Data"
top: "label"
include {
phase: TRAIN
}
data_param {
source: "../../data/pascal-context/pascal-context-train-gt59-lmdb"
batch_size: 1
backend: LMDB
}
}
layer {
name: "data"
type: "Data"
top: "data"
include {
phase: TEST
}
transform_param {
mean_value: 104.00699
mean_value: 116.66877
mean_value: 122.67892
}
data_param {
source: "../../data/pascal-context/pascal-context-val-lmdb"
batch_size: 1
backend: LMDB
}
}
layer {
name: "label"
type: "Data"
top: "label"
include {
phase: TEST
}
data_param {
source: "../../data/pascal-context/pascal-context-val-gt59-lmdb"
batch_size: 1
backend: LMDB
}
}
On Tue, 15 Mar 2016 at 22:39, AdrianLsk [email protected] wrote:
@bittnt https://github.com/bittnt
I am training the net with your prototxt file, where I just modified the
data layer in order to use my data. It looks like this:`name: 'TVG_CRF_RNN_BONE_SEG'
force_backward: true
layers { top: "data" top: "label_argmax" name: "train_data" type:
HDF5_DATA include { phase: TRAIN }
hdf5_data_param { source: "caffe/model/train_data.txt" batch_size: 1 } }layers { top: "data" top: "label_argmax" name: "val_data" type: HDF5_DATA
include { phase: TEST }
hdf5_data_param { source: "caffe/model/val_data.txt" batch_size: 1 } }layers { bottom: 'data' top: 'conv1_1' name: 'conv1_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 64 pad: 100 kernel_size: 3 }
}
layers { bottom: 'conv1_1' top: 'conv1_1' name: 'relu1_1' type: RELU }
layers { bottom: 'conv1_1' top: 'conv1_2' name: 'conv1_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 64 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv1_2' top: 'conv1_2' name: 'relu1_2' type: RELU }
layers { name: 'pool1' bottom: 'conv1_2' top: 'pool1' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { name: 'conv2_1' bottom: 'pool1' top: 'conv2_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 128 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv2_1' top: 'conv2_1' name: 'relu2_1' type: RELU }
layers { bottom: 'conv2_1' top: 'conv2_2' name: 'conv2_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 128 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv2_2' top: 'conv2_2' name: 'relu2_2' type: RELU }
layers { bottom: 'conv2_2' top: 'pool2' name: 'pool2' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool2' top: 'conv3_1' name: 'conv3_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 256 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv3_1' top: 'conv3_1' name: 'relu3_1' type: RELU }
layers { bottom: 'conv3_1' top: 'conv3_2' name: 'conv3_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 256 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv3_2' top: 'conv3_2' name: 'relu3_2' type: RELU }
layers { bottom: 'conv3_2' top: 'conv3_3' name: 'conv3_3' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 256 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv3_3' top: 'conv3_3' name: 'relu3_3' type: RELU }
layers { bottom: 'conv3_3' top: 'pool3' name: 'pool3' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool3' top: 'conv4_1' name: 'conv4_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv4_1' top: 'conv4_1' name: 'relu4_1' type: RELU }
layers { bottom: 'conv4_1' top: 'conv4_2' name: 'conv4_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv4_2' top: 'conv4_2' name: 'relu4_2' type: RELU }
layers { bottom: 'conv4_2' top: 'conv4_3' name: 'conv4_3' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv4_3' top: 'conv4_3' name: 'relu4_3' type: RELU }
layers { bottom: 'conv4_3' top: 'pool4' name: 'pool4' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool4' top: 'conv5_1' name: 'conv5_1' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv5_1' top: 'conv5_1' name: 'relu5_1' type: RELU }
layers { bottom: 'conv5_1' top: 'conv5_2' name: 'conv5_2' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv5_2' top: 'conv5_2' name: 'relu5_2' type: RELU }
layers { bottom: 'conv5_2' top: 'conv5_3' name: 'conv5_3' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 512 pad: 1 kernel_size: 3 } }
layers { bottom: 'conv5_3' top: 'conv5_3' name: 'relu5_3' type: RELU }
layers { bottom: 'conv5_3' top: 'pool5' name: 'pool5' type: POOLING
pooling_param { pool: MAX kernel_size: 2 stride: 2 } }
layers { bottom: 'pool5' top: 'fc6' name: 'fc6' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE kernel_size: 7 num_output: 4096 } }
layers { bottom: 'fc6' top: 'fc6' name: 'relu6' type: RELU }
layers { bottom: 'fc6' top: 'fc6' name: 'drop6' type: DROPOUT
dropout_param { dropout_ratio: 0.5 } }
layers { bottom: 'fc6' top: 'fc7' name: 'fc7' type: CONVOLUTION
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE kernel_size: 1 num_output: 4096 } }
layers { bottom: 'fc7' top: 'fc7' name: 'relu7' type: RELU }
layers { bottom: 'fc7' top: 'fc7' name: 'drop7' type: DROPOUT
dropout_param { dropout_ratio: 0.5 } }
layers { name: 'score-fr' type: CONVOLUTION bottom: 'fc7' top: 'score'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 3 kernel_size: 1 } }layers { type: DECONVOLUTION name: 'score2' bottom: 'score' top: 'score2'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { kernel_size: 4 stride: 2 num_output: 3 } }layers { name: 'score-pool4' type: CONVOLUTION bottom: 'pool4' top:
'score-pool4'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 3 kernel_size: 1 } }layers { type: CROP name: 'crop' bottom: 'score-pool4' bottom: 'score2'
top: 'score-pool4c' }layers { type: ELTWISE name: 'fuse' bottom: 'score2' bottom: 'score-pool4c'
top: 'score-fused'
eltwise_param { operation: SUM } }layers { type: DECONVOLUTION name: 'score4' bottom: 'score-fused'
top: 'score4'
blobs_lr: 1 weight_decay: 1
convolution_param { bias_term: false kernel_size: 4 stride: 2 num_output:
3 } }layers { name: 'score-pool3' type: CONVOLUTION bottom: 'pool3' top:
'score-pool3'
blobs_lr: 1 blobs_lr: 2 weight_decay: 1 weight_decay: 0
convolution_param { engine: CAFFE num_output: 3 kernel_size: 1 } }layers { type: CROP name: 'crop' bottom: 'score-pool3' bottom: 'score4'
top: 'score-pool3c' }layers { type: ELTWISE name: 'fuse' bottom: 'score4' bottom: 'score-pool3c'
top: 'score-final'
eltwise_param { operation: SUM } }layers { type: DECONVOLUTION name: 'upsample'
bottom: 'score-final' top: 'bigscore'
blobs_lr: 0
convolution_param { bias_term: false num_output: 3 kernel_size: 16 stride:
8 } }layers { type: CROP name: 'crop' bottom: 'bigscore' bottom: 'data' top:
'coarse' }layers { type: SPLIT name: 'splitting'
bottom: 'coarse' top: 'unary' top: 'Q0'
}layers {
name: "inference1"
type: MULTI_STAGE_MEANFIELD
bottom: "unary"
bottom: "Q0"
bottom: "data"
top: "pred"
blobs_lr: 10
blobs_lr: 10
blobs_lr: 10 #new parameter
multi_stage_meanfield_param {
num_iterations: 10
compatibility_mode: POTTS
threshold: 2
theta_alpha: 160
theta_beta: 3
theta_gamma: 3
spatial_filter_weight: 3
bilateral_filter_weight: 5
}
}
layers { bottom: "pred" bottom: "label_argmax" top: "loss" name: "loss"
type: SOFTMAX_LOSS }
}`After the snapshot is saved during training, I load this prototxt together
with the caffemodel file in python to check the parameters. The weights
have the initial zero values.I've tried to train the net with smaller and also bigger values for crf's
learning rate and also withou the crf itself. The outcome is always the
same as described above..—
You are receiving this because you were mentioned.
Reply to this email directly or view it on GitHub
#35 (comment)
from crfasrnn.
Closing old issues with no recent activity.
from crfasrnn.
Related Issues (20)
- Cannot align apparently disconnected blobs. HOT 1
- Dense InfogainLoss Function
- Feature extraction from CRF-RNN vs FCN8s
- CRFasRNN output vs score_final (FCN)
- Identify only small set of classes
- cuDNN compile error on Tesla K80 GPU - error: too few arguments to function
- Mean vector
- Python demo script on Ubuntu 14.04: "Check failed: *ptr host allocation of size 282240000 failed" HOT 1
- This implementation has not been tested batch size > 1 HOT 2
- Error while building the custom caffe repo HOT 1
- Performance!!!
- Training with my own data : Multiple problems arising HOT 1
- Training on Pascal VOC
- Processing multiple images in parallel
- How to select the theta_alpha, theta_beta and theta_gamma?
- AttributeError: 'dict' object has no attribute 'itervalues' in /caffe/python/caffe/pycaffe.py line 260 HOT 1
- how to change the image size? HOT 1
- prototext sample- weight initialization HOT 1
- eltwise_layer.cpp:34] Check failed: bottom[0]->shape() == bottom[i]->shape() bottom[0]: 1 21 333 500 (3496500), bottom[1]: 1 21 375 500 (3937500) HOT 2
- FCN pretrained weights on COCO data
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from crfasrnn.