Get desktop application:
View/edit binary Protocol Buffers messages
Message that stores parameters used by AccuracyLayer
Used in:
,When computing accuracy, count as correct by comparing the true label to the top k scoring classes. By default, only compare to the top scoring class (i.e. argmax).
Used in:
Used in:
Used in:
Used in:
Message that stores parameters used by ArgMaxLayer
Used in:
,If true produce pairs (argmax, maxval)
Message that stores parameters used by ArgSortLayer
Used in:
Added 2015/7/7
Used in:
The initial filler for the scale
The initial filler for the shift
The epsilon added to variance
whether or not using moving average for inference
The decay factor for moving average
Used in:
NOTE: disparity_num must be divided by 4
0 mean only calculate once, non 0 mean accumulate sum
SOBEL_AND_IMAGE is default correlation
Used in:
Used in:
Used in:
How much does the moving average decay each iteration?
Used in:
If false, accumulate global mean/variance values via a moving average. If true, use those accumulated values instead of computing mean/variance across the batch.
How much does the moving average decay each iteration?
Small value to add to the variance estimate so that we don't divide by zero.
Used in:
The first axis of bottom[0] (the first input Blob) along which to apply bottom[1] (the second input Blob). May be negative to index from the end (e.g., -1 for the last axis). For example, if bottom[0] is 4D with shape 100x3x40x60, the output top[0] will have the same shape, and bottom[1] may have any of the following shapes (for the given value of axis): (axis == 0 == -4) 100; 100x3; 100x3x40; 100x3x40x60 (axis == 1 == -3) 3; 3x40; 3x40x60 (axis == 2 == -2) 40; 40x60 (axis == 3 == -1) 60 Furthermore, bottom[1] may have the empty shape (regardless of the value of "axis") -- a scalar bias.
(num_axes is ignored unless just one bottom is given and the bias is a learned parameter of the layer. Otherwise, num_axes is determined by the number of axes by the second bottom.) The number of axes of the input (bottom[0]) covered by the bias parameter, or -1 to cover all axes of bottom[0] starting from `axis`. Set num_axes := 0, to add a zero-axis Blob: a scalar.
(filler is ignored unless just one bottom is given and the bias is a learned parameter of the layer.) The initialization for the learned bias parameter. Default is the zero (0) initialization, resulting in the BiasLayer initially performing the identity operation.
Used in:
Used in:
Used in:
, , , ,5D dimensions -- deprecated. Use "shape" instead.
The BlobProtoVector is simply a way to pass multiple blobproto instances around.
Specifies the shape (dimensions) of a Blob.
Used in:
, , ,CTC parameter
Used in:
Used in:
Used in:
Used in:
Used in:
Message that stores parameters used by ConcatLayer
Used in:
,The axis along which to concatenate -- may be negative to index from the end (e.g., -1 for the last axis). Other axes must have the same dimension for all the bottom blobs. By default, ConcatLayer concatenates blobs along the "channels" axis (1).
DEPRECATED: alias for "axis" -- does not support negative indexing.
Message that stores parameters used by ContrastiveLossLayer
Used in:
,margin for dissimilar pair
Used in:
The number of outputs for the layer
whether to have bias terms
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding depth
The padding height
The padding width
The kernel size (square)
The kernel depth
The kernel height
The kernel width
The group size for group conv
The stride (equal in Y, X)
The stride depth
The stride height
The stride width
Message that stores parameters used by ConvolutionLayer
Used in:
,The number of outputs for the layer
whether to have bias terms
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding height
The padding width
The kernel size (square)
The kernel height
The kernel width
The group size for group conv
The stride (equal in Y, X)
The stride height
The stride width
The filler for the weight
The filler for the bias
repeated QuantizeParameter quantize_param = 74;
fusion relu for ReflectionPadConvolution [ocl]
out pad for pytorch deconv add by jianfei 20200824
The out_padding size (equal in Y, X)
The out_padding height
The out_padding width
Used in:
Used in:
The number of outputs for the layer
whether to have bias terms
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding height
The padding width
The kernel size (square)
The kernel height
The kernel width
The group size for group conv
The stride (equal in Y, X)
The stride height
The stride width
Used in:
groups size per batch
Used in:
Message that stores parameters used by CorrelationLayer
Used in:
The padding size (equal in Y, X)
The kernel size (square)
The maximum displacement (square)
The stride in blob 1 (equal in Y, X)
The stride in blob 2 (equal in Y, X)
For Correlation1D:
Correlate only to the left (-1) or right (1)
Use absolute value of result
Multiplicative is normal correlation
Used in:
Used in:
Used in:
Message that stores parameters used by CumProdLayer
Used in:
Message that stores parameters used by DataLayer
Used in:
,Specify the data source.
Specify the batch size.
The rand_skip variable is for the data layer to skip a few data points to avoid all asynchronous sgd clients to start at the same point. The skip point would be set as rand_skip * rand(0,1). Note that rand_skip should not be larger than the number of keys in the database.
DEPRECATED. See TransformationParameter. For data pre-processing, we can do simple scaling and subtracting the data mean, if provided. Note that the mean subtraction is always carried out before scaling.
DEPRECATED. See TransformationParameter. Specify if we would like to randomly crop an image.
DEPRECATED. See TransformationParameter. Specify if we want to randomly mirror data.
Force the encoded image to have 3 color channels
Used in:
the actual image data, in bytes
Optionally, the datum could also hold float data.
If true data contains an encoded image that need to be decoded
Used in:
The number of outputs for the layer
whether to have bias terms
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding depth
The padding height
The padding width
The kernel size (square)
The kernel depth
The kernel height
The kernel width
The group size for group conv
The stride (equal in Y, X)
The stride depth
The stride height
The stride width
The padding size (equal in Y, X)
The padding depth
The padding height
The padding widt
Message that store parameters used by DetectionOutputLayer
Used in:
Number of classes to be predicted. Required!
If true, bounding box are shared among different classes.
Background label id. If there is no background class, set it as -1.
Parameters used for non maximum suppression.
Parameters used for saving detection results.
Type of coding method for bbox.
If true, variance is encoded in target; otherwise we need to adjust the predicted offset accordingly.
Number of total bboxes to be kept per image after nms step. -1 means keeping all bboxes after nms step.
Only consider detections whose confidences are larger than a threshold. If not provided, consider all boxes.
If true, visualize the detection results.
The threshold used to visualize the detection results.
Message that stores parameters used by DropoutLayer
Used in:
,dropout ratio
Message that stores parameters used by DummyDataLayer. DummyDataLayer fills any number of arbitrarily shaped blobs with random (or constant) data generated by "Fillers" (see "message FillerParameter").
Used in:
,This layer produces N >= 1 top blobs. DummyDataParameter must specify 1 or N shape fields, and 0, 1 or N data_fillers. If 0 data_fillers are specified, ConstantFiller with a value of 0 is used. If 1 data_filler is specified, it is applied to all top blobs. If N are specified, the ith is applied to the ith top blob.
4D dimensions -- deprecated. Use "shape" instead.
Used in:
(message has no fields)
Message that stores parameters used by EltwiseLayer
Used in:
,element-wise operation
blob-wise coefficient for SUM operation
Whether to use an asymptotically slower (for >2 inputs) but stabler method of computing the gradient for the PROD operation. (No effect for SUM op.)
Used in:
Message that stores parameters used by ExpLayer
Used in:
,ExpLayer computes outputs y = base ^ (shift + scale * x), for base > 0. Or if base is set to the default (-1), base is set to e, so y = exp(shift + scale * x).
Used in:
, , , , , , , , , , ,The filler type.
the value in constant filler
the min value in uniform filler
the max value in uniform filler
the mean value in Gaussian filler
the std value in Gaussian filler
The expected number of non-zero output weights for a given input in Gaussian filler -- the default -1 means don't perform sparsification.
/ Message that stores parameters used by FlattenLayer
Used in:
The first axis to flatten: all preceding axes are retained in the output. May be negative to index from the end (e.g., -1 for the last axis).
The last axis to flatten: all following axes are retained in the output. May be negative to index from the end (e.g., the default -1 for the last axis).
Used in:
Used in:
Used in:
Used in:
Used in:
Used in:
Message that stores parameters used by HDF5DataLayer
Used in:
,Specify the data source.
Specify the batch size.
Message that stores parameters used by HDF5OutputLayer
Used in:
, ,Used in:
Used in:
,Specify the Norm to use L1 or L2
Used in:
Message that stores parameters used by ImageDataLayer
Used in:
,Specify the data source.
Specify the batch size.
The rand_skip variable is for the data layer to skip a few data points to avoid all asynchronous sgd clients to start at the same point. The skip point would be set as rand_skip * rand(0,1). Note that rand_skip should not be larger than the number of keys in the database.
Whether or not ImageLayer should shuffle the list of files at every epoch.
It will also resize images if new_height or new_width are not zero.
Specify if the images are color or gray
DEPRECATED. See TransformationParameter. For data pre-processing, we can do simple scaling and subtracting the data mean, if provided. Note that the mean subtraction is always carried out before scaling.
DEPRECATED. See TransformationParameter. Specify if we would like to randomly crop an image.
DEPRECATED. See TransformationParameter. Specify if we want to randomly mirror data.
Message that stores parameters InfogainLossLayer
Used in:
,Specify the infogain matrix source.
Message that stores parameters used by InnerProductLayer
Used in:
,The number of outputs for the layer
whether to have bias terms
The filler for the weight
The filler for the bias
The first axis to be lumped into a single inner product computation; all preceding axes are retained in the output. May be negative to index from the end (e.g., -1 for the last axis).
Used in:
Message that stores parameters used by Interp3dLayer
Used in:
Depth of output
Height of output
Width of output
zoom factor
shrink factor
padding at begin of input
padding at end of input
Message that stores parameters used by InterpLayer
Used in:
Height of output
Width of output
zoom factor
shrink factor
padding at begin of input
padding at end of input
Used in:
Message that stores parameters used by LRNLayer
Used in:
,Used in:
Used in:
The number of outputs for the layer
The filler for weight
The filler for the bias
NOTE Update the next available ID when you add a new LayerParameter field. LayerParameter next available layer-specific ID: 143 (last added: detection_output_param)
Used in:
the layer name
the layer type
the name of each bottom blob
the name of each top blob
The train / test phase for computation.
The amount of weight to assign each top blob in the objective. Each layer assigns a default value, usually of either 0 or 1, to each top blob.
Specifies training parameters (multipliers on global learning constants, and the name and other settings used for weight sharing).
The blobs containing the numeric parameters of the layer.
Rules controlling whether and when a layer is included in the network, based on the current NetState. You may specify a non-zero number of rules to include OR exclude, but not both. If no include or exclude rules are specified, the layer is always included. If the current NetState meets ANY (i.e., one or more) of the specified rules, the layer is included/excluded.
set data type of top blobs;(fp32 fp16 uchar int8 ...)
Parameters for data pre-processing.
Parameters shared by loss layers.
Layer type-specific parameters. Note: certain layers may have more than one computational engine for their implementation. These layers include an Engine type and engine parameter for selecting the implementation. The default for the engine is set by the ENGINE switch at compile-time.
! add by linan 2017.11.23 ! original number is 141
! <<<
! add by linan 2017.4.20
! add by linan 2016.12.26
! add by pengcuo 2017.02.07
! add by linan 2017.5.3
! add by pengcuo 2017.09.07
! add pengcuo 2017.8.17
! add by pengcuo 2017.09.21
! add by pengcuo 2017.08
! add by linan 2017.10.17
! <<<
! add by yingrui 2017.12.22
! add by yingrui 2017.10.25
! add by yingrui 2017.11.07
! add by linan 2019.11.15, 3d layer parameters
optional ConvolutionTranspose3dParameter convolution_transpose3d_param = 2009;
! add by linan 2018.5.28
! add by linjin 2020.02.19
! add by liangjiexin 2020.3.2
! add by liangjiexin 2020.4.15
! add by linjin 2020.4.27
! add by jianfei 2020.8.14
! add by jianfei 2020.7.14
! add by linjin 2020.8.19
! add by linjin 2020.8.6
! add by linjin 2020.9.3
! add by liangjiexin 2020.9.10
! add by jianfei 2020.9.27
! add by jianfei 2020.10.15
! add by liangjiexin 2020.10.28
! added by tianzichen 2020.10.29 hexagon dsp uint16 quantize bits, used only when PrecisionType = UINT16, range [1, 16] default = 0 means this layer's quantize bits is not specified, will use global quantize bits.
Used in:
Message that stores parameters used by ExpLayer
Used in:
LogLayer computes outputs y = log_base(shift + scale * x), for base > 0. Or if base is set to the default (-1), base is set to e, so y = ln(shift + scale * x) = log_e(shift + scale * x)
Message that stores parameters shared by loss layers
Used in:
,If specified, ignore instances with the given label.
If true, normalize each batch across all instances (including spatial dimesions, but not ignored instances); else, divide by batch size only.
Message that stores parameters used by MVNLayer
Used in:
,This parameter can be set to false to normalize mean only
This parameter can be set to true to perform DNN-like MVN
Used in:
Used in:
gemm mode
Used in:
Used in:
Message that stores parameters used by MemoryDataLayer
Used in:
,Used in:
Used in:
Used in:
consider giving the network a name
The input blobs to the network.
The shape of the input blobs.
4D input dimensions -- deprecated. Use "shape" instead. If specified, for each input blob there should be four values specifying the num, channels, height and width of the input blob. Thus, there should be a total of (4 * #input) numbers.
Whether the network will force every layer to carry out backward operation. If set False, then whether to carry out backward is determined automatically according to the net structure and learning rates.
The current "state" of the network, including the phase, level, and stage. Some layers may be included/excluded depending on this state and the states specified in the layers' include and exclude fields.
Print debugging information about results while running Net::Forward, Net::Backward, and Net::Update.
The layers that make up the net. Each of their configurations, including connectivity and behavior, is specified as a LayerParameter.
ID 100 so layers are printed last.
DEPRECATED: use 'layer' instead.
Used in:
,Used in:
,Set phase to require the NetState have a particular phase (TRAIN or TEST) to meet this rule.
Set the minimum and/or maximum levels in which the layer should be used. Leave undefined to meet the rule regardless of level.
Customizable sets of stages to include or exclude. The net must have ALL of the specified stages and NONE of the specified "not_stage"s to meet the rule. (Use multiple NetStateRules to specify conjunctions of stages.)
Used in:
Threshold to be used in nms.
Maximum number of results to be kept.
Used in:
Used in:
Parametric ReLU described in K. He et al, Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification, 2015.
Used in:
Initial value of a_i. Default is a_i=0.25 for all i.
Whether or not slope paramters are shared across channels.
Used in:
output channel number
equal to pooled_size
Message that stores parameters used by PSROIPoolingLayer
Used in:
output channel number
equal to pooled_size
scale for roi
Used in:
Used in:
Specifies training parameters (multipliers on global learning constants, and the name and other settings used for weight sharing).
Used in:
The names of the parameter blobs -- useful for sharing parameters among layers, but never required otherwise. To share a parameter between two layers, give it a (non-empty) name.
Whether to require shared weights to have the same shape, or just the same count -- defaults to STRICT if unspecified.
The multiplier on the global learning rate for this parameter.
The multiplier on the global weight decay for this parameter.
Used in:
STRICT (default) requires that num, channels, height, width each match.
PERMISSIVE requires only the count (num*channels*height*width) to match.
Used in:
Used in:
The new orders of the axes of data. Notice it should be with in the same range as the input data, and it starts from 0. Do not provide repeated order.
Used in:
, ,Used in:
Used in:
The filler for the slope
The filler for the bias
Message that stores parameters used by PoolingLayer
Used in:
The pooling method
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding depth
The padding height
The padding width
The kernel size (square)
The kernel depth
The kernel height
The kernel width
The stride (equal in Y, X)
The stride depth
The stride height
The stride width
If global_pooling then it will pool over the size of the bottom by doing kernel_h = bottom->height and kernel_w = bottom->width
Used in:
Message that stores parameters used by PoolingLayer
Used in:
,The pooling method
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding height
The padding width
The kernel size (square)
The kernel height
The kernel width
The stride (equal in Y, X)
The stride height
The stride width
If global_pooling then it will pool over the size of the bottom by doing kernel_h = bottom->height and kernel_w = bottom->width
Used in:
Used in:
Used in:
Used in:
Message that stores parameters used by PowerLayer
Used in:
,PowerLayer computes outputs y = (shift + scale * x) ^ power.
Message that store parameters used by PriorBoxLayer
Used in:
Minimum box size (in pixels). Required!
Maximum box size (in pixels). Required!
Various of aspect ratios. Duplicate ratios will be ignored. If none is provided, we use default ratio 1.
If true, will flip each aspect ratio. For example, if there is aspect ratio "r", we will generate aspect ratio "1.0/r" as well.
If true, will clip the prior so that it is within [0, 1]
Variance for adjusting the prior bboxes.
Encode/decode type.
Used in:
Used in:
Encode/decode type. Minimum box size (in pixels). Required!
If true, will clip the prior so that it is within [0, 1]
Variance for adjusting the prior bboxes.
By default, we calculate img_height, img_width, step_x, step_y based on bottom[0] (feat) and bottom[1] (img). Unless these values are explicitely provided. Explicitly provide the img_size.
Either img_size or img_h/img_w should be specified; not both.
Explicitly provide the step size.
Either step or step_h/step_w should be specified; not both.
Offset to the top left corner of each cell.
Message that stores parameters used by PythonLayer
Used in:
Used in:
, , ,each one in uint8_t represents how much in float.
what 0 represents.
what 255 represents.
what 0.0f is represented in uint8
whether to adjust range_min and range_max such that zero_point can be precisely represented by integer.
Used in:
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The pooled output height
The pooled output width
Multiplicative spatial scale factor to translate ROI coords from their input scale to the scale used when pooling
Used in:
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The pooled output height
The pooled output width
Multiplicative spatial scale factor to translate ROI coords from their input scale to the scale used when pooling
Used in:
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The pooled output height
The pooled output width
Multiplicative spatial scale factor to translate ROI coords from their input scale to the scale used when pooling
Used in:
Used in:
Message that stores parameters used by ROIPoolingLayer
Used in:
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The pooled output height
The pooled output width
Multiplicative spatial scale factor to translate ROI coords from their input scale to the scale used when pooling
Used in:
Used in:
Used in:
Allow non-zero slope for negative inputs to speed up optimization Described in: Maas, A. L., Hannun, A. Y., & Ng, A. Y. (2013). Rectifier nonlinearities improve neural network acoustic models. In ICML Workshop on Deep Learning for Audio, Speech, and Language Processing.
Message that stores parameters used by ReLULayer
Used in:
,Allow non-zero slope for negative inputs to speed up optimization Described in: Maas, A. L., Hannun, A. Y., & Ng, A. Y. (2013). Rectifier nonlinearities improve neural network acoustic models. In ICML Workshop on Deep Learning for Audio, Speech, and Language Processing.
Used in:
Message that stores parameters used by RecurrentLayer
Used in:
The dimension of the output (and usually hidden state) representation -- must be explicitly set to not-zero.
The filler for the weight
The filler for the bias
Whether to enable displaying debug_info in the unrolled recurrent net.
Used in:
Used in:
The index of the axis to tile.
The number of copies (tiles) of the blob to output.
Used in:
Specify the output dimensions. If some of the dimensions are set to 0, the corresponding dimension from the bottom layer is used (unchanged). Exactly one dimension may be set to -1, in which case its value is inferred from the count of the bottom blob and the remaining dimensions. For example, suppose we want to reshape a 2D blob "input" with shape 2 x 8: layer { type: "Reshape" bottom: "input" top: "output" reshape_param { ... } } If "input" is 2D with shape 2 x 8, then the following reshape_param specifications are all equivalent, producing a 3D blob "output" with shape 2 x 2 x 4: reshape_param { shape { dim: 2 dim: 2 dim: 4 } } reshape_param { shape { dim: 0 dim: 2 dim: 4 } } reshape_param { shape { dim: 0 dim: 2 dim: -1 } } reshape_param { shape { dim: -1 dim: 0 dim: 2 } }
axis and num_axes control the portion of the bottom blob's shape that are replaced by (included in) the reshape. By default (axis == 0 and num_axes == -1), the entire bottom blob shape is included in the reshape, and hence the shape field must specify the entire output shape. axis may be non-zero to retain some portion of the beginning of the input shape (and may be negative to index from the end; e.g., -1 to begin the reshape after the last axis, including nothing in the reshape, -2 to include only the last axis, etc.). For example, suppose "input" is a 2D blob with shape 2 x 8. Then the following ReshapeLayer specifications are all equivalent, producing a blob "output" with shape 2 x 2 x 4: reshape_param { shape { dim: 2 dim: 2 dim: 4 } } reshape_param { shape { dim: 2 dim: 4 } axis: 1 } reshape_param { shape { dim: 2 dim: 4 } axis: -3 } num_axes specifies the extent of the reshape. If num_axes >= 0 (and axis >= 0), the reshape will be performed only on input axes in the range [axis, axis+num_axes]. num_axes may also be -1, the default, to include all remaining axes (starting from axis). For example, suppose "input" is a 2D blob with shape 2 x 8. Then the following ReshapeLayer specifications are equivalent, producing a blob "output" with shape 1 x 2 x 8. reshape_param { shape { dim: 1 dim: 2 dim: 8 } } reshape_param { shape { dim: 1 dim: 2 } num_axes: 1 } reshape_param { shape { dim: 1 } num_axes: 0 } On the other hand, these would produce output blob shape 2 x 1 x 8: reshape_param { shape { dim: 2 dim: 1 dim: 8 } } reshape_param { shape { dim: 1 } axis: 1 num_axes: 0 }
Used in:
Used in:
Used in:
Used in:
Used in:
Output directory. If not empty, we will save the results.
Output name prefix.
Output format. VOC - PASCAL VOC output format. COCO - MS COCO output format.
If you want to output results, must also provide the following two files. Otherwise, we will ignore saving results. label map file.
A file which contains a list of names and sizes with same order of the input DB. The file is in the following format: name height width ...
Number of test images. It can be less than the lines specified in name_size_file. For example, when we only want to evaluate on part of the test images.
Used in:
Used in:
Message that stores parameters used by SigmoidLayer
Used in:
,Used in:
Message that stores parameters used by SliceLayer
Used in:
,The axis along which to slice -- may be negative to index from the end (e.g., -1 for the last axis). By default, SliceLayer concatenates blobs along the "channels" axis (1).
DEPRECATED: alias for "axis" -- does not support negative indexing.
Used in:
the grid depth to be considered.
scale in range [0, 1] along width
scale in range [0, 1] along height
padding(or offset) along width
padding(or offset) along height
padding(or offset) along depth
Message that stores parameters used by SoftmaxLayer, SoftmaxWithLossLayer
Used in:
,The axis along which to perform the softmax -- may be negative to index from the end (e.g., -1 for the last axis). Any other axes will be evaluated as independent softmaxes.
Used in:
NOTE Update the next available ID when you add a new SolverParameter field. SolverParameter next available ID: 36 (last added: clip_gradients)
//////////////////////////////////////////////////////////////////////////// Specifying the train and test networks Exactly one train net must be specified using one of the following fields: train_net_param, train_net, net_param, net One or more test nets may be specified using any of the following fields: test_net_param, test_net, net_param, net If more than one test net field is specified (e.g., both net and test_net are specified), they will be evaluated in the field order given above: (1) test_net_param, (2) test_net, (3) net_param/net. A test_iter must be specified for each test_net. A test_level and/or a test_stage may also be specified for each test_net. ////////////////////////////////////////////////////////////////////////////
Proto filename for the train net, possibly combined with one or more test nets.
Inline train net param, possibly combined with one or more test nets.
Proto filename for the train net.
Proto filenames for the test nets.
Inline train net params.
Inline test net params.
The states for the train/test nets. Must be unspecified or specified once per net. By default, all states will have solver = true; train_state will have phase = TRAIN, and all test_state's will have phase = TEST. Other defaults are set according to the NetState defaults.
The number of iterations for each test net.
The number of iterations between two testing phases.
If true, run an initial test pass before the first iteration, ensuring memory availability and printing the starting value of the loss.
The base learning rate
the number of iterations between displaying info. If display = 0, no info will be displayed.
Display the loss averaged over the last average_loss iterations
the maximum number of iterations
The learning rate decay policy.
The parameter to compute the learning rate.
The parameter to compute the learning rate.
The momentum value.
The weight decay.
regularization types supported: L1 and L2 controlled by weight_decay
the stepsize for learning rate policy "step"
the stepsize for learning rate policy "multistep"
Set clip_gradients to >= 0 to clip parameter gradients to that L2 norm, whenever their actual L2 norm is larger.
The snapshot interval
The prefix for the snapshot.
whether to snapshot diff in the results or not. Snapshotting diff will help debugging but the final protocol buffer size will be much larger.
the device_id will that be used in GPU mode. Use device_id = 0 in default.
If non-negative, the seed with which the Solver will initialize the Caffe random number generator -- useful for reproducible results. Otherwise, (and by default) initialize using a seed derived from the system clock.
numerical stability for AdaGrad
If true, print information about the state of the net that may help with debugging learning problems.
If false, don't save a snapshot after training finishes.
the mode solver will use: 0 for CPU and 1 for GPU. Use GPU in default.
Used in:
Solver type
Used in:
A message that stores the solver snapshots
The current iteration
The file that stores the learned net.
The history for sgd solvers
The current step for learning rate
Used in:
Used in:
Used in:
Used in:
Message that stores parameters used by TanHLayer
Used in:
,Used in:
Message that stores parameters used by ThresholdLayer
Used in:
,Strictly positive values
Message that stores parameters used by TileLayer
Used in:
The index of the axis to tile.
The number of copies (tiles) of the blob to output.
Message that stores parameters used to apply transformation to the data layer's data
Used in:
,For data pre-processing, we can do simple scaling and subtracting the data mean, if provided. Note that the mean subtraction is always carried out before scaling.
Specify if we want to randomly mirror data.
Specify if we would like to randomly crop an image.
mean_file and mean_value cannot be specified at the same time
if specified can be repeated once (would substract it from all the channels) or can be repeated the same number of times as channels (would subtract them from the corresponding channel)
Used in:
Used in:
The pooling method
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding height
The padding width
The kernel size (square)
The kernel height
The kernel width
The stride (equal in Y, X)
The stride height
The stride width
destined unpooled map size
destined unpooled map height
destined unpooled map width
Used in:
Used in:
DEPRECATED: V0LayerParameter is the old way of specifying layer parameters in Caffe. We keep this message type around for legacy support.
Used in:
the layer name
the string to specify the layer type
Parameters to specify layers with inner products.
The number of outputs for the layer
whether to have bias terms
The filler for the weight
The filler for the bias
The padding size
The kernel size
The group size for group conv
The stride
The pooling method
dropout ratio
for local response norm
for local response norm
for local response norm
For data layers, specify the data source
For data pre-processing, we can do simple scaling and subtracting the data mean, if provided. Note that the mean subtraction is always carried out before scaling.
For data layers, specify the batch size.
For data layers, specify if we would like to randomly crop an image.
For data layers, specify if we want to randomly mirror data.
The blobs containing the numeric parameters of the layer
The ratio that is multiplied on the global learning rate. If you want to set the learning ratio for one blob, you need to set it for all blobs.
The weight decay that is multiplied on the global weight decay.
The rand_skip variable is for the data layer to skip a few data points to avoid all asynchronous sgd clients to start at the same point. The skip point would be set as rand_skip * rand(0,1). Note that rand_skip should not be larger than the number of keys in the database.
Fields related to detection (det_*) foreground (object) overlap threshold
background (non-object) overlap threshold
Fraction of batch that should be foreground objects
Amount of contextual padding to add around a window (used only by the window_data_layer)
Mode for cropping out a detection window warp: cropped window is warped to a fixed size and aspect ratio square: the tightest square around the window is cropped
For ReshapeLayer, one needs to specify the new dimensions.
Whether or not ImageLayer should shuffle the list of files at every epoch. It will also resize images if new_height or new_width are not zero.
For ConcatLayer, one needs to specify the dimension for concatenation, and the other dimensions must be the same for all the bottom blobs. By default it will concatenate blobs along the channels dimension.
Used in:
DEPRECATED: use LayerParameter.
Used in:
Used in:
Used in:
Used in:
Message that stores parameters used by WindowDataLayer
Used in:
,Specify the data source.
For data pre-processing, we can do simple scaling and subtracting the data mean, if provided. Note that the mean subtraction is always carried out before scaling.
Specify the batch size.
Specify if we would like to randomly crop an image.
Specify if we want to randomly mirror data.
Foreground (object) overlap threshold
Background (non-object) overlap threshold
Fraction of batch that should be foreground objects
Amount of contextual padding to add around a window (used only by the window_data_layer)
Mode for cropping out a detection window warp: cropped window is warped to a fixed size and aspect ratio square: the tightest square around the window is cropped
cache_images: will load all images in memory for faster access
append root_folder to locate images
Used in:
! PPL Will ignore FillerParameter ! But caffe use Filler to init blobs_.
Used in: