Get desktop application:
View/edit binary Protocol Buffers messages
Used in:
If true produce pairs (argmax, maxval)
The axis along which to maximise -- may be negative to index from the end (e.g., -1 for the last axis). By default ArgMaxLayer maximizes over the flattened trailing dimensions for each index of the first / num dimension.
Used in:
If true, will use the moving average mean and std for training and test. Will override the lr_param and freeze all the parameters. Make sure to initialize the layer properly with pretrained parameters.
Used in:
Used in:
If false, accumulate global mean/variance values via a moving average. If true, use those accumulated values instead of computing mean/variance across the batch.
How much does the moving average decay each iteration?
Small value to add to the variance estimate so that we don't divide by zero.
Used in:
The first axis of bottom[0] (the first input Blob) along which to apply bottom[1] (the second input Blob). May be negative to index from the end (e.g., -1 for the last axis). For example, if bottom[0] is 4D with shape 100x3x40x60, the output top[0] will have the same shape, and bottom[1] may have any of the following shapes (for the given value of axis): (axis == 0 == -4) 100; 100x3; 100x3x40; 100x3x40x60 (axis == 1 == -3) 3; 3x40; 3x40x60 (axis == 2 == -2) 40; 40x60 (axis == 3 == -1) 60 Furthermore, bottom[1] may have the empty shape (regardless of the value of "axis") -- a scalar bias.
(num_axes is ignored unless just one bottom is given and the bias is a learned parameter of the layer. Otherwise, num_axes is determined by the number of axes by the second bottom.) The number of axes of the input (bottom[0]) covered by the bias parameter, or -1 to cover all axes of bottom[0] starting from `axis`. Set num_axes := 0, to add a zero-axis Blob: a scalar.
(filler is ignored unless just one bottom is given and the bias is a learned parameter of the layer.) The initialization for the learned bias parameter. Default is the zero (0) initialization, resulting in the BiasLayer initially performing the identity operation.
Used in:
, ,4D dimensions -- deprecated. Use "shape" instead.
The BlobProtoVector is simply a way to pass multiple blobproto instances around.
Specifies the shape (dimensions) of a Blob.
Used in:
, , , ,Used in:
,The axis along which to concatenate -- may be negative to index from the end (e.g., -1 for the last axis). Other axes must have the same dimension for all the bottom blobs. By default, ConcatLayer concatenates blobs along the "channels" axis (1).
DEPRECATED: alias for "axis" -- does not support negative indexing.
Used in:
,The number of outputs for the layer
whether to have bias terms
Pad, kernel size, and stride are all given as a single value for equal dimensions in all spatial dimensions, or once per spatial dimension.
The padding size; defaults to 0
The kernel size
The stride; defaults to 1
Factor used to dilate the kernel, (implicitly) zero-filling the resulting holes. (Kernel dilation is sometimes referred to by its use in the algorithme à trous from Holschneider et al. 1987.)
The dilation; defaults to 1
For 2D convolution only, the *_h and *_w versions may also be used to specify both spatial dimensions.
The padding height (2D only)
The padding width (2D only)
The kernel height (2D only)
The kernel width (2D only)
The stride height (2D only)
The stride width (2D only)
The group size for group conv
The filler for the weight
The filler for the bias
The axis to interpret as "channels" when performing convolution. Preceding dimensions are treated as independent inputs; succeeding dimensions are treated as "spatial". With (N, C, H, W) inputs, and axis == 1 (the default), we perform N independent 2D convolutions, sliding C-channel (or (C/g)-channels, for groups g>1) filters across the spatial axes (H, W) of the input. With (N, C, D, H, W) inputs, and axis == 1, we perform N independent 3D convolutions, sliding (C/g)-channels filters across the spatial axes (D, H, W) of the input.
Whether to force use of the general ND convolution, even if a specific implementation for blobs of the appropriate number of spatial dimensions is available. (Currently, there is only a 2D-specific convolution implementation; for input blobs with num_axes != 2, this option is ignored and the ND implementation will be used.)
Used in:
Used in:
To crop, elements of the first bottom are selected to fit the dimensions of the second, reference bottom. The crop is configured by - the crop `axis` to pick the dimensions for cropping - the crop `offset` to set the shift for all/each dimension to align the cropped bottom with the reference bottom. All dimensions up to but excluding `axis` are preserved, while the dimensions including and trailing `axis` are cropped. If only one `offset` is set, then all dimensions are offset by this amount. Otherwise, the number of offsets must equal the number of cropped axes to shift the crop in each dimension accordingly. Note: standard dimensions are N,C,H,W so the default is a spatial crop, and `axis` may be negative to index from the end (e.g., -1 for the last axis).
Message that store parameters used by DetectionOutputLayer
Used in:
Number of classes to be predicted. Required!
If true, bounding box are shared among different classes.
Background label id. If there is no background class, set it as -1.
Parameters used for non maximum suppression.
Type of coding method for bbox.
If true, variance is encoded in target; otherwise we need to adjust the predicted offset accordingly.
Number of total bboxes to be kept per image after nms step. -1 means keeping all bboxes after nms step.
Only consider detections whose confidences are larger than a threshold. If not provided, consider all boxes.
If true, visualize the detection results.
The threshold used to visualize the detection results.
If provided, save outputs to video file.
Used in:
,dropout ratio
Message that stores parameters used by ELULayer
Used in:
Described in: Clevert, D.-A., Unterthiner, T., & Hochreiter, S. (2015). Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). arXiv
Used in:
,element-wise operation
blob-wise coefficient for SUM operation
Whether to use an asymptotically slower (for >2 inputs) but stabler method of computing the gradient for the PROD operation. (No effect for SUM op.)
Used in:
Message that stores parameters used by EmbedLayer
Used in:
The number of outputs for the layer
The input is given as integers to be interpreted as one-hot vector indices with dimension num_input. Hence num_input should be 1 greater than the maximum possible input value.
Whether to use a bias term
The filler for the weight
The filler for the bias
Message that stores parameters used by ExpLayer
Used in:
,ExpLayer computes outputs y = base ^ (shift + scale * x), for base > 0. Or if base is set to the default (-1), base is set to e, so y = exp(shift + scale * x).
Used in:
, , , , , , ,The filler type.
the value in constant filler
the min value in uniform filler
the max value in uniform filler
the mean value in Gaussian filler
the std value in Gaussian filler
The expected number of non-zero output weights for a given input in Gaussian filler -- the default -1 means don't perform sparsification.
Normalize the filler variance by fan_in, fan_out, or their average. Applies to 'xavier' and 'msra' fillers.
Used in:
/ Message that stores parameters used by FlattenLayer
Used in:
The first axis to flatten: all preceding axes are retained in the output. May be negative to index from the end (e.g., -1 for the last axis).
The last axis to flatten: all following axes are retained in the output. May be negative to index from the end (e.g., the default -1 for the last axis).
Used in:
,The number of outputs for the layer
whether to have bias terms
The filler for the weight
The filler for the bias
The first axis to be lumped into a single inner product computation; all preceding axes are retained in the output. May be negative to index from the end (e.g., -1 for the last axis).
Specify whether to transpose the weight matrix or not. If transpose == true, any operations will be performed on the transpose of the weight matrix. The weight matrix itself is not going to be transposed but rather the transfer flag of operations will be toggled accordingly.
Used in:
This layer produces N >= 1 top blob(s) to be assigned manually. Define N shapes to set a shape for each top. Define 1 shape to set the same shape for every top. Define no shape to defer to reshaping manually.
Message that stores parameters used by LRNLayer
Used in:
,Used in:
Used in:
NOTE Update the next available ID when you add a new LayerParameter field. LayerParameter next available layer-specific ID: 147 (last added: recurrent_param)
Used in:
the layer name
the layer type
the name of each bottom blob
the name of each top blob
The train / test phase for computation.
The amount of weight to assign each top blob in the objective. Each layer assigns a default value, usually of either 0 or 1, to each top blob.
Specifies training parameters (multipliers on global learning constants, and the name and other settings used for weight sharing).
The blobs containing the numeric parameters of the layer.
Specifies whether to backpropagate to each bottom. If unspecified, Caffe will automatically infer whether each input needs backpropagation to compute parameter gradients. If set to true for some inputs, backpropagation to those inputs is forced; if set false for some inputs, backpropagation to those inputs is skipped. The size must be either 0 or equal to the number of bottoms.
Rules controlling whether and when a layer is included in the network, based on the current NetState. You may specify a non-zero number of rules to include OR exclude, but not both. If no include or exclude rules are specified, the layer is always included. If the current NetState meets ANY (i.e., one or more) of the specified rules, the layer is included/excluded.
Parameters shared by loss layers.
Layer type-specific parameters. Note: certain layers may have more than one computational engine for their implementation. These layers include an Engine type and engine parameter for selecting the implementation. The default for the engine is set by the ENGINE switch at compile-time.
Message that stores parameters used by LogLayer
Used in:
LogLayer computes outputs y = log_base(shift + scale * x), for base > 0. Or if base is set to the default (-1), base is set to e, so y = ln(shift + scale * x) = log_e(shift + scale * x)
Message that stores parameters shared by loss layers
Used in:
,If specified, ignore instances with the given label.
Deprecated. Ignored if normalization is specified. If normalization is not specified, then setting this to false will be equivalent to normalization = BATCH_SIZE to be consistent with previous behavior.
How to normalize the loss for loss layers that aggregate across batches, spatial dimensions, or other dimensions. Currently only implemented in SoftmaxWithLoss layer.
Used in:
Divide by the number of examples in the batch times spatial dimensions. Outputs that receive the ignore label will NOT be ignored in computing the normalization factor.
Divide by the total number of output locations that do not take the ignore_label. If ignore_label is not set, this behaves like FULL.
Divide by the batch size.
Do not normalize the loss.
Used in:
,This parameter can be set to false to normalize mean only
This parameter can be set to true to perform DNN-like MVN
Epsilon for not dividing by zero while normalizing variance
consider giving the network a name
DEPRECATED. See InputParameter. The input blobs to the network.
DEPRECATED. See InputParameter. The shape of the input blobs.
4D input dimensions -- deprecated. Use "input_shape" instead. If specified, for each input blob there should be four values specifying the num, channels, height and width of the input blob. Thus, there should be a total of (4 * #input) numbers.
Whether the network will force every layer to carry out backward operation. If set False, then whether to carry out backward is determined automatically according to the net structure and learning rates.
The current "state" of the network, including the phase, level, and stage. Some layers may be included/excluded depending on this state and the states specified in the layers' include and exclude fields.
Print debugging information about results while running Net::Forward, Net::Backward, and Net::Update.
The layers that make up the net. Each of their configurations, including connectivity and behavior, is specified as a LayerParameter.
ID 100 so layers are printed last.
DEPRECATED: use 'layer' instead.
Used in:
Used in:
,Set phase to require the NetState have a particular phase (TRAIN or TEST) to meet this rule.
Set the minimum and/or maximum levels in which the layer should be used. Leave undefined to meet the rule regardless of level.
Customizable sets of stages to include or exclude. The net must have ALL of the specified stages and NONE of the specified "not_stage"s to meet the rule. (Use multiple NetStateRules to specify conjunctions of stages.)
Used in:
Threshold to be used in nms.
Maximum number of results to be kept.
Parameter for adaptive nms.
Message that stores parameters used by NormalizeLayer
Used in:
Initial value of scale. Default is 1.0 for all
Whether or not scale parameters are shared across channels.
Epsilon for not dividing by zero while normalizing variance
Parametric ReLU described in K. He et al, Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification, 2015.
Used in:
Initial value of a_i. Default is a_i=0.25 for all i.
Whether or not slope paramters are shared across channels.
Used in:
output channel number
equal to pooled_size
Specifies training parameters (multipliers on global learning constants, and the name and other settings used for weight sharing).
Used in:
The names of the parameter blobs -- useful for sharing parameters among layers, but never required otherwise. To share a parameter between two layers, give it a (non-empty) name.
Whether to require shared weights to have the same shape, or just the same count -- defaults to STRICT if unspecified.
The multiplier on the global learning rate for this parameter.
The multiplier on the global weight decay for this parameter.
Used in:
STRICT (default) requires that num, channels, height, width each match.
PERMISSIVE requires only the count (num*channels*height*width) to match.
Used in:
Used in:
The new orders of the axes of data. Notice it should be with in the same range as the input data, and it starts from 0. Do not provide repeated order.
Used in:
, ,Used in:
,The pooling method
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The padding size (equal in Y, X)
The padding height
The padding width
The kernel size (square)
The kernel height
The kernel width
The stride (equal in Y, X)
The stride height
The stride width
If global_pooling then it will pool over the size of the bottom by doing kernel_h = bottom->height and kernel_w = bottom->width
Used in:
Used in:
Used in:
,PowerLayer computes outputs y = (shift + scale * x) ^ power.
Message that store parameters used by PriorBoxLayer
Used in:
Minimum box size (in pixels). Required!
Maximum box size (in pixels). Required!
Various of aspect ratios. Duplicate ratios will be ignored. If none is provided, we use default ratio 1.
If true, will flip each aspect ratio. For example, if there is aspect ratio "r", we will generate aspect ratio "1.0/r" as well.
If true, will clip the prior so that it is within [0, 1]
Variance for adjusting the prior bboxes.
By default, we calculate img_height, img_width, step_x, step_y based on bottom[0] (feat) and bottom[1] (img). Unless these values are explicitely provided. Explicitly provide the img_size.
Either img_size or img_h/img_w should be specified; not both.
Explicitly provide the step size.
Either step or step_h/step_w should be specified; not both.
Offset to the top left corner of each cell.
Encode/decode type.
Used in:
Used in:
Used in:
Pad, kernel size, and stride are all given as a single value for equal dimensions in height and width or as Y, X pairs.
The pooled output height
The pooled output width
Multiplicative spatial scale factor to translate ROI coords from their input scale to the scale used when pooling
Message that stores parameters used by ReLULayer
Used in:
,Allow non-zero slope for negative inputs to speed up optimization Described in: Maas, A. L., Hannun, A. Y., & Ng, A. Y. (2013). Rectifier nonlinearities improve neural network acoustic models. In ICML Workshop on Deep Learning for Audio, Speech, and Language Processing.
Used in:
Message that stores parameters used by ReductionLayer
Used in:
reduction operation
The first axis to reduce to a scalar -- may be negative to index from the end (e.g., -1 for the last axis). (Currently, only reduction along ALL "tail" axes is supported; reduction of axis M through N, where N < num_axes - 1, is unsupported.) Suppose we have an n-axis bottom Blob with shape: (d0, d1, d2, ..., d(m-1), dm, d(m+1), ..., d(n-1)). If axis == m, the output Blob will have shape (d0, d1, d2, ..., d(m-1)), and the ReductionOp operation is performed (d0 * d1 * d2 * ... * d(m-1)) times, each including (dm * d(m+1) * ... * d(n-1)) individual data. If axis == 0 (the default), the output Blob always has the empty shape (count 1), performing reduction across the entire input -- often useful for creating new loss functions.
coefficient for output
Used in:
Used in:
Specify the output dimensions. If some of the dimensions are set to 0, the corresponding dimension from the bottom layer is used (unchanged). Exactly one dimension may be set to -1, in which case its value is inferred from the count of the bottom blob and the remaining dimensions. For example, suppose we want to reshape a 2D blob "input" with shape 2 x 8: layer { type: "Reshape" bottom: "input" top: "output" reshape_param { ... } } If "input" is 2D with shape 2 x 8, then the following reshape_param specifications are all equivalent, producing a 3D blob "output" with shape 2 x 2 x 4: reshape_param { shape { dim: 2 dim: 2 dim: 4 } } reshape_param { shape { dim: 0 dim: 2 dim: 4 } } reshape_param { shape { dim: 0 dim: 2 dim: -1 } } reshape_param { shape { dim: 0 dim:-1 dim: 4 } }
axis and num_axes control the portion of the bottom blob's shape that are replaced by (included in) the reshape. By default (axis == 0 and num_axes == -1), the entire bottom blob shape is included in the reshape, and hence the shape field must specify the entire output shape. axis may be non-zero to retain some portion of the beginning of the input shape (and may be negative to index from the end; e.g., -1 to begin the reshape after the last axis, including nothing in the reshape, -2 to include only the last axis, etc.). For example, suppose "input" is a 2D blob with shape 2 x 8. Then the following ReshapeLayer specifications are all equivalent, producing a blob "output" with shape 2 x 2 x 4: reshape_param { shape { dim: 2 dim: 2 dim: 4 } } reshape_param { shape { dim: 2 dim: 4 } axis: 1 } reshape_param { shape { dim: 2 dim: 4 } axis: -3 } num_axes specifies the extent of the reshape. If num_axes >= 0 (and axis >= 0), the reshape will be performed only on input axes in the range [axis, axis+num_axes]. num_axes may also be -1, the default, to include all remaining axes (starting from axis). For example, suppose "input" is a 2D blob with shape 2 x 8. Then the following ReshapeLayer specifications are equivalent, producing a blob "output" with shape 1 x 2 x 8. reshape_param { shape { dim: 1 dim: 2 dim: 8 } } reshape_param { shape { dim: 1 dim: 2 } num_axes: 1 } reshape_param { shape { dim: 1 } num_axes: 0 } On the other hand, these would produce output blob shape 2 x 1 x 8: reshape_param { shape { dim: 2 dim: 1 dim: 8 } } reshape_param { shape { dim: 1 } axis: 1 num_axes: 0 }
Used in:
The pooling method
Used in:
Used in:
Used in:
The first axis of bottom[0] (the first input Blob) along which to apply bottom[1] (the second input Blob). May be negative to index from the end (e.g., -1 for the last axis). For example, if bottom[0] is 4D with shape 100x3x40x60, the output top[0] will have the same shape, and bottom[1] may have any of the following shapes (for the given value of axis): (axis == 0 == -4) 100; 100x3; 100x3x40; 100x3x40x60 (axis == 1 == -3) 3; 3x40; 3x40x60 (axis == 2 == -2) 40; 40x60 (axis == 3 == -1) 60 Furthermore, bottom[1] may have the empty shape (regardless of the value of "axis") -- a scalar multiplier.
(num_axes is ignored unless just one bottom is given and the scale is a learned parameter of the layer. Otherwise, num_axes is determined by the number of axes by the second bottom.) The number of axes of the input (bottom[0]) covered by the scale parameter, or -1 to cover all axes of bottom[0] starting from `axis`. Set num_axes := 0, to multiply with a zero-axis Blob: a scalar.
(filler is ignored unless just one bottom is given and the scale is a learned parameter of the layer.) The initialization for the learned scale parameter. Default is the unit (1) initialization, resulting in the ScaleLayer initially performing the identity operation.
Whether to also learn a bias (equivalent to a ScaleLayer+BiasLayer, but may be more efficient). Initialized with bias_filler (defaults to 0).
Used in:
The number of group
Used in:
,Used in:
Used in:
,The axis along which to slice -- may be negative to index from the end (e.g., -1 for the last axis). By default, SliceLayer concatenates blobs along the "channels" axis (1).
DEPRECATED: alias for "axis" -- does not support negative indexing.
Message that stores parameters used by SoftmaxLayer, SoftmaxWithLossLayer
Used in:
,The axis along which to perform the softmax -- may be negative to index from the end (e.g., -1 for the last axis). Any other axes will be evaluated as independent softmaxes.
Used in:
Used in:
,Used in:
Message that stores parameters used by ThresholdLayer
Used in:
,Strictly positive values
Message that stores parameters used by TileLayer
Used in:
The index of the axis to tile.
The number of copies (tiles) of the blob to output.
DEPRECATED: use LayerParameter.
Used in:
Used in:
Used in: