Get desktop application:
View/edit binary Protocol Buffers messages
Configuration message for the AdamOptimizer See: https://www.tensorflow.org/api_docs/python/tf/train/AdamOptimizer
Used in:
Configuration proto for the anchor generator to use in the object detection pipeline. See core/anchor_generator.py for details.
Used in:
,Configuration proto for ArgMaxMatcher. See matchers/argmax_matcher.py for details.
Used in:
Threshold for positive matches.
Threshold for negative matches.
Whether to construct ArgMaxMatcher without thresholds.
If True then negative matches are the ones below the unmatched_threshold, whereas ignored matches are in between the matched and umatched threshold. If False, then negative matches are in between the matched and unmatched threshold, and everything lower than unmatched is ignored.
Whether to ensure each row is matched to at least one column.
Force constructed match objects to use matrix multiplication based gather instead of standard tf.gather
Configuration proto for non-max-suppression operation on a batch of detections.
Used in:
Scalar threshold for score (low scoring boxes are removed).
Scalar threshold for IOU (boxes that have high IOU overlap with previously selected boxes are removed).
Maximum number of detections to retain per class.
Maximum number of detections to retain across all classes.
Configuration proto for batch norm to apply after convolution op. See https://www.tensorflow.org/api_docs/python/tf/contrib/layers/batch_norm
Used in:
Whether to train the batch norm variables. If this is set to false during training, the current value of the batch_norm variables are used for forward pass but they are never updated.
Configuration proto for bipartite matcher. See matchers/bipartite_matcher.py for details.
Used in:
Force constructed match objects to use matrix multiplication based gather instead of standard tf.gather
Classification loss using a sigmoid function over the class prediction with the highest prediction score.
Used in:
Interpolation weight between 0 and 1.
Whether hard boot strapping should be used or not. If true, will only use one class favored by model. Othewise, will use all predicted class probabilities.
DEPRECATED, do not use. Output loss per anchor.
Configuration proto for the box coder to be used in the object detection pipeline. See core/box_coder.py for details.
Used in:
Configuration proto for box predictor. See core/box_predictor.py for details.
Used in:
,Configuration for class prediction loss function.
Used in:
,Configuration message for a constant learning rate.
Used in:
Configuration proto for Convolutional box predictor.
Used in:
Hyperparameters for convolution ops used in the box predictor.
Minumum feature depth prior to predicting box encodings and class predictions.
Maximum feature depth prior to predicting box encodings and class predictions. If max_depth is set to 0, no additional feature map will be inserted before location and class predictions.
Number of the additional conv layers before the predictor.
Whether to use dropout for class prediction.
Keep probability for dropout
Size of final convolution kernel. If the spatial resolution of the feature map is smaller than the kernel size, then the kernel size is set to min(feature_width, feature_height).
Size of the encoding for boxes.
Whether to apply sigmoid to the output of class predictions. TODO(jonathanhuang): Do we need this since we have a post processing module.?
Whether to use depthwise separable convolution for box predictor layers.
Configuration message for a cosine decaying learning rate as defined in object_detection/utils/learning_schedules.py
Used in:
Top level configuration for DetectionModels.
Used in:
Message for configuring DetectionModel evaluation jobs (eval.py).
Used in:
Number of visualization images to generate.
Number of examples to process of evaluation.
How often to run evaluation.
Maximum number of times to run evaluation. If set to 0, will run forever.
Whether the TensorFlow graph used for evaluation should be saved to disk.
Path to directory to store visualizations in. If empty, visualization images are not exported (only shown on Tensorboard).
BNS name of the TensorFlow master.
Type of metrics to use for evaluation.
Path to export detections to COCO compatible JSON format.
Option to not read groundtruth labels and only export detections to COCO-compatible JSON file.
Use exponential moving averages of variables for evaluation. TODO(rathodv): When this is false make sure the model is constructed without moving averages in restore_fn.
Whether to evaluate instance masks. Note that since there is no evaluation code currently for instance segmenation this option is unused.
Minimum score threshold for a detected object box to be visualized
Maximum number of detections to visualize
When drawing a single detection, each label is by default visualized as <label name> : <label score>. One can skip the name or/and score using the following fields:
Whether to show groundtruth boxes in addition to detected boxes in visualizations.
Box color for visualizing groundtruth boxes.
Whether to keep image identifier in filename when exported to visualization_export_dir.
Configuration message for an exponentially decaying learning rate. See https://www.tensorflow.org/versions/master/api_docs/python/train/ \ decaying_the_learning_rate#exponential_decay
Used in:
An externally defined input reader. Users may define an extension to this proto to interface their own input readers.
Used in:
(message has no fields)
Configuration for Faster R-CNN models. See meta_architectures/faster_rcnn_meta_arch.py and models/model_builder.py Naming conventions: Faster R-CNN models have two stages: a first stage region proposal network (or RPN) and a second stage box classifier. We thus use the prefixes `first_stage_` and `second_stage_` to indicate the stage to which each parameter pertains when relevant.
Used in:
Whether to construct only the Region Proposal Network (RPN).
Number of classes to predict.
Image resizer for preprocessing the input image.
Feature extractor config.
Anchor generator to compute RPN anchors.
Atrous rate for the convolution op applied to the `first_stage_features_to_crop` tensor to obtain box predictions.
Hyperparameters for the convolutional RPN box predictor.
Kernel size to use for the convolution op just prior to RPN box predictions.
Output depth for the convolution op just prior to RPN box predictions.
The batch size to use for computing the first stage objectness and location losses.
Fraction of positive examples per image for the RPN.
Non max suppression score threshold applied to first stage RPN proposals.
Non max suppression IOU threshold applied to first stage RPN proposals.
Maximum number of RPN proposals retained after first stage postprocessing.
First stage RPN localization loss weight.
First stage RPN objectness loss weight.
Output size (width and height are set to be the same) of the initial bilinear interpolation based cropping during ROI pooling.
Kernel size of the max pool op on the cropped feature map during ROI pooling.
Stride of the max pool op on the cropped feature map during ROI pooling.
Hyperparameters for the second stage box predictor. If box predictor type is set to rfcn_box_predictor, a R-FCN model is constructed, otherwise a Faster R-CNN model is constructed.
The batch size per image used for computing the classification and refined location loss of the box classifier. Note that this field is ignored if `hard_example_miner` is configured.
Fraction of positive examples to use per image for the box classifier.
Post processing to apply on the second stage box classifier predictions. Note: the `score_converter` provided to the FasterRCNNMetaArch constructor is taken from this `second_stage_post_processing` proto.
Second stage refined localization loss weight.
Second stage classification loss weight
Second stage instance mask loss weight. Note that this is only applicable when `MaskRCNNBoxPredictor` is selected for second stage and configured to predict instance masks.
If not left to default, applies hard example mining only to classification and localization loss..
Loss for second stage box classifers, supports Softmax and Sigmoid. Note that score converter must be consistent with loss type. When there are multiple labels assigned to the same boxes, recommend to use sigmoid loss and enable merge_multiple_label_boxes. If not specified, Softmax loss is used as default.
Configuration proto for FasterRCNNBoxCoder. See box_coders/faster_rcnn_box_coder.py for details.
Used in:
Scale factor for anchor encoded box center.
Scale factor for anchor encoded box height.
Scale factor for anchor encoded box width.
Used in:
Type of Faster R-CNN model (e.g., 'faster_rcnn_resnet101'; See builders/model_builder.py for expected types).
Output stride of extracted RPN feature map.
Whether to update batch norm parameters during training or not. When training with a relative large batch size (e.g. 8), it could be desirable to enable batch norm update.
Configuration proto for image resizer that resizes to a fixed shape.
Used in:
Desired height of image in pixels.
Desired width of image in pixels.
Desired method when resizing image.
Whether to also resize the image channels from 3 to 1 (RGB to grayscale).
Configuration proto for GridAnchorGenerator. See anchor_generators/grid_anchor_generator.py for details.
Used in:
Anchor height in pixels.
Anchor width in pixels.
Anchor stride in height dimension in pixels.
Anchor stride in width dimension in pixels.
Anchor height offset in pixels.
Anchor width offset in pixels.
List of scales for the anchors.
List of aspect ratios for the anchors.
Configuation for hard example miner.
Used in:
,Maximum number of hard examples to be selected per image (prior to enforcing max negative to positive ratio constraint). If set to 0, all examples obtained after NMS are considered.
Minimum intersection over union for an example to be discarded during NMS.
Maximum number of negatives to retain for each positive anchor. If num_negatives_per_positive is 0 no prespecified negative:positive ratio is enforced.
Minimum number of negative anchors to sample for a given image. Setting this to a positive number samples negatives in an image without any positive anchors and thus not bias the model towards having at least one detection per image.
Whether to use classification losses ('cls', default), localization losses ('loc') or both losses ('both'). In the case of 'both', cls_loss_weight and loc_loss_weight are used to compute weighted sum of the two losses.
Used in:
Configuration proto for the convolution op hyperparameters to use in the object detection pipeline.
Used in:
, , , , ,Regularizer for the weights of the convolution op.
Initializer for the weights of the convolution op.
BatchNorm hyperparameters. If this parameter is NOT set then BatchNorm is not applied!
Type of activation to apply after convolution.
Used in:
Use None (no activation)
Use tf.nn.relu
Use tf.nn.relu6
Operations affected by hyperparameters.
Used in:
Convolution, Separable Convolution, Convolution transpose.
Fully connected
Configuration proto for image resizing operations. See builders/image_resizer_builder.py for details.
Used in:
,Proto with one-of field for initializers.
Used in:
Used in:
Path to StringIntLabelMap pbtxt file specifying the mapping from string labels to integer ids.
Whether data should be processed in the order they are read in, or shuffled randomly.
Buffer size to be used when shuffling.
Buffer size to be used when shuffling file names.
Maximum number of records to keep in reader queue.
Minimum number of records to keep in reader queue. A large value is needed to generate a good random shuffle.
The number of times a data source is read. If set to zero, the data source will be reused indefinitely.
Number of reader instances to create.
Number of decoded records to prefetch before batching.
Number of parallel decode ops to apply.
Whether to load groundtruth instance masks.
Type of instance mask.
Instance mask format. Note that PNG masks are much more space efficient.
Used in:
Default implementation, currently NUMERICAL_MASKS
[num_masks, H, W] float32 binary masks.
Encoded PNG masks.
Configuration for intersection-over-area (IOA) similarity calculator.
Used in:
(message has no fields)
Configuration for intersection-over-union (IOU) similarity calculator.
Used in:
(message has no fields)
Configuration proto for image resizer that keeps aspect ratio.
Used in:
Desired size of the smaller image dimension in pixels.
Desired size of the larger image dimension in pixels.
Desired method when resizing image.
Whether to pad the image with zeros so the output spatial size is [max_dimension, max_dimension]. Note that the zeros are padded to the bottom and the right of the resized image.
Whether to also resize the image channels from 3 to 1 (RGB to grayscale).
Configuration proto for KeypointBoxCoder. See box_coders/keypoint_box_coder.py for details.
Used in:
Scale factor for anchor encoded box center and keypoints.
Scale factor for anchor encoded box height.
Scale factor for anchor encoded box width.
Configuration proto for L1 Regularizer. See https://www.tensorflow.org/api_docs/python/tf/contrib/layers/l1_regularizer
Used in:
Configuration proto for L2 Regularizer. See https://www.tensorflow.org/api_docs/python/tf/contrib/layers/l2_regularizer
Used in:
Configuration message for optimizer learning rate.
Used in:
, ,Configuration for bounding box localization loss function.
Used in:
Message for configuring the localization loss, classification loss and hard example miner used for training object detection models. See core/losses.py for details
Used in:
Localization loss to use.
Classification loss to use.
If not left to default, applies hard example mining.
Classification loss weight.
Localization loss weight.
Configuration message for a manually defined learning rate schedule.
Used in:
Used in:
Used in:
Hyperparameters for fully connected ops used in the box predictor.
Whether to use dropout op prior to the both box and class predictions.
Keep probability for dropout. This is only used if use_dropout is true.
Size of the encoding for the boxes.
Hyperparameters for convolution ops used in the box predictor.
Whether to predict instance masks inside detection boxes.
The depth for the first conv2d_transpose op applied to the image_features in the mask prediction branch. If set to 0, the value will be set automatically based on the number of channels in the image features and the number of classes.
Whether to predict keypoints inside detection boxes.
The height and the width of the predicted mask.
The number of convolutions applied to image_features in the mask prediction branch.
Configuration proto for the matcher to be used in the object detection pipeline. See core/matcher.py for details.
Used in:
Configuration proto for MeanStddevBoxCoder. See box_coders/mean_stddev_box_coder.py for details.
Used in:
(message has no fields)
Configuration message for the MomentumOptimizer See: https://www.tensorflow.org/api_docs/python/tf/train/MomentumOptimizer
Used in:
Configuration proto for RetinaNet anchor generator described in https://arxiv.org/abs/1708.02002. See anchor_generators/multiscale_grid_anchor_generator.py for details.
Used in:
minimum level in feature pyramid
maximum level in feature pyramid
Scale of anchor to feature stride
Aspect ratios for anchors at each grid point.
Number of intermediate scale each scale octave
Whether to produce anchors in normalized coordinates.
Configuration for negative squared distance similarity calculator.
Used in:
(message has no fields)
Normalizes pixel values in an image. For every channel in the image, moves the pixel values from the range [original_minval, original_maxval] to [target_minval, target_maxval].
Used in:
Top level optimizer message.
Used in:
Configuration proto for post-processing predicted boxes and scores.
Used in:
,Non max suppression parameters.
Score converter to use.
Scale logit (input) value before conversion in post-processing step. Typically used for softmax distillation, though can be used to scale for other reasons.
Enum to specify how to convert the detection scores.
Used in:
Input scores equals output scores.
Applies a sigmoid on input scores.
Applies a softmax on input scores
Message for defining a preprocessing operation on input data. See: //third_party/tensorflow_models/object_detection/core/preprocessor.py
Used in:
Converts the RGB image to a grayscale image. This also converts the image depth from 3 to 1, unlike RandomRGBtoGray which does not change the image depth.
Used in:
(message has no fields)
Configuration message for the RMSPropOptimizer See: https://www.tensorflow.org/api_docs/python/tf/train/RMSPropOptimizer
Used in:
Randomly changes image brightness by up to max_delta. Image outputs will be saturated between 0 and 1.
Used in:
Randomly scales contract by a value between [min_delta, max_delta].
Used in:
Randomly alters hue by a value of up to max_delta.
Used in:
Randomly changes saturation by a value between [min_delta, max_delta].
Used in:
Randomly adds black square patches to an image.
Used in:
The maximum number of black patches to add.
The probability of a black patch being added to an image.
Ratio between the dimension of the black patch to the minimum dimension of the image (patch_width = patch_height = min(image_height, image_width)).
Randomly crops the image and bounding boxes.
Used in:
Cropped image must cover at least one box by this fraction.
Aspect ratio bounds of cropped image.
Allowed area ratio of cropped image to original image.
Minimum overlap threshold of cropped boxes to keep in new image. If the ratio between a cropped bounding box and the original is less than this value, it is removed from the new image.
Probability of keeping the original image.
Randomly crops an image followed by a random pad.
Used in:
Cropping operation must cover at least one box by this fraction.
Aspect ratio bounds of image after cropping operation.
Allowed area ratio of image after cropping operation.
Minimum overlap threshold of cropped boxes to keep in new image. If the ratio between a cropped bounding box and the original is less than this value, it is removed from the new image.
Probability of keeping the original image during the crop operation.
Maximum dimensions for padded image. If unset, will use double the original image dimension as a lower bound. Both of the following fields should be length 2.
Color of the padding. If unset, will pad using average color of the input image. This field should be of length 3.
Randomly crops an iamge to a given aspect ratio.
Used in:
Aspect ratio.
Minimum overlap threshold of cropped boxes to keep in new image. If the ratio between a cropped bounding box and the original is less than this value, it is removed from the new image.
Performs a random color distortion. color_orderings should either be 0 or 1.
Used in:
Randomly horizontally flips the image and detections 50% of the time.
Used in:
Specifies a mapping from the original keypoint indices to horizontally flipped indices. This is used in the event that keypoints are specified, in which case when the image is horizontally flipped the keypoints will need to be permuted. E.g. for keypoints representing left_eye, right_eye, nose_tip, mouth, left_ear, right_ear (in that order), one might specify the keypoint_flip_permutation below: keypoint_flip_permutation: 1 keypoint_flip_permutation: 0 keypoint_flip_permutation: 2 keypoint_flip_permutation: 3 keypoint_flip_permutation: 5 keypoint_flip_permutation: 4
Randomly enlarges or shrinks image (keeping aspect ratio).
Used in:
Randomly jitters corners of boxes in the image determined by ratio. ie. If a box is [100, 200] and ratio is 0.02, the corners can move by [1, 4].
Used in:
Configuration proto for random normal initializer. See https://www.tensorflow.org/api_docs/python/tf/random_normal_initializer
Used in:
Randomly adds padding to the image.
Used in:
Minimum dimensions for padded image. If unset, will use original image dimension as a lower bound.
Maximum dimensions for padded image. If unset, will use double the original image dimension as a lower bound.
Color of the padding. If unset, will pad using average color of the input image.
Randomly scales the values of all pixels in the image by some constant value between [minval, maxval], then clip the value to a range between [0, 1.0].
Used in:
Randomly convert entire image to grey scale.
Used in:
Randomly resizes the image up to [target_height, target_width].
Used in:
Randomly rotates the image and detections by 90 degrees counter-clockwise 50% of the time.
Used in:
(message has no fields)
Randomly vertically flips the image and detections 50% of the time.
Used in:
Specifies a mapping from the original keypoint indices to vertically flipped indices. This is used in the event that keypoints are specified, in which case when the image is vertically flipped the keypoints will need to be permuted. E.g. for keypoints representing left_eye, right_eye, nose_tip, mouth, left_ear, right_ear (in that order), one might specify the keypoint_flip_permutation below: keypoint_flip_permutation: 1 keypoint_flip_permutation: 0 keypoint_flip_permutation: 2 keypoint_flip_permutation: 3 keypoint_flip_permutation: 5 keypoint_flip_permutation: 4
Configuration proto for region similarity calculators. See core/region_similarity_calculator.py for details.
Used in:
Proto with one-of field for regularizers.
Used in:
Resizes images to [new_height, new_width].
Used in:
Used in:
Enumeration type for image resizing methods provided in TensorFlow.
Used in:
,Corresponds to tf.image.ResizeMethod.BILINEAR
Corresponds to tf.image.ResizeMethod.NEAREST_NEIGHBOR
Corresponds to tf.image.ResizeMethod.BICUBIC
Corresponds to tf.image.ResizeMethod.AREA
Used in:
Hyperparameters for convolution ops used in the box predictor.
Bin sizes for RFCN crops.
Target depth to reduce the input image features to.
Size of the encoding for the boxes.
Size to resize the rfcn crops to.
Randomly crops a image according to: Liu et al., SSD: Single shot multibox detector. This preprocessing step defines multiple SSDRandomCropOperations. Only one operation (chosen at random) is actually performed on an image.
Used in:
Randomly crops a image to a fixed aspect ratio according to: Liu et al., SSD: Single shot multibox detector. Multiple SSDRandomCropFixedAspectRatioOperations are defined by this preprocessing step. Only one operation (chosen at random) is actually performed on an image.
Used in:
Aspect ratio to crop to. This value is used for all crop operations.
Used in:
Cropped image must cover at least this fraction of one original bounding box.
The area of the cropped image must be within the range of [min_area, max_area].
Cropped box area ratio must be above this threhold to be kept.
Probability a crop operation is skipped.
Used in:
Cropped image must cover at least this fraction of one original bounding box.
The aspect ratio of the cropped image must be within the range of [min_aspect_ratio, max_aspect_ratio].
The area of the cropped image must be within the range of [min_area, max_area].
Cropped box area ratio must be above this threhold to be kept.
Probability a crop operation is skipped.
Randomly crops and pads an image according to: Liu et al., SSD: Single shot multibox detector. This preprocessing step defines multiple SSDRandomCropPadOperations. Only one operation (chosen at random) is actually performed on an image.
Used in:
Randomly crops and pads an image to a fixed aspect ratio according to: Liu et al., SSD: Single shot multibox detector. Multiple SSDRandomCropPadFixedAspectRatioOperations are defined by this preprocessing step. Only one operation (chosen at random) is actually performed on an image.
Used in:
Aspect ratio to pad to. This value is used for all crop and pad operations.
Min ratio of padded image height and width to the input image's height and width. Two entries per operation.
Max ratio of padded image height and width to the input image's height and width. Two entries per operation.
Used in:
Cropped image must cover at least this fraction of one original bounding box.
The aspect ratio of the cropped image must be within the range of [min_aspect_ratio, max_aspect_ratio].
The area of the cropped image must be within the range of [min_area, max_area].
Cropped box area ratio must be above this threhold to be kept.
Probability a crop operation is skipped.
Used in:
Cropped image must cover at least this fraction of one original bounding box.
The aspect ratio of the cropped image must be within the range of [min_aspect_ratio, max_aspect_ratio].
The area of the cropped image must be within the range of [min_area, max_area].
Cropped box area ratio must be above this threhold to be kept.
Probability a crop operation is skipped.
Min ratio of padded image height and width to the input image's height and width. Two entries per operation.
Max ratio of padded image height and width to the input image's height and width. Two entries per operation.
Padding color.
Scales boxes from normalized coordinates to pixel coordinates.
Used in:
(message has no fields)
Sigmoid Focal cross entropy loss as described in https://arxiv.org/abs/1708.02002
Used in:
DEPRECATED, do not use.
modulating factor for the loss.
alpha weighting factor for the loss.
Configuration proto for SquareBoxCoder. See box_coders/square_box_coder.py for details.
Used in:
Scale factor for anchor encoded box center.
Scale factor for anchor encoded box length.
Configuration for Single Shot Detection (SSD) models.
Used in:
Number of classes to predict.
Image resizer for preprocessing the input image.
Feature extractor config.
Box coder to encode the boxes.
Matcher to match groundtruth with anchors.
Region similarity calculator to compute similarity of boxes.
Whether background targets are to be encoded as an all zeros vector or a one-hot vector (where background is the 0th class).
classification weight to be associated to negative anchors (default: 1.0). The weight must be in [0., 1.].
Box predictor to attach to the features.
Anchor generator to compute anchors.
Post processing to apply on the predictions.
Whether to normalize the loss by number of groundtruth boxes that match to the anchors.
Whether to normalize the localization loss by the code size of the box encodings. This is applied along with other normalization factors.
Loss configuration for training.
Configuration proto for SSD anchor generator described in https://arxiv.org/abs/1512.02325. See anchor_generators/multiple_grid_anchor_generator.py for details.
Used in:
Number of grid layers to create anchors for.
Scale of anchors corresponding to finest resolution.
Scale of anchors corresponding to coarsest resolution
Can be used to override min_scale->max_scale, with an explicitly defined set of scales. If empty, then min_scale->max_scale is used.
Aspect ratios for anchors at each grid point.
When this aspect ratio is greater than 0, then an additional anchor, with an interpolated scale is added with this aspect ratio.
Whether to use the following aspect ratio and scale combination for the layer with the finest resolution : (scale=0.1, aspect_ratio=1.0), (scale=min_scale, aspect_ration=2.0), (scale=min_scale, aspect_ratio=0.5).
The base anchor size in height dimension.
The base anchor size in width dimension.
Anchor stride in height dimension in pixels for each layer. The length of this field is expected to be equal to the value of num_layers.
Anchor stride in width dimension in pixels for each layer. The length of this field is expected to be equal to the value of num_layers.
Anchor height offset in pixels for each layer. The length of this field is expected to be equal to the value of num_layers.
Anchor width offset in pixels for each layer. The length of this field is expected to be equal to the value of num_layers.
Used in:
Type of ssd feature extractor.
The factor to alter the depth of the channels in the feature extractor.
Minimum number of the channels in the feature extractor.
Hyperparameters for the feature extractor.
The nearest multiple to zero-pad the input height and width dimensions to. For example, if pad_to_multiple = 2, input dimensions are zero-padded until the resulting dimensions are even.
Whether to update batch norm parameters during training or not. When training with a relative small batch size (e.g. 1), it is desirable to disable batch norm update and use pretrained batch norm params. Note: Some feature extractors are used with canned arg_scopes (e.g resnet arg scopes). In these cases training behavior of batch norm variables may depend on both values of `batch_norm_trainable` and `is_training`. When canned arg_scopes are used with feature extractors `conv_hyperparams` will apply only to the additional layers that are added and are outside the canned arg_scope.
Whether to use explicit padding when extracting SSD multiresolution features. Note that this does not apply to the base feature extractor.
Whether to use depthwise separable convolutions for to extract additional feature maps added by SSD.
Used in:
String name. The most common practice is to set this to a MID or synsets id.
Integer id that maps to the string name above. Label ids should start from 1.
Human readable string label.
Normalizes an image by subtracting a mean from each channel.
Used in:
The mean to subtract from each channel. Should be of same dimension of channels in the input image.
An input reader that reads TF Example protos from local TFRecord files.
Used in:
Path(s) to `TFRecordFile`s.
Message for configuring DetectionModel training jobs (train.py).
Used in:
Input queue batch size.
Data augmentation options.
Whether to synchronize replicas during training.
How frequently to keep checkpoints.
Optimizer used to train the DetectionModel.
If greater than 0, clips gradients by this value.
Checkpoint to restore variables from. Typically used to load feature extractor variables trained outside of object detection.
Specifies if the finetune checkpoint is from an object detection model. If from an object detection model, the model being trained should have the same parameters with the exception of the num_classes parameter. If false, it assumes the checkpoint was a object classification model.
Whether to load all checkpoint vars that match model variable names and sizes. This option is only available if `from_detection_checkpoint` is True.
Number of steps to train the DetectionModel for. If 0, will train the model indefinitely.
Number of training steps between replica startup. This flag must be set to 0 if sync_replicas is set to true.
If greater than 0, multiplies the gradient of bias variables by this amount.
Variables that should not be updated during training.
Number of replicas to aggregate before making parameter updates.
Maximum number of elements to store within a queue.
Number of threads to use for batching.
Maximum capacity of the queue used to prefetch assembled batches.
If true, boxes with the same coordinates will be merged together. This is useful when each box can have multiple labels. Note that only Sigmoid classification losses should be used.
Whether to add regularization loss to `total_loss`. This is true by default and adds all regularization losses defined in the model to `total_loss`. Setting this option to false is very useful while debugging the model and losses.
Maximum number of boxes used during training. Set this to at least the maximum amount of boxes in the input data. Otherwise, it may cause "Data loss: Attempted to pad to a smaller size than the input element" errors.
Whether to remove padding along `num_boxes` dimension of the groundtruth tensors.
Convenience message for configuring a training and eval pipeline. Allows all of the pipeline parameters to be configured from one file.
Configuration proto for truncated normal initializer. See https://www.tensorflow.org/api_docs/python/tf/truncated_normal_initializer
Used in:
Configuration proto for variance scaling initializer. See https://www.tensorflow.org/api_docs/python/tf/contrib/layers/ variance_scaling_initializer
Used in:
Used in:
Configuration proto for weight shared convolutional box predictor.
Used in:
Hyperparameters for convolution ops used in the box predictor.
Number of the additional conv layers before the predictor.
Output depth for the convolution ops prior to predicting box encodings and class predictions.
Size of final convolution kernel. If the spatial resolution of the feature map is smaller than the kernel size, then the kernel size is set to min(feature_width, feature_height).
Size of the encoding for boxes.
Bias initialization for class prediction. It has been show to stabilize training where there are large number of negative boxes. See https://arxiv.org/abs/1708.02002 for details.
Intersection over union location loss: 1 - IOU
Used in:
(message has no fields)
L2 location loss: 0.5 * ||weight * (a - b)|| ^ 2
Used in:
DEPRECATED, do not use. Output loss per anchor.
Classification loss using a sigmoid function over class predictions.
Used in:
DEPRECATED, do not use. Output loss per anchor.
SmoothL1 (Huber) location loss. The smooth L1_loss is defined elementwise as .5 x^2 if |x| <= delta and 0.5 x^2 + delta * (|x|-delta) otherwise, where x is the difference between predictions and target.
Used in:
DEPRECATED, do not use. Output loss per anchor.
Delta value for huber loss.
Classification loss using a softmax function over class predictions.
Used in:
DEPRECATED, do not use. Output loss per anchor.
Scale logit (input) value before calculating softmax classification loss. Typically used for softmax distillation.