Get desktop application:
View/edit binary Protocol Buffers messages
Header for the random forest model.
Next ID: 7
Used in:
Number of shards used to store the nodes.
Number of trees.
Whether the vote of individual trees are distributions or winner-take-all.
Evaluation of the model, on the out-of-bag examples, during the training.
Variable importance measures.
Container used to store the trees' nodes.
Number of nodes trained and then pruned during the training. The classical random forest learning algorithm does not prune nodes.
Used in:
Individual random seeds user to train the trees. Is specified, the number of seeds should be equal to "num_trees".
Next ID: 3
Used in:
Number of trees available in the model when evaluated.
Training configuration for the Random Forest algorithm.
Next ID: 17
Number of trees in the random forest.
Decision tree specific parameters.
Whether the vote of individual trees are distributions or winner-take-all. With winner_take_all_inference=true, each tree cast a vote for a single label value. With winner_take_all_inference=false, each tree cast a weighted vote for each label values. The original random forest implementation uses winner_take_all_inference=true (default value). However, "winner_take_all_inference=false" often leads to better results and smaller models.
Computes and report the OOB performances of the model during training. The added computing cost is relatively small.
Computes the importance of each variable (i.e. each input feature) during training. Computing the variable importance is expensive and can significantly slow down the training.
Number of time the dataset is shuffled for each tree when computing the variable importances. Increasing this number can increase significantly the training time (if "compute_oob_variable_importances:true") and increase the stability of the oob variable importance metrics.
The Out-of-bag evaluation is computed if one of the condition is true: - This is the last tree of the model. - The last OOB was computed more than "oob_evaluation_interval_in_seconds" ago. - This last OOB was computed more than "oob_evaluation_interval_in_trees" trees ago.
If true, each tree is trained on a separate dataset sampled with replacement from the original dataset. If false, all the trees are trained on the same dataset. Note: If bootstrap_training_dataset:false, OOB metrics are not available. bootstrap_training_dataset:true is the default value for Random Forest. bootstrap_training_dataset:false can be used to simulate related decision forest algorithms (e.g. "Extremely randomized trees" https://link.springer.com/content/pdf/10.1007%2Fs10994-006-6226-1.pdf).
If true, the training examples are sampled with replacement. If false, the training samples are sampled without replacement. Only used when "bootstrap_training_dataset=true". If false (sampling without replacement) and if "bootstrap_size_ratio=1" (default), all the examples are used to train all the trees (you probably do not want that).
Number of example in each bootstrap expressed as a ratio of the training dataset size.
If true, the "bootstrap_size_ratio" parameter will be adapted dynamically such that the "num_trees" will be trained in the "maximum_training_duration" time. "bootstrap_size_ratio" can only be reduced i.e. enabling this feature can only reduce the training time.
Maximum impact of the "adapt_bootstrap_size_ratio_for_maximum_training_duration" parameter.
Total maximum of nodes in the model. If specified, and if the total number of nodes is exceeded, the training stops and the forest is truncated.
If set, and if "compute_oob_performances" is true, export the out-of-bag predictions of the model on the training dataset in the file specified by "export_oob_prediction_path". Note that "export_oob_prediction_path" is a typed-path e.g. a path with a format prefix. The writer implementation of the format should be linked in the binary. For example, to export the predictions to the csv or tfrecord+tfe format, you need to make sure that dataset:csv_example_writer or dataset:tf_example_io_tfrecord are respectively linked. Example: export_oob_prediction_path = "csv:/tmp/oob_predictions.csv"
Fields used for low level and/or internal API. In most cases, the user should not care about this field.