Proto commits in hughperkins/tf-coriander

These commits are when the Protocol Buffers files have changed: (only the last 100 relevant commits are shown)

Commit:421de4a
Author:Allen Lavoie
Committer:GitHub

Note that RewriterConfig is not subject to API guarantees

Commit:deed76c
Author:David Norman
Committer:Rasmus Munk Larsen

[XLA] Add F16 support to the Literal protobuf and LiteralUtils class. (#9913) * Add F16 support to the Literal protobuf and LiteralUtils class. No support has been added to any public backend, however the unit tests demonstrate that the literals can store and retreive data correctly. * Changes after code review * Use an alternative form of local initialization * Change a few more C-style casts to C++ casts

Commit:365b5c1
Author:Eugene Brevdo
Committer:TensorFlower Gardener

Bugfix for proto file changes introduced in recent while_loop/stop_gradient PR. PiperOrigin-RevId: 155886984

Commit:a5b9ef0
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add timeline support in tfprof This CL mainly adds timeline support in three views of tfprof. It includes a few other small changes: 1. Handle the case that one Op fires multiple kernels. 2. Remove the requirements for CostGraph for easier user adoption, for now. 3. Some speed improvements in graph view. 4. Consolidate the all kinds of tfprof output into one -output option. PiperOrigin-RevId: 155822542

Commit:98ccdcd
Author:Benoit Steiner
Committer:TensorFlower Gardener

Automated g4 rollback of changelist 155779520 PiperOrigin-RevId: 155811666

Commit:b37da7a
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add timeline support in tfprof This CL mainly adds timeline support in three views of tfprof. It includes a few other small changes: 1. Handle the case that one Op fires multiple kernels. 2. Remove the requirements for CostGraph for easier user adoption, for now. 3. Some speed improvements in graph view. 4. Consolidate the all kinds of tfprof output into one -output option. PiperOrigin-RevId: 155779520

Commit:96980d5
Author:Yuan Yu
Committer:TensorFlower Gardener

Fix a bug when applying stop_gradient() on the gradient of a while loop. It turned out that I really needed to have tf.gradients() to backprop correctly on the gradient loop even if we have a stop_gradient on it. So, with this CL, I suspect that we are epsilon away to support higher-order gradients for while loops. PiperOrigin-RevId: 155568884

Commit:697f34c
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Extend tfprof to associate op stats with Python codes. It's backward compatible. Stats of a source code line are aggregated from all ops created by that line. A example. _TFProfRoot (0us/22.44ms) model_analyzer_test.py:149:run_filename_as_m...:none (0us/22.44ms) model_analyzer_test.py:33:_run_code_in_main:none (0us/22.44ms) model_analyzer_test.py:208:<module>:test.main() (0us/22.44ms) model_analyzer_test.py:132:testComplexCodeView:x = lib.BuildFull... (0us/22.44ms) model_analyzer_testlib.py:63:BuildFullModel:return sgd_op.min... (0us/21.83ms) model_analyzer_testlib.py:54:BuildFullModel:seq.append(array_... (0us/254us) model_analyzer_testlib.py:42:BuildSmallModel:x = nn_ops.conv2d... (0us/134us) ... model_analyzer_testlib.py:61:BuildFullModel:loss = nn_ops.l2_... (0us/28us) model_analyzer_test.py:134:testComplexCodeView:sess.run(variable... (0us/0us) Change: 155393864

Commit:7cac7f2
Author:Yao Zhang
Committer:TensorFlower Gardener

Specify a different outer class name for the generation of java output. Change: 155296402

Commit:b96fa4f
Author:Anna R
Committer:TensorFlower Gardener

Automated rollback of change 155258346 Change: 155292907

Commit:87ba9f5
Author:Benoit Steiner
Committer:TensorFlower Gardener

Return more complete device information from the GetDevices() method of grappler clusters. Change: 155264843

Commit:1e59f00
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Extend tfprof to associate op stats with Python codes. It's backward compatible. Stats of a source code line are aggregated from all ops created by that line. A example. _TFProfRoot (0us/22.44ms) model_analyzer_test.py:149:run_filename_as_m...:none (0us/22.44ms) model_analyzer_test.py:33:_run_code_in_main:none (0us/22.44ms) model_analyzer_test.py:208:<module>:test.main() (0us/22.44ms) model_analyzer_test.py:132:testComplexCodeView:x = lib.BuildFull... (0us/22.44ms) model_analyzer_testlib.py:63:BuildFullModel:return sgd_op.min... (0us/21.83ms) model_analyzer_testlib.py:54:BuildFullModel:seq.append(array_... (0us/254us) model_analyzer_testlib.py:42:BuildSmallModel:x = nn_ops.conv2d... (0us/134us) ... model_analyzer_testlib.py:61:BuildFullModel:loss = nn_ops.l2_... (0us/28us) model_analyzer_test.py:134:testComplexCodeView:sess.run(variable... (0us/0us) Change: 155258346

Commit:692fad2
Author:Dan Ringwalt
Committer:TensorFlower Gardener

Merge changes from github. Change: 155209832

Commit:15b8f3d
Author:Vijay Vasudevan

Merge commit for internal changes

Commit:f28935a
Author:Brennan Saeta
Committer:TensorFlower Gardener

Implement ClusterSpec Propagation in TF Master ClusterSpec propagation is a capability upgrade for TensorFlow that should make it much easier to (1) build distributed TensorFlow clusters, and (2) handle node failures. The ClusterSpec propagation capability allows TensorFlow workers to be booted independently of each other, and with no knowledge about others. The client can then construct a ClusterDef (ClusterSpec), and then send it to the TF master at session creation. The master in turn then propagates the ClusterDef along to all of the workers. Change: 155159972

Commit:99edb91
Author:Li Chen
Committer:Vijay Vasudevan

Add ibverbs support for RoCEv2 networks (#9432) * add gid to enable running on roce * fix typo * add gid definition in rdma.h * fix typo * gid changed to snp and iid * typo * increase hop_limit to its upperbound

Commit:a25509e
Author:Shanqing Cai
Committer:TensorFlower Gardener

Add TFDBG support to GrpcSession * Along the way, unify the way the debugger works in DirectSession (non-distributed Sessions) and MasterSession (for distributed Sessions). * The SummarizDebugTensorWatches method is invoked in DirectSession::GetOrCreateExecutors() and MasterSession::HashBuildGraphOptions() method to generate keys for partition graphs and executors. * The DebugStateInterface::PublishDebugMetadata() method is used to send metadata about the debugged Session::Run() call to debug URLs. This happens in DirectSession::Run() and MasterSession::DoRunWithLocalExecution() respectively. * The DebugGraphDecoratorInterface::DecorateGraph() and DebugGraphDecoratorInterface::PublishGraph() methods are used to insert debug ops to the debugged graph and send the modified graph to debug URLs. This happens in DirectSession::GetOrCreateExecutors() and GraphMgr::InitItem(), respectively. Change: 154631802

Commit:39f51e6
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Store the number of trees and layers that were attempted to be built. Useful to terminate training. Also can be used as seeds for random operations. Change: 154472242

Commit:3269423
Author:Shanqing Cai
Committer:TensorFlower Gardener

Merge changes from github. Change: 153925676

Commit:a2be7b1
Author:Shanqing Cai

Merge commit for internal changes

Commit:225f4e3
Author:Peter Hawkins
Committer:TensorFlower Gardener

[XLA] Change types of InfeedRequest and OutfeedRequest configuration fields to "bytes" instead of "string". Change: 153665746

Commit:dd40e98
Author:Jun Shi
Committer:drpngx

Ibverbs-based RDMA path (#8943) * initial check-in * modifications according to review feedback * fix BUILD file * Update BUILD Add missing comma in verbs/BUILD file * Buildifier fix * Buildifier fix #2 * added #define TENSORFLOW_USE_VERBS * fix defines() * buildifier fix * link option fix * fix TENSORFLOW_USE_VERBS

Commit:94a2da3
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add a GPUOPTIONS option to force all tensors to be gpu_compatible Change: 153386455

Commit:a880533
Author:Yuefeng Zhou
Committer:TensorFlower Gardener

Fold the content a const op into its consumer node. Change: 153204827

Commit:c11ea29
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add serving code for set membership split type. Change: 153101982

Commit:7d4aadd
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Automated rollback of change 153039058 Change: 153045639

Commit:b439663
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add a GPUOPTIONS option to force all tensors to be gpu_compatible Change: 153039058

Commit:1bd5769
Author:Jeffrey A. Dean
Committer:TensorFlower Gardener

Minor internal change. Change: 152949383

Commit:181823f
Author:Andrew Harp
Committer:TensorFlower Gardener

Add run_mode field to TestResults proto for differentiating between test types. Change: 152740961

Commit:ecb5266
Author:Shanqing Cai
Committer:TensorFlower Gardener

tfdbg core: allow gRPC debug server to remotely disable/enable debug ops Synopsis of changes: * The EventReply protobuf is expanded: a new field called "debug_op_state_change" is added to allow the debug server to remotely enable and disable debug ops. * At the end of every debug gRPC stream, the server sends all the queued EventReply protos to the client. The client (i.e., the debugged TF runtime) receives them and toggle the enabled status of the debug ops accordingly. * Added gated_grpc attribute to existing debug ops. This new boolean attribute is set to False by default, ensuring backward compatibility in behavior. If set to True, the debug ops will send the output tensors through grpc:// streams if and only if they are currently enabled. Otherwise we say that the debug op is "gated off" at the grpc:// URL. * If a debug op is gated off at all URLs it possesses, it will perform no expensive computation and instead just emit an empty (size {0}) output tensor. Other detailed change descriptions: * All debug ops now share the same base class "BaseDebugOp" to reduce the amount of boilerplate, which has grown in size due to the new gRPC gating logic. Change: 152733779

Commit:9e7bf40
Author:drpngx
Committer:GitHub

Branch 152703253 (#9112) * Improve py_func error handling. Automatically translate some python errors into corresponding TF errors at runtime. Change: 152156821 * Update interaction with libpng so that we use the public API instead of knowledge of the internal libpng data structures. Change: 152167754 * TensorBoard plugins now contain their own name/route prefix. Change: 152167807 * Passes trainable flag to separable_conv2d biases. Change: 152170239 * Saving resource variables with a caching device. Change: 152171539 * Drop loss from estimator_spec.eval_metric_ops, as required by core Estimator. Change: 152179924 * sample_stats.percentile DOCFIX. Change: 152182295 * Added a memory optimizer to grappler. Change: 152184170 * Change default behavior of the tf runs selector: - If there are fewer than 41 runs, enable them all by default - If there are 41 runs or more, disable them all by default This is in response to user complaints that having it enable only the first ten runs by default was confusing, because it was not obvious to users that some runs had been disabled. However, it still solves the initial user complaint that having very many runs simultaneously enabled would lag the UI. I also changed the "toggle all runs" button to try to turn everything off before turning everything on. Also, I improved the logic for detecting when the runs selection is back in the default state, so that we can avoid generating long URI strings wherever possible. Change: 152188948 * Autogenerated Change: Change TensorBoard TAG to 52 Change: 152189000 * Remove warning that only happening with config cuda. Change: 152189205 * Make resource variable shared name consistent with non-resource variables. Remove colocation constraint from resource variable cached value with the variable itself. Change: 152192203 * Add a way to specify the optimization order; refactor and add constant folding to meta optimizer. Change: 152193646 * Backport fixes and improvements from external Keras. Change: 152198296 * Merge changes from github. Change: 152200430 * Go: Update generated wrapper functions for TensorFlow ops. Change: 152200754 * Update ops-related pbtxt files. Change: 152203174 * Make ImportGraphDef() work with functions. In addition to modify graph_constructor.cc, this patch adds some other functionality to enable importing fucntions: * Ability to add FunctionDefLibraries to Graphs and FunctionLibraryDefinitions (in addition to existing functions) * FunctionDefsEqual() utility function Change: 152205258 * Expand contrib test to more than just test targets. Change: 152206822 * Preserve graph version during optimization Change: 152213262 * Exclude enter and exit nodes from shape refiner's constant folding. Change: 152213637 * Allow reshape_mover and algebraic_simplifier to make multiple mutations, by avoiding the short-circuit std::any_of. Change: 152232810 * Fix dynamic_rnn transpose bug (can input/output non-3d tensors). Also a few cleanups to RNN code. Change: 152267628 * Fix flaky tests Change: 152272801 * Add an auto parallelization grappler optimization pass. Change: 152276787 * Change json.decode.JSONDecodeError to ValueError. JSONDecodeError seems to be the exception used in the simplejson module, not the json module. Change: 152278012 * Internal change. Change: 152281471 * [XLA] Force buffer sharing of separate while instructions. Change: 152288540 * replica_device_setter should work for resource variables Change: 152289915 * Fix ./configure script 1. Add %workspace% in .bazelrc file when using import statement 2. Write action_env into bazelrc file for required environment variables for OpenCL support Change: 152290700 * Pointing a number of Tensorboard graph visualization-related help links to the new locations for the correspondent API documentation. Change: 152293459 * Restore most of pull request #8606 Pull request #8606 added str(Label(...)) for most dependencies in tensorflow.bzl, allowing most functions to be used from repositories which include TensorFlow as a submodule. Unfortunately, it broke when pulled into Google and was removed in cl/152200430. This CL restores the change, except for two Android-only functions; these were the only problematic bits. Change: 152297413 * Removed dead code in Estimator. Change: 152297597 * Assert rank is at least equal to new_rank for `_sparse_inner_flatten`. Change: 152303319 * Extend quantization ranges to include 0.0f. Change: 152304380 * Remove Keras config file saving. Change: 152306552 * API backwards compatibility tests. Change: 152310869 * [TF:XLA] Add a test for an R3 -> R4 broadcast. Change: 152313967 * Fix the problem that no enough placeholders for persistent tensor batch delete The deleter_key is always a device_name, hence there is only one of it. Hence, we cannot delete >1 handles at one time. In the fix, it creates delete placeholder on demand, the max number of placeholders is _DEAD_HANDLES_THRESHOLD. Change: 152322770 * [XLA] Add several reduction tests. Change: 152323510 * Added the memory optimizer to the meta optimizer. Change: 152323689 * Started a set of utilities to categorize op types Change: 152329057 * Add AudioSpectrogram op to TensorFlow for audio feature generation Change: 152332221 * Update ops-related pbtxt files. Change: 152332812 * Automated rollback of change 152332221 Change: 152333917 * Call Py_CLEAR on dead fields during TF_RESOURCE-to-ndarray conversion Change: 152338333 * [TF contrib seq2seq] Initial, incomplete implementation of beam search decoder. **DOES NOT WORK, pushed for collaboration only** Change: 152343927 * [XLA] Change HloPassPipeline to disallow Add* calls after Run. Change: 152345578 * Automated rollback of change 152332812 Change: 152349057 * Remove all 64/32 bit compiler warnings from core/ops. Change: 152353506 * libtensorflow.so: Don't export private symbols. With this change, libtensorflow.so will only export functions defined in c_api.h. This also results in a decreased binary size of libtensorflow.so. On Linux the decrease was from roughly 150MB to 67MB. On OS X it was from roughly 101MB to 82MB. Also fixes #8923 Change: 152366053 * Add Elu ops in XLA. Change: 152383201 * Fixed test. ('broadcast_dims' has size 1) Change: 152383633 * Add more detailed error message for rank assertion in _sparse_inner_flatten. Change: 152397909 * tensor_bundle: propagrates errors related to directory creation. Change: 152401909 * matrix_adjoint added to contrib/linalg/linear_operator_util Change: 152404828 * Add an is_active method to plugins This method determines whether a plugin is active. A plugin may be inactive if say it lacks data. This new is_active method allows us to add a route to TensorBoard noting which plugins are active. The frontend could then avoid querying routes of inactive plugins. Change: 152406232 * Replace a gather op for shapes by a stack op so dilated convolutions can be placed on GPU even with strict placing (before the gather went to CPU). Change: 152411159 * [TF:XLA] Implement BatchToSpace, BatchToSpaceND, SpaceToBatch, SpaceToBatchND. Fix crashes in core implementations of the same operators for zero-sized blocks. Change: 152416903 * Estimator saves relative paths in checkpoint. Change: 152420211 * Fix layers_test exception regex matching. Change: 152422855 * Unhide bijectors. Correct TransformedDistribution docstring. Change: 152424418 * Choosing a saner default for min_eval_frequency in the constructor for Experiment for the GCS file system, because the default of 1 causes performance problems. Change: 152439984 * Inherit use_resource from scope for partitioned variables. Change: 152442103 * Support quantized reshape in hexagon runtime Change: 152445539 * tfdbg CLI: add command list_source (ls) + UI fixes and improvements The new list_source (shorthand: ls) command lists Python source files responsible for constructing the nodes and tensors encountered in the run() call. It divides the source files into two categories and list them separately. 1) files that are not part of the TensorFlow Python library, and 2) files that are a part of it. The list contains information about how many nodes, tensors and dumps of tensors the files is responsible for. The file paths contain clickable links to the existing print_source/ps command. The list_source/ls command supports filtering by file-path and node-name regex patterns. UI fixes: * Fixed inconsistent black vs. transparent background color that made the layout look messy on some terminal types. Now using the transparent color for default font color consistently. * In the print_source command output, add clickable links to expand source lines and graph elements. Change: 152446002 * tfcompile: Be a little more verbose about missing required flags. Fixes #9014 Change: 152446338 * Disable failing test cases in pooling_ops_test. Change: 152447322 * Register more types for tf.image_crop_and_resize(). Resolves #9020. Change: 152448160 * Automated rollback of change 152439984 Change: 152450929 * Add a route to TensorBoard for fetching plugin names Specifically, we add a /data/plugins_listing route to the TensorBoard application. This route responds with an object mapping the name of each initialized plugin to whether it is active. This route could help the frontend avoid issuing requests to inactive plugins. Ordered the listing of routes within application.py so there is a little more organization. Refactored the test for application to use a fake plugin. Change: 152451390 * Added the ability to retrieve the amount of usable gpu memory Change: 152453470 * Allow to set session ConfigProto in RunConfig and use it in Estimator. Change: 152454548 * Colocate ResourceVariable reads with their handles. Change: 152455939 * tfdbg: update doc for new command list_source/ls Change: 152456128 * Make rnn directions slightly easier to follow. Change: 152456296 * Internal change Change: 152458104 * Adds batch renormalization. NOTE: if you use renormalization, you might want to use faster moving average updates, i.e. lower `decay` values. Change: 152458872 * When using ImportGraphDef with a passed in ShapeRefiner, use the producer version of the GraphDef when importing; the ShapeRefiner may be initialized with a different graph_def_version, so we need to be able to override it. The test failed without the change to graph_constructor and passes with it. The test uses a legacy graph that is supported (reduction shape). Change: 152459169 * Allow any iterable for `export_strategies` arg. Change: 152461826 * Log steps/sec every 100 steps in MonitoredSession, as before. Change: 152465320 * Fixes documentation to note that the in case of ties the identity of the return value of ArgMin and ArgMaxis not guaranteed . Change: 152465346 * Automated rollback of change 152465346 Change: 152465844 * Fix shape inference fn on _ParallelConcatStart. Change: 152466076 * Fix getting started guide Explain numerical differences in loss fix one example to print Change: 152466119 * Remove superfluous mode argument. Change: 152467334 * Add a tool that converts HLO computations to tensorflow GraphDef which can be visualized on Tensorboard. This CL defines basic tensorflow::OpDef for each HLO instruction/node. More attributes (e.g. shapes, colors) will be added in the future. Change: 152477918 * [TF:XLA] Increase shard count of //third_party/tensorflow/compiler/tests:spacetobatch_test to reduce flakiness when built under ASAN. Change: 152496244 * Make projector plugin backend read assets saved via the PluginAssets API. At the same time, keep backwards compatibility with the old way of looking up assets. Change: 152504793 * Move MNIST pointers to mirror hosted by the CVDF on Google Cloud. Fixes: #9031 Change: 152504901 * Merge changes from github. Change: 152508170 * Update API after changing default step couter frequency before. Change: 152517535 * Move a few random op helper functions to header files 1. shape_inference::RandomShape 2. OpKernel::MakeShape(Tensor, TensorShape*) Change: 152522156 * addresses the divide by zero bug Change: 152522488 * Clarify doc on tf.assign. Change: 152523909 * Sparse adam for resource variables. Change: 152525327 * Automated rollback of change 152310869 Change: 152528732 * Add an env_var tf_sync_on_finish_bool that block until device has finished all queued operations in a step if true. Change: 152533676 * Add more node attributes for HloInstruction on Tensorboard e.g. shape and layout etc. Change: 152534472 * Add tf.complex64 GPU support to tf.gather. Also add ldg specializations for std::complex. Change: 152537848 * Formatting changes Change: 152544842 * Upgrade TensorBoard TypeScript to 2.2.1 See also: #8326 Change: 152545950 * TEST: Getting reasonable test sizes on linalg library, removing need for sharding. Change: 152546409 * Disabling _testSourceUtilModuleReturnsTrue as its causing opensource issues. Change: 152548721 * Fix race due to unsafe buffer forwarding in maxpooling second order gradients added in #6664. Re-enable previously flaky tests. Clean up a few minor things in maxpooling_op_gpu.cu.cc Change: 152550050 * LinearOperator: adjoint_arg kwarg added to all operators. Now, operator.apply(x, adjoint_arg=True) means that the adjoint of 'x' is taken before application of operator. Sometimes this is done more efficiently than simply taking adjoint. Change: 152560471 * Adds weighted_average_loss metric key. Change: 152560999 * Documentation: Fix bug in manual device placement example Change: 152563392 * Change for internal compatibility. * Use std::vector for storage instead of map. Do the sorting inplace and return the same vector to avoid any copies. On larger streams it is about 50% faster. Change: 152576112 * Add tf.add_n GPU support for complex64/complex128. Also adds a unit test for tf.add_n. Change: 152577190 * - Adds support for nested types in tf.case and tf.cond. - Adds a "strict" mode which disables silent unpacking of singleton lists. - Adds shape inference to tf.case. - Adds a lot of unit tests. Change: 152581097 * [XLA] Add support for folding transpose into convolution Change: 152581336 * Add a smoke test to ensure that the doc generator runs. Change: 152592164 * Add tensorboard to the _do_not_descend_map of the PublicAPIVisitor. Change: 152592268 * Add auto parallelization to meta optimizer. Enable MetaOptimizer if any one of the optimizers is on. Change: 152598517 * Update ops-related pbtxt files. Change: 152629248 * Prevent the renorm_weight from being updated too early. Change: 152631776 * Automated rollback of change 152528732 Change: 152652473 * Construct TensorBoard dashboards in a JS list Previously, adding a dashboard to TensorBoard involved changing logic in several places. As part of this effort, added constructors to dashboards. Tweaked logic in various dashboards to preserve original behavior. For instance, the graph dashboard can only perform fitting after the dashboard is attached to the DOM. Change: 152658532 * Make CheckpointSaverListener visible next to CheckpointSaverHook. Change: 152662945 * tfdbg CLI: minor bug fixes 1: The calculation of the scroll command in the scroll bar didn't take into account that the y-coordinate of the scroll block is in the ScrollBar coordinate system, while the mouse click y-coordinate is in the screen coordinate system. 2: The y position of the ScrollBar was off by one. 3: The command box is not re-created after mouse-triggered commands, leading to strange-looking cursor position. Change: 152684294 * Remove obsolete use of validate_indices from embedding_ops.py validate_indices is ignored, so it shouldn't appear in new code. Change: 152691948 * Preparation of using GMock matchers in XLA tests. Change: 152691970 * Replace RuntimeException by RuntimeError in coordinator documentation. Change: 152697758 * Move the TensorBoard debugger plugin to be internal. This feature is currently not open-source anyway. Change: 152700267 * Add a single-machine tf.learn Estimator implementation for the WALS solver. Change: 152700915 * Add tf.contrib.training.python_input -- making it easy to feed data into TensorFlow from python coroutines. Change: 152701623 * Show that QuantizeToFloat consistently introduces a small error. The error is equal to range_min - round(range_min / range_scale) * range_scale Change: 152702015 * Internal Changes Change: 152703253 * Remove tensorflow/tensorboard/plugins/debugger, as part of merge resolution.

Commit:cc97ef9
Author:Gunhan Gulsoy
Committer:TensorFlower Gardener

Automated rollback of change 152528732 Change: 152652473

Commit:df7f248
Author:Yao Zhang
Committer:TensorFlower Gardener

Add auto parallelization to meta optimizer. Enable MetaOptimizer if any one of the optimizers is on. Change: 152598517

Commit:52dcb25
Author:Rohan Jain
Committer:GitHub

Branch 152550050 (#9059) * Improve py_func error handling. Automatically translate some python errors into corresponding TF errors at runtime. Change: 152156821 * Update interaction with libpng so that we use the public API instead of knowledge of the internal libpng data structures. Change: 152167754 * TensorBoard plugins now contain their own name/route prefix. Change: 152167807 * Passes trainable flag to separable_conv2d biases. Change: 152170239 * Saving resource variables with a caching device. Change: 152171539 * Drop loss from estimator_spec.eval_metric_ops, as required by core Estimator. Change: 152179924 * sample_stats.percentile DOCFIX. Change: 152182295 * Added a memory optimizer to grappler. Change: 152184170 * Change default behavior of the tf runs selector: - If there are fewer than 41 runs, enable them all by default - If there are 41 runs or more, disable them all by default This is in response to user complaints that having it enable only the first ten runs by default was confusing, because it was not obvious to users that some runs had been disabled. However, it still solves the initial user complaint that having very many runs simultaneously enabled would lag the UI. I also changed the "toggle all runs" button to try to turn everything off before turning everything on. Also, I improved the logic for detecting when the runs selection is back in the default state, so that we can avoid generating long URI strings wherever possible. Change: 152188948 * Autogenerated Change: Change TensorBoard TAG to 52 Change: 152189000 * Remove warning that only happening with config cuda. Change: 152189205 * Make resource variable shared name consistent with non-resource variables. Remove colocation constraint from resource variable cached value with the variable itself. Change: 152192203 * Add a way to specify the optimization order; refactor and add constant folding to meta optimizer. Change: 152193646 * Backport fixes and improvements from external Keras. Change: 152198296 * Merge changes from github. Change: 152200430 * Go: Update generated wrapper functions for TensorFlow ops. Change: 152200754 * Update ops-related pbtxt files. Change: 152203174 * Make ImportGraphDef() work with functions. In addition to modify graph_constructor.cc, this patch adds some other functionality to enable importing fucntions: * Ability to add FunctionDefLibraries to Graphs and FunctionLibraryDefinitions (in addition to existing functions) * FunctionDefsEqual() utility function Change: 152205258 * Expand contrib test to more than just test targets. Change: 152206822 * Preserve graph version during optimization Change: 152213262 * Exclude enter and exit nodes from shape refiner's constant folding. Change: 152213637 * Allow reshape_mover and algebraic_simplifier to make multiple mutations, by avoiding the short-circuit std::any_of. Change: 152232810 * Fix dynamic_rnn transpose bug (can input/output non-3d tensors). Also a few cleanups to RNN code. Change: 152267628 * Fix flaky tests Change: 152272801 * Add an auto parallelization grappler optimization pass. Change: 152276787 * Change json.decode.JSONDecodeError to ValueError. JSONDecodeError seems to be the exception used in the simplejson module, not the json module. Change: 152278012 * Internal change. Change: 152281471 * [XLA] Force buffer sharing of separate while instructions. Change: 152288540 * replica_device_setter should work for resource variables Change: 152289915 * Fix ./configure script 1. Add %workspace% in .bazelrc file when using import statement 2. Write action_env into bazelrc file for required environment variables for OpenCL support Change: 152290700 * Pointing a number of Tensorboard graph visualization-related help links to the new locations for the correspondent API documentation. Change: 152293459 * Restore most of pull request #8606 Pull request #8606 added str(Label(...)) for most dependencies in tensorflow.bzl, allowing most functions to be used from repositories which include TensorFlow as a submodule. Unfortunately, it broke when pulled into Google and was removed in cl/152200430. This CL restores the change, except for two Android-only functions; these were the only problematic bits. Change: 152297413 * Removed dead code in Estimator. Change: 152297597 * Assert rank is at least equal to new_rank for `_sparse_inner_flatten`. Change: 152303319 * Extend quantization ranges to include 0.0f. Change: 152304380 * Remove Keras config file saving. Change: 152306552 * API backwards compatibility tests. Change: 152310869 * [TF:XLA] Add a test for an R3 -> R4 broadcast. Change: 152313967 * Fix the problem that no enough placeholders for persistent tensor batch delete The deleter_key is always a device_name, hence there is only one of it. Hence, we cannot delete >1 handles at one time. In the fix, it creates delete placeholder on demand, the max number of placeholders is _DEAD_HANDLES_THRESHOLD. Change: 152322770 * [XLA] Add several reduction tests. Change: 152323510 * Added the memory optimizer to the meta optimizer. Change: 152323689 * Started a set of utilities to categorize op types Change: 152329057 * Add AudioSpectrogram op to TensorFlow for audio feature generation Change: 152332221 * Update ops-related pbtxt files. Change: 152332812 * Automated rollback of change 152332221 Change: 152333917 * Call Py_CLEAR on dead fields during TF_RESOURCE-to-ndarray conversion Change: 152338333 * [TF contrib seq2seq] Initial, incomplete implementation of beam search decoder. **DOES NOT WORK, pushed for collaboration only** Change: 152343927 * [XLA] Change HloPassPipeline to disallow Add* calls after Run. Change: 152345578 * Automated rollback of change 152332812 Change: 152349057 * Remove all 64/32 bit compiler warnings from core/ops. Change: 152353506 * libtensorflow.so: Don't export private symbols. With this change, libtensorflow.so will only export functions defined in c_api.h. This also results in a decreased binary size of libtensorflow.so. On Linux the decrease was from roughly 150MB to 67MB. On OS X it was from roughly 101MB to 82MB. Also fixes #8923 Change: 152366053 * Add Elu ops in XLA. Change: 152383201 * Fixed test. ('broadcast_dims' has size 1) Change: 152383633 * Add more detailed error message for rank assertion in _sparse_inner_flatten. Change: 152397909 * tensor_bundle: propagrates errors related to directory creation. Change: 152401909 * matrix_adjoint added to contrib/linalg/linear_operator_util Change: 152404828 * Add an is_active method to plugins This method determines whether a plugin is active. A plugin may be inactive if say it lacks data. This new is_active method allows us to add a route to TensorBoard noting which plugins are active. The frontend could then avoid querying routes of inactive plugins. Change: 152406232 * Replace a gather op for shapes by a stack op so dilated convolutions can be placed on GPU even with strict placing (before the gather went to CPU). Change: 152411159 * [TF:XLA] Implement BatchToSpace, BatchToSpaceND, SpaceToBatch, SpaceToBatchND. Fix crashes in core implementations of the same operators for zero-sized blocks. Change: 152416903 * Estimator saves relative paths in checkpoint. Change: 152420211 * Fix layers_test exception regex matching. Change: 152422855 * Unhide bijectors. Correct TransformedDistribution docstring. Change: 152424418 * Choosing a saner default for min_eval_frequency in the constructor for Experiment for the GCS file system, because the default of 1 causes performance problems. Change: 152439984 * Inherit use_resource from scope for partitioned variables. Change: 152442103 * Support quantized reshape in hexagon runtime Change: 152445539 * tfdbg CLI: add command list_source (ls) + UI fixes and improvements The new list_source (shorthand: ls) command lists Python source files responsible for constructing the nodes and tensors encountered in the run() call. It divides the source files into two categories and list them separately. 1) files that are not part of the TensorFlow Python library, and 2) files that are a part of it. The list contains information about how many nodes, tensors and dumps of tensors the files is responsible for. The file paths contain clickable links to the existing print_source/ps command. The list_source/ls command supports filtering by file-path and node-name regex patterns. UI fixes: * Fixed inconsistent black vs. transparent background color that made the layout look messy on some terminal types. Now using the transparent color for default font color consistently. * In the print_source command output, add clickable links to expand source lines and graph elements. Change: 152446002 * tfcompile: Be a little more verbose about missing required flags. Fixes #9014 Change: 152446338 * Disable failing test cases in pooling_ops_test. Change: 152447322 * Register more types for tf.image_crop_and_resize(). Resolves #9020. Change: 152448160 * Automated rollback of change 152439984 Change: 152450929 * Add a route to TensorBoard for fetching plugin names Specifically, we add a /data/plugins_listing route to the TensorBoard application. This route responds with an object mapping the name of each initialized plugin to whether it is active. This route could help the frontend avoid issuing requests to inactive plugins. Ordered the listing of routes within application.py so there is a little more organization. Refactored the test for application to use a fake plugin. Change: 152451390 * Added the ability to retrieve the amount of usable gpu memory Change: 152453470 * Allow to set session ConfigProto in RunConfig and use it in Estimator. Change: 152454548 * Colocate ResourceVariable reads with their handles. Change: 152455939 * tfdbg: update doc for new command list_source/ls Change: 152456128 * Make rnn directions slightly easier to follow. Change: 152456296 * Internal change Change: 152458104 * Adds batch renormalization. NOTE: if you use renormalization, you might want to use faster moving average updates, i.e. lower `decay` values. Change: 152458872 * When using ImportGraphDef with a passed in ShapeRefiner, use the producer version of the GraphDef when importing; the ShapeRefiner may be initialized with a different graph_def_version, so we need to be able to override it. The test failed without the change to graph_constructor and passes with it. The test uses a legacy graph that is supported (reduction shape). Change: 152459169 * Allow any iterable for `export_strategies` arg. Change: 152461826 * Log steps/sec every 100 steps in MonitoredSession, as before. Change: 152465320 * Fixes documentation to note that the in case of ties the identity of the return value of ArgMin and ArgMaxis not guaranteed . Change: 152465346 * Automated rollback of change 152465346 Change: 152465844 * Fix shape inference fn on _ParallelConcatStart. Change: 152466076 * Fix getting started guide Explain numerical differences in loss fix one example to print Change: 152466119 * Remove superfluous mode argument. Change: 152467334 * Add a tool that converts HLO computations to tensorflow GraphDef which can be visualized on Tensorboard. This CL defines basic tensorflow::OpDef for each HLO instruction/node. More attributes (e.g. shapes, colors) will be added in the future. Change: 152477918 * [TF:XLA] Increase shard count of //third_party/tensorflow/compiler/tests:spacetobatch_test to reduce flakiness when built under ASAN. Change: 152496244 * Make projector plugin backend read assets saved via the PluginAssets API. At the same time, keep backwards compatibility with the old way of looking up assets. Change: 152504793 * Move MNIST pointers to mirror hosted by the CVDF on Google Cloud. Fixes: #9031 Change: 152504901 * Merge changes from github. Change: 152508170 * Update API after changing default step couter frequency before. Change: 152517535 * Move a few random op helper functions to header files 1. shape_inference::RandomShape 2. OpKernel::MakeShape(Tensor, TensorShape*) Change: 152522156 * addresses the divide by zero bug Change: 152522488 * Clarify doc on tf.assign. Change: 152523909 * Sparse adam for resource variables. Change: 152525327 * Automated rollback of change 152310869 Change: 152528732 * Add an env_var tf_sync_on_finish_bool that block until device has finished all queued operations in a step if true. Change: 152533676 * Add more node attributes for HloInstruction on Tensorboard e.g. shape and layout etc. Change: 152534472 * Add tf.complex64 GPU support to tf.gather. Also add ldg specializations for std::complex. Change: 152537848 * Formatting changes Change: 152544842 * Upgrade TensorBoard TypeScript to 2.2.1 See also: #8326 Change: 152545950 * TEST: Getting reasonable test sizes on linalg library, removing need for sharding. Change: 152546409 * Disabling _testSourceUtilModuleReturnsTrue as its causing opensource issues. Change: 152548721 * Fix race due to unsafe buffer forwarding in maxpooling second order gradients added in #6664. Re-enable previously flaky tests. Clean up a few minor things in maxpooling_op_gpu.cu.cc Change: 152550050

Commit:2238fd7
Author:Martin Wicke
Committer:TensorFlower Gardener

Automated rollback of change 152310869 Change: 152528732

Commit:dc14caf
Author:Benoit Steiner
Committer:TensorFlower Gardener

Added the memory optimizer to the meta optimizer. Change: 152323689

Commit:06c622b
Author:Gunhan Gulsoy
Committer:TensorFlower Gardener

API backwards compatibility tests. Change: 152310869

Commit:e69f717
Author:Rohan Jain
Committer:GitHub

Branch 152232810 (#8988) * Improve py_func error handling. Automatically translate some python errors into corresponding TF errors at runtime. Change: 152156821 * Update interaction with libpng so that we use the public API instead of knowledge of the internal libpng data structures. Change: 152167754 * TensorBoard plugins now contain their own name/route prefix. Change: 152167807 * Passes trainable flag to separable_conv2d biases. Change: 152170239 * Saving resource variables with a caching device. Change: 152171539 * Drop loss from estimator_spec.eval_metric_ops, as required by core Estimator. Change: 152179924 * sample_stats.percentile DOCFIX. Change: 152182295 * Added a memory optimizer to grappler. Change: 152184170 * Change default behavior of the tf runs selector: - If there are fewer than 41 runs, enable them all by default - If there are 41 runs or more, disable them all by default This is in response to user complaints that having it enable only the first ten runs by default was confusing, because it was not obvious to users that some runs had been disabled. However, it still solves the initial user complaint that having very many runs simultaneously enabled would lag the UI. I also changed the "toggle all runs" button to try to turn everything off before turning everything on. Also, I improved the logic for detecting when the runs selection is back in the default state, so that we can avoid generating long URI strings wherever possible. Change: 152188948 * Autogenerated Change: Change TensorBoard TAG to 52 Change: 152189000 * Remove warning that only happening with config cuda. Change: 152189205 * Make resource variable shared name consistent with non-resource variables. Remove colocation constraint from resource variable cached value with the variable itself. Change: 152192203 * Add a way to specify the optimization order; refactor and add constant folding to meta optimizer. Change: 152193646 * Backport fixes and improvements from external Keras. Change: 152198296 * Merge changes from github. Change: 152200430 * Go: Update generated wrapper functions for TensorFlow ops. Change: 152200754 * Update ops-related pbtxt files. Change: 152203174 * Make ImportGraphDef() work with functions. In addition to modify graph_constructor.cc, this patch adds some other functionality to enable importing fucntions: * Ability to add FunctionDefLibraries to Graphs and FunctionLibraryDefinitions (in addition to existing functions) * FunctionDefsEqual() utility function Change: 152205258 * Expand contrib test to more than just test targets. Change: 152206822 * Preserve graph version during optimization Change: 152213262 * Exclude enter and exit nodes from shape refiner's constant folding. Change: 152213637 * Allow reshape_mover and algebraic_simplifier to make multiple mutations, by avoiding the short-circuit std::any_of. Change: 152232810 * fixing workspace.bzl * workspace.bzl further fixes * fixing tensorflow.bzl merge conflicts * fixing typo in dnn.h * fixing bad merge for dnn.h

Commit:8f74d59
Author:Yao Zhang
Committer:TensorFlower Gardener

Add a way to specify the optimization order; refactor and add constant folding to meta optimizer. Change: 152193646

Commit:663eaa0
Author:Dan Smilkov
Committer:TensorFlower Gardener

Fix the circular dependency tf/contrib/tensorboard <--> tf/tensorboard. tf/tensorboard should not depend on tf/contrib/.. This is resolved by moving projector_config.proto from tf/contrib/tensorboard/plugins/projector to tf/tensorboard/plugins/projector. The user-facing API hasn't changed. `ProjectorConfig` is still exposed as `tensorflow.contrib.tensorboard.plugins.projector.ProjectorConfig`. Change: 152030381

Commit:50be7aa
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Migrate trees, models, testutils, and resources libraries to boosted_trees. Change: 151832033

Commit:e4c4dc2
Author:Vijay Vasudevan
Committer:TensorFlower Gardener

Remove stale documtation about node colocation Change: 151778994

Commit:39af895
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Move all parameters of remote graph execution to NodeDef instead of additional serialized data Change: 151771414

Commit:396b6bd
Author:Brennan Saeta
Committer:TensorFlower Gardener

Consolidate worker state behind a session-centric abstraction. State in workers is currently splayed across graph_mgr, rendezvous_mgr, and additional components. This has resulted in it being difficult to ensure proper cleanup and shut down of the worker components. In addition to paving the way for a more reliable shut down, this CL also sets up the beginnings of ClusterSpec propagation. ClusterSpec propagation is a capability upgrade for TensorFlow that should make it much easier to (1) build distributed TensorFlow clusters, and (2) handle node failures. After the ClusterSpec propagation capability is fully implemented, the TensorFlow workers can be booted independently of each other, and with no knowledge about others. A client can then query a central cluster scheduler or other API to find all of the workers, and then send the ClusterDef (ClusterSpec) to the TF master, which then propagates that along to all of the workers. This change is only the first of a sequence to fully implement ClusterSpec propagation in TensorFlow. Change: 151229111

Commit:0d259df
Author:Benoit Steiner
Committer:TensorFlower Gardener

Added a new OpPerformanceList message Change: 151191936

Commit:4bbd781
Author:Benoit Steiner
Committer:TensorFlower Gardener

Added a config option to control model pruning Change: 151130707

Commit:2bb0a62
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Migrate utils and protos to contrib/boosted_trees. Change: 150897748

Commit:0b5a6f5
Author:A. Unique TensorFlower
Committer:gunan

Adds java package and outer class name to rewriter_config.proto. Change: 150664674

Commit:a7e5032
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Adds java package and outer class name to rewriter_config.proto. Change: 150664674

Commit:95cdab3
Author:Benoit Steiner
Committer:TensorFlower Gardener

Created a proto to configure the amount of graph rewriting taking place. Change: 150648084

Commit:dcd71f6
Author:Peter Hawkins
Committer:TensorFlower Gardener

[XLA] Give Transpose its own Request, rather than piggybacking on ReshapeRequest. Avoids building unnecessary Reshape operators when Transpose was called by the client. Also avoids building Transpose operators when Reshape has identity transpose dimensions, for example when the client called the variant of ComputationBuilder::Reshape() that does not transpose. Makes the HLO graph emitted by the TF bridge more readable. Change: 150253949

Commit:1b881b7
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

First version of BigQuery Reader. Change: 150016997

Commit:c559153
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Attempt to specify tensorflow::Reset(). Change: 149714094

Commit:00d0347
Author:Brennan Saeta
Committer:TensorFlower Gardener

[TF:XLA] Add debug metadata to HLO ops. In order to support end-to-end debugging and performance profiling tooling for the TensorFlow::XLA toolchain, this change adds a DebugMetadata proto to the HloInstruction class, and pipes it through the tf2xla stack. Change: 149703349

Commit:96cb8f8
Author:Benoit Steiner
Committer:TensorFlower Gardener

Started to open source Grappler. First application is the GPU layout optimizer. Change: 149558284

Commit:3b59692
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Factor-out dryrun functionalities from graph_transfer_utils to remote_fused_graph_execute_op in order to cache shapes in RemtoeFusedGraphExecuteInfo Change: 149143066

Commit:54c0ff4
Author:Eugene Brevdo
Committer:TensorFlower Gardener

Add BenchmarkType to TestResults proto and fix logging of C++ microbenchmarks. C++ microbenchmarks results are now normalized by iters (like python benchmarks). Change: 149045367

Commit:4f5c6bb
Author:Yuefeng Zhou
Committer:TensorFlower Gardener

Add interfaces in OpKernelContext and ResourceBase for tracking persistent storage. Record persistent tensor memory and persistent memory (originally auxiliary memory) in the cost model. Change: 148920117

Commit:1296e56
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add optional transport_options to RecvTensorRequest. Change: 148827175

Commit:e3d1c7d
Author:Shanqing Cai
Committer:TensorFlower Gardener

tfdbg: add option to tolerate debug op creation failures Change: 148788753

Commit:efc8f98
Author:Jacques Pienaar
Committer:TensorFlower Gardener

[XLA] Add basic outfeed support. Change: 148699787

Commit:cc8bfc4
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add remote_fused_graph_execute_info to abstract hexagon specific parameters Change: 148489487

Commit:8120e2a
Author:Peter Hawkins
Committer:TensorFlower Gardener

[XLA] Add an IsFinite operation that tests elementwise whether values are finite (i.e., not NaN or Inf). Change: 148485205

Commit:4fac560
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Open source HParams object and proto. Change: 148474914

Commit:b3c75d3
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

tensorflow::Reset() now honors the device_filter in SessionOptions the same way as does tensorflow::NewSession(). This allows Reset() to succeed, even if an unrelated worker happens to be down. Change: 148343973

Commit:11f11c8
Author:Jonathan Hseu
Committer:TensorFlower Gardener

Move tensorflow::ReaderBase to tensorflow/core/framework so that it can be used in custom op libraries and contrib. Change: 148181260

Commit:7817ac8
Author:Mark Heffernan
Committer:TensorFlower Gardener

[XLA] Properly version outfeed and send operations in UserComputation. Previously outfeed and send operations were unconditionally emitted during UserComputation lowering even if the outfeed/send was not in the requested version (computation snapshot). This CL versions these operations. Also, opportunistically improve logging in UserComputation, Service, and ComputationTracker which was used to root cause the underlying bug. Change: 148170893

Commit:57559fd
Author:Yao Zhang
Committer:TensorFlower Gardener

Add compute_time and memory_time to CostGraphDef::Node. Change: 147871989

Commit:f128935
Author:Yuefeng Zhou
Committer:TensorFlower Gardener

Record persistent memory in tracking allocator and cost model. Change: 147123056

Commit:78be42e
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Tolerate unrecognized top-level fields in fast Example parsing. Some internal serialization routines use unrecognized field numbers. Change: 147039073

Commit:13a5be6
Author:Yuefeng Zhou
Committer:TensorFlower Gardener

Record kernel's host memory and device memory in cost model. Change: 146997688

Commit:aabc797
Author:Shanqing Cai
Committer:TensorFlower Gardener

tfdbg core: add core metadata to debugger data stream + better support of concurrent debugged runs * Let the debugger send/dump an Event proto holding a JSON string in its log_message.message field. The JSON metadata includes, 1) An optional, client-specified global_step field that defaults to -1 if not supplied 2) A session run count 3) An executor invocation step count 4) Input names (feed keys) 5) Output names (fetched Tensor names) 6) Target node names * grpc_debug_server.EventListenerBaseServicer now requires a constructor of the type EventListenerBaseStreamHandler and will construct a new handler object from it, for every stream. This leads to better support of concurrent debugged Session::Run() calls. * Add support for path names in grpc:// URLs, such as "grpc://localhost:6000/thread1". Different path names will lead to separate gRPC streams being opened to the same server:port, supporting concurrent debugged Session::Run() calls. Change: 146896481

Commit:ad67415
Author:A. Unique TensorFlower
Committer:gunan

Rename the DEBUG logging level to DEBUGGING. Some software projects compile C/C++ code with -DDEBUG (equivalent to -DDEBUG=1) in debug builds. Notably, this is done by default in debug builds on iOS. Therefore, we should avoid using DEBUG as an identifier in C/C++ code. Change: 146840049

Commit:3db4be7
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Rename the DEBUG logging level to DEBUGGING. Some software projects compile C/C++ code with -DDEBUG (equivalent to -DDEBUG=1) in debug builds. Notably, this is done by default in debug builds on iOS. Therefore, we should avoid using DEBUG as an identifier in C/C++ code. Change: 146840049

Commit:2525297
Author:Peter Hawkins
Committer:gunan

[XLA] Update comment on RNG_UNIFORM in xla_data.proto Change: 145810242

Commit:8f6e925
Author:Justin Lebar
Committer:gunan

[XLA] Rename CompilationOptions proto to ExecutionOptions. We want to put fields in this proto that aren't strictly related to compilation. Change: 146477500

Commit:bf7470f
Author:A. Unique TensorFlower
Committer:gunan

Added optional string argument to infeed HLO op. Change: 145188452

Commit:3af171a
Author:Tayo Oguntebi
Committer:gunan

Addition of Outfeed HLO op. Change: 145772331

Commit:bcfe10e
Author:Justin Lebar
Committer:gunan

[XLA] Move shape_with_output_layout and seed fields into ExecutionOptions. This simplifies the execution API by getting rid of two default params. Also change HloModuleConfig so it stores each of the fields of ExecutionOptions individually, instead of keeping an instance of the ExecutionOptions proto. This is necessary because HloModuleConfig already has a field derived from shape_with_output_layout -- if we stored the ExecutionOptions proto in HloModuleConfig, its shape wouldn't necessarily match the shape we already have. Change: 146477669

Commit:8a0eb24
Author:Justin Lebar
Committer:gunan

[XLA] Move fast-math flags into HLO module config. Previously, XLA controlled the presence/absence of fast-math flags (FMF) via a command-line flag. This patch changes things so we use a new CompileOptions proto instead. This proto lives in HloModuleConfig, and is passed to the service via ExecuteRequest. This change lets us entirely remove llvm_backend_flags.{h,cc}. In addition, this change takes us from two to one fast-math flags. Previously we tried to control "unsafe FP transformations" separately from "full fast math". It turns out that LLVM is misleadingly inconsistent in how it handles these. In the backend, they are indeed two separate options that can be enabled/disabled independently. In the frontend, however, unsafe-fp-math implies all the other FMFs. As a result, it doesn't really make sense for XLA to attempt to split out these two flags, at least not until LLVM changes how it handles them. Change: 146183994

Commit:d878488
Author:Alexandre Passos
Committer:gunan

Metagraph round-tripping for resource variables. Change: 146519414

Commit:d428d23
Author:A. Unique TensorFlower
Committer:gunan

Add an op to execute a subgraph fused and transferred by GraphTransferer in a remote processor Change: 146527671

Commit:8aa897e
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add an op to execute a subgraph fused and transferred by GraphTransferer in a remote processor Change: 146527671

Commit:987e3d9
Author:Alexandre Passos
Committer:TensorFlower Gardener

Metagraph round-tripping for resource variables. Change: 146519414

Commit:7754ec4
Author:Justin Lebar
Committer:TensorFlower Gardener

[XLA] Move shape_with_output_layout and seed fields into ExecutionOptions. This simplifies the execution API by getting rid of two default params. Also change HloModuleConfig so it stores each of the fields of ExecutionOptions individually, instead of keeping an instance of the ExecutionOptions proto. This is necessary because HloModuleConfig already has a field derived from shape_with_output_layout -- if we stored the ExecutionOptions proto in HloModuleConfig, its shape wouldn't necessarily match the shape we already have. Change: 146477669

Commit:84e1a0b
Author:Justin Lebar
Committer:TensorFlower Gardener

[XLA] Rename CompilationOptions proto to ExecutionOptions. We want to put fields in this proto that aren't strictly related to compilation. Change: 146477500

Commit:d72465a
Author:A. Unique TensorFlower
Committer:gunan

Add a mechanism for hiding, skipping, and modifying the generated op functions for C++. A souped-up version of the hidden_ops mechanism in Python, the intent is to use this for most or all of the client languages, with a common list of changes to make in a common file and per-language overrides. Also: * include the documentation for outputs in the generated comments * several updates to C++ API to match Python * fix C++ shape function for ConcatV2 now that we use it by default * split op_gen_lib out of core:framework, since it is only used by the op generators, and I don't want to add another proto to mobile builds Change: 146267344

Commit:4fe798c
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Use graph transfer info proto in GraphTransferer Change: 146421389

Commit:f0fa1e4
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Document the unit used to store times in the cost graph. Change: 146416639

Commit:b44959b
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add proto for graph transfer to serialize soc node Change: 146409586

Commit:a201aff
Author:Zongheng Yang
Committer:TensorFlower Gardener

Remove the "Experimental" markers of RunOptions and RunMetadata protos. This seems like a reasonable thing to do, as (1) they have been in existence for ~1 year, (2) people have added new fields to them since their inception, and (3) the C API supports them. Change: 146271432

Commit:8fe3202
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Add a mechanism for hiding, skipping, and modifying the generated op functions for C++. A souped-up version of the hidden_ops mechanism in Python, the intent is to use this for most or all of the client languages, with a common list of changes to make in a common file and per-language overrides. Also: * include the documentation for outputs in the generated comments * several updates to C++ API to match Python * fix C++ shape function for ConcatV2 now that we use it by default * split op_gen_lib out of core:framework, since it is only used by the op generators, and I don't want to add another proto to mobile builds Change: 146267344

Commit:d45505f
Author:Justin Lebar
Committer:TensorFlower Gardener

[XLA] Move fast-math flags into HLO module config. Previously, XLA controlled the presence/absence of fast-math flags (FMF) via a command-line flag. This patch changes things so we use a new CompileOptions proto instead. This proto lives in HloModuleConfig, and is passed to the service via ExecuteRequest. This change lets us entirely remove llvm_backend_flags.{h,cc}. In addition, this change takes us from two to one fast-math flags. Previously we tried to control "unsafe FP transformations" separately from "full fast math". It turns out that LLVM is misleadingly inconsistent in how it handles these. In the backend, they are indeed two separate options that can be enabled/disabled independently. In the frontend, however, unsafe-fp-math implies all the other FMFs. As a result, it doesn't really make sense for XLA to attempt to split out these two flags, at least not until LLVM changes how it handles them. Change: 146183994

Commit:67fe19d
Author:Peter Hawkins
Committer:TensorFlower Gardener

[XLA] Update comment on RNG_UNIFORM in xla_data.proto Change: 145810242

Commit:9113e98
Author:Tayo Oguntebi
Committer:TensorFlower Gardener

Addition of Outfeed HLO op. Change: 145772331

Commit:4fe280c
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Added optional string argument to infeed HLO op. Change: 145188452

Commit:3f45361
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Expose the two constants that control GPU Event polling pacing as GPUOptions values that are user settable. Change: 144992308

Commit:ff151c9
Author:Andrew Harp
Committer:TensorFlower Gardener

Android demo: read MultiBox priors from a txt file in Java rather than reading from a proto file in C++ code, in order to reduce code/build complexity. New model file archive with corresponding changes has been uploaded to https://storage.googleapis.com/download.tensorflow.org/models/mobile_multibox_v1a.zip Resolves #6670 Change: 144872035

Commit:873473e
Author:A. Unique TensorFlower
Committer:TensorFlower Gardener

Delete the FunctionDef.Node code paths, now that we have switched to the NodeDef representation. Change: 144281952

Commit:d4a9d91
Author:Peter Hawkins
Committer:TensorFlower Gardener

Add support for list(func) AttrValues. Change: 144211572