Sanjoy Das [Tue, 29 May 2018 06:55:19 +0000 (23:55 -0700)]
Fix an incorrect precondition check in IndexedArrayAnalysis
PiperOrigin-RevId:
198354001
Sanjoy Das [Tue, 29 May 2018 06:03:39 +0000 (23:03 -0700)]
Pass HloOpcode instead of HloInstruction; NFC
Minor code cleanup change.
PiperOrigin-RevId:
198351045
Sanjoy Das [Tue, 29 May 2018 05:16:46 +0000 (22:16 -0700)]
Make IndexedArrayAnalysis behave well around StatusOr
PiperOrigin-RevId:
198348355
Jiri Simsa [Mon, 28 May 2018 16:33:49 +0000 (09:33 -0700)]
Adding tf.data optimization for rewriting `map(...).batch(...)` to `map_and_batch(...)`.
PiperOrigin-RevId:
198310806
Tom Hennigan [Mon, 28 May 2018 13:32:04 +0000 (06:32 -0700)]
Sort variables in C++ instead of Python.
PiperOrigin-RevId:
198298103
A. Unique TensorFlower [Mon, 28 May 2018 12:21:13 +0000 (05:21 -0700)]
Relax compatibility checks for Protocol Buffer classes, to not rely on the exact layout of the generated classes. The definition of the message is checked instead.
PiperOrigin-RevId:
198292780
Justin Lebar [Sun, 27 May 2018 20:27:28 +0000 (13:27 -0700)]
[XLA] Don't display metadata inline in HLO graph dump.
We only want to display it in the tooltip.
PiperOrigin-RevId:
198235268
A. Unique TensorFlower [Sun, 27 May 2018 17:49:12 +0000 (10:49 -0700)]
TPUEstimator.export_savedmodel() saves a SavedModel with both TPU and CPU graphs.
PiperOrigin-RevId:
198229550
A. Unique TensorFlower [Sat, 26 May 2018 15:25:12 +0000 (08:25 -0700)]
Extracts the 'remove random shuffle node' optimization into its own method.
PiperOrigin-RevId:
198169790
Guangda Lai [Sat, 26 May 2018 06:44:11 +0000 (23:44 -0700)]
Fix the issue where returned Status is not used.
PiperOrigin-RevId:
198146500
A. Unique TensorFlower [Sat, 26 May 2018 05:46:46 +0000 (22:46 -0700)]
DepthwiseConv Optimizations
PiperOrigin-RevId:
198144118
A. Unique TensorFlower [Sat, 26 May 2018 04:38:56 +0000 (21:38 -0700)]
Use dict(locals()) instead of distribution_util.parent_frame_arguments. This will be much
faster at object construction time (going forward we'll figure out a way to make this a function to call).
PiperOrigin-RevId:
198141184
Justin Lebar [Sat, 26 May 2018 03:23:31 +0000 (20:23 -0700)]
[XLA] Don't compute relative error when the expected value is 0.
In literal_comparison, don't try to compute a relative error when the
expected value is 0, because doing so would mean that the only
acceptable value *is* zero, which probably isn't what you mean.
PiperOrigin-RevId:
198137414
Sanjoy Das [Sat, 26 May 2018 02:21:57 +0000 (19:21 -0700)]
Add support for unary and binary ops to indexed tensor analysis
I've added a TODO to clean up the use of ValueOrDie which I will address in an
immediately following CL.
PiperOrigin-RevId:
198134579
Yunxing Dai [Sat, 26 May 2018 02:18:30 +0000 (19:18 -0700)]
[Hlo Graphviz] Always show metadata as tooltips
Always show metadata as tooltips.
PiperOrigin-RevId:
198134430
Sanjoy Das [Sat, 26 May 2018 00:46:19 +0000 (17:46 -0700)]
Add a type-erased broadcast implementation to xla::Literal
And use this in HLO evaluator. Since broadcast only moves bytes around we don't
need a type specialized implementation.
I'll use this in a later change.
PiperOrigin-RevId:
198128524
Akshay Modi [Sat, 26 May 2018 00:37:01 +0000 (17:37 -0700)]
Minor eager performance improvements
- remove linear regression dependence on global step.
This speeds things up a lot for the benchmark (since it removes a bunch of
unnecessary code), but is obviously not a fair comparison.
I think its worth doing, since I don't see any reason to have a global step
in eager.
- nn_ops dropout had an unnecessary convert_to_tensor, convert back to numpy
(with a GPU this would copy out, copy back).
- cudnn_recurrent reshape would always fallback to the slow path - so I just
converted it to be in the fastpath - this will be low impact.
- tensor_shape should not generate a new object every time
- remove unnecessary list creation and searching in some dtypes functions
PiperOrigin-RevId:
198127757
Peter Hawkins [Sat, 26 May 2018 00:29:37 +0000 (17:29 -0700)]
[TF:XLA] Add direct implementation of AssignVariableOp for XLA devices.
This allows us to avoid an XLA compilation and tensor copies when assigning to a variable placed on an XLA device.
PiperOrigin-RevId:
198127062
A. Unique TensorFlower [Sat, 26 May 2018 00:22:17 +0000 (17:22 -0700)]
Turn on heuristic (mostly-NHWC) convolution layout assignment for (V100, fp16) by default.
Also increase the column reduction tile size to reduce atomic operations.
PiperOrigin-RevId:
198126505
Sanjoy Das [Sat, 26 May 2018 00:22:11 +0000 (17:22 -0700)]
Enable while loop constant sinking for GPU
To avoid keeping constants in while loop bodies after optimization (where they
may cause extra copies) we run a late pass of LICM that has been asked to hoist
constants when it can.
PiperOrigin-RevId:
198126497
Akshay Agrawal [Sat, 26 May 2018 00:12:49 +0000 (17:12 -0700)]
Provide a step container to OpKernelContexts when running eagerly.
This lets us run some ops that require step containers (e.g.
TensorArray-related ops).
Before change:
Benchmark Time(ns) CPU(ns) Iterations
-------------------------------------------------------------------
BM_CreateGraph 61292 80915 8581
BM_RunGraph 7899 13398 51251
BM_CreateAndDestroySession 2588 2594 269838
BM_KernelAndDeviceInit 2971 2976 235908
BM_KernelAndDeviceRun 505 506 1000000
After change:
Benchmark Time(ns) CPU(ns) Iterations
-------------------------------------------------------------------
BM_CreateGraph 78295 105539 8698
BM_RunGraph 9907 16988 47908
BM_CreateAndDestroySession 2773 2778 247635
BM_KernelAndDeviceInit 2678 2682 270054
BM_KernelAndDeviceRun 553 554 1000000
PiperOrigin-RevId:
198125630
Jiri Simsa [Sat, 26 May 2018 00:05:33 +0000 (17:05 -0700)]
[tf.data] Fixing concurrency issue in `map_and_batch`.
PiperOrigin-RevId:
198124860
Alexandre Passos [Fri, 25 May 2018 23:43:40 +0000 (16:43 -0700)]
Ignore while loops instead of mangling them in the automatic control dependencies.
PiperOrigin-RevId:
198122188
A. Unique TensorFlower [Fri, 25 May 2018 23:43:29 +0000 (16:43 -0700)]
Extracts the 'remove reverse node' optimization into its own method.
PiperOrigin-RevId:
198122165
A. Unique TensorFlower [Fri, 25 May 2018 23:07:25 +0000 (16:07 -0700)]
Automated g4 rollback of changelist
198087342
PiperOrigin-RevId:
198117552
Nick Felt [Fri, 25 May 2018 22:11:46 +0000 (15:11 -0700)]
Add warning to LookupOrCreate about reentrancy issue
PiperOrigin-RevId:
198110382
Igor Ganichev [Fri, 25 May 2018 20:58:51 +0000 (13:58 -0700)]
Add EagerTensor profiler and device shape utilities
This change includes the following steps to make
EagerTensor profiler work:
- Add a PaddedShapeFn to XlaDevice::Metadata. We need a
backend-independent way to get a fully-padded shape and
its layout on the device. This function is set during
device construction. CPU and GPU devices effectively get
an identity function since they neither change the layout
nor pad. TPU gets the appropriate function.
- Add TFE_TensorDebugInfo struct and C API methods for it.
These methods are necessary to fetch the shape and layout
from under the C API to the Python level. This can be a home
for more debug information later.
- Make EagerTensor weak referencable. This involves adding a
pointer to the list of current weak references. This addition
should have negligible overhead when profiler is not used.
The only operations on this field are setting it to null on
construction and checking if it is null on destruction.
- Adding C++ functions callable from Python to register an instance
of EagerTensorProfiler and retrieve debug information for a given
EagerTensor. These functions are used in the new "inspect" module.
- Finally, writing the actual profiler.
PiperOrigin-RevId:
198098380
A. Unique TensorFlower [Fri, 25 May 2018 20:44:36 +0000 (13:44 -0700)]
Disable //tensorflow/contrib/lite/python:lite_test on Windows
PiperOrigin-RevId:
198096344
A. Unique TensorFlower [Fri, 25 May 2018 20:39:25 +0000 (13:39 -0700)]
[tpu:profiler] Capture the data for generating a memory viewer of the profiling results.
PiperOrigin-RevId:
198095564
Sanjoy Das [Fri, 25 May 2018 20:38:24 +0000 (13:38 -0700)]
[TF:XLA] Bump open source llvm revision to r333273
PiperOrigin-RevId:
198095416
Alexandre Passos [Fri, 25 May 2018 20:20:13 +0000 (13:20 -0700)]
Public API to switch between eager execution and graph building.
Now, after tf.enable_eager_execution() has been executed, entering the context
manager of a tf.Graph will enable graph mode. So, for example
```
tf.enable_eager_execution()
with tf.Graph().as_default():
c = tf.constant(1.0) # this is a graph tensor
c2 = tf.constant(1.0) # this is an eager tensor
```
The main use-case of this is allowing documentation writers to make a single
notebook which starts with eager execution and seamlessly transitions to
building graphs.
This also makes many explicit enablings of graph mode in the code redundant
(a cleanup cl will follow).
PiperOrigin-RevId:
198092991
A. Unique TensorFlower [Fri, 25 May 2018 19:58:55 +0000 (12:58 -0700)]
Use functions to build dense splits. Tensorflow Function invocations share the same graph so using them reduces the graph construction overhead.
PiperOrigin-RevId:
198090110
A. Unique TensorFlower [Fri, 25 May 2018 19:56:40 +0000 (12:56 -0700)]
[tpu:profiler] Minor change in the description of tool name proto.
PiperOrigin-RevId:
198089875
A. Unique TensorFlower [Fri, 25 May 2018 19:54:49 +0000 (12:54 -0700)]
Add ScopedAllocatorOptimizer in support of CollectiveReduce.
The efficiency of CollectiveReduce is greatly improved by merging
multiple parallel reductions over smaller tensors into a single
reduction over a larger tensor that is the concatentation of the
smaller tensors. Because CollectiveReduce is essentially an
element-wise array operation which operates on a 1-D reshape of
the input tensor it is eligible for a ScopedAllocation optimization.
The optimization works by looking for serially independent instances
of CollectiveReduce that lie within the same name-scope tier and
have the same control-flow (e.g. loop) embedding structure. Where
two or more such nodes are found the upstream nodes that generate
their inputs are modified to write their outputs into consecutive
regions of a single tensor buffer maintained by a ScopedAllocator.
The multiple CollectiveReduce nodes are then replaced by a single
CollectiveReduce that operates in-place on the backing buffer.
The effectiveness of the optimization depends on there being candidate
CollectiveReduce nodes with these characteristics that become eligible
for execution at close to the same time. If the name scope is too
large, and includes nodes that become execution eligible at very different
times, this graph rewrite could result in a slowdown.
Note that this optimization is experimental: it is not guaranteed to
work, especially for ops other than CollectiveReduce.
PiperOrigin-RevId:
198089642
A. Unique TensorFlower [Fri, 25 May 2018 19:35:50 +0000 (12:35 -0700)]
enhance Tensorflow GBDT and GBRT model by exposing a new two dimensional output in prediction ops (example id, tree leaf node index id) for input as other model features
PiperOrigin-RevId:
198087342
A. Unique TensorFlower [Fri, 25 May 2018 19:22:45 +0000 (12:22 -0700)]
Extracts the 'simplify slice' optimization into its own method.
PiperOrigin-RevId:
198085532
Peter Hawkins [Fri, 25 May 2018 19:04:49 +0000 (12:04 -0700)]
[TF:XLA] Register Switch and Merge ops on XLA devices.
PiperOrigin-RevId:
198083156
Derek Murray [Fri, 25 May 2018 18:42:33 +0000 (11:42 -0700)]
Automated g4 rollback of changelist
192848921
PiperOrigin-RevId:
198079927
A. Unique TensorFlower [Fri, 25 May 2018 18:34:30 +0000 (11:34 -0700)]
Extracts the 'simplify strided slice' optimization into its own method.
PiperOrigin-RevId:
198078724
Igor Ganichev [Fri, 25 May 2018 18:27:39 +0000 (11:27 -0700)]
Bump TPU batch size and wrap apply_grads in defun
PiperOrigin-RevId:
198077643
Akshay Modi [Fri, 25 May 2018 18:02:42 +0000 (11:02 -0700)]
Release C++ lock before calling back into python
PiperOrigin-RevId:
198073059
A. Unique TensorFlower [Fri, 25 May 2018 17:54:38 +0000 (10:54 -0700)]
DepthwiseConv optimizations.
PiperOrigin-RevId:
198071709
Mark Daoust [Fri, 25 May 2018 17:45:27 +0000 (10:45 -0700)]
Link to tf.estimator docs for premade estimators.
PiperOrigin-RevId:
198070157
A. Unique TensorFlower [Fri, 25 May 2018 15:55:24 +0000 (08:55 -0700)]
Code simplification in dump_graphviz.cc:
Just output all arrays, before writing edges, so we don't
need to keep track of which arrays we've already output.
PiperOrigin-RevId:
198055327
Shanqing Cai [Fri, 25 May 2018 13:56:38 +0000 (06:56 -0700)]
Minor clarification to model_to_estimator() doc string
PiperOrigin-RevId:
198044106
Asim Shankar [Fri, 25 May 2018 09:23:06 +0000 (02:23 -0700)]
eager: Update introduction notebooks.
PiperOrigin-RevId:
198022387
Asim Shankar [Fri, 25 May 2018 08:36:23 +0000 (01:36 -0700)]
Fix typo, fix build.
PiperOrigin-RevId:
198017870
A. Unique TensorFlower [Fri, 25 May 2018 03:36:45 +0000 (20:36 -0700)]
Extracts the 'simplify tile node' optimization into its own method.
PiperOrigin-RevId:
197996636
A. Unique TensorFlower [Fri, 25 May 2018 02:49:05 +0000 (19:49 -0700)]
Go: Update generated wrapper functions for TensorFlow ops.
PiperOrigin-RevId:
197993384
A. Unique TensorFlower [Fri, 25 May 2018 02:45:27 +0000 (19:45 -0700)]
Initialize the score threshold to -inf to avoid filtering out negative logits
PiperOrigin-RevId:
197993147
A. Unique TensorFlower [Fri, 25 May 2018 02:20:31 +0000 (19:20 -0700)]
Update ops-related pbtxt files.
PiperOrigin-RevId:
197991672
Yifei Feng [Fri, 25 May 2018 02:12:26 +0000 (19:12 -0700)]
Merge changes from github.
Revert #18413. Too many internal test failures due to the name scope change caused by this change.
Revert #18192. Cannot use re2::StringPiece internally. Need alternative for set call. Will pull and clean this up in a separate change.
PiperOrigin-RevId:
197991247
A. Unique TensorFlower [Fri, 25 May 2018 01:55:30 +0000 (18:55 -0700)]
Extracts the 'simplify pad node' optimization into its own method.
PiperOrigin-RevId:
197989813
Sanjoy Das [Fri, 25 May 2018 01:23:48 +0000 (18:23 -0700)]
Rename TileLoader to MemoryTile; NFC
In a later change I will expand MemoryTile to store tiles and load "3d" tiles
(where we broadcast along one dimension as we load).
PiperOrigin-RevId:
197987185
A. Unique TensorFlower [Fri, 25 May 2018 00:48:21 +0000 (17:48 -0700)]
Add heuristic on picking NHWC layout for (V100, fp16) convolutions.
Also move AlgorithmPicker after layout assignment, as now
cudnn_convolution_runner will return failures on invalid input layouts.
Also add a backend debug option to switch the layout heuristic. By default
it has the old behavior (all NCHW).
PiperOrigin-RevId:
197983747
A. Unique TensorFlower [Fri, 25 May 2018 00:06:34 +0000 (17:06 -0700)]
Enabling some potential optimization using the restrict qualifier.
PiperOrigin-RevId:
197979118
Akshay Modi [Thu, 24 May 2018 23:53:33 +0000 (16:53 -0700)]
When converting a numpy float64 to an EagerTensor, always ensure that it
becomes a float64 tensor.
Earlier py_seq_tensor would fall back to a float32 if not explicitly requesting
a float64 (which would not happen if we had no other information).
PiperOrigin-RevId:
197977260
Igor Ganichev [Thu, 24 May 2018 23:44:17 +0000 (16:44 -0700)]
Don't XLA-compile naked variable reads
Before this change, when we executed a naked variable read (i.e. outside of
a defun, directly running <xla_device>->Compute()), tf2xla kernel would
copy the variable's tensor leading to many unnecessary copies.
This change uses the regular non-tf2xla kernel for naked variable reads
and marks the tf2xla one for CompilationOnly().
PiperOrigin-RevId:
197976146
A. Unique TensorFlower [Thu, 24 May 2018 23:31:48 +0000 (16:31 -0700)]
move wide string manipulations out of windows_file_system
PiperOrigin-RevId:
197974385
Akshay Modi [Thu, 24 May 2018 23:20:31 +0000 (16:20 -0700)]
Remove _get_backward_fn and depend on _gradient_function directly.
(_magic_gradient_function was renamed to _gradient_function)
Before:
entry {
name: "MicroBenchmarks.benchmark_tf_gradient_forward_identity"
iters: 30000
wall_time: 5.
88456789653
extras {
key: "examples_per_sec"
value {
double_value: 169936.011885
}
}
}
After:
entry {
name: "MicroBenchmarks.benchmark_tf_gradient_forward_identity"
iters: 30000
wall_time: 5.
04853725433
extras {
key: "examples_per_sec"
value {
double_value: 198077.175551
}
}
}
PiperOrigin-RevId:
197972668
Yu-Cheng Ling [Thu, 24 May 2018 23:02:11 +0000 (16:02 -0700)]
Fix the generated builtin_ops enum header.
PiperOrigin-RevId:
197969642
A. Unique TensorFlower [Thu, 24 May 2018 22:53:44 +0000 (15:53 -0700)]
Extracts the 'simplify squeeze node' optimization into its own method.
PiperOrigin-RevId:
197968452
David Majnemer [Thu, 24 May 2018 22:45:25 +0000 (15:45 -0700)]
[XLA] Remove maps with a single instruction
These maps aren't really pulling their weight, fold them to the instruction
that they compute.
PiperOrigin-RevId:
197967117
Priya Gupta [Thu, 24 May 2018 22:35:13 +0000 (15:35 -0700)]
Avoid infinite recursion when checking for indexed slices.
PiperOrigin-RevId:
197965508
Igor Saprykin [Thu, 24 May 2018 22:28:03 +0000 (15:28 -0700)]
Allow combinations to be used on the class level. Make "mode" optional.
Applying a generator to a class is the same as applying that generator to every member of that class. It is meant to allow avoiding repetition in some cases.
The implementation relies on some internals of parameterized tests and how it works with a class level declaration: https://github.com/abseil/abseil-py/blob/master/absl/testing/parameterized.py#L319.
The "mode" argument is required before this change. To accommodate cases where execution mode isn't the point of the test, "mode" became optional with "graph" mode being default. Another idea I had was to pick a random mode by default.
PiperOrigin-RevId:
197964501
A. Unique TensorFlower [Thu, 24 May 2018 22:27:00 +0000 (15:27 -0700)]
Add local_init_run_options to SessionManager and Supervisor so that
collective_graph_key can be passed in when collective ops are used
in variable initialization.
PiperOrigin-RevId:
197964316
Sanjoy Das [Thu, 24 May 2018 22:19:40 +0000 (15:19 -0700)]
Rename getInt64 to GetInt64 to follow Google style
PiperOrigin-RevId:
197963232
A. Unique TensorFlower [Thu, 24 May 2018 21:59:29 +0000 (14:59 -0700)]
Windows build script change for release job
PiperOrigin-RevId:
197959602
A. Unique TensorFlower [Thu, 24 May 2018 21:59:05 +0000 (14:59 -0700)]
Small fix so that GDN can run on TPU
PiperOrigin-RevId:
197959536
Francois Chollet [Thu, 24 May 2018 21:58:15 +0000 (14:58 -0700)]
Raise ValueError when calling model.summary() before it is built
PiperOrigin-RevId:
197959372
Peter Hawkins [Thu, 24 May 2018 21:20:39 +0000 (14:20 -0700)]
[TF:XLA] Avoid buffer copy when copying a Tensor onto an XLA device.
PiperOrigin-RevId:
197952565
Chris Leary [Thu, 24 May 2018 21:03:41 +0000 (14:03 -0700)]
[XLA] Convert infeed call to take a LiteralSlice.
PiperOrigin-RevId:
197949637
Shanqing Cai [Thu, 24 May 2018 21:02:30 +0000 (14:02 -0700)]
tfdbg: fix issue where total source file size exceeds gRPC message size limit
* Source file content is now sent one by one, making it less likely that individual
messages will have sizes above the 4-MB gRPC message size limit.
* In case the message for a single source file exceeds the limit, the client handles
it gracefully by skipping the sending and print a warning message.
Fixes: https://github.com/tensorflow/tensorboard/issues/1118
PiperOrigin-RevId:
197949416
Akshay Agrawal [Thu, 24 May 2018 20:30:15 +0000 (13:30 -0700)]
Fix bugs with the code blocks in defun's docstring.
PiperOrigin-RevId:
197943921
A. Unique TensorFlower [Thu, 24 May 2018 20:19:47 +0000 (13:19 -0700)]
Automated g4 rollback of changelist
197868028
PiperOrigin-RevId:
197942379
A. Unique TensorFlower [Thu, 24 May 2018 20:18:32 +0000 (13:18 -0700)]
add maxpoolgrad transposer for layout optimizer.
PiperOrigin-RevId:
197942180
Amit Patankar [Thu, 24 May 2018 20:15:37 +0000 (13:15 -0700)]
Removing outdated links.
PiperOrigin-RevId:
197941740
A. Unique TensorFlower [Thu, 24 May 2018 20:13:42 +0000 (13:13 -0700)]
Extracts the Simplify Pack optimization into its own method.
PiperOrigin-RevId:
197941474
Nick Felt [Thu, 24 May 2018 20:07:50 +0000 (13:07 -0700)]
Ensure ResourceMgr::LookupOrCreate calls create fn just once
This addresses a race condition where LookupOrCreate is called at the same time from two threads, and both Lookup()s fail, so the creator() function is run twice, even though only a single Create() will then succeed.
The motivation is that some creator() functions have side-effects, e.g. tf.contrib.summary.create_file_writer()'s init op opens an events file. This change ensures that if two init ops for file writers with the same resource name are run in the same session.run() call, only one events file will be created. (Current behavior will often open two files; typically the second one overwrites the first but this won't happen if the filename_suffix values are different or the timestamps happen to straddle a second boundary.)
PiperOrigin-RevId:
197940607
A. Unique TensorFlower [Thu, 24 May 2018 20:03:10 +0000 (13:03 -0700)]
Updated documentation for tf.reduce_join.
PiperOrigin-RevId:
197939808
A. Unique TensorFlower [Thu, 24 May 2018 20:00:07 +0000 (13:00 -0700)]
Only wait for one of the input tensors to be ready.
The waiting was implemented to avoid reading stale models as much as possible.
However with this dependency, each input column creates a Send/Recv to PS0
which slows down training significantly.
Colocate Quantile and Stats accumulators for the same handler.
PiperOrigin-RevId:
197939327
A. Unique TensorFlower [Thu, 24 May 2018 19:34:02 +0000 (12:34 -0700)]
Modify tf.image.central_crop to support batched-input.
Currently central_crop works on singular images with dynamic dimensions. For large image classification models, it would be nice if central_crop can be modified to support batched input. This CL makes that change.
PiperOrigin-RevId:
197935606
Benoit Steiner [Thu, 24 May 2018 19:23:32 +0000 (12:23 -0700)]
Mark queue related ops as having side effect
PiperOrigin-RevId:
197933941
Jacques Pienaar [Thu, 24 May 2018 19:22:04 +0000 (12:22 -0700)]
Don't use hex floats.
Hex float literals are in C11 and C++17, but not in C++11, so use plain float notation.
PiperOrigin-RevId:
197933744
A. Unique TensorFlower [Thu, 24 May 2018 19:11:41 +0000 (12:11 -0700)]
Fix doc: "--input_arrays" instead of "--input_array".
PiperOrigin-RevId:
197932202
Sanjoy Das [Thu, 24 May 2018 18:54:56 +0000 (11:54 -0700)]
[TF:XLA] Bump open source llvm revision to r333167
PiperOrigin-RevId:
197929434
Mark Daoust [Thu, 24 May 2018 18:44:15 +0000 (11:44 -0700)]
Fix `tf_inspect.getargspec` callable objects other than functions.
PiperOrigin-RevId:
197927601
Derek Murray [Thu, 24 May 2018 18:37:12 +0000 (11:37 -0700)]
[tf.data] Add `tf.contrib.data.choose_from_datasets()`.
This is a deterministic counterpart to `tf.contrib.data.sample_from_datasets()`.
PiperOrigin-RevId:
197926386
A. Unique TensorFlower [Thu, 24 May 2018 18:29:34 +0000 (11:29 -0700)]
Extracts the 'Move Constants Past Enter Node' optimization into its own method.
PiperOrigin-RevId:
197924962
Allen Lavoie [Thu, 24 May 2018 18:23:18 +0000 (11:23 -0700)]
Make the existing checkpointable data structure a CheckpointableDataStructure
Gives it better/more consistent handling of Layers.
PiperOrigin-RevId:
197923880
A. Unique TensorFlower [Thu, 24 May 2018 18:18:45 +0000 (11:18 -0700)]
boosted_trees: used double precision instead of single precision while accumulating batches within MakeStatsSummary, as float type faces numerical precision problems when batch gets larger and stats gets smaller.
PiperOrigin-RevId:
197923022
Derek Murray [Thu, 24 May 2018 18:14:17 +0000 (11:14 -0700)]
Deprecate `DeviceBase::GetStepAllocator()` and replace with calls to `GetAllocator()`.
The `GetStepAllocator()` API relied on the existence of a "step resource manager",
which is no longer a concept in the runtime (it was replaced by "step containers").
Since the additional flexibility does not appear to be used in the codebase, the
`GetScopedAllocator()` seems to provide a similar extension point (based on step IDs),
and the `OpKernelContext::get_allocator()` method is called frequently, this change
simplifies the implementation somewhat.
The `GetStepAllocator()` method is retained as a non-virtual stub that forwards to
`GetAllocator()`, because at least one third-party library (libxsmm) calls this
interface directly.
PiperOrigin-RevId:
197922154
Francois Chollet [Thu, 24 May 2018 18:11:42 +0000 (11:11 -0700)]
Add shape validation for symbolic tensors passed to fit (only graph mode).
PiperOrigin-RevId:
197921675
Akshay Agrawal [Thu, 24 May 2018 17:58:47 +0000 (10:58 -0700)]
Fix convert_to_tensor logic in GradientDescentOptimizer's _prepare method
Previously, eagerly executing an optimizer that had been used in a `defun`
led to a cryptic error because the learning rate tensor supplied to the update
op was in fact a vestigial graph Tensor.
PiperOrigin-RevId:
197919104
Nupur Garg [Thu, 24 May 2018 17:53:28 +0000 (10:53 -0700)]
Improve TOCO Python API.
PiperOrigin-RevId:
197918102
A. Unique TensorFlower [Thu, 24 May 2018 17:52:18 +0000 (10:52 -0700)]
Fix build failure introduced by cl/
197457316
PiperOrigin-RevId:
197917867
Alexandre Passos [Thu, 24 May 2018 17:38:48 +0000 (10:38 -0700)]
Warn about tf.Variable semantics
PiperOrigin-RevId:
197915380
Allen Lavoie [Thu, 24 May 2018 17:30:41 +0000 (10:30 -0700)]
Add a checkpointable map data structure
PiperOrigin-RevId:
197913890
Justin Lebar [Thu, 24 May 2018 17:05:10 +0000 (10:05 -0700)]
[XLA] Speed up slice_test again.
Previous patch missed one instance of creating a constant inside of
slice_test.
PiperOrigin-RevId:
197909685
Benjamin Kramer [Thu, 24 May 2018 16:50:19 +0000 (09:50 -0700)]
[XLA] Devectorize constant-sized arrays
A sufficiently smart compiler could promote these from heap to stack, in
practice no compiler does that. Remove the superfluous heap allocations
manually.
PiperOrigin-RevId:
197907388