1. 26 2月, 2016 9 次提交
    • A
      Fix an error message in tf.sparse_to_dense to include the possibility that... · 73d557cc
      A. Unique TensorFlower 提交于
      Fix an error message in tf.sparse_to_dense to include the possibility that indices are invalid because they are out of bounds.
      Change: 115522264
      73d557cc
    • E
      Added TestReporter and test / benchmark reporting tools. · fcfa866d
      Eugene Brevdo 提交于
      These tools are meant to allow recording of benchmark & unit test
      structured output to pbtxt files in a directory only when the
      environment variable TEST_REPORT_FILE_PREFIX is set.  For now,
      only saving of C++  microbenchmark output is supported.
      Change: 115518303
      fcfa866d
    • S
      Added unit test for max_to_keep being None. · 4ecd2a70
      Sherry Moore 提交于
      Change: 115516426
      4ecd2a70
    • K
      Move all Tensorflow WORKSPACE rules to a skylark macro · 77da168d
      Kiril Gorovoy 提交于
      Change: 115515678
      77da168d
    • J
      Remove no-longer-needed RequireDefaultOps(). · 9ba55d8a
      Josh Levenberg 提交于
      Change: 115511835
      9ba55d8a
    • J
      Remove no-longer-needed RequireDefaultOps(). · ab286e09
      Josh Levenberg 提交于
      Change: 115511794
      ab286e09
    • V
      Switch nn.moments() to using a one-pass stable algorithm. · bce62166
      Vincent Vanhoucke 提交于
      Helps with: https://github.com/tensorflow/tensorflow/issues/917
      Also fixes https://github.com/tensorflow/tensorflow/issues/1162
      
      The main benefit is that the computation of the sufficient statistics is now decoupled of the aggregation of the moments, which means that if you want to perform the accumulation incrementally, you don't have to keep all the inputs around, and can instead keep the much more compact sum and sum-of-squares. Accumulation could also be performed locally if you aggregate across multiple devices.
      Computing sum and sum-of-squares can also theoretically be performed in parallel now.
      
      Tested running inception: same performance, same step time.
      Batch normalization benchmark is a bit faster on CPU, a bit slower on GPU:
      
      Before:
      cpu shape:4/3 #layers:10 mode:py scale:True train:False - 1.139310 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.021970 secs
      cpu shape:4/3 #layers:10 mode:py scale:True train:True - 2.767147 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:True - 0.074531 secs
      cpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.742835 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.013473 secs
      cpu shape:4/3 #layers:10 mode:py scale:True train:True - 1.738806 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:True - 0.052777 secs
      cpu shape:2/1 #layers:10 mode:py scale:True train:False - 0.119180 secs
      gpu shape:2/1 #layers:10 mode:py scale:True train:False - 0.011201 secs
      cpu shape:2/1 #layers:10 mode:py scale:True train:True - 0.218297 secs
      gpu shape:2/1 #layers:10 mode:py scale:True train:True - 0.048526 secs
      
      After:
      cpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.998944 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.025828 secs
      cpu shape:4/3 #layers:10 mode:py scale:True train:True - 2.657428 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:True - 0.086614 secs
      cpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.603137 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:False - 0.017668 secs
      cpu shape:4/3 #layers:10 mode:py scale:True train:True - 1.519533 secs
      gpu shape:4/3 #layers:10 mode:py scale:True train:True - 0.055214 secs
      cpu shape:2/1 #layers:10 mode:py scale:True train:False - 0.071344 secs
      gpu shape:2/1 #layers:10 mode:py scale:True train:False - 0.016440 secs
      cpu shape:2/1 #layers:10 mode:py scale:True train:True - 0.222093 secs
      gpu shape:2/1 #layers:10 mode:py scale:True train:True - 0.039967 secs
      Change: 115507032
      bce62166
    • J
      Execute TODO to explain graph-consumer usage of · 2cc5ed87
      Josh Levenberg 提交于
      RemoveNewDefaultAttrsFromGraphDef().
      Change: 115506523
      2cc5ed87
    • A
      Switch sdca_ops to use tf.load_library mechanism. · 8041c546
      A. Unique TensorFlower 提交于
      Change: 115505008
      8041c546
  2. 25 2月, 2016 18 次提交
  3. 24 2月, 2016 13 次提交