Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 23 for _backprop (3.81 sec)

  1. samples/bookinfo/src/productpage/static/tailwind/tailwind.css

    s":" ","--tw-backdrop-contrast":" ","--tw-backdrop-grayscale":" ","--tw-backdrop-hue-rotate":" ","--tw-backdrop-invert":" ","--tw-backdrop-opacity":" ","--tw-backdrop-saturate":" ","--tw-backdrop-sepia":" "}),e({".backdrop-filter":{"@defaults backdrop-filter":{},"backdrop-filter":Fe},".backdrop-filter-none":{"backdrop-filter":"none"}})},transitionProperty:({matchUtilities:i,theme:e})=>{let t=e("transitionTimingFunction.DEFAULT"),r=e("transitionDuration.DEFAULT");i({transition:n=>({"transition-pr...
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Tue May 28 14:48:01 UTC 2024
    - 357.1K bytes
    - Viewed (1)
  2. tensorflow/compiler/mlir/lite/stablehlo/transforms/fuse_convolution_pass.cc

          });
        }
        filter_value = filter.getValue();
        mul_value = multiplier.getValue();
        // In MHLO, Conv filter is in HWIO format, Depthwise conv filter is in HW1O
        // format and backprop input conv filter is in HWOI format.
        // Only fuses multiplier if all dimensions other than the out channel
        // dimension are equal to 1.
        if (!TFL::IsDimensionsDegenerateExceptLastOne(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 22:21:19 UTC 2024
    - 8.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/tpu_space_to_depth_pass.mlir

        // CHECK: %[[BACKPROP:.*]] = "tf.Conv2DBackpropFilter"
        // CHECK-SAME: strides = [1, 1, 1, 1]
        // CHECK-SAME: (tensor<2x115x115x12xf32>, tensor<4xi32>, tensor<2x112x112x64xf32>) -> tensor<4x4x12x64xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 37.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir

      // CHECK-NEXT: %[[offset_backprop:.*]] = mhlo.convert %[[red2]] : tensor<8xf32>
    
      // CHECK-NEXT: %[[x_backprop:.*]] = mhlo.convert %[[mul3]] : tensor<8x8x8x8xf32>
      // CHECK-NEXT: return %[[x_backprop]] : tensor<8x8x8x8xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 335.5K bytes
    - Viewed (0)
  5. tensorflow/c/while_loop_test.cc

          Add(params_->body_inputs[0], {one, 0}, params_->body_graph, s_);
      ASSERT_EQ(TF_OK, TF_GetCode(s_)) << TF_Message(s_);
      params_->body_outputs[0] = {add, 0};
    
      ExpectOK();
    
      // Create backprop graph
      TF_Output grad_output;
      TF_AddGradients(graph_, outputs_.data(), outputs_.size(), inputs_.data(), 1,
                      nullptr, s_, &grad_output);
      ASSERT_EQ(TF_OK, TF_GetCode(s_)) << TF_Message(s_);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 11 06:05:56 UTC 2024
    - 15.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

          Value scratch2 =
              ApplyReduction(loc, weighted_grad, reduce_dims, &rewriter);
    
          // x_backprop = y_backprop * (scale * scratch1)
          auto scaled_grad =
              rewriter.create<mhlo::MulOp>(loc, op.getScale(), scratch1);
          x_backprop = rewriter.create<mhlo::MulOp>(
              loc, grad,
              Broadcast1DToFeatureDim(loc, act, scaled_grad, feature_dim,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

    example, suppose y = f(x) and we wish to apply a custom function g for backprop
    such that dx = g(dy). In Python,
    
    ```python
    with tf.get_default_graph().gradient_override_map(
        {'IdentityN': 'OverrideGradientWithG'}):
      y, _ = identity_n([f(x), x])
    
    @tf.RegisterGradient('OverrideGradientWithG')
    def ApplyG(op, dy, _):
      return [None, g(dy)]  # Do not backprop to f(x).
    ```
      }];
    
      let arguments = (ins
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/tests/lower_tf.mlir

      // CHECK-DAG: %[[SOFTMAX:.*]] = "tf.Div"(%[[SOFTMAX_EXP]], %[[SOFTMAX_SUM]]) : (tensor<2x3xf32>, tensor<2x1xf32>) -> tensor<2x3xf32>
    
      // CHECK-DAG: %[[BACKPROP:.*]] = "tf.Sub"(%[[SOFTMAX]], %[[LABELS]]) : (tensor<2x3xf32>, tensor<2x3xf32>) -> tensor<2x3xf32>
      // CHECK: return %[[LOSS]], %[[BACKPROP]]
    
      %0:2 = "tf.SoftmaxCrossEntropyWithLogits"(%features, %labels) : (tensor<2x3xf32>, tensor<2x3xf32>) -> (tensor<2xf32>, tensor<2x3xf32>)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 92K bytes
    - Viewed (0)
  9. RELEASE.md

            *   `tf.compat.v1.nn.fused_batch_norm` backprop to `offset` when
                `is_training=False`
            *   `tf.image.adjust_contrast` forward
            *   `tf.nn.depthwise_conv2d` backprop to `filter` when not using cuDNN
                convolution
            *   `tf.image.resize` with `method=ResizeMethod.NEAREST` backprop
            *   `tf.math.bincount` - TODO: confirm exception added
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td

      let description = [{
        Automatic space to depth transform is done by adding space to depth transform op after host input
        and applying space to depth transform for the first convolution and its backprop filter on TPU.
    
        For example, original program:
    
        ```mlir
        module {
          func @while_body {
            %input = "tf.IteratorGetNext"(...) {device = "/CPU:0"}: -> tensor<2x224x224x3xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 99.6K bytes
    - Viewed (0)
Back to top