- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for BroadcastGradientArgs (0.53 sec)
-
tensorflow/compiler/mlir/tensorflow/tests/constant-fold.mlir
%s1 = "tf.Const"() {value = dense<[1, 2]> : tensor<2xi32>} : () -> tensor<2xi32> %r0, %r1 = "tf.BroadcastGradientArgs"(%s0, %s1) {} : (tensor<2xi32>, tensor<2xi32>) -> (tensor<0xi32>, tensor<0xi32>) // CHECK-DAG: %[[R:.*]] = "tf.Const"() <{value = dense<> : tensor<0xi32>}> : () -> tensor<0xi32> // CHECK-NOT: tf.BroadcastGradientArgs // CHECK: return %[[R]], %[[R]] func.return %r0, %r1 : tensor<0xi32>, tensor<0xi32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jan 31 23:22:24 UTC 2024 - 36.7K bytes - Viewed (0) -
tensorflow/compiler/jit/tests/opens2s_gnmt_mixed_precision.golden_summary
Clustered nodes: 2385 Unclustered nodes: 4221 Number of clusters: 30 unclustered size 4221 Add 17 AddN 1 All 1 ApplyAdam 38 Assert 7 Assign 47 AssignAdd 2 AssignSub 2 BroadcastGradientArgs 44 Cast 38 ConcatV2 3 Const 875 ControlTrigger 5 Enter 874 Equal 4 Exit 69 ExpandDims 9 Fill 5 FloorMod 1 GreaterEqual 7 Identity 113 IsVariableInitialized 1 IteratorGetNext 1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 06 10:38:14 UTC 2023 - 5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/constant-folding-hook.mlir
module attributes {tf.versions = {producer = 179 : i32}} { func.func @main() -> (tensor<0xi32>, tensor<0xi32>) { %0 = "tf.Const"() {value = dense<[]> : tensor<0xi32>} : () -> tensor<0xi32> %r0, %r1 = "tf.BroadcastGradientArgs"(%0, %0) {T = i32} : (tensor<0xi32>, tensor<0xi32>) -> (tensor<0xi32>, tensor<0xi32>) func.return %r0, %r1 : tensor<0xi32>, tensor<0xi32> } } // CHECK-LABEL: HloModule main
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 23 18:56:13 UTC 2022 - 1.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/examples/mnist/ops_defs.py
elif act == 'TANH': y = math_ops.conj(y) grad = gen_math_ops.tanh_grad(y, grad) broadcast_shape = tf.shape(y) input_value_shape = tf.shape(op.inputs[2]) _, reduction_axes = tf.raw_ops.BroadcastGradientArgs( s0=broadcast_shape, s1=input_value_shape) updates_grad_reshaped = tf.reduce_sum( grad, axis=reduction_axes, keepdims=True) bias_grad = tf.reshape(updates_grad_reshaped, input_value_shape)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Aug 31 20:23:51 UTC 2023 - 6.8K bytes - Viewed (0) -
tensorflow/cc/gradients/math_grad.cc
Output s_max = Shape(scope, max); Output min_mask = Less(scope, x, min); Output max_mask = Greater(scope, x, max); auto r_min = internal::BroadcastGradientArgs(scope, s_x, s_min); auto r_max = internal::BroadcastGradientArgs(scope, s_x, s_max); Output grad = grad_inputs[0]; Output zeros = ZerosLike(scope, grad); Output x_grad =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Aug 25 18:20:20 UTC 2023 - 50.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-prefer-tf2xla.mlir
} // ----- // CHECK-LABEL: func @testBroadcastGradientArgs func.func @testBroadcastGradientArgs(%s0: tensor<4xi32>, %s1: tensor<4xi32>) -> (tensor<1xi32>, tensor<0xi32>) { // CHECK: tf.BroadcastGradientArgs %r0, %r1 = "tf.BroadcastGradientArgs"(%s0, %s1) : (tensor<4xi32>, tensor<4xi32>) -> (tensor<1xi32>, tensor<0xi32>) func.return %r0, %r1 : tensor<1xi32>, tensor<0xi32> } // ----- // CHECK-LABEL: @acos
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 15.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/mark_ops_for_outside_compilation.mlir
%s1 = "tf.Const"() {value = dense<[ 1, 1, 1, 1280]> : tensor<4xi32>} : () -> tensor<4xi32> // CHECK: "tf.BroadcastGradientArgs" // CHECK-NOT: _xla_outside_compilation %r0, %r1 = "tf.BroadcastGradientArgs"(%s0, %s1) {} : (tensor<4xi32>, tensor<4xi32>) -> (tensor<1xi32>, tensor<3xi32>) tf_device.return
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 16:22:32 UTC 2024 - 29.5K bytes - Viewed (0) -
tensorflow/cc/gradients/array_grad.cc
REGISTER_NO_GRADIENT_OP("InvertPermutation"); REGISTER_NO_GRADIENT_OP("Shape"); REGISTER_NO_GRADIENT_OP("ShapeN"); REGISTER_NO_GRADIENT_OP("Rank"); REGISTER_NO_GRADIENT_OP("Size"); REGISTER_NO_GRADIENT_OP("BroadcastGradientArgs"); REGISTER_NO_GRADIENT_OP("OneHot"); Status PackGrad(const Scope& scope, const Operation& op, const std::vector<Output>& grad_inputs, std::vector<Output>* grad_outputs) { int N;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Oct 10 23:33:32 UTC 2023 - 31.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-with-tf2xla-hlo-importer.mlir
} // CHECK-LABEL: func @testBroadcastGradientArgs func.func @testBroadcastGradientArgs(%s0: tensor<4xi32>, %s1: tensor<4xi32>) -> (tensor<1xi32>, tensor<0xi32>) { // CHECK: tf.BroadcastGradientArgs %r0, %r1 = "tf.BroadcastGradientArgs"(%s0, %s1) : (tensor<4xi32>, tensor<4xi32>) -> (tensor<1xi32>, tensor<0xi32>) func.return %r0, %r1 : tensor<1xi32>, tensor<0xi32> } // CHECK-LABEL: @acos
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 38.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tf-ops.mlir
func.return %0 : tensor<3xf32> } // ----- // Test valid tf.BroadcastGradientArgs // CHECK-LABEL: func @testBroadcastGradientArgs func.func @testBroadcastGradientArgs(%s0: tensor<4xi32>, %s1: tensor<4xi32>) -> (tensor<1xi32>, tensor<0xi32>) { %r0, %r1 = "tf.BroadcastGradientArgs"(%s0, %s1) : (tensor<4xi32>, tensor<4xi32>) -> (tensor<1xi32>, tensor<0xi32>)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 23 14:40:35 UTC 2023 - 236.4K bytes - Viewed (0)