- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 13 for broadcast_or (0.33 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/tests/replace_cast_hacks_with_tf_xla_ops.mlir
// CHECK: %[[BROADCAST_1:.*]] = "tf.BroadcastTo"(%[[CAST]], %[[CONCAT_1]]) : (tensor<?x1x1024xi8>, tensor<3xi64>) -> tensor<?x1x1024xi8> // CHECK: %[[BROADCAST_2:.*]] = "tf.BroadcastTo"(%[[WEIGHT]], %[[CONCAT_2]]) : (tensor<1024x3xi8>, tensor<3xi64>) -> tensor<?x1024x3xi8> // CHECK: %[[DOT:.*]] = "tf.XlaDotV2"(%[[BROADCAST_1]], %[[BROADCAST_2]]) } // ----- module attributes {} {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 81K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/prepare-tf.mlir
} func.func @broadcast_to_i16_low_dim(%input: tensor<3xi16>, %shape: tensor<2xi32>) -> tensor<3x3xi16> { %0 = "tf.BroadcastTo"(%input, %shape) : (tensor<3xi16>, tensor<2xi32>) -> tensor<3x3xi16> func.return %0: tensor<3x3xi16> // CHECK-LABEL: broadcast_to_i16_low_dim // CHECK: %0 = "tf.BroadcastTo"(%arg0, %arg1) : (tensor<3xi16>, tensor<2xi32>) -> tensor<3x3xi16> // CHECK: return %0 : tensor<3x3xi16> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 07:26:59 UTC 2024 - 59.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_model_test.cc
readonly_model_->subgraphs()->Get(subgraph_idx); // There should be a single broadcast_to op. EXPECT_THAT(*readonly_subgraph->operators(), SizeIs(1)); EXPECT_THAT(subgraph->operators, SizeIs(1)); const auto& broadcast_to = subgraph->operators[0]; EXPECT_THAT(model_.operator_codes[broadcast_to->opcode_index]->builtin_code, Eq(BuiltinOperator_BROADCAST_TO));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 73.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/decompose_resource_ops.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 19:47:48 UTC 2024 - 51.3K bytes - Viewed (0) -
pkg/controller/nodelifecycle/node_lifecycle_controller.go
defer utilruntime.HandleCrash() // Start events processing pipeline. nc.broadcaster.StartStructuredLogging(3) logger := klog.FromContext(ctx) logger.Info("Sending events to api server") nc.broadcaster.StartRecordingToSink( &v1core.EventSinkImpl{ Interface: v1core.New(nc.kubeClient.CoreV1().RESTClient()).Events(""), }) defer nc.broadcaster.Shutdown() // Close node update queue to cleanup go routine.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 51.6K bytes - Viewed (0) -
pkg/controller/podautoscaler/horizontal.go
cpuInitializationPeriod, delayOfInitialReadinessStatus time.Duration, ) *HorizontalController { broadcaster := record.NewBroadcaster(record.WithContext(ctx)) broadcaster.StartStructuredLogging(3) broadcaster.StartRecordingToSink(&v1core.EventSinkImpl{Interface: evtNamespacer.Events("")}) recorder := broadcaster.NewRecorder(scheme.Scheme, v1.EventSource{Component: "horizontal-pod-autoscaler"})
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 63.6K bytes - Viewed (0) -
pkg/controller/job/job_controller.go
defer utilruntime.HandleCrash() logger := klog.FromContext(ctx) // Start events processing pipeline. jm.broadcaster.StartStructuredLogging(3) jm.broadcaster.StartRecordingToSink(&v1core.EventSinkImpl{Interface: jm.kubeClient.CoreV1().Events("")}) defer jm.broadcaster.Shutdown() defer jm.queue.ShutDown() defer jm.orphanQueue.ShutDown() logger.Info("Starting job controller")
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Jun 10 23:56:37 UTC 2024 - 77.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td
(SelectOp $condition, $input2, $input1)>; // Fuse select(broadcast_to(input, shape), x, y) -> selectV2(input, x, y) // Also, fuse selectv2(broadcast_to(input, shape), x, y) -> selectV2(input, x, y) // It is safe to perform this transform here because- // the shapes of `pre_broadcast` and `dim` must be broadcast // compatible for the `broadcast_to` op to be valid.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 20:31:41 UTC 2024 - 66.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/lower_tf.mlir
// CHECK: "tf.BroadcastTo"(%[[ARG1]], %[[ARG0]]) %0 = "tf.Fill"(%arg0, %arg1) : (tensor<*xi64>, tensor<*xf32>) -> tensor<*xf32> func.return %0 : tensor<*xf32> } func.func @empty(%arg0: tensor<?xi32>) -> tensor<*xf32> { // CHECK-DAG: [[CST:%.+]] = "tf.Const"() <{value = dense<0.000000e+00> : tensor<f32>}> // CHECK-DAG: [[RES:%.+]] = "tf.BroadcastTo"([[CST]], %arg0)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 05 18:35:42 UTC 2024 - 92K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tensor_array_ops_decomposition.mlir
// Test read and write on a tensor list. // CHECK-LABEL: func @main func.func @main() -> tensor<3xf32> { %size = "tf.Const"() {value = dense<5> : tensor<i32>} : () -> tensor<i32> // CHECK: %[[BUFFER:.*]] = "tf.BroadcastTo" // CHECK-SAME: -> tensor<5x3xf32> // CHECK: %[[VAR:.*]] = "tf.MlirLocalVarOp"() : () -> tensor<!tf_type.resource<tensor<5x3xf32>>> // CHECK: "tf.AssignVariableOp"(%[[VAR]], %[[BUFFER]])
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 49K bytes - Viewed (0)