Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 1,151 for shake (0.07 sec)

  1. build-logic/packaging/src/main/kotlin/gradlebuild.shaded-jar.gradle.kts

    import gradlebuild.basics.decapitalize
    import gradlebuild.shade.ArtifactTypes.buildReceiptType
    import gradlebuild.shade.ArtifactTypes.classTreesType
    import gradlebuild.shade.ArtifactTypes.entryPointsType
    import gradlebuild.shade.ArtifactTypes.manifestsType
    import gradlebuild.shade.ArtifactTypes.relocatedClassesAndAnalysisType
    import gradlebuild.shade.ArtifactTypes.relocatedClassesType
    import gradlebuild.shade.extension.ShadedJarExtension
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Sat Sep 30 16:17:28 UTC 2023
    - 8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util.cc

    // Extracts shape from XlaArgument as TensorShape. If shape is a xla::Shape,
    // that is converted to a TensorShape.
    absl::StatusOr<TensorShape> GetTensorShapeFromXlaArgument(
        const XlaArgument& arg) {
      if (absl::holds_alternative<xla::Shape>(arg.shape)) {
        TensorShape arg_shape;
        TF_RETURN_IF_ERROR(
            XLAShapeToTensorShape(std::get<xla::Shape>(arg.shape), &arg_shape));
        return arg_shape;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 17:24:39 UTC 2024
    - 45.3K bytes
    - Viewed (0)
  3. tensorflow/c/kernels/ops/bitcast.cc

      if (input_type_size < output_type_size) {
        TF_ShapeInferenceContextWithRankAtLeast(ctx, shape, 1, shape, status);
    
        if (TF_GetCode(status) == TF_OK) {
          TF_DimensionHandle* last_dim = TF_NewDimensionHandle();
          size_t divisor_val = output_type_size / input_type_size;
          TF_ShapeInferenceContextDim(ctx, shape, -1, last_dim);
          if (!TF_DimensionHandleValueKnown(last_dim) ||
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 07:51:50 UTC 2024
    - 5.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/lstm.json

        {
          "tensors": [
            {
              "shape": [1, 5, 2],
              "name": "input0"
            },
            {
              "shape": [2, 5],
              "buffer": 1,
              "name": "input2input_weights1"
            },
            {
              "shape": [2, 5],
              "buffer": 2,
              "name": "input2forget_weights2"
            },
            {
              "shape": [2, 5],
              "buffer": 3,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 01 06:25:50 UTC 2024
    - 9.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/shape_inference.cc

        // Merge node causes a loop so we remove NextIteration->Merge edge before
        // performing shape inference. But removing those edges also prevents us
        // from inferring output shape for Merge node (we need shapes for all its
        // inputs).
        // For loop invariant resource input's Merge node, we set output resource
        // shape as Enter node's resource shape.
        // TODO(b/129367850): clean this up.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 00:41:19 UTC 2024
    - 13K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/xla_host_send_device_context.h

    //  se::DeviceMemoryBase gpu_dst{device_tensor.data(), 4 * sizeof(float)};
    //  xla::Shape shape(xla::F32, {2, 2}, {}, {})
    //  tsl::AsyncValueRef<std::unique_ptr<se::Event>> done_event =
    //      tsl::MakeConstructedAsyncValueRef<std::unique_ptr<se::Event>>(stream.parent());
    //  done_event->Init();
    //
    //  XlaHostSendDeviceContext device_context(&stream, &gpu_dst,
    //    shape, done_event);
    //  device_context.CopyCPUTensorToDeviceSync(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 3.7K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tfr/tests/end2end.mlir

    // CHECK-NEXT: %[[SHAPE:.*]] = "tf.RiscShape"(%arg0) {T = i32} : (tensor<2x3xf32>) -> tensor<*xi32>
    // CHECK-NEXT: %[[ALPHA1:.*]] = "tf.RiscBroadcast"(%[[ALPHA]], %[[SHAPE]]) : (tensor<f32>, tensor<*xi32>) -> tensor<*xf32>
    // CHECK-NEXT: %[[MAX:.*]] = "tf.RiscMaximum"(%arg0, %[[ALPHA1]]) : (tensor<2x3xf32>, tensor<*xf32>) -> tensor<*xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 13.4K bytes
    - Viewed (0)
  8. tensorflow/c/eager/abstract_tensor_handle.cc

    namespace tensorflow {
    
    std::string AbstractTensorHandle::DebugString() const {
      PartialTensorShape shape;
      Status s = Shape(&shape);
      std::string shape_string;
      if (!s.ok()) {
        shape_string = "<error computing shape>";
      } else {
        shape_string = shape.DebugString();
      }
      return absl::StrCat("TensorHandle(shape=", shape_string,
                          ", dtype=", DataType_Name(DataType()),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 15 09:49:45 UTC 2024
    - 1.4K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/prepare_tpu_computation_for_tf_export.mlir

      // CHECK-SAME: recv_key = "host_compute_channel_recv"
      // CHECK-SAME: send_key = "host_compute_channel_send"
      // CHECK-SAME: shape_inference_graph = @host_func
      // CHECK-SAME: shapes = [#tf_type.shape<*>, #tf_type.shape<3x?>]
      // CHECK-SAME: tpu_core = 0 : i64
      // CHECK: func @host_func
      // CHECK: %[[RECV_OUTPUT:[0-9]*]]:2 = "tf._XlaRecvAtHost"
      // CHECK-SAME: key = "host_compute_channel_send"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 14 18:46:36 UTC 2024
    - 9.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/utils/convert_tensor.h

    // Converts a shape from MLIR to a TensorFlow tensor shape proto.
    void ConvertToTensorShapeProto(llvm::ArrayRef<int64_t> shape,
                                   TensorShapeProto* output_shape);
    
    // Converts an MLIR type to a TensorFlow tensor shape.
    PartialTensorShape ConvertTypeToTensorShape(const mlir::Type& type);
    
    // Converts an MLIR shaped type to a TensorFlow shape attribute.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 2.9K bytes
    - Viewed (0)
Back to top