Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 529 for shade (0.08 sec)

  1. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/lstm.json

        {
          "tensors": [
            {
              "shape": [1, 5, 2],
              "name": "input0"
            },
            {
              "shape": [2, 5],
              "buffer": 1,
              "name": "input2input_weights1"
            },
            {
              "shape": [2, 5],
              "buffer": 2,
              "name": "input2forget_weights2"
            },
            {
              "shape": [2, 5],
              "buffer": 3,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 01 06:25:50 UTC 2024
    - 9.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/shape_inference.cc

        // Merge node causes a loop so we remove NextIteration->Merge edge before
        // performing shape inference. But removing those edges also prevents us
        // from inferring output shape for Merge node (we need shapes for all its
        // inputs).
        // For loop invariant resource input's Merge node, we set output resource
        // shape as Enter node's resource shape.
        // TODO(b/129367850): clean this up.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 00:41:19 UTC 2024
    - 13K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_host_send_device_context.h

    //  se::DeviceMemoryBase gpu_dst{device_tensor.data(), 4 * sizeof(float)};
    //  xla::Shape shape(xla::F32, {2, 2}, {}, {})
    //  tsl::AsyncValueRef<std::unique_ptr<se::Event>> done_event =
    //      tsl::MakeConstructedAsyncValueRef<std::unique_ptr<se::Event>>(stream.parent());
    //  done_event->Init();
    //
    //  XlaHostSendDeviceContext device_context(&stream, &gpu_dst,
    //    shape, done_event);
    //  device_context.CopyCPUTensorToDeviceSync(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 3.7K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/utils/convert_tensor.h

    // Converts a shape from MLIR to a TensorFlow tensor shape proto.
    void ConvertToTensorShapeProto(llvm::ArrayRef<int64_t> shape,
                                   TensorShapeProto* output_shape);
    
    // Converts an MLIR type to a TensorFlow tensor shape.
    PartialTensorShape ConvertTypeToTensorShape(const mlir::Type& type);
    
    // Converts an MLIR shaped type to a TensorFlow shape attribute.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 2.9K bytes
    - Viewed (0)
  5. src/runtime/pprof/protomem_test.go

    	const expectedLocation = "runtime/pprof.nonRecursiveGenericAllocFunction[go.shape.struct {},go.shape.struct { runtime/pprof.buf [128]uint8 }];runtime/pprof.nonRecursiveGenericAllocFunction[go.shape.struct...
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue May 21 14:38:45 UTC 2024
    - 6.7K bytes
    - Viewed (0)
  6. src/math/big/arith_decl.go

    // Notable members of the hall of shame include:
    //   - github.com/remyoudompheng/bigfft
    //
    // Do not remove or change the type signature.
    // See go.dev/issue/67401.
    //
    //go:linkname addVV
    //go:noescape
    func addVV(z, x, y []Word) (c Word)
    
    // subVV should be an internal detail,
    // but widely used packages access it using linkname.
    // Notable members of the hall of shame include:
    //   - github.com/remyoudompheng/bigfft
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:15:13 UTC 2024
    - 2.6K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/replace_stablehlo_ops_in_main_function_with_xla_call_module_ops.mlir

      // CHECK: %[[CUSTOM_AGGREGATOR_2:.*]], {{.*}}, {{.*}}, {{.*}} = "tf.CustomAggregator"(%[[XLA_CALL_MODULE:.*]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 01:09:50 UTC 2024
    - 39.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/xla_tpu_device.cc

    // Given a tensor of `shape` and `type`, as what shape should it be stored on
    // the TPU device? This function tranposes or flattens the excessively-padded
    // tensors to rank 1, but leaves other tensor shapes alone.
    absl::StatusOr<xla::Shape> TpuShapeRepresentation(
        const TensorShape& shape, DataType type, bool use_fast_memory,
        XlaLayoutPreference layout_preference) {
      xla::Shape xla_shape;
      TF_RETURN_IF_ERROR(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 22:53:47 UTC 2024
    - 20.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/transforms/shape_inference.cc

      }
    
      SmallVector<int64_t> shape;
      bool refined_shape = false;
      // Build the shape of the refined type, if lhs is unranked it
      // will be directly the shape of the refined type, otherwise we merged by
      // taking the most specialized. This combines `10x?x?` and `?x?x8` into
      // `10x?x8`.
      if (!lhs_shape_type.hasRank()) {
        if (rhs_shape_type.hasRank()) {
          shape.append(rhs_shape_type.getShape().begin(),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Jun 08 07:28:49 UTC 2024
    - 134.1K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/shape_inference_test.cc

      auto c = ops::Add(root.WithOpName("C"), a, b);
      auto d = ops::Neg(root.WithOpName("D"), c);
    
      a.node()->AddAttr("_index", 0);
      b.node()->AddAttr("_index", 1);
    
      std::unique_ptr<Graph> graph(new Graph(OpRegistry::Global()));
      TF_CHECK_OK(root.ToGraph(graph.get()));
    
      std::map<int, InferredShape> arg_shapes;
      arg_shapes[0].shape = TensorShape({2, 3});
      arg_shapes[1].shape = TensorShape({2, 3});
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 00:41:19 UTC 2024
    - 10.3K bytes
    - Viewed (0)
Back to top