Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 137 for Toutput_types (0.27 sec)

  1. tensorflow/compiler/mlir/tensorflow/tests/graphdef2mlir/tf-data-pipeline.pbtxt

          }
        }
      }
    }
    node {
      name: "TensorSliceDataset"
      op: "TensorSliceDataset"
      input: "tensors/normalize_tensors/component_0"
      attr {
        key: "Toutput_types"
        value {
          list {
            type: DT_INT32
          }
        }
      }
      attr {
        key: "output_shapes"
        value {
          list {
            shape {
            }
          }
        }
      }
    }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 29 04:41:05 UTC 2021
    - 4K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/utils/perception_ops_utils_test.cc

      auto indices_type = RankedTensorType::get(input_shape, builder->getI64Type());
      auto output_type = RankedTensorType::get(output_shape, builder->getF32Type());
      SmallVector<mlir::Type, 2> input_types{input_type, indices_type};
      SmallVector<mlir::Type, 1> output_types{output_type};
      return createMaxUnpoolingFunc<2, 1>(builder, input_types, output_types);
    }
    
    template <int N>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Sep 29 21:02:21 UTC 2022
    - 7.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/utils/lstm_utils_test.cc

      auto output_types = fused_lstm_func_.getFunctionType().getResults();
      SmallVector<int64_t, 2> output_shape{1, mlir::ShapedType::kDynamic};
      EXPECT_EQ(mlir::cast<RankedTensorType>(output_types[0]).getShape().size(),
                output_shape.size());
      for (int i = 0; i < output_shape.size(); i++) {
        EXPECT_EQ(mlir::cast<RankedTensorType>(output_types[0]).getDimSize(i),
                  output_shape[i]);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 10K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/transforms/post_quantize.cc

          Value dequantized_result = dequantize_op.getInput();
          output_types.push_back(dequantized_result.getType());
          terminator->setOperand(i, dequantized_result);
          returned_op->erase();
        } else {
          output_types.push_back(returned_value.getType());
        }
      }
      auto new_func_type = builder.getFunctionType(input_types, output_types);
      func.setType(new_func_type);
    }
    
    enum RemoveVolatileOpsType {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 17.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

          shape[split_dimension] = shape[split_dimension] / num_split;
          output_type =
              mlir::RankedTensorType::get(shape, input_type.getElementType());
        }
      } else {
        output_type = input_type;
      }
    
      // Creates a split op that splits |src_input| along |split_dimension|.
      llvm::SmallVector<mlir::Type, 4> output_types(num_split, output_type);
      *split_op = builder->create<mlir::TF::SplitOp>(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/encapsulate_xla_computations_pass.cc

        }
    
        // Outputs.
        const int num_outputs = launch->output_types().size();
        absl::flat_hash_set<Node*> control_outputs;
        std::vector<std::vector<std::pair<Node*, int>>> data_outputs(num_outputs);
        const DataTypeVector& output_types(launch->output_types());
    
        for (const Edge* le : launch->out_edges()) {
          if (le->IsControlEdge()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 12 06:33:33 UTC 2024
    - 15.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/build_xla_ops_pass.cc

      MemoryTypeVector input_mtypes, output_mtypes;
      DeviceType device_type("");
      TF_RETURN_IF_ERROR(
          DeviceNameToDeviceType(n->assigned_device_name(), &device_type));
      TF_RETURN_IF_ERROR(MemoryTypesForNode(root.graph()->op_registry(),
                                            device_type, n->def(), &input_mtypes,
                                            &output_mtypes));
      return output_mtypes;
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 12 06:33:33 UTC 2024
    - 24.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/utils/nms_utils.cc

      Value iou_threshold = func_.getArgument(3);
      Value score_threshold = func_.getArgument(4);
      auto output_type0 = func_.getFunctionType().getResult(0);
      auto output_type1 = func_.getFunctionType().getResult(1);
    
      OpBuilder builder(func_.getBody());
      auto op = builder.create<mlir::TFL::NonMaxSuppressionV4Op>(
          func_.getLoc(), output_type0, output_type1, boxes, scores,
          max_output_size, iou_threshold, score_threshold);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/utils/lstm_utils.cc

          size_of_splits = dim / splits;
          output_shape.push_back(size_of_splits);
        } else {
          output_shape.push_back(dim);
        }
      }
    
      SmallVector<mlir::Type, 4> output_types;
      for (int i = 0; i < splits; ++i) {
        output_types.push_back(
            mlir::RankedTensorType::get(output_shape, input_type.getElementType()));
      }
      auto size_of_splits_op = Create1DConstantOp(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 36.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc

          }
          output_types = {new_output_type};
        } else {
          output_types = {output_type.clone(elem_type.getStorageType())};
        }
    
        SmallVector<Value> args = {q_op.getArg(), scale, zero_point};
        FlatSymbolRefAttr func_name =
            FlatSymbolRefAttr::get(rewriter.getStringAttr(kQuantizeFuncName));
    
        auto quantize_call = rewriter.create<TF::PartitionedCallOp>(
            loc, output_types, args, func_name,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 54.5K bytes
    - Viewed (0)
Back to top