Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 58 of 58 for do (0.02 sec)

  1. tensorflow/cc/framework/while_gradients_test.cc

          [](const Scope& s, const std::vector<Output>& inputs,
             std::vector<Output>* outputs) {
            // Use AddN, rather than Add, because the gradient function doesn't
            // depend on the input shapes, and thus we do not need to store
            // intermediate values in a stack.
            outputs->push_back(ops::AddN(s, {inputs[0], 1}));
            return s.status();
          });
      CreateBackprop();
    
      Run<int>({1}, {1});
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 27 20:32:17 UTC 2017
    - 7.7K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/xla_rewrite.cc

          non_resource_args.push_back(arg);
          if (has_resources) in_order = false;
        } else {
          resource_args.push_back(arg);
          has_resources = true;
        }
      }
    
      if (!in_order) {
        // Functions do not get reused in practice, so skip the check for if the
        // callee has been updated.
        StringAttr callee_sym = cluster_func_op.getFuncAttr().getAttr();
        MoveResourceArgsToEnd(symtab.lookup<func::FuncOp>(callee_sym));
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 4.2K bytes
    - Viewed (0)
  3. tensorflow/c/experimental/saved_model/core/revived_types/variable.cc

          dtype_(dtype),
          shape_(shape),
          ctx_(ctx) {}
    
    Variable::~Variable() {
      // If the handle is null (perhaps because variable was std::moved from), then
      // we don't have to do anything.
      if (handle_ == nullptr) {
        return;
      }
    
      Status status = internal::DestroyResource(ctx_, handle_.get());
      if (!status.ok()) {
        LOG(ERROR) << "Error destroying variable: " << name_
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Oct 08 20:55:40 UTC 2020
    - 4.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/passes/unwrap_xla_call_module_op.cc

      explicit UnwrapXlaCallModuleOpPass() = default;
    
     private:
      void runOnOperation() override;
    };
    
    void UnwrapXlaCallModuleOp(TF::XlaCallModuleOp call_op,
                               SymbolTable& symbol_table) {
      // Do not inline lifted quantized functions used for fusing patterns.
      // TODO - b/310539922: Remove reference to TF/TFL utils.
      if (call_op->hasAttr(kQuantTraitAttrName)) {
        return;
      }
    
      auto function_name = call_op
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 05 07:39:40 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/cluster_scoping_pass.cc

      starts.push_back(start);
      auto enter = [&](Node* n) { AddOrAppendXlaInternalScope(n, unique_suffix); };
      DFSFrom(*graph_, starts, enter, /*leave=*/nullptr,
              /*stable_comparator=*/NodeComparatorName(),
              // Do not filter any edges to better capture the semantics of
              // transitive closure of successors.  We may revisit this when
              // we see more cases needing cluster scoping in the future.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 5.7K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/transforms/constant_fold.cc

    // Policy:
    //
    // Disable constant folding if operands size is greater than a certain
    // threshold (`kOperandsSizeThreshold`).
    //
    // Otherwise, allow folding if we do not know the shape of an operand or
    // result i.e., one of these values has non-static shape. If we know all the
    // shapes, find the total size of the operands and results. Folding of the op is
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/transforms/tpu_reorder_replicate_and_partitioned_inputs.cc

        const std::optional<::llvm::StringRef> op_xla_sharding =
            partitioned_input.get_XlaSharding();
        const auto op_partition_dims = partitioned_input.getPartitionDims();
        // Abort if TPUPartitionedInputV2(s) do not have the same attributes.
        if (!llvm::equal(partition_dims, op_partition_dims)) {
          return partitioned_input->emitOpError()
                 << "expects partition_dims = " << partition_dims << " but found "
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jan 24 23:08:55 UTC 2023
    - 7.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/ir/tf_ops_tensor_helper.cc

          int64_t &common_dim = common_dims[i];
          if (common_dim == ShapedType::kDynamic) {
            common_dim = dim;
          } else if (common_dim != dim) {
            // If mask_one_dim is true, do not emit an error if this is the only
            // dimension with mismatches. Note down the dimension to mask it from
            // the following types.
            if (mask_one_dim && dim_to_mask == ShapedType::kDynamic) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.7K bytes
    - Viewed (0)
Back to top