Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 80 for created (0.16 sec)

  1. tensorflow/c/eager/c_api_unified_experimental.cc

    using tensorflow::tracing::TracingTensorHandle;
    
    void TF_SetTracingImplementation(const char* name, TF_Status* s) {
      tsl::Set_TF_Status_from_Status(s, SetDefaultTracingEngine(name));
    }
    
    // Creates a new TensorFlow function, it is an execution context attached to a
    // given tracing context.
    TF_ExecutionContext* TF_CreateFunction(const char* fn_name, TF_Status* s) {
      return wrap(CreateTracingExecutionContext(fn_name, s));
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 10:15:17 UTC 2024
    - 9K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/debugging/dump_tensor_op.cc

    }
    
    // `DumpTensor` op saves entire value of input to as a tensor proto into a
    // specified directory and filename. When enabled is set to false, op is
    // disabled and won't save any value. It also creates `QuantizationUnit` proto
    // with `func_name` and `node_name` to identify the op.
    REGISTER_OP("DumpTensor")
        .Input("tensor_data: T")
        .Attr("log_dir_path: string")
        .Attr("file_name: string")
        .Attr("T: type")
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 03:12:17 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.cc

        module_op.emitError() << "quant-convert-tpu-model-to-cpu pattern "
                                 "conversion did not converge.";
        signalPassFailure();
        return;
      }
    }
    
    }  // namespace
    
    // Creates an instance of `ConvertTpuModelToCpuPass`.
    std::unique_ptr<OperationPass<ModuleOp>> CreateConvertTpuModelToCpuPass() {
      return std::make_unique<ConvertTpuModelToCpuPass>();
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/tests/device_compiler_test_helper.cc

    #include "tensorflow/compiler/jit/xla_compilation_cache.pb.h"
    #include "xla/service/hlo.pb.h"
    #include "tensorflow/core/platform/path.h"
    #include "tensorflow/core/public/session.h"
    
    namespace tensorflow {
    namespace {
    
    // Creates a float tensor of linearly increasing values, starting from offset.
    Tensor CreateInputTensor(const TensorShape& shape, float offset) {
      Tensor tensor(DT_FLOAT, shape);
      for (int64 i = 0; i < tensor.flat<float>().size(); ++i) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Feb 09 08:24:16 UTC 2024
    - 6.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/stablehlo/passes/prepare_quantize.cc

                ctx);
        if (failed(applyPatternsAndFoldGreedily(func_op, std::move(patterns_2)))) {
          signalPassFailure();
        }
      }
    }
    
    }  // namespace
    
    // Creates an instance of the TensorFlow dialect PrepareQuantize pass.
    std::unique_ptr<OperationPass<ModuleOp>> CreatePrepareQuantizePass(
        const bool enable_per_channel_quantized_weight, const int bit_width) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 03 05:11:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/stablehlo/passes/lift_quantizable_spots_as_functions.cc

        if (failed(ApplyQuantizationSpec(spec, module_op))) {
          signalPassFailure();
          return;
        }
      }
    }
    
    }  // namespace
    
    // Creates `LiftQuantizableSpotsAsFunctionsPass` with user-defined
    // `QuantizationSpecs`.
    std::unique_ptr<OperationPass<ModuleOp>>
    CreateLiftQuantizableSpotsAsFunctionsPass(
        const QuantizationSpecs& quantization_specs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 9.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_quantized_functions.cc

                                   "attribute in the quantized composite function.";
          signalPassFailure();
        }
      }
    }
    
    }  // namespace
    
    // Creates an instance of the pass for inserting quantized functions.
    std::unique_ptr<OperationPass<ModuleOp>> CreateInsertQuantizedFunctionsPass(
        QuantMethod quantization_method, OpSet target_opset) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 05:52:39 UTC 2024
    - 8.7K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/transforms/modify_io_nodes.cc

        return;
      }
    
      auto new_func_type =
          builder.getFunctionType(new_input_types, new_output_types);
      func.setType(new_func_type);
    }
    }  // namespace
    
    // Creates an instance of the TensorFlow Lite modify io nodes pass.
    std::unique_ptr<OperationPass<func::FuncOp>> CreateModifyIONodesPass(
        Type input_type, Type output_type) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/transforms/xla_call_module_serialization.cc

        return WalkResult::advance();
      });
      if (result.wasInterrupted()) {
        return failure();
      }
    
      return builder.getArrayAttr(function_list);
    }
    
    // Creates a pruned module containing the XlaCallModule's entry function and
    // other functions transitively called by the entry function.
    FailureOr<OwningOpRef<ModuleOp>> PruneStablehloModule(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/transforms/default_quant_params.cc

      } else {
        builder.setInsertionPointToStart(&block);
      }
      TypeAttr type_attr = TypeAttr::get(new_type);
      auto quantize = builder.create<TFL::QuantizeOp>(value.getLoc(), new_type,
                                                      value, type_attr);
      auto dequantize = builder.create<TFL::DequantizeOp>(
          value.getLoc(), expressed_type, quantize.getOutput());
      value.replaceAllUsesWith(dequantize);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.4K bytes
    - Viewed (0)
Back to top