Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 37 for more (0.04 sec)

  1. tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.h

    #include "absl/status/status.h"
    #include "absl/types/variant.h"
    #include "xla/client/compile_only_client.h"
    #include "xla/pjrt/compile_options.pb.h"
    #include "tensorflow/core/framework/tensor_shape.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile.pb.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace v1 {
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 13 08:08:57 UTC 2024
    - 2.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/utils/const_tensor_utils.h

    #include "mlir/IR/Builders.h"  // from @llvm-project
    #include "mlir/IR/BuiltinTypes.h"  // from @llvm-project
    #include "tensorflow/compiler/mlir/lite/schema/schema_generated.h"
    #include "tensorflow/core/framework/tensor.pb.h"
    #include "tensorflow/core/framework/tensor_shape.pb.h"
    
    namespace mlir {
    namespace TFL {
    
    bool IsQuantized(const tflite::TensorT& tensor);
    
    absl::StatusOr<mlir::quant::QuantizedType> GetQuantizedType(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 07 23:04:40 UTC 2024
    - 2.9K bytes
    - Viewed (0)
  3. tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util.h

    #define TENSORFLOW_C_EXPERIMENTAL_NEXT_PLUGGABLE_DEVICE_TENSOR_PJRT_BUFFER_UTIL_H_
    
    #include "xla/pjrt/c/pjrt_c_api.h"
    #include "xla/pjrt/pjrt_c_api_client.h"
    #include "tensorflow/core/framework/tensor.h"
    
    namespace tensorflow {
    
    absl::StatusOr<PJRT_Buffer*> GetPjRtCBufferFromTensor(const Tensor* tensor);
    
    absl::Status SetPjRtCBufferToTensor(PJRT_Buffer* c_buffer,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 05:48:24 UTC 2024
    - 1.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/quantize_preprocess.h

    #include "absl/strings/string_view.h"
    #include "mlir/IR/BuiltinOps.h"  // from @llvm-project
    #include "mlir/IR/MLIRContext.h"  // from @llvm-project
    #include "mlir/Pass/PassManager.h"  // from @llvm-project
    #include "tensorflow/core/public/session.h"
    
    namespace tensorflow {
    namespace quantization {
    
    // Default MLIR dump file prefix for TensorFlow quantization passes.
    inline constexpr absl::string_view kDefaultTfQuantMlirDumpFilePrefix =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 12:49:45 UTC 2024
    - 3.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/aot/embedded_protocol_buffers.h

    #ifndef TENSORFLOW_COMPILER_AOT_EMBEDDED_PROTOCOL_BUFFERS_H_
    #define TENSORFLOW_COMPILER_AOT_EMBEDDED_PROTOCOL_BUFFERS_H_
    
    #include "absl/types/span.h"
    #include "xla/statusor.h"
    #include "tensorflow/core/platform/protobuf.h"
    
    namespace tensorflow {
    namespace tfcompile {
    using absl::StatusOr;
    
    // Represents a set of protocol buffers embedded into an object file and
    // describes how to access them at runtime.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 10 18:19:50 UTC 2024
    - 3.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/python/converter_python_api.h

    namespace tflite {
    
    // Convert a model represented in `input_contents`. `model_flags_proto`
    // describes model parameters. `flags_proto` describes conversion
    // parameters (see relevant .protos for more information). Returns a string
    // representing the contents of the converted model. When extended_return
    // flag is set to true returns a dictionary that contains string representation
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/translate/import_model.h

    #include "tensorflow/compiler/mlir/tensorflow/translate/mlir_roundtrip_flags.h"
    #include "tensorflow/core/framework/function.h"
    #include "tensorflow/core/framework/graph.pb.h"
    #include "tensorflow/core/framework/graph_debug_info.pb.h"
    #include "tensorflow/core/graph/graph.h"
    
    namespace tensorflow {
    
    inline constexpr absl::string_view kImportModelDefaultGraphFuncName = "main";
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 01 11:17:36 UTC 2024
    - 6.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/utils/convert_attr.h

    #define TENSORFLOW_COMPILER_MLIR_TENSORFLOW_UTILS_CONVERT_ATTR_H_
    
    #include "mlir/IR/Attributes.h"  // from @llvm-project
    #include "mlir/IR/Builders.h"  // from @llvm-project
    #include "tensorflow/core/framework/attr_value.pb.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    
    using tsl::StatusOr;
    
    // Converts non func AttrValue proto into an MLIR attribute. Func attribute is
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 1.6K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.h

    #include "llvm/ADT/StringRef.h"
    #include "mlir/Pass/Pass.h"  // from @llvm-project
    #include "tensorflow/compiler/tf2xla/xla_helpers.h"
    #include "xla/client/compile_only_client.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace internal {
    
    // Legalize the given MLIR module to XLA HLO using a combination of the MLIR
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.h

    #include <string>
    #include <vector>
    
    #include "llvm/ADT/StringRef.h"
    #include "mlir/Pass/Pass.h"  // from @llvm-project
    #include "tensorflow/compiler/tf2xla/xla_helpers.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace internal {
    
    // Compiles a serialized MLIR module and returns a serialized MLIR module of the
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 2.8K bytes
    - Viewed (0)
Back to top