- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 37 for Implementation (0.16 sec)
-
tensorflow/compiler/jit/device_compiler.h
} DeviceCompilerClient<ExecutableType, ClientType>* compiler_client() { return compiler_client_.get(); } string DebugString() const override; private: // Common implementation of Compile and CompileSingleOp. The `OpKernelContext` // parameter is always null for the former. Status CompileImpl( const XlaCompiler::CompileOptions& compile_options,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 22.1K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util.h
private: xla::LocalClient* client_; se::DeviceMemoryAllocator* xla_allocator_; bool allocate_xla_tensors_; bool use_multiple_streams_; int device_ordinal_; }; // A simple TensorBuffer implementation that allows us to create Tensors that // take ownership of pre-allocated memory. class XlaTensorBuffer : public TensorBuffer { public: XlaTensorBuffer(const void* ptr, size_t expected_size, size_t actual_size,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 11.8K bytes - Viewed (0) -
tensorflow/compiler/jit/flags.h
// from an XLA cluster. bool tf_xla_check_cluster_output_numerics; // Disables all constant folding. The primary use for this is for testing to // guarantee that tests are run on XLA and not on TF's CPU implementation. bool tf_xla_disable_constant_folding; // Disables full embedding pipelining when true. Instead, strict SparseCore // TensorCore sequencing will be used. bool tf_xla_disable_full_embedding_pipelining;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 17 18:52:57 UTC 2024 - 14.5K bytes - Viewed (0) -
testing/performance/src/templates/native-dependents-resources/googleTest/libs/googleTest/1.7.0/include/gtest/gtest-printers.h
// A user can override this behavior for a class type Foo by defining // a << operator in the namespace where Foo is defined. // // We put this operator in namespace 'internal2' instead of 'internal' // to simplify the implementation, as much code in 'internal' needs to // use << in STL, which would conflict with our own << were it defined // in 'internal'. // // Note that this operator<< takes a generic std::basic_ostream<Char,
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 04 07:21:38 UTC 2024 - 30.9K bytes - Viewed (0) -
testing/performance/src/templates/native-dependents-resources/googleTest/libs/googleTest/1.7.0/include/gtest/internal/gtest-death-test-internal.h
public: virtual ~DeathTestFactory() { } virtual bool Create(const char* statement, const RE* regex, const char* file, int line, DeathTest** test) = 0; }; // A concrete DeathTestFactory implementation for normal use. class DefaultDeathTestFactory : public DeathTestFactory { public: virtual bool Create(const char* statement, const RE* regex, const char* file, int line, DeathTest** test); };
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 04 07:21:38 UTC 2024 - 13.1K bytes - Viewed (0) -
testing/performance/src/templates/native-dependents-resources/googleTest/libs/googleTest/1.7.0/include/gtest/internal/gtest-param-util.h
mutable scoped_ptr<const T> value_; }; // class ValuesInIteratorRangeGenerator::Iterator // No implementation - assignment is unsupported. void operator=(const ValuesInIteratorRangeGenerator& other); const ContainerType container_; }; // class ValuesInIteratorRangeGenerator // INTERNAL IMPLEMENTATION - DO NOT USE IN USER CODE. // // Stores a parameter value and later creates tests parameterized with that
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 04 07:21:38 UTC 2024 - 23.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/calibration/component.h
// running calibration. absl::StatusOr<ModuleOp> ImportCalibratedSavedModel( absl::string_view calibrated_saved_model_path); absl::Nonnull<MLIRContext*> ctx_; // Contains function implementations from the python layer. Should be injected // from the python level using pybind11. absl::Nonnull<const tensorflow::quantization::PyFunctionLibrary*> py_function_lib_; // Path to the pre-calibrated SavedModel.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 5.4K bytes - Viewed (0) -
tensorflow/c/kernels.h
#endif typedef struct TF_Tensor TF_Tensor; // -------------------------------------------------------------------------- // C API for TensorFlow Kernels. // // This API allows developers to register custom kernel implementations for // TensorFlow. // // See c_api.h header comments for a discussion about API conventions. // // Users wishing to extend TensorFlow with new kernels will call
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 09 22:46:22 UTC 2024 - 24.6K bytes - Viewed (0) -
testing/performance/src/templates/native-dependents-resources/googleTest/libs/googleTest/1.7.0/include/gtest/internal/gtest-port.h
// Determines whether Google Test's own tr1 tuple implementation // should be used. #ifndef GTEST_USE_OWN_TR1_TUPLE // The user didn't tell us, so we need to figure it out. // We use our own TR1 tuple if we aren't sure the user has an // implementation of it already. At this time, libstdc++ 4.0.0+ and // MSVC 2010 are the only mainstream standard libraries that come // with a TR1 tuple implementation. NVIDIA's CUDA NVCC compiler
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 04 07:21:38 UTC 2024 - 67.2K bytes - Viewed (0) -
testing/performance/src/templates/native-dependents-resources/googleTest/libs/googleTest/1.7.0/include/gtest/internal/gtest-tuple.h
typename T6 = void, typename T7 = void, typename T8 = void, typename T9 = void> class tuple; // Anything in namespace gtest_internal is Google Test's INTERNAL // IMPLEMENTATION DETAIL and MUST NOT BE USED DIRECTLY in user code. namespace gtest_internal { // ByRef<T>::type is T if T is a reference; otherwise it's const T&. template <typename T>
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 04 07:21:38 UTC 2024 - 27.6K bytes - Viewed (0)