- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 472 for compute (0.15 sec)
-
configure.py
native_cuda_compute_capabilities = get_native_cuda_compute_capabilities( environ_cp) if not native_cuda_compute_capabilities: default_cuda_compute_capabilities = _DEFAULT_CUDA_COMPUTE_CAPABILITIES else: default_cuda_compute_capabilities = native_cuda_compute_capabilities ask_cuda_compute_capabilities = ( 'Please specify a list of comma-separated CUDA compute capabilities '
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 18:25:36 GMT 2024 - 53.8K bytes - Viewed (1) -
tensorflow/c/experimental/gradients/math_grad.cc
"Expected numeric or variant tensor, got dtype ", dtype); } return tensorflow::ops::Conj(ctx, input, output, name); } class AddGradientFunction : public GradientFunction { public: Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { // TODO(b/161805092): Support broadcasting.
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 15.2K bytes - Viewed (0) -
tensorflow/c/c_test.c
TF_Status* s = TF_NewStatus(); TF_OpKernelConstruction_GetAttrType(ctx, "foobar", &type, s); TF_DeleteStatus(s); return NULL; } // A compute function. This will never actually get called in this test, it's // just nice to know that it compiles. void compute(void* kernel, TF_OpKernelContext* ctx) { TF_Tensor* input; TF_Status* s = TF_NewStatus(); TF_GetInput(ctx, 0, &input, s); TF_DeleteTensor(input);
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Apr 24 20:50:35 GMT 2024 - 2.8K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad.cc
: forward_outputs_(f_outputs) { for (auto output : forward_outputs_) { if (output) { output->Ref(); } } } Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { AbstractTensorHandle* upstream_grad = grad_outputs[0];
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
api/maven-api-core/src/main/java/org/apache/maven/api/SessionData.java
*/ @Nullable <T> T get(@Nonnull Key<T> key); /** * Retrieve of compute the data associated with the specified key. * * @param key the key for which to retrieve the session data, must not be {@code null} * @param supplier the supplier will compute the new value * @return the session data associated with the key */ @Nullable
Java - Registered: Sun May 05 03:35:11 GMT 2024 - Last Modified: Wed Feb 28 23:31:09 GMT 2024 - 4.7K bytes - Viewed (0) -
.bazelrc
build:cross_compile_linux_arm64_xla --crosstool_top=//tools/toolchains/cross_compile/cc:cross_compile_toolchain_suite # RBE cross-compile configs for Linux Aarch64 build:rbe_cross_compile_linux_arm64 --config=cross_compile_linux_arm64 build:rbe_cross_compile_linux_arm64 --config=rbe_cross_compile_base test:rbe_cross_compile_linux_arm64 --config=rbe_cross_compile_base # RBE cross-compile configs for XLA Linux Aarch64
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu May 02 19:34:20 GMT 2024 - 52.8K bytes - Viewed (2) -
api/maven-api-core/src/main/java/org/apache/maven/api/ExtensibleEnum.java
* * @since 4.0.0 */ @Experimental public interface ExtensibleEnum { /** * The {@code id} uniquely represents a value for this extensible enum. * This id should be used to compute the equality and hash code for the instance. * * @return the id */ @Nonnull String id();
Java - Registered: Sun May 05 03:35:11 GMT 2024 - Last Modified: Fri Mar 01 17:18:13 GMT 2024 - 1.5K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/array_grad.cc
#include "tensorflow/c/eager/abstract_context.h" namespace tensorflow { namespace gradients { namespace { class IdentityNGradientFunction : public GradientFunction { public: Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { for (int i = 0; i < grad_outputs.size(); i++) {
C++ - Registered: Tue Apr 09 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 1.6K bytes - Viewed (0) -
api/maven-api-core/src/main/java/org/apache/maven/api/DependencyScope.java
Java - Registered: Sun May 05 03:35:11 GMT 2024 - Last Modified: Wed Mar 27 14:46:12 GMT 2024 - 3.7K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
// If provided, a non-null `forward_function` will be used instead of the // backward function (`backward_function_getter` / // `backward_function_deleter`) to compute jvps for this operation. If // `forward_function` is null, a GradientTape is used on the backward function // to compute the jvp, which will waste computation when executing eagerly. // // Unlike GradientTape::RecordOperation, Accumulate runs gradient computation
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1)